[one-users] OpenNebula 3.8.3 and VMWare ESXi 5.0: internal error HTTP response code 403 for upload

chenxiang at aquala-tech.com chenxiang at aquala-tech.com
Wed Apr 24 08:08:46 PDT 2013


A lot of new things to learn. After getting ttylinux work with ESXi, I
headed for CentOS 6.2 (also imported from Marketplace), and it did not
work as expected.

This time I already have a datastore with DS_MAD="vmware". I imported
CentOS 6.2 from marketplace, and had the following in that datastore
directory, which looked alright:

oneadmin at opennebula:~/var/datastores/100/c5610951183fa21ac0900ce147a677b7$
ls -l
total 8388616
-rw------- 1 oneadmin cloud 8589934592 Jul  2  2012 CentOS-6.2-flat.vmdk
-rw------- 1 oneadmin cloud        515 Jul  3  2012 disk.vmdk

Then I instantiated a VM, but it failed to boot (prolog seemed to be OK).
Below is the VM log:

Wed Apr 24 22:40:41 2013 [DiM][I]: New VM state is ACTIVE.
Wed Apr 24 22:40:41 2013 [LCM][I]: New VM state is PROLOG.
Wed Apr 24 22:40:41 2013 [VM][I]: Virtual Machine has no context
Wed Apr 24 22:47:36 2013 [TM][I]: clone: Cloning
opennebula:/srv/cloud/one/var/datastores/100/c5610951183fa21ac0900ce147a677b7
in /vmfs/volumes/101/135/disk.0
Wed Apr 24 22:47:36 2013 [TM][I]: ExitCode: 0
Wed Apr 24 22:47:37 2013 [LCM][I]: New VM state is BOOT
Wed Apr 24 22:47:37 2013 [VMM][I]: Generating deployment file:
/srv/cloud/one/var/vms/135/deployment.1
Wed Apr 24 22:47:37 2013 [VMM][I]: ExitCode: 0
Wed Apr 24 22:47:37 2013 [VMM][I]: Successfully execute network driver
operation: pre.
Wed Apr 24 22:47:47 2013 [VMM][I]: Command execution fail:
/srv/cloud/one/var/remotes/vmm/vmware/deploy
/srv/cloud/one/var/vms/135/deployment.1 vmware01 135 vmware01
Wed Apr 24 22:47:47 2013 [VMM][D]: deploy: Successfully defined domain
one-135.
Wed Apr 24 22:47:47 2013 [VMM][E]: deploy: Error executing: virsh -c
'esx://vmware01/?no_verify=1&auto_answer=1' start one-135 err: ExitCode: 1
Wed Apr 24 22:47:47 2013 [VMM][I]: out:
Wed Apr 24 22:47:47 2013 [VMM][I]: error: Failed to start domain one-135
Wed Apr 24 22:47:47 2013 [VMM][I]: error: internal error Could not start
domain: FileNotFound - File [101] 135/disk.0/disk.vmdk was not found
Wed Apr 24 22:47:47 2013 [VMM][I]:
Wed Apr 24 22:47:47 2013 [VMM][I]: ExitCode: 1
Wed Apr 24 22:47:47 2013 [VMM][I]: Failed to execute virtualization driver
operation: deploy.
Wed Apr 24 22:47:47 2013 [VMM][E]: Error deploying virtual machine
Wed Apr 24 22:47:47 2013 [DiM][I]: New VM state is FAILED

I double check on the ESXi box, File [101] 135/disk.0/disk.vmdk was
actually there.

/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0 # ls -l
-rw-------    1 oneadmin root         8589934592 Apr 24 14:47
CentOS-6.2-flat.vmdk
-rw-------    1 oneadmin root                515 Apr 24 14:47 disk.vmdk
-rw-r--r--    1 root     root                  0 Apr 24 14:47 one-135.vmsd
-rw-r--r--    1 root     root                875 Apr 24 14:47 one-135.vmx
-rw-r--r--    1 root     root                262 Apr 24 14:47 one-135.vmxf
-rw-r--r--    1 root     root              38580 Apr 24 14:47 vmware.log

So I looked into vmware.log, and found that it was looking for
CentOS-6.2.vmdk. I did not have this file anywhere.

2013-04-24T14:47:35.839Z| vmx| DISK: OPEN scsi0:0
'/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk'
persistent R[]
2013-04-24T14:47:35.842Z| vmx| AIOGNRC: Failed to open
'/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/CentOS-6.2.vmdk'
: Could not find the file (600000003) (0x2013).
2013-04-24T14:47:35.842Z| vmx| DISKLIB-VMFS  :
"/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/CentOS-6.2.vmdk"
: failed to open (The system cannot find the file specified): AIOMgr_Open
failed. Type 3
2013-04-24T14:47:35.842Z| vmx| DISKLIB-LINK  :
"/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk" :
failed to open (The system cannot find the file specified).
2013-04-24T14:47:35.842Z| vmx| DISKLIB-CHAIN :
"/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk" :
failed to open (The system cannot find the file specified).
2013-04-24T14:47:35.842Z| vmx| DISKLIB-LIB   : Failed to open
'/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk'
with flags 0xa The system cannot find the file specified (25).
2013-04-24T14:47:35.842Z| vmx| Msg_Post: Error
2013-04-24T14:47:35.842Z| vmx| [msg.disk.fileNotFound] VMware ESX cannot
find the virtual disk
"/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk".
Verify the path is valid and try again.
2013-04-24T14:47:35.842Z| vmx| [msg.disk.noBackEnd] Cannot open the disk
'/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk'
or one of the snapshot disks it depends on.
2013-04-24T14:47:35.842Z| vmx| [msg.disk.configureDiskError] Reason: The
system cannot find the file specified.

So, the behavior of the CentOS 6.2 image is different from the ttylinux
image. What should I do?


Chen Xiang


> Ok, I see what happened then. Datastores to be used with VMwaqre ESX
> hosts must be wither DS_MAD="vmware" or DS_MAD="vmfs", otherwise the
> drivers won't know how to handle VMware vmdk disks.
>
> Glad is working now.
>
> -T
> --
> Constantino Vázquez Blanco, PhD, MSc
> Project Engineer
> OpenNebula - The Open-Source Solution for Data Center Virtualization
> www.OpenNebula.org | @tinova79 | @OpenNebula
>
>
> On Wed, Apr 24, 2013 at 12:28 PM,  <chenxiang at aquala-tech.com> wrote:
>> The disk.vmdk problem seems to be associated with the DS_MAD setting. I
>> created two datastores to test this, the first one with DS_MAD="vmware"
>> and the second one with DS_MAD="fs". When I imported the images from
>> marketplace, in the first datastore I have disk.vmdk, and in the other
>> one
>> I have ttylinux.vmdk.
>>
>> When I encountered the problem in this thread, I was importing the
>> ttylinux image into the default datastore, which had DS_MAD="fs". So,
>> that
>> caused the problem.
>>
>> A lot of things to learn. Thank you so much for your assistance. You are
>> so helpful for me.
>>
>> Chen Xiang
>>
>>
>>> On Wed, Apr 24, 2013 at 7:39 AM,  <chenxiang at aquala-tech.com> wrote:
>>>> I double checked the VNC issue. I did set the VNC stuff according to
>>>> the
>>>> instructions, but that setting got lost across reboots, so I was not
>>>> able
>>>> to connect to VNC.
>>>>
>>>> I searched on the web, and found a solution. That is to copy the
>>>> modified
>>>> version of service.xml to the localdisk of ESXi as a backup, and then
>>>> modified /etc/rc.local to recover that file from the backup version,
>>>> and
>>>> to refresh the firewall settings. (I need to do the same for the
>>>> oneadmin
>>>> SSH key, otherwise the front end won't be able to connect to the ESXi
>>>> box
>>>> when the box reboots.) Then VNC worked.
>>>
>>> Glad you got it working. A similar approach is defined here [1]
>>> (search for "persistency")
>>>
>>> [1] http://opennebula.org/documentation:rel3.8:evmwareg
>>>
>>>>
>>>> Regarding the disk.vmdk. I checked into the oned.log. No I did not
>>>> find
>>>> the error message you mentioned.
>>>
>>> Can you send us the relevant files in oned.log at the time of
>>> registering a VMware disk folder _without_ the manual renaming to
>>> disk.vmdk?
>>>
>>> Thanks a lot for your feedback,
>>>
>>> -T
>>>
>>>>
>>>> Chen Xiang
>>>>
>>>>
>>>>> Hi,
>>>>>
>>>>> comments inline,
>>>>>
>>>>> On Tue, Apr 23, 2013 at 8:54 AM,  <chenxiang at aquala-tech.com> wrote:
>>>>>> Now I am able to get VM's running with ESXi nodes. I would like to
>>>>>> share
>>>>>> what I did to make things work, and would like you guys to help me
>>>>>> further.
>>>>>>
>>>>>> From Marketplace I imported the ttylinux-vmware image to local
>>>>>> infrastructure. When the images becomes READY, login to the front
>>>>>> end
>>>>>> server and browse to var/datastores/1 and further browse to the
>>>>>> directory
>>>>>> holding that particular image. Below was what I had when things did
>>>>>> not
>>>>>> work:
>>>>>>
>>>>>> oneadmin at opennebula:~/var/datastores/1/10fc21f21a452add3838d76d63052457$
>>>>>> ls -l
>>>>>> total 104864
>>>>>> -rw------- 1 oneadmin cloud 107374080 Jul  3  2012
>>>>>> ttylinux-flat.vmdk
>>>>>> -rw------- 1 oneadmin cloud       509 Jul  3  2012 ttylinux.vmdk
>>>>>>
>>>>>> Base on the error message I got from VM log, I decided to copy
>>>>>> ttylinux.vmdk to disk.vmdk. So I did that and then I had the
>>>>>> following:
>>>>>>
>>>>>> oneadmin at opennebula:~/var/datastores/1/10fc21f21a452add3838d76d63052457$
>>>>>> cp ttylinux.vmdk disk.vmdk
>>>>>> oneadmin at opennebula:~/var/datastores/1/10fc21f21a452add3838d76d63052457$
>>>>>> ls -l
>>>>>> total 104868
>>>>>> -rw------- 1 oneadmin cloud       509 Apr 23 14:33 disk.vmdk
>>>>>> -rw------- 1 oneadmin cloud 107374080 Jul  3  2012
>>>>>> ttylinux-flat.vmdk
>>>>>> -rw------- 1 oneadmin cloud       509 Jul  3  2012 ttylinux.vmdk
>>>>>>
>>>>>> Now I went back to the same template, and instantiated a VM
>>>>>> instance,
>>>>>> now
>>>>>> it worked.
>>>>>
>>>>> The vmware/cp script should automatically rename the file. Can you
>>>>> see
>>>>> any line in /var/log/one/oned.log similar to:
>>>>>
>>>>> --
>>>>> Error renaming disk file $BASE_DISK_FILE to disk.vmdk
>>>>> --
>>>>>
>>>>>
>>>>>>
>>>>>> Now I have a new problem. This front-end has been tested for KVM
>>>>>> with
>>>>>> hundreds of VM instances before I tried ESXi. Now I am trying VM-ID
>>>>>> above
>>>>>> 100. Now the VM instances were running, but I was not able to
>>>>>> connect
>>>>>> to
>>>>>> the VM console via VNC from SunStone. What should I do? (VM Console
>>>>>> in
>>>>>> vSphere Client still works.)
>>>>>
>>>>>
>>>>> Have you configured the ESX host to allow VNC connections? See
>>>>> http://opennebula.org/documentation:rel3.8:evmwareg#vnc
>>>>>
>>>>> Regards,
>>>>>
>>>>> -Tino
>>>>>
>>>>>>
>>>>>> Thanks a lot for your assistance.
>>>>>>
>>>>>> Chen Xiang
>>>>>>
>>>>>>
>>>>>>
>>>>>>> With the proper NFS settings, I was able to define the domain, but
>>>>>>> fail
>>>>>>> to
>>>>>>> create the VM. The error message says "File [0]
>>>>>>> 120/disk.0/disk.vmdk
>>>>>>> was
>>>>>>> not found".
>>>>>>>
>>>>>>> I login to the ESXi box, and verified that I do have
>>>>>>> /vmfs/volumes/0
>>>>>>> pointing to the right NFS mount. And inside /vmfs/volumes/0 I do
>>>>>>> have
>>>>>>> 120/disk.0/disk.vmdk .
>>>>>>>
>>>>>>> Any ideas?
>>>>>>>
>>>>>>> Below is a copy of the VM log.
>>>>>>>
>>>>>>> Tue Apr 23 13:12:58 2013 [DiM][I]: New VM state is ACTIVE.
>>>>>>> Tue Apr 23 13:12:58 2013 [LCM][I]: New VM state is PROLOG.
>>>>>>> Tue Apr 23 13:12:58 2013 [VM][I]: Virtual Machine has no context
>>>>>>> Tue Apr 23 13:13:12 2013 [TM][I]: clone: Cloning
>>>>>>> /vmfs/volumes/1/10fc21f21a452add3838d76d63052457 in
>>>>>>> vmware02:/vmfs/volumes/0/120/disk.0
>>>>>>> Tue Apr 23 13:13:12 2013 [TM][I]: ExitCode: 0
>>>>>>> Tue Apr 23 13:13:12 2013 [LCM][I]: New VM state is BOOT
>>>>>>> Tue Apr 23 13:13:12 2013 [VMM][I]: Generating deployment file:
>>>>>>> /srv/cloud/one/var/vms/120/deployment.0
>>>>>>> Tue Apr 23 13:13:12 2013 [VMM][I]: ExitCode: 0
>>>>>>> Tue Apr 23 13:13:12 2013 [VMM][I]: Successfully execute network
>>>>>>> driver
>>>>>>> operation: pre.
>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: Command execution fail:
>>>>>>> /srv/cloud/one/var/remotes/vmm/vmware/deploy
>>>>>>> /srv/cloud/one/var/vms/120/deployment.0 vmware02 120 vmware02
>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][D]: deploy: Successfully defined
>>>>>>> domain
>>>>>>> one-120.
>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][E]: deploy: Error executing: virsh
>>>>>>> -c
>>>>>>> 'esx://vmware02/?no_verify=1&auto_answer=1' start one-120 err:
>>>>>>> ExitCode: 1
>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: out:
>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: error: Failed to start domain
>>>>>>> one-120
>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: error: internal error Could not
>>>>>>> start
>>>>>>> domain: FileNotFound - File [0] 120/disk.0/disk.vmdk was not found
>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]:
>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: ExitCode: 1
>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: Failed to execute virtualization
>>>>>>> driver
>>>>>>> operation: deploy.
>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][E]: Error deploying virtual machine
>>>>>>> Tue Apr 23 13:13:24 2013 [DiM][I]: New VM state is FAILED
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Chen Xiang
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> Please use root_squash instead of no_root_squash
>>>>>>>>
>>>>>>>> --
>>>>>>>> /srv/cloud/one/var/datastores/0
>>>>>>>> *(rw,sync,no_subtree_check,no_root_squash,anonuid=10000,anongid=10000)
>>>>>>>> /srv/cloud/one/var/datastores/1
>>>>>>>> *(rw,sync,no_subtree_check,no_root_squash,anonuid=10000,anongid=10000)
>>>>>>>> --
>>>>>>>>
>>>>>>>> You will need to force the nfs server to re-read the conf file.
>>>>>>>>
>>>>>>>> Regards,
>>>>>>>>
>>>>>>>> -Tino
>>>>>>>> --
>>>>>>>> Constantino Vázquez Blanco, PhD, MSc
>>>>>>>> Project Engineer
>>>>>>>> OpenNebula - The Open-Source Solution for Data Center
>>>>>>>> Virtualization
>>>>>>>> www.OpenNebula.org | @tinova79 | @OpenNebula
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Apr 22, 2013 at 3:46 PM,  <chenxiang at aquala-tech.com>
>>>>>>>> wrote:
>>>>>>>>> Here is what I got when trying to define the VM via virsh:
>>>>>>>>>
>>>>>>>>> oneadmin at opennebula:~/images$ virsh -c
>>>>>>>>> 'esx://vmware02/?no_verify=1&auto_answer=1'
>>>>>>>>> Enter username for vmware02 [root]:
>>>>>>>>> Enter root's password for vmware02:
>>>>>>>>> Welcome to virsh, the virtualization interactive terminal.
>>>>>>>>>
>>>>>>>>> Type:  'help' for help with commands
>>>>>>>>>        'quit' to quit
>>>>>>>>>
>>>>>>>>> virsh # define /srv/cloud/one/var/vms/111/deployment.0
>>>>>>>>> 2013-04-22 13:24:49.391+0000: 17332: info : libvirt version:
>>>>>>>>> 0.9.10
>>>>>>>>> 2013-04-22 13:24:49.391+0000: 17332: warning :
>>>>>>>>> virVMXFormatVNC:3224
>>>>>>>>> :
>>>>>>>>> VNC
>>>>>>>>> port 6011 it out of [5900..5964] range
>>>>>>>>> error: Failed to define domain from
>>>>>>>>> /srv/cloud/one/var/vms/111/deployment.0
>>>>>>>>> error: internal error HTTP response code 403 for upload to
>>>>>>>>> 'https://vmware02:443/folder/111%2fdisk%2e0/one%2d111.vmx?dcPath=ha%2ddatacenter&dsName=0'
>>>>>>>>>
>>>>>>>>> virsh # exit
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>>
>>>>>>>>>> I am running OpenNebula 3.8.3 on Ubuntu 12.04 (front end) with
>>>>>>>>>> VMWare
>>>>>>>>>> ESXi
>>>>>>>>>> 5.0 (node). I am able to create a VMWare node in SunStone,
>>>>>>>>>> register
>>>>>>>>>> the
>>>>>>>>>> ttylinux images (ttylinux.vmdk.tar download from C12G.com), but
>>>>>>>>>> failed
>>>>>>>>>> to
>>>>>>>>>> instantiate a VM.
>>>>>>>>>>
>>>>>>>>>> Here is what I have as the NFS exports:
>>>>>>>>>>
>>>>>>>>>> /srv/cloud/one/var/datastores/0
>>>>>>>>>> *(rw,sync,no_subtree_check,no_root_squash,anonuid=10000,anongid=10000)
>>>>>>>>>> /srv/cloud/one/var/datastores/1
>>>>>>>>>> *(rw,sync,no_subtree_check,no_root_squash,anonuid=10000,anongid=10000)
>>>>>>>>>>
>>>>>>>>>> One the ESXi node I mounted the NFS exports to /vmfs/volumes/0
>>>>>>>>>> and
>>>>>>>>>> /vmfs/volumes/1 respectively.
>>>>>>>>>>
>>>>>>>>>> Yes I configured libvirt-0.9.10 and make it worked with VMWare,
>>>>>>>>>> verified
>>>>>>>>>> by command lines such as the following command (where vmware02
>>>>>>>>>> is
>>>>>>>>>> my
>>>>>>>>>> ESXi
>>>>>>>>>> hostname):
>>>>>>>>>>
>>>>>>>>>> virsh -c 'esx://vmware02/?no_verify=1&auto_answer=1'
>>>>>>>>>>
>>>>>>>>>> Below is my VM template:
>>>>>>>>>>
>>>>>>>>>> CPU="1"
>>>>>>>>>> DISK=[
>>>>>>>>>>   IMAGE="tty_vmdk",
>>>>>>>>>>   IMAGE_UNAME="oneadmin" ]
>>>>>>>>>> GRAPHICS=[
>>>>>>>>>>   LISTEN="0.0.0.0",
>>>>>>>>>>   TYPE="vnc" ]
>>>>>>>>>> MEMORY="512"
>>>>>>>>>> NAME="ttylinux"
>>>>>>>>>>
>>>>>>>>>> Below is what I got when trying to instantiate a VM:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Mon Apr 22 21:27:11 2013 [DiM][I]: New VM state is ACTIVE.
>>>>>>>>>> Mon Apr 22 21:27:11 2013 [LCM][I]: New VM state is PROLOG.
>>>>>>>>>> Mon Apr 22 21:27:11 2013 [VM][I]: Virtual Machine has no context
>>>>>>>>>> Mon Apr 22 21:27:17 2013 [TM][I]: clone: Cloning
>>>>>>>>>> /vmfs/volumes/1/43352fb75cee9bbc1da3c1e7ff474e26 in
>>>>>>>>>> vmware02:/vmfs/volumes/0/112/disk.0
>>>>>>>>>> Mon Apr 22 21:27:17 2013 [TM][I]: ExitCode: 0
>>>>>>>>>> Mon Apr 22 21:27:17 2013 [LCM][I]: New VM state is BOOT
>>>>>>>>>> Mon Apr 22 21:27:17 2013 [VMM][I]: Generating deployment file:
>>>>>>>>>> /srv/cloud/one/var/vms/112/deployment.0
>>>>>>>>>> Mon Apr 22 21:27:17 2013 [VMM][I]: ExitCode: 0
>>>>>>>>>> Mon Apr 22 21:27:17 2013 [VMM][I]: Successfully execute network
>>>>>>>>>> driver
>>>>>>>>>> operation: pre.
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: Command execution fail:
>>>>>>>>>> /srv/cloud/one/var/remotes/vmm/vmware/deploy
>>>>>>>>>> /srv/cloud/one/var/vms/112/deployment.0 vmware02 112 vmware02
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][E]: deploy: Error executing:
>>>>>>>>>> virsh
>>>>>>>>>> -c
>>>>>>>>>> 'esx://vmware02/?no_verify=1&auto_answer=1' define
>>>>>>>>>> /srv/cloud/one/var/vms/112/deployment.0 err: ExitCode: 1
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: out:
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: 2013-04-22 13:27:21.858+0000:
>>>>>>>>>> 17586:
>>>>>>>>>> info : libvirt version: 0.9.10
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: 2013-04-22 13:27:21.858+0000:
>>>>>>>>>> 17586:
>>>>>>>>>> warning : virVMXFormatVNC:3224 : VNC port 6012 it out of
>>>>>>>>>> [5900..5964]
>>>>>>>>>> range
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: error: Failed to define
>>>>>>>>>> domain
>>>>>>>>>> from
>>>>>>>>>> /srv/cloud/one/var/vms/112/deployment.0
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: error: internal error HTTP
>>>>>>>>>> response
>>>>>>>>>> code 403 for upload to
>>>>>>>>>> 'https://vmware02:443/folder/112%2fdisk%2e0/one%2d112.vmx?dcPath=ha%2ddatacenter&dsName=0'
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]:
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: ExitCode: 255
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: Failed to execute
>>>>>>>>>> virtualization
>>>>>>>>>> driver
>>>>>>>>>> operation: deploy.
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][E]: Error deploying virtual
>>>>>>>>>> machine
>>>>>>>>>> Mon Apr 22 21:27:22 2013 [DiM][I]: New VM state is FAILED
>>>>>>>>>>
>>>>>>>>>> What might be wrong? This looks like that I did not have write
>>>>>>>>>> access
>>>>>>>>>> to
>>>>>>>>>> some place. So I tried both the oneadmin and root acount in
>>>>>>>>>> etc/vmwarerc,
>>>>>>>>>> with not much luck.
>>>>>>>>>>
>>>>>>>>>> On the front end the oneadmin user belongs to the following
>>>>>>>>>> groups:
>>>>>>>>>> cloud,
>>>>>>>>>> adm, sudo, libvirtd.
>>>>>>>>>>
>>>>>>>>>> Best regards,
>>>>>>>>>>
>>>>>>>>>> Chen Xiang
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> _______________________________________________
>>>>>>>>>> Users mailing list
>>>>>>>>>> Users at lists.opennebula.org
>>>>>>>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> Users mailing list
>>>>>>>>> Users at lists.opennebula.org
>>>>>>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Users mailing list
>>>>>>> Users at lists.opennebula.org
>>>>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> Users mailing list
>>>>>> Users at lists.opennebula.org
>>>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users at lists.opennebula.org
>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>>
>>
>>
>> _______________________________________________
>> Users mailing list
>> Users at lists.opennebula.org
>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>




More information about the Users mailing list