[one-users] Problem creating disk images

Jon three18ti at gmail.com
Mon May 13 10:33:34 PDT 2013


Hello All,

I've got an answer of sorts... It seems that OpenNebula doesn't like the
"qcow2" type of driver.

If I use the following machine definition to define the vm and start it, I
get the same "Invalid argument" error:

>> oneadmin at loki:~$ virsh define /var/lib/one/vms/99/deployment.0
>> oneadmin at loki:~$ virsh start one-99
>> error: Failed to start domain one-99
>> error: internal error process exited while connecting to monitor: kvm:
-drive
>>
file=/var/lib/one//datastores/0/99/disk.1,if=none,id=drive-ide0-1-0,format=qcow2:
>> could not open disk image /var/lib/one//datastores/0/99/disk.1: Invalid
argument


However, changing "qcow2" to "raw" in the below deployment allows me to
define the VM with virsh and start it. (Or edit it directly in libvirt)

This would indicate to me that

Deployment file in question:
oneadmin at loki:~$ cat /var/lib/one/vms/99/deployment.0
<domain type='kvm' xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'>
        <name>one-99</name>
        <cputune>
                <shares>1024</shares>
        </cputune>
        <memory>1048576</memory>
        <os>
                <type arch='x86_64'>hvm</type>
                <boot dev='hd'/>
        </os>
        <devices>
                <emulator>/usr/bin/kvm</emulator>
                <disk type='file' device='cdrom'>
                        <source
file='/var/lib/one//datastores/0/99/disk.0'/>
                        <target dev='hdb'/>
                        <readonly/>
                        <driver name='qemu' type='raw' cache='none'/>
                </disk>
                <disk type='file' device='disk'>
                        <source
file='/var/lib/one//datastores/0/99/disk.1'/>
                        <target dev='hdc'/>
                        <driver name='qemu' type='qcow2' cache='default'/>
                </disk>
                <disk type='file' device='disk'>
                        <source
file='/var/lib/one//datastores/0/99/disk.2'/>
                        <target dev='hdd'/>
                        <driver name='qemu' type='qcow2' cache='default'/>
                </disk>
                <disk type='file' device='cdrom'>
                        <source
file='/var/lib/one//datastores/0/99/disk.3'/>
                        <target dev='hda'/>
                        <readonly/>
                        <driver name='qemu' type='raw'/>
                </disk>
                <interface type='bridge'>
                        <source bridge='ovsbr0'/>
                        <virtualport type='openvswitch'/>
                        <mac address='02:00:44:47:83:43'/>
                </interface>
        </devices>
        <features>
                <acpi/>
        </features>
</domain>

Template used to create deployment file in question:

TEMPLATE 15 INFORMATION

ID             : 15
NAME           : ubuntu-13.04-x86_64-with_storage
USER           : oneadmin
GROUP          : oneadmin
REGISTER TIME  : 04/27 17:47:17

PERMISSIONS

OWNER          : um-
GROUP          : ---
OTHER          : ---

TEMPLATE CONTENTS

CONTEXT=[
  ETH0_DNS="$NETWORK[DNS,NETWORK_ID=\"3\"]",
  ETH0_GATEWAY="$NETWORK[GATEWAY,NETWORK_ID=\"3\"]",
  ETH0_IP="$NIC[IP,NETWORK_ID=\"3\"]",
  ETH0_MASK="$NETWORK[NETWORK_MASK,NETWORK_ID=\"3\"]",
  ETH0_NETWORK="$NETWORK[NETWORK_ADDRESS,NETWORK_ID=\"3\"]",
  SSH_PUBLIC_KEY="ssh-rsa abcd foo at bar.com ]
CPU="1"
DISK=[
  IMAGE_ID="11" ]
DISK=[
  CACHE="default",
  DRIVER="qcow2",
  FORMAT="ext4",
  SIZE="5120",
  TYPE="fs" ]
DISK=[
  CACHE="default",
  DRIVER="qcow2",
  SIZE="2048",
  TYPE="swap" ]
MEMORY="1024"
NIC=[
  NETWORK_ID="0" ]
OS=[
  ARCH="x86_64" ]

Here is the template of the VM with RAW disks that starts successfully:

oneadmin at loki:~$ onetemplate show 25
TEMPLATE 25 INFORMATION

ID             : 25
NAME           : ubuntu-13.04-test_raw_storage
USER           : oneadmin
GROUP          : oneadmin
REGISTER TIME  : 05/13 11:11:24

PERMISSIONS

OWNER          : um-
GROUP          : ---
OTHER          : ---

TEMPLATE CONTENTS

CONTEXT=[
  ETH0_CONTEXT_FORCE_IPV4="$NETWORK[CONTEXT_FORCE_IPV4,NETWORK_ID=\"0\"]",
  ETH0_DNS="$NETWORK[DNS,NETWORK_ID=\"0\"]",
  ETH0_GATEWAY="$NETWORK[GATEWAY,NETWORK_ID=\"0\"]",
  ETH0_GATEWAY6="$NETWORK[GATEWAY6,NETWORK_ID=\"0\"]",
  ETH0_IP="$NIC[IP,NETWORK_ID=\"0\"]",
  ETH0_IPV6="$NIC[IP6_GLOBAL,NETWORK_ID=\"0\"]",
  ETH0_MASK="$NETWORK[NETWORK_MASK,NETWORK_ID=\"0\"]",
  ETH0_NETWORK="$NETWORK[NETWORK_ADDRESS,NETWORK_ID=\"0\"]",
  SSH_PUBLIC_KEY="ssh-rsa  ]
CPU="2"
DISK=[
  DRIVER="raw",
  FORMAT="ext4",
  SIZE="5120",
  TYPE="fs" ]
DISK=[
  DRIVER="raw",
  SIZE="2048",
  TYPE="swap" ]
DISK=[
  IMAGE_ID="11" ]
GRAPHICS=[
  LISTEN="0.0.0.0",
  TYPE="VNC" ]
MEMORY="2042"
NIC=[
  NETWORK_ID="0" ]
OS=[
  ARCH="x86_64" ]
VCPU="2"

(actually, updating template #15 to reflect "raw" drivers in place of
"qcow2" drivers also allows the VM to boot)

So, I guess the question then becomes, how do I tell OpenNebula to create
qcow2 images instead of raw disk images?

Thanks,
Jon A



On Sun, May 12, 2013 at 1:45 PM, Jon <three18ti at gmail.com> wrote:

> Hello All,
>
> After much aggravation, I finally came up with the idea to disable
> apparmor for libvirtd.  So I copied the profile
> from /etc/apparmor.d/usr.sbin.libvirtd
> to /etc/apparmor.d/disabled/usr.sbin.libvirtd
>
> And now I am not receiving the error message:
>
> >> error: internal error cannot load AppArmor profile
> 'libvirt-8a633864-4564-a814-8dc8-73bca05476a1'
>
> While this doesn't seem ideal, OpenNebula doesn't seem to work with the
> libvirt profile active after changing owner / group that libvirt runs as.
>  Has anyone been able to get libvirt to run with an active apparmor profile
> under the oneadmin / cloud user / group on an Ubuntu system?
>
> What I find even more puzzling, even after I cleared
> out /etc/apparmor.d/libvirt/:
>
> >> root at loki:~# ls -lah /etc/apparmor.d/libvirt/
> >> total 20K
> >> drwxr-xr-x  2 root root  12K May 12 13:18 .
> >> drwxr-xr-x 10 root root 4.0K May 12 13:22 ..
> >> -rw-r--r--  1 root root  164 May 12 13:17 TEMPLATE
>
> after tearing down apparmor the profile
> for libvirt-39122dda-bedb-4808-fd25-5e68446224ff is still active:
>
> >> root at loki:~# /etc/init.d/apparmor teardown
> >>  * Unloading AppArmor profiles        [ OK ]
> >> root at loki:~# apparmor_status
> >> apparmor module is loaded.
> >> 1 profiles are loaded.
> >> 1 profiles are in enforce mode.
> >>    libvirt-39122dda-bedb-4808-fd25-5e68446224ff
> >> 0 profiles are in complain mode.
> >> 1 processes have profiles defined.
> >> 1 processes are in enforce mode.
> >>    libvirt-39122dda-bedb-4808-fd25-5e68446224ff (12050)
> >> 0 processes are in complain mode.
> >> 0 processes are unconfined but have a profile defined
>
> All that said, I'm still having trouble attaching a disk.  The good news
> is I get the same "invalid argument" error when attaching a volatile disk
> or a persistent disk, however, the error prevents the VM from booting:
>
> >> Sun May 12 13:29:14 2013 [DiM][I]: New VM state is ACTIVE.
> >> Sun May 12 13:29:15 2013 [LCM][I]: New VM state is PROLOG.
> >> Sun May 12 13:29:29 2013 [LCM][E]: monitor_done_action, VM in a wrong
> state
> >> Sun May 12 13:30:28 2013 [TM][I]: clone: Cloning
> loki:/var/lib/one/datastores/103/4a657983e37df7e79b59110748f58ce8 in
> /var/lib/one/datastores/0/98/disk.0
> >> Sun May 12 13:30:28 2013 [TM][I]: ExitCode: 0
> >> Sun May 12 13:31:00 2013 [TM][I]: mkimage: Making filesystem of 5120M
> and type ext4 at loki:/var/lib/one//datastores/0/98/disk.1
> >> Sun May 12 13:31:00 2013 [TM][I]: ExitCode: 0
> >> Sun May 12 13:31:02 2013 [TM][I]: mkimage: Making filesystem of 2048M
> and type swap at loki:/var/lib/one//datastores/0/98/disk.2
> >> Sun May 12 13:31:02 2013 [TM][I]: ExitCode: 0
> >> Sun May 12 13:31:03 2013 [TM][I]: context: Generating context block
> device at loki:/var/lib/one//datastores/0/98/disk.3
> >> Sun May 12 13:31:03 2013 [TM][I]: ExitCode: 0
> >> Sun May 12 13:31:03 2013 [LCM][I]: New VM state is BOOT
> >> Sun May 12 13:31:03 2013 [VMM][I]: Generating deployment file:
> /var/lib/one/vms/98/deployment.0
> >> Sun May 12 13:31:03 2013 [VMM][I]: ExitCode: 0
> >> Sun May 12 13:31:03 2013 [VMM][I]: Successfully execute network driver
> operation: pre.
> >> Sun May 12 13:31:07 2013 [VMM][I]: Command execution fail: cat << EOT |
> /var/tmp/one/vmm/kvm/deploy /var/lib/one//datastores/0/98/deployment.0 loki
> 98 loki
> >> Sun May 12 13:31:07 2013 [VMM][I]: error: Failed to create domain from
> /var/lib/one//datastores/0/98/deployment.0
> >> Sun May 12 13:31:07 2013 [VMM][I]: error: internal error process exited
> while connecting to monitor: kvm: -drive
> file=/var/lib/one//datastores/0/98/disk.1,if=none,id=drive-ide0-1-0,format=qcow2,cache=none:
> could not open disk image /var/lib/one//datastores/0/98/disk.1: Invalid
> argument
> >> Sun May 12 13:31:07 2013 [VMM][I]:
> >> Sun May 12 13:31:07 2013 [VMM][E]: Could not create domain from
> /var/lib/one//datastores/0/98/deployment.0
> >> Sun May 12 13:31:07 2013 [VMM][I]: ExitCode: 255
> >> Sun May 12 13:31:07 2013 [VMM][I]: Failed to execute virtualization
> driver operation: deploy.
> >> Sun May 12 13:31:07 2013 [VMM][E]: Error deploying virtual machine:
> Could not create domain from /var/lib/one//datastores/0/98/deployment.0
> >> Sun May 12 13:31:08 2013 [DiM][I]: New VM state is FAILED
>
> What is the invalid argument here?
>
> Thanks for any suggestions.
>
> Best Regards,
> Jon A
>
>
>
> On Wed, May 8, 2013 at 11:40 AM, Jon <three18ti at gmail.com> wrote:
>
>> Hey All,
>>
>> So I've just upgraded to 4.0Final, however, I am still experiencing the
>> previous errors Cannot load AppArmor Profile when attempting to use a
>> volatile disk and an invalid argument when attempting to use a datablock as
>> a disk image.
>>
>> Given that this is the the same error I've been plagued with since
>> OpenNebula 3.8, I'm fairly certain there is something wrong with my
>> configuration, but I'm at a loss as to what it could be.
>>
>> Any ideas are greatly appreciated.
>>
>> thanks,
>> Jon A
>>
>>
>> On Mon, May 6, 2013 at 11:51 AM, Jon <three18ti at gmail.com> wrote:
>>
>>> Hello Ruben,
>>>
>>> Thanks for this.  I actually remember performing this step several
>>> months ago when I installed 3.8, however, I did not perform this step when
>>> I installed 4.0.
>>>
>>> However, it seems that something is still not quite right.  I am now
>>> unable to deploy any VM with a failure message:
>>>
>>> >> error: internal error cannot load AppArmor profile
>>> 'libvirt-76e12b0b-7d3c-c71a-46f0-bfe44917aedb'
>>>
>>> And a relevant entry in /var/log/libvirt/libvirtd.log (followed by a
>>> cascading series of errors)
>>>
>>> >> 2013-05-06 17:06:39.665+0000: 1577: error : virCommandWait:2314 :
>>> internal error Child process (/usr/lib/libvirt/virt-aa-helper -p 0 -c -u
>>> libvirt-76e12b0b-7d3c-c71a-46f0-bfe44917aedb) status unexpected: exit
>>> status 1
>>> >> 2013-05-06 17:06:39.665+0000: 1577: error :
>>> AppArmorGenSecurityLabel:446 : internal error cannot load AppArmor profile
>>> 'libvirt-76e12b0b-7d3c-c71a-46f0-bfe44917aedb'
>>>
>>>
>>> The docs say I need to add the following line to
>>> /etc/apparmor.d/libvirt-qemu , I think this mean I need to add the
>>> following line to /etc/apparmor.d/abstractions/libvirt-qemu since this is
>>> the file that is loaded by the libvirt/TEMPLATE
>>>
>>> >> owner /var/lib/one/** rw,
>>>
>>> After adding this entry to /etc/apparmor.d/abstractions/libvirt-qemu and
>>> restarting apparmor my vm fails to start with the same error message.
>>>
>>> So I found an old mailing list article about the same issue:
>>>
>>> >>
>>> http://lists.opennebula.org/pipermail/users-opennebula.org/2012-June/019427.html
>>>
>>> This thread also indicates that the following line is required in
>>> /etc/apparmor.d/usr.sbin.libvirtd
>>>
>>> >> /var/lib/one/** lrwk,
>>>
>>> After doing this, I am still getting the error.
>>>
>>> The last reply in this thread says the final solution was killing the
>>> apparmor process and restarting the apparmor service.
>>>
>>> I performed a teardown, which didn't help so I performed a full system
>>> reboot.
>>>
>>> After the reboot, I still receive the error "internal error cannot load
>>> AppArmor profile"
>>>
>>> It appears that this issue only occurs when I am attempting to use
>>> volitile disks.
>>>
>>> However, when I attempt to attach a blank, persistent datablock to a
>>> virtual machine, I get the error:
>>>
>>> >> error: internal error process exited while connecting to monitor:
>>> kvm: -drive
>>> file=/var/lib/one//datastores/0/95/disk.0,if=none,id=drive-ide0-1-0,format=qcow2,cache=none:
>>> could not open disk image /var/lib/one//datastores/0/95/disk.0: Invalid
>>> argument
>>>
>>> I'm not sure if this is related to disabling dynamic ownership and
>>> setting the ownership and group to oneadmin / cloud or if it is an entierly
>>> different error altogether.
>>>
>>> I will say that I am still able to clone an imported disk image from the
>>> marketplace, and it appears to launch and run successfully.
>>>
>>> Thanks for all your help.
>>> Jon A
>>>
>>> Logs are attached below
>>>
>>> Failed VM with volatile disk image:
>>>
>>> Mon May 6 11:04:42 2013 [DiM][I]: New VM state is ACTIVE.
>>> Mon May 6 11:04:42 2013 [LCM][I]: New VM state is PROLOG.
>>> Mon May 6 11:06:05 2013 [TM][I]: clone: Cloning
>>> loki:/var/lib/one/datastores/103/4a657983e37df7e79b59110748f58ce8 in
>>> /var/lib/one/datastores/0/91/disk.0
>>> Mon May 6 11:06:05 2013 [TM][I]: ExitCode: 0
>>> Mon May 6 11:06:33 2013 [TM][I]: mkimage: Making filesystem of 5M and
>>> type ext4 at loki:/var/lib/one//datastores/0/91/disk.1
>>> Mon May 6 11:06:33 2013 [TM][I]: ExitCode: 0
>>> Mon May 6 11:06:35 2013 [TM][I]: mkimage: Making filesystem of 2048M and
>>> type swap at loki:/var/lib/one//datastores/0/91/disk.2
>>> Mon May 6 11:06:35 2013 [TM][I]: ExitCode: 0
>>> Mon May 6 11:06:36 2013 [TM][I]: context: Generating context block
>>> device at loki:/var/lib/one//datastores/0/91/disk.3
>>> Mon May 6 11:06:36 2013 [TM][I]: ExitCode: 0
>>> Mon May 6 11:06:37 2013 [LCM][I]: New VM state is BOOT
>>> Mon May 6 11:06:37 2013 [VMM][I]: Generating deployment file:
>>> /var/lib/one/vms/91/deployment.6
>>> Mon May 6 11:06:37 2013 [VMM][I]: ExitCode: 0
>>> Mon May 6 11:06:37 2013 [VMM][I]: Successfully execute network driver
>>> operation: pre.
>>> Mon May 6 11:06:39 2013 [VMM][I]: Command execution fail: cat << EOT |
>>> /var/tmp/one/vmm/kvm/deploy /var/lib/one//datastores/0/91/deployment.6 loki
>>> 91 loki
>>> Mon May 6 11:06:39 2013 [VMM][I]: error: Failed to create domain from
>>> /var/lib/one//datastores/0/91/deployment.6
>>> Mon May 6 11:06:39 2013 [VMM][I]: error: internal error cannot load
>>> AppArmor profile 'libvirt-76e12b0b-7d3c-c71a-46f0-bfe44917aedb'
>>> Mon May 6 11:06:39 2013 [VMM][E]: Could not create domain from
>>> /var/lib/one//datastores/0/91/deployment.6
>>> Mon May 6 11:06:39 2013 [VMM][I]: ExitCode: 255
>>> Mon May 6 11:06:39 2013 [VMM][I]: Failed to execute virtualization
>>> driver operation: deploy.
>>> Mon May 6 11:06:39 2013 [VMM][E]: Error deploying virtual machine: Could
>>> not create domain from /var/lib/one//datastores/0/91/deployment.6
>>> Mon May 6 11:06:40 2013 [DiM][I]: New VM state is FAILED
>>>
>>> Error log from VM with attached datablock:
>>>
>>> Mon May 6 11:40:18 2013 [DiM][I]: New VM state is ACTIVE.
>>> Mon May 6 11:40:19 2013 [LCM][I]: New VM state is PROLOG.
>>> Mon May 6 11:40:20 2013 [TM][I]: ln: Linking
>>> /var/lib/one/datastores/1/27ce5578d5cc9748cc44c8fedb75758b in
>>> loki:/var/lib/one//datastores/0/95/disk.0
>>> Mon May 6 11:40:20 2013 [TM][I]: ExitCode: 0
>>> Mon May 6 11:41:35 2013 [TM][I]: clone: Cloning
>>> loki:/var/lib/one/datastores/103/4a657983e37df7e79b59110748f58ce8 in
>>> /var/lib/one/datastores/0/95/disk.1
>>> Mon May 6 11:41:35 2013 [TM][I]: ExitCode: 0
>>> Mon May 6 11:41:37 2013 [TM][I]: context: Generating context block
>>> device at loki:/var/lib/one//datastores/0/95/disk.2
>>> Mon May 6 11:41:37 2013 [TM][I]: ExitCode: 0
>>> Mon May 6 11:41:37 2013 [LCM][I]: New VM state is BOOT
>>> Mon May 6 11:41:37 2013 [VMM][I]: Generating deployment file:
>>> /var/lib/one/vms/95/deployment.0
>>> Mon May 6 11:41:38 2013 [VMM][I]: ExitCode: 0
>>> Mon May 6 11:41:38 2013 [VMM][I]: Successfully execute network driver
>>> operation: pre.
>>> Mon May 6 11:41:40 2013 [VMM][I]: Command execution fail: cat << EOT |
>>> /var/tmp/one/vmm/kvm/deploy /var/lib/one//datastores/0/95/deployment.0 loki
>>> 95 loki
>>> Mon May 6 11:41:40 2013 [VMM][I]: error: Failed to create domain from
>>> /var/lib/one//datastores/0/95/deployment.0
>>> Mon May 6 11:41:40 2013 [VMM][I]: error: internal error process exited
>>> while connecting to monitor: kvm: -drive
>>> file=/var/lib/one//datastores/0/95/disk.0,if=none,id=drive-ide0-1-0,format=qcow2,cache=none:
>>> could not open disk image /var/lib/one//datastores/0/95/disk.0: Invalid
>>> argument
>>> Mon May 6 11:41:40 2013 [VMM][I]:
>>> Mon May 6 11:41:40 2013 [VMM][E]: Could not create domain from
>>> /var/lib/one//datastores/0/95/deployment.0
>>> Mon May 6 11:41:40 2013 [VMM][I]: ExitCode: 255
>>> Mon May 6 11:41:40 2013 [VMM][I]: Failed to execute virtualization
>>> driver operation: deploy.
>>> Mon May 6 11:41:40 2013 [VMM][E]: Error deploying virtual machine: Could
>>> not create domain from /var/lib/one//datastores/0/95/deployment.0
>>> Mon May 6 11:41:41 2013 [DiM][I]: New VM state is FAILED
>>>
>>>
>>> On Mon, May 6, 2013 at 5:23 AM, Ruben S. Montero <
>>> rsmontero at opennebula.org> wrote:
>>>
>>>> Have you tried to set the KVM configuration /etc/libvirt/qemu.conf as in
>>>>
>>>> $ grep -vE '^($|#)' /etc/libvirt/qemu.conf
>>>> user = "oneadmin"
>>>> group = "oneadmin"
>>>> dynamic_ownership = 0
>>>>
>>>>
>>>> more details here:
>>>>
>>>> http://opennebula.org/documentation:rel3.8:kvmg#configuration
>>>>
>>>>
>>>> On Sun, May 5, 2013 at 9:40 PM, Jon <three18ti at gmail.com> wrote:
>>>>
>>>>> Hello,
>>>>>
>>>>> I am trying to create a virtual machine by installing the OS from an
>>>>> iso to a qemu image.
>>>>>
>>>>> However, whenever the disk image is created, it is created with a user
>>>>> and group of oneadmin:cloud, however, once the VM attempts to start it is
>>>>> changed to libvirt-qemu:kvm then root:root, and I get an entry in the log:
>>>>>
>>>>> >> error: internal error process exited while connecting to monitor:
>>>>> kvm: -drive
>>>>> file=/var/lib/one//datastores/0/87/disk.1,if=none,id=drive-ide0-1-0,format=qcow2,cache=none:
>>>>> could not open disk image
>>>>>
>>>>> However, this disk image -does- exist, although with incorrect
>>>>> permissions:
>>>>>
>>>>> >> root at loki:~# ls -lah /var/lib/one//datastores/0/87
>>>>> >> -rw-r--r-- 1 root         root  5.1G Apr 27 19:24 disk.1
>>>>>
>>>>> What is changing the ownership of my disk images?
>>>>>
>>>>> Thanks,
>>>>> Jon A
>>>>>
>>>>>
>>>>> The full VM log is below:
>>>>>
>>>>> Sat Apr 27 19:23:31 2013 [DiM][I]: New VM state is ACTIVE.
>>>>> Sat Apr 27 19:23:31 2013 [LCM][I]: New VM state is PROLOG.
>>>>> Sat Apr 27 19:24:29 2013 [TM][I]: clone: Cloning
>>>>> loki:/var/lib/one/datastores/103/4a657983e37df7e79b59110748f58ce8 in
>>>>> /var/lib/one/datastores/0/87/disk.0
>>>>> Sat Apr 27 19:24:29 2013 [TM][I]: ExitCode: 0
>>>>> Sat Apr 27 19:24:51 2013 [TM][I]: mkimage: Making filesystem of 5120M
>>>>> and type ext4 at loki:/var/lib/one//datastores/0/87/disk.1
>>>>> Sat Apr 27 19:24:51 2013 [TM][I]: ExitCode: 0
>>>>> Sat Apr 27 19:24:52 2013 [TM][I]: context: Generating context block
>>>>> device at loki:/var/lib/one//datastores/0/87/disk.2
>>>>> Sat Apr 27 19:24:52 2013 [TM][I]: ExitCode: 0
>>>>> Sat Apr 27 19:24:53 2013 [LCM][I]: New VM state is BOOT
>>>>> Sat Apr 27 19:24:53 2013 [VMM][I]: Generating deployment file:
>>>>> /var/lib/one/vms/87/deployment.0
>>>>> Sat Apr 27 19:24:53 2013 [VMM][I]: ExitCode: 0
>>>>> Sat Apr 27 19:24:53 2013 [VMM][I]: Successfully execute network driver
>>>>> operation: pre.
>>>>> Sat Apr 27 19:24:56 2013 [VMM][I]: Command execution fail: cat << EOT
>>>>> | /var/tmp/one/vmm/kvm/deploy /var/lib/one//datastores/0/87/deployment.0
>>>>> loki 87 loki
>>>>> Sat Apr 27 19:24:56 2013 [VMM][I]: error: Failed to create domain from
>>>>> /var/lib/one//datastores/0/87/deployment.0
>>>>> Sat Apr 27 19:24:56 2013 [VMM][I]: error: internal error process
>>>>> exited while connecting to monitor: kvm: -drive
>>>>> file=/var/lib/one//datastores/0/87/disk.1,if=none,id=drive-ide0-1-0,format=qcow2,cache=none:
>>>>> could not open disk image /var/lib/one//datastores/0/87/disk.1: Invalid
>>>>> argument
>>>>> Sat Apr 27 19:24:56 2013 [VMM][I]:
>>>>> Sat Apr 27 19:24:56 2013 [VMM][E]: Could not create domain from
>>>>> /var/lib/one//datastores/0/87/deployment.0
>>>>> Sat Apr 27 19:24:56 2013 [VMM][I]: ExitCode: 255
>>>>> Sat Apr 27 19:24:56 2013 [VMM][I]: Failed to execute virtualization
>>>>> driver operation: deploy.
>>>>> Sat Apr 27 19:24:56 2013 [VMM][E]: Error deploying virtual machine:
>>>>> Could not create domain from /var/lib/one//datastores/0/87/deployment.0
>>>>> Sat Apr 27 19:24:57 2013 [DiM][I]: New VM state is FAILED
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list
>>>>> Users at lists.opennebula.org
>>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Ruben S. Montero, PhD
>>>> Project co-Lead and Chief Architect
>>>> OpenNebula - The Open Source Solution for Data Center Virtualization
>>>> www.OpenNebula.org | rsmontero at opennebula.org | @OpenNebula
>>>>
>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.opennebula.org/pipermail/users-opennebula.org/attachments/20130513/8edd9ab4/attachment-0002.htm>


More information about the Users mailing list