[one-users] VM stuck in boot state
Javier Fontan
jfontan at gmail.com
Thu May 12 04:32:45 PDT 2011
Can you check /var/log/libvirt/qemu/one-54.log? There should be some info there.
On Tue, May 10, 2011 at 5:26 PM, Karthik Mallavarapu
<karthik.mallavarapu at gmail.com> wrote:
> Hello All,
> I have an opennebula installation with 1 frontend and two nodes. The base OS
> is ubuntu 10.10 64 bit edition. When I try to deploy a VM using the command
> onevm create ubuntu.template the VM gets deployed successfully on one node
> but it gets stuck in BOOT state on the other node. I am enclosing
> the log file and deployment.0 file of that VM and transcript of the log from
> $ONE_LOCATION/var/oned.log.
>
> vm.log
> Tue May 10 18:01:54 2011 [DiM][I]: New VM state is ACTIVE.
> Tue May 10 18:01:54 2011 [LCM][I]: New VM state is PROLOG.
> Tue May 10 18:01:54 2011 [VM][I]: Virtual Machine has no context
> Tue May 10 18:01:55 2011 [TM][I]: tm_clone.sh:
> cloud-3:/srv/cloud/images/ubuntu/ubu64.img
> 192.168.2.5:/srv/cloud/one/var//54/images/disk.0
> Tue May 10 18:01:55 2011 [TM][I]: tm_clone.sh: DST:
> /srv/cloud/one/var//54/images/disk.0
> Tue May 10 18:01:55 2011 [TM][I]: tm_clone.sh: Creating directory
> /srv/cloud/one/var//54/images
> Tue May 10 18:01:55 2011 [TM][I]: tm_clone.sh: Executed "mkdir -p
> /srv/cloud/one/var//54/images".
> Tue May 10 18:01:55 2011 [TM][I]: tm_clone.sh: Executed "chmod a+w
> /srv/cloud/one/var//54/images".
> Tue May 10 18:01:55 2011 [TM][I]: tm_clone.sh: Cloning
> /srv/cloud/images/ubuntu/ubu64.img
> Tue May 10 18:01:55 2011 [TM][I]: tm_clone.sh: Executed "cp -r
> /srv/cloud/images/ubuntu/ubu64.img /srv/cloud/one/var//54/images/disk.0".
> Tue May 10 18:01:55 2011 [TM][I]: tm_clone.sh: Executed "chmod a+rw
> /srv/cloud/one/var//54/images/disk.0".
> Tue May 10 18:01:57 2011 [LCM][I]: New VM state is BOOT
> Tue May 10 18:01:57 2011 [VMM][I]: Generating deployment file:
> /srv/cloud/one/var/54/deployment.0
> deployment.0
> <domain type='kvm'>
> <name>one-54</name>
> <memory>614400</memory>
> <os>
> <type arch='x86_64'>hvm</type>
> <boot dev='hd'/>
> </os>
> <devices>
> <emulator>/usr/bin/kvm</emulator>
> <disk type='file' device='disk'>
> <source
> file='/srv/cloud/one/var//54/images/disk.0'/>
> <target dev='sda'/>
> <driver name='qemu' type='qcow2'/>
> </disk>
> <interface type='bridge'>
> <source bridge='br0'/>
> <mac address='02:00:c0:a8:02:03'/>
> </interface>
> </devices>
> <features>
> <acpi/>
> </features>
> </domain>
> /srv/cloud/one/var/oned.log
> Tue May 10 18:01:54 2011 [DiM][D]: Deploying VM 54
> Tue May 10 18:01:55 2011 [TM][D]: Message received: LOG - 54 tm_clone.sh:
> cloud-3:/srv/cloud/images/ubuntu/ubu64.img
> 192.168.2.5:/srv/cloud/one/var//54/images/disk.0
> Tue May 10 18:01:55 2011 [TM][D]: Message received: LOG - 54 tm_clone.sh:
> DST: /srv/cloud/one/var//54/images/disk.0
> Tue May 10 18:01:55 2011 [TM][D]: Message received: LOG - 54 tm_clone.sh:
> Creating directory /srv/cloud/one/var//54/images
> Tue May 10 18:01:55 2011 [TM][D]: Message received: LOG - 54 tm_clone.sh:
> Executed "mkdir -p /srv/cloud/one/var//54/images".
> Tue May 10 18:01:55 2011 [TM][D]: Message received: LOG - 54 tm_clone.sh:
> Executed "chmod a+w /srv/cloud/one/var//54/images".
> Tue May 10 18:01:55 2011 [TM][D]: Message received: LOG - 54 tm_clone.sh:
> Cloning /srv/cloud/images/ubuntu/ubu64.img
> Tue May 10 18:01:55 2011 [TM][D]: Message received: LOG - 54 tm_clone.sh:
> Executed "cp -r /srv/cloud/images/ubuntu/ubu64.img
> /srv/cloud/one/var//54/images/disk.0".
> Tue May 10 18:01:55 2011 [TM][D]: Message received: LOG - 54 tm_clone.sh:
> Executed "chmod a+rw /srv/cloud/one/var//54/images/disk.0".
> Tue May 10 18:01:55 2011 [TM][D]: Message received: TRANSFER SUCCESS 54 -
> onevm show command for that particular VM gives the following output.
> VIRTUAL MACHINE 54 INFORMATION
>
> ID : 54
> NAME : ubuntu
> STATE : ACTIVE
> LCM_STATE : BOOT
> START TIME : 05/10 18:01:39
> END TIME : -
> DEPLOY ID: : -
> The strange thing with the issue at hand is that, we had successfully
> deployed VMs before on this particular node. But recently we had upgraded
> the libvirt version to 0.9 on both
> the nodes. This particular node was operational even after the libvirt
> version upgrade. Deployment on this node stopped working from the time we
> upgraded the libvirt version
> of the 2nd node. By the way, the host monitoring seems to be working as it
> is evident from the oned.log file.
> Could some one please throw some light on this issue. I have tried to dig up
> the old threads but the suggested fixes did not really work in my case.
> Thanks a lot for your time.
> Regards,
> Karthik
> _______________________________________________
> Users mailing list
> Users at lists.opennebula.org
> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>
>
--
Javier Fontan, Grid & Virtualization Technology Engineer/Researcher
DSA Research Group: http://dsa-research.org
Globus GridWay Metascheduler: http://www.GridWay.org
OpenNebula Virtual Infrastructure Engine: http://www.OpenNebula.org
More information about the Users
mailing list