[one-users] ubuntu-cloud boot (?) problems

Xasima xasima at gmail.com
Wed Nov 14 06:41:12 PST 2012


Hello, Thank you. I have successfully updated  opennebula up to 3.8.1, and
this fix cdrom mapping issue.
However, I had a long time attempts to manage to set up ubuntu 12.04 with
serial access (virsh console enabled). I have included my steps in the
bottom of the mail if that helps anyone else.

Besides I was able to set up my quest, I have a little question on the
configuration.

*1) What is the proper correspondence between DISK DEV_PREFIX , VM OS ROOT
and actual image partition mappings. *
My already prepared qcow2 image has /dev/sda1 partition mapping inside.
But I leave unchanged the opennebula image templates with
DEV_PREFIX="hd"
DRIVER = qcow2

While opennebula vm template has
OS     = [ ARCH = x86_64,
           BOOT = hd,
           ROOT = sda1,
...]

While "onevm show" display
  TARGET="hda"
against this disk

Don't I need to place DEV_PREFIX to "sd" instead ?

*2)  Where is my contextualization cdrom ? *
My contextualization cd-rom is automatically mapped to  TARGET="hdb" (as
displayed by "onevm show")
But guest do know only about sda devices
weblab at vm3:~$ ls /dev/ | grep hd
weblab at vm3:~$ ls /dev/ | grep sd
sda
sda1
sda2

weblab at vm3:~$ sudo dmesg | grep -i cd
[    0.328988] ehci_hcd: USB 2.0 'Enhanced' Host Controller (EHCI) Driver
[    0.329619] ohci_hcd: USB 1.1 'Open' Host Controller (OHCI) Driver
[    0.330202] uhci_hcd: USB Universal Host Controller Interface driver
[    0.330828] uhci_hcd 0000:00:01.2: PCI INT D -> Link[LNKD] -> GSI 11
(level, high) -> IRQ 11
[    0.331665] uhci_hcd 0000:00:01.2: setting latency timer to 64
[    0.331674] uhci_hcd 0000:00:01.2: UHCI Host Controller
[    0.336156] uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus
number 1
[    0.336936] uhci_hcd 0000:00:01.2: irq 11, io base 0x0000c100
[    0.495645] scsi 0:0:1:0: CD-ROM            QEMU     QEMU DVD-ROM
1.0  PQ: 0 ANSI: 5
[    0.497029] sr0: scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray
[    0.497686] cdrom: Uniform CD-ROM driver Revision: 3.20
[    0.501810] sr 0:0:1:0: Attached scsi CD-ROM sr0

weblab at vm3:~$ cat /boot/config-3.2.0-32-virtual | grep -i iso9660
CONFIG_ISO9660_FS=m

weblab at vm3:~$ sudo modprobe iso9660
FATAL: Could not load /lib/modules/3.2.0-32-generic/modules.dep

Seems that either contextualization init.sh is not ok, or do I need to use
generic / not a virtual kernel?

The part of my init.sh contextualization script (just copied from ttylinux
template)
if [ -f /mnt/context/context.sh ]
then
  . /mnt/context/context.sh
else
  mount -t iso9660 /dev/sr0 /mnt
  ./mnt/context/context.sh
fi


--------------
Steps to enable serial console on ubuntu 12.04 with opennebula
--------------
*Preparation*
1) Install (apt-get install) questfish + libguestfs-tools with dependencies
to be able to manage easily quest fs, since some changes to guest grub and
init.d are required to enable serial console
2) Create an qcow2 image using vm-builder with predefined ip/gateway (on
the case if contextualization script fails)
3) Start image with libvirt, change xml to include serial console and test
if it works. Ensure if it behaves well both with ssh and virsh console
access.

*Test the type of the image*
weblab at metrics:~/ubuntu-kvm$ qemu-img info base-m1s.img
image: base-m1s.img
file format: raw
virtual size: 5.0G (5368709120 bytes)
disk size: 565M
weblab at metrics:~/ubuntu-kvm$

*Check if the grub is configured to use serial console*
weblab at metrics:~/ubuntu-kvm$ sudo virt-edit base-m1s.img /boot/grub/menu.lst
...
title           Ubuntu 12.04.1 LTS, kernel 3.2.0-32-virtual
uuid            c645f23d-9d48-43d3-b042-7b06ae9f56b3
kernel          /boot/vmlinuz-3.2.0-32-virtual
root=UUID=c645f23d-9d48-43d3-b042-7b06ae9f56b3 ro quiet splash console=tty1
console=ttyS0,115200n8
initrd          /boot/initrd.img-3.2.0-32-virtual
...

*Check if initrs and vmlinuz are presented so will explicitly point to them
in opennebula vm template*
weblab at metrics:~/ubuntu-kvm$ sudo virt-ls -la base-m1s.img / | grep boot
drwxr-xr-x  3 0 0  4096 Nov 12 14:58 boot
lrwxrwxrwx  1 0 0    33 Nov 12 14:58 initrd.img ->
/boot/initrd.img-3.2.0-32-virtual
lrwxrwxrwx  1 0 0    29 Nov 12 14:58 vmlinuz ->
boot/vmlinuz-3.2.0-32-virtual

*Double check that ttyS0 service will be up *(just followed some
instructions)
weblab at metrics:~/ubuntu-kvm$ sudo virt-cat -a base-m1s.img
/etc/init/ttyS0.conf
# ttyS0 - getty
#
# This service maintains a getty on ttyS0 from the point the system is
# started until it is shut down again.
start on stopped rc or RUNLEVEL=[2345]
stop on runlevel [!2345]

respawn
exec /sbin/getty -L 115200 ttyS0 vt102

or use guestfish to create such file.

*Important! Figure out that partition is sda, not hda, so... will change
this correspondingly in opennebula image and vm templates*
weblab at metrics:~/ubuntu-kvm$ sudo virt-filesystems -a base-m1s.img --all
--long --uuid -h
Name      Type       VFS  Label MBR Size Parent   UUID
/dev/sda1 filesystem ext4 -     -   3.8G -
 c645f23d-9d48-43d3-b042-7b06ae9f56b3
/dev/sda2 filesystem swap -     -   976M -
 6ad7b5f6-9503-413b-a660-99dfb7686459
/dev/sda1 partition  -    -     83  3.8G /dev/sda -
/dev/sda2 partition  -    -     82  976M /dev/sda -
/dev/sda  device     -    -     -   5.0G -        -

*Check if network settings are already in place, so even if the opennebula
contextualization script fails, it will be brought up under the predefined
ip. *
weblab at metrics:~/ubuntu-kvm$ sudo virt-cat -a base-m1s.img
/etc/network/interfaces
# This file describes the network interfaces available on your system
# and how to activate them. For more information, see interfaces(5).

# The loopback network interface
auto lo
iface lo inet loopback

# The primary network interface
auto eth0
iface eth0 inet static
        address 10.0.0.95
        netmask 255.128.0.0
        network 10.0.0.0
        broadcast 10.127.255.255
        gateway 10.0.0.1
        # dns-* options are implemented by the resolvconf package, if
installed
        dns-nameservers 10.0.0.1
        dns-search defaultdomain

*Opennebula Image template *
weblab at metrics:~/ubuntu-kvm$  cat base-m1s.image.template
NAME = "base-m1.small - qcow"
PATH = /home/weblab/ubuntu-kvm/base-m1s.qcow2
TYPE = OS
DRIVER = qcow2

sudo -u oneadmin oneimage create base-m1s.image.template -d default
sudo -u oneadmin oneimage show 18
..
DEV_PREFIX="hd"
DRIVER = qcow2

*Opennebula VM template*
weblab at metrics:~/ubuntu-kvm$ cat base-m1s.vm.template
NAME   = vm3-on-qcow
CPU    = 0.6
MEMORY = 512

OS     = [ ARCH = x86_64,
           BOOT = hd,
           ROOT = sda1,
           KERNEL = /vmlinuz,
           INITRD = /initrd.img,
           KERNEL_CMD = "ro console=tty1 console=ttyS0,115200n8" ]

DISK   = [ IMAGE_ID = 18,
           DRIVER = qcow2,
           READONLY = no ]

NIC    = [ NETWORK_ID = 14 ]

FEATURES = [ acpi = yes ]

REQUIREMENTS = "FALSE"


CONTEXT = [
           HOSTNAME    = "$NAME",
           IP_PUBLIC   = "$NIC[IP]",
           DNS =  "$NETWORK[DNS,  NETWORK_ID=9]",
           GATEWAY = "$NETWORK[GATEWAY, NETWORK_ID=9]",
           NETMASK = "$NETWORK[NETWORK_MASK,  NETWORK_ID=9]",
           FILES     = "/tmp/ttylinux/init.sh /tmp/ttylinux/id_rsa.pub",
           ROOT_PUBKEY = "id_rsa.pub" ]

RAW      = [ type = "kvm",
             data = "<devices><serial type=\"pty\"><target
port=\"0\"/></serial><console type=\"pty\"><target port=\"0\"
type=\"serial\"/></console></devices>" ]


*Show*
-----------
sudo -u oneadmin onevm show 77
VIRTUAL MACHINE TEMPLATE
CONTEXT=[
  DISK_ID="1",
  FILES="/tmp/ttylinux/init.sh /tmp/ttylinux/id_rsa.pub",
  HOSTNAME="vm3-on-qcow",
  IP_PUBLIC="10.0.0.95",
  ROOT_PUBKEY="id_rsa.pub",
  TARGET="hdb" ]
CPU="0.6"
DISK=[
  CLONE="YES",
  DATASTORE="default",
  DATASTORE_ID="1",
  DEV_PREFIX="hd",
  DISK_ID="0",
  DRIVER="qcow2",
  IMAGE="base-m1.small - qcow",
  IMAGE_ID="18",
  READONLY="NO",
  SAVE="NO",
  SOURCE="/var/lib/one/datastores/1/3fdc724b56b20346ed18687e677d6ae8",
  TARGET="hda",
  TM_MAD="ssh",
  TYPE="FILE" ]
FEATURES=[
   ACPI="yes" ]
MEMORY="512"
NAME="vm3-on-qcow"
NIC=[
  BRIDGE="br0",
  IP="10.0.0.95",
  MAC="02:00:0a:00:00:5f",
  NETWORK="m1 network",
  NETWORK_ID="14",
  VLAN="NO" ]
OS=[
  ARCH="x86_64",
  BOOT="hd",
  INITRD="/initrd.img",
  KERNEL="/vmlinuz",
  KERNEL_CMD="ro console=tty1 console=ttyS0,115200n8",
  ROOT="sda1" ]
RAW=[
  DATA="<devices><serial type=\"pty\"><target port=\"0\"/></serial><console
type=\"pty\"><target port=\"0\" type=\"serial\"/></console></devices>",
  TYPE="kvm" ]
REQUIREMENTS="FALSE"
VMID="77"

*Checking with virsh on nodehost*
frontend >> ssh nodehost
nodehost >> sudo virsh --connect qemu:///system

virsh # list --all
 Id Name                 State
----------------------------------
 11 one-77               running
  - vm3                  shut off

virsh # ttyconsole 11
/dev/pts/0

virsh # console 11
Connected to domain one-77
Escape character is ^]
(--- Press Enter)
Ubuntu 12.04.1 LTS vm3 ttyS0

vm3 login:
...
(-- Press "Ctr + ]"   to logout from vm to  virsh )

virsh # dumpxml 11
<domain type='kvm' id='11'>
  <name>one-77</name>
  <uuid>9f5fe3e7-5abd-1a45-6df8-84c91fb0af9e</uuid>
  <memory>524288</memory>
  <currentMemory>524288</currentMemory>
  <vcpu>1</vcpu>
  <cputune>
    <shares>615</shares>
  </cputune>
  <os>
    <type arch='x86_64' machine='pc-1.0'>hvm</type>
    <kernel>/vmlinuz</kernel>
    <initrd>/initrd.img</initrd>
    <cmdline>root=/dev/sda1 ro console=tty1 console=ttyS0,115200n8</cmdline>
    <boot dev='hd'/>
  </os>
  <features>
    <acpi/>
  </features>
  <clock offset='utc'/>
  <on_poweroff>destroy</on_poweroff>
  <on_reboot>restart</on_reboot>
  <on_crash>destroy</on_crash>
  <devices>
    <emulator>/usr/bin/kvm</emulator>
    <disk type='file' device='disk'>
      <driver name='qemu' type='qcow2'/>
      <source file='/var/lib/one/datastores/0/77/disk.0'/>
      <target dev='hda' bus='ide'/>
      <alias name='ide0-0-0'/>
      <address type='drive' controller='0' bus='0' unit='0'/>
    </disk>
    <disk type='file' device='cdrom'>
      <driver name='qemu' type='raw'/>
      <source file='/var/lib/one/datastores/0/77/disk.1'/>
      <target dev='hdb' bus='ide'/>
      <readonly/>
      <alias name='ide0-0-1'/>
      <address type='drive' controller='0' bus='0' unit='1'/>
    </disk>
    <controller type='ide' index='0'>
      <alias name='ide0'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x01'
function='0x1'/>
    </controller>
    <interface type='bridge'>
      <mac address='02:00:0a:00:00:5f'/>
      <source bridge='br0'/>
      <target dev='vnet0'/>
      <alias name='net0'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
function='0x0'/>
    </interface>
    <serial type='pty'>
      <source path='/dev/pts/0'/>
      <target port='0'/>
      <alias name='serial0'/>
    </serial>
    <console type='pty' tty='/dev/pts/0'>
      <source path='/dev/pts/0'/>
      <target type='serial' port='0'/>
      <alias name='serial0'/>
    </console>
    <memballoon model='virtio'>
      <alias name='balloon0'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x04'
function='0x0'/>
    </memballoon>
  </devices>
</domain>



On Fri, Nov 2, 2012 at 1:50 PM, Jaime Melis <j.melis at gmail.com> wrote:

> Hello,
>
> I believe you are affected by the bug that incorrectly maps the context
> cdroms. I recommend you update to 3.8.1 where this bug is fixed.
>
> More info on the problem: http://dev.opennebula.org/issues/1594
>
> cheers,
> Jaime
>
>
> On Fri, Nov 2, 2012 at 11:05 AM, Xasima <xasima at gmail.com> wrote:
>
>> Hello. I have some problems with boot of cloud-based Ubuntu. There are
>> two ubuntu 12.04 server (front-end and node) with openebula upgraded up to
>> 3.8. I have successfully deployed opennebula-ttylinux with qemu / kvm
>> for the first time to try. I want now to deploy already prepared
>> EC2-compatible image of recent ubuntu.
>>
>> Actually the image and VM are deployed with no error (logs are ok), but
>>  VM doesn't consume CPU at all. I think it doesn't boot properly.
>>
>> *> sudo -u oneadmin onevm list*
>>     ID USER     GROUP    NAME                       STAT  UCPU  UMEM
>>  HOST         TIME
>>     61 oneadmin oneadmin ttylinux                   runn    6         64M
>>     metrics-ba   0d 01h29
>>     62 oneadmin oneadmin ubuntu-cloud64-    runn    *0         *512M*   *metrics-ba
>>   0d 00h10
>>
>> The only thing that seems strange for me in logs is the drive mapping
>> (available from libvirt-qemu log on the node).
>>
>> *> ssh node && cat /var/log/libvirt/qemu/one-62.log*
>> 2012-11-02 09:16:56.096+0000: starting up
>> LC_ALL=C
>> PATH=/usr/local/sbin:/usr/local/bin:/usr/bin:/usr/sbin:/sbin:/bin
>> /usr/bin/kvm -S -M pc-1.0 -enable-kvm -m 512 -smp
>> 1,sockets=1,cores=1,threads=1 -name one-62 -uuid
>> 2c15ca04-7d5f-ab4c-8bdb-43d2add1a2fe -nographic -nodefconfig -nodefaults
>> -chardev
>> socket,id=charmonitor,path=/var/lib/libvirt/qemu/one-62.monitor,server,nowait
>> -mon chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown
>> -kernel /vmlinuz -initrd /initrd.img *-drive
>> file=/var/lib/one/datastores/0/62/disk.0,if=none,id=drive-ide0-0-0,format=qcow2
>> * -device ide-drive,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0
>> -drive *
>> file=/var/lib/one/datastores/0/62/disk.0,if=none,media=cdrom,id=drive-ide0-1-0,readonly=on,format=raw
>> * -device ide-drive,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0
>> -netdev tap,fd=19,id=hostnet0 -device
>> rtl8139,netdev=hostnet0,id=net0,mac=02:00:0a:00:00:5d,bus=pci.0,addr=0x3
>> -usb -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x4
>>
>> Could anyone help to determine what is the cause of the failure and how
>> to resolve this?
>>
>> -----------------
>> Here is the full information on my steps
>>
>> *1. Image specific information*
>> Download
>> http://cloud-images.ubuntu.com/releases/precise/release/ubuntu-12.04-server-cloudimg-amd64-disk1.img to
>> front-end. Manifest and ovf are available
>> http://cloud-images.ubuntu.com/releases/precise/release/ as well to
>> check what is installed on the image,
>>
>> *2. Image file format information*
>> > *qemu-img info precise-server-cloudimg-amd64-disk1.img*
>> image: precise-server-cloudimg-amd64-disk1.img
>> file format: qcow2
>> virtual size: 2.0G (2147483648 bytes)
>> disk size: 222M
>> cluster_size: 65536
>>
>> *3. Content of the image*
>> Using *qemu-img convert (to raw) && **kpartx -a -v precise...img &&  mount
>> /dev/mapper/loop1p1 /mnt/*
>> I have ensured the content of the image
>> *> ls /mnt/*
>> bin   dev  home        lib    lost+found  mnt  proc  run   selinux  sys
>>  usr  *vmlinuz*
>> boot  etc  *initrd.img*  lib64  media       opt  root  sbin  srv
>>  tmp  var
>>
>> *> cat /mnt/etc/fstab*
>>  LABEL=cloudimg-rootfs   /        ext4   defaults        0 0
>>
>> *> umount && kpartx -d*
>>
>> 4. *Opennebula Image template*
>> * > cat 64base-image.one*
>> NAME = ubuntu-cloud64-qcow2
>> PATH = "/tmp/ttylinux/precise-server-cloudimg-amd64-disk1.img"
>> TYPE = OS
>> FSTYPE= "qcow2"
>>
>> The state of drive on opennebula
>> *> sudo -u oneadmin oneimage show 12*
>>  IMAGE 12 INFORMATION
>> ID             : 12
>> NAME           : ubuntu-cloud64-qcow2
>> USER           : oneadmin
>> GROUP          : oneadmin
>> DATASTORE      : default
>> TYPE           : OS
>> REGISTER TIME  : 11/02 12:04:47
>> PERSISTENT     : No
>> SOURCE         :
>> /var/lib/one/datastores/1/a4d9b6af3313f826d9113b4e3b0ac25b
>> PATH           : /tmp/ttylinux/precise-server-cloudimg-amd64-disk1.img
>> SIZE           : 223M
>> STATE          : used
>> RUNNING_VMS    : 1
>>
>> PERMISSIONS
>> OWNER          : um-
>> GROUP          : ---
>> OTHER          : ---
>>
>> IMAGE TEMPLATE
>> DEV_PREFIX="hd"
>> FSTYPE="qcow2"
>>
>> 5. *Opennebula VM template*
>> *> cat 64base.one*
>> NAME   = ubuntu-cloud64-on-qcow2
>> CPU    = 0.6
>> MEMORY = 512
>>
>> OS     = [ ARCH = x86_64,
>>            BOOT = hd,
>>            KERNEL = /vmlinuz,
>>            INITRD = /initrd.img ]
>>
>> DISK   = [ IMAGE_ID = 12,
>>            DRIVER = qcow2,
>>            TYPE = disk,
>>            READONLY = no ]
>>
>> NIC    = [ NETWORK_ID = 9 ]
>>
>> FEATURES = [ acpi = yes ]
>>
>> REQUIREMENTS = "FALSE"
>>
>> CONTEXT = [
>>      HOSTNAME    = "$NAME",
>>      IP_PUBLIC   = "$NIC[IP]",
>>      DNS =  "$NETWORK[DNS,  NETWORK_ID=9]",
>>      GATEWAY = "$NETWORK[GATEWAY, NETWORK_ID=9]",
>>      NETMASK = "$NETWORK[NETWORK_MASK,  NETWORK_ID=9]",
>>      FILES     = "/tmp/ttylinux/init.sh /tmp/ttylinux/id_rsa.pub",
>>      TARGET      = "hdc",
>>      ROOT_PUBKEY = "id_rsa.pub"
>> ]
>>
>> 6. *Log of VM deployment (on front-end) *
>> *> sudo -u oneadmin onevm deploy 62 5*
>> *> tail -f /var/log/one/62.log*
>> Fri Nov  2 12:11:01 2012 [DiM][I]: New VM state is ACTIVE.
>> Fri Nov  2 12:11:02 2012 [LCM][I]: New VM state is PROLOG.
>> Fri Nov  2 12:17:05 2012 [TM][I]: clone: Cloning
>> metrics:/var/lib/one/datastores/1/a4d9b6af3313f826d9113b4e3b0ac25b in
>> /var/lib/one/datastores/0/62/disk.0
>> Fri Nov  2 12:17:05 2012 [TM][I]: ExitCode: 0
>> Fri Nov  2 12:17:09 2012 [TM][I]: context: Generating context block
>> device at metrics-backend:/var/lib/one/datastores/0/62/disk.1
>> Fri Nov  2 12:17:09 2012 [TM][I]: ExitCode: 0
>> Fri Nov  2 12:17:09 2012 [LCM][I]: New VM state is BOOT
>> Fri Nov  2 12:17:09 2012 [VMM][I]: Generating deployment file:
>> /var/lib/one/62/deployment.0
>> Fri Nov  2 12:17:11 2012 [VMM][I]: ExitCode: 0
>> Fri Nov  2 12:17:11 2012 [VMM][I]: Successfully execute network driver
>> operation: pre.
>> Fri Nov  2 12:17:13 2012 [VMM][I]: ExitCode: 0
>> Fri Nov  2 12:17:13 2012 [VMM][I]: Successfully execute virtualization
>> driver operation: deploy.
>> Fri Nov  2 12:17:13 2012 [VMM][I]: ExitCode: 0
>> Fri Nov  2 12:17:13 2012 [VMM][I]: Successfully execute network driver
>> operation: post.
>> Fri Nov  2 12:17:13 2012 [LCM][I]: New VM state is RUNNING
>>
>> *> ssh node && cat /var/log/libvirt/qemu/one-62.log*
>> 2012-11-02 09:16:56.096+0000: starting up
>> LC_ALL=C
>> PATH=/usr/local/sbin:/usr/local/bin:/usr/bin:/usr/sbin:/sbin:/bin
>> /usr/bin/kvm -S -M pc-1.0 -enable-kvm -m 512 -smp
>> 1,sockets=1,cores=1,threads=1 -name one-62 -uuid
>> 2c15ca04-7d5f-ab4c-8bdb-43d2add1a2fe -nographic -nodefconfig -nodefaults
>> -chardev
>> socket,id=charmonitor,path=/var/lib/libvirt/qemu/one-62.monitor,server,nowait
>> -mon chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown
>> -kernel /vmlinuz -initrd /initrd.img -drive
>> file=/var/lib/one/datastores/0/62/disk.0,if=none,id=drive-ide0-0-0,format=qcow2
>> -device ide-drive,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 -drive
>> file=/var/lib/one/datastores/0/62/disk.0,if=none,media=cdrom,id=drive-ide0-1-0,readonly=on,format=raw
>> -device ide-drive,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -netdev
>> tap,fd=19,id=hostnet0 -device
>> rtl8139,netdev=hostnet0,id=net0,mac=02:00:0a:00:00:5d,bus=pci.0,addr=0x3
>> -usb -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x4
>> ~
>>
>> 7. *Status of the machine on opennebula*
>> *> sudo -u oneadmin onevm list*
>>     ID USER     GROUP    NAME            STAT UCPU    UMEM HOST
>>   TIME
>>     61 oneadmin oneadmin ttylinux        runn    6     64M metrics-ba
>> 0d 01h29
>>     62 oneadmin oneadmin ubuntu-cloud64- runn    *0      *512M* *metrics-ba
>>   0d 00h10
>>
>> *> sudo -u oneadmin onevm show 62*
>> VIRTUAL MACHINE 62 INFORMATION
>> ID                  : 62
>> NAME                : ubuntu-cloud64-on-qcow2
>> USER                : oneadmin
>> GROUP               : oneadmin
>> STATE               : ACTIVE
>> LCM_STATE           : RUNNING
>> RESCHED             : No
>> HOST                : metrics-backend
>> START TIME          : 11/02 12:08:37
>> END TIME            : -
>> DEPLOY ID           : one-62
>>
>> VIRTUAL MACHINE MONITORING
>> USED CPU            : 0
>> NET_RX              : 1M
>> USED MEMORY         : 512M
>> NET_TX              : 0K
>>
>> PERMISSIONS
>> OWNER               : um-
>> GROUP               : ---
>> OTHER               : ---
>>
>> VIRTUAL MACHINE TEMPLATE
>> CONTEXT=[
>>   DISK_ID="1",
>>   DNS="10.0.0.20",
>>   FILES="/tmp/ttylinux/init.sh /tmp/ttylinux/id_rsa.pub",
>>   GATEWAY="10.0.0.1",
>>   HOSTNAME="ubuntu-cloud64-on-qcow2",
>>   IP_PUBLIC="10.*.*.*" ,
>>   NETMASK="255.128.0.0",
>>   ROOT_PUBKEY="id_rsa.pub",
>>   TARGET="hdc" ]
>> CPU="0.6"
>> DISK=[
>>   CLONE="YES",
>>   DATASTORE="default",
>>   DATASTORE_ID="1",
>>   DEV_PREFIX="hd",
>>   DISK_ID="0",
>>   DRIVER="qcow2",
>>   IMAGE="ubuntu-cloud64-qcow2",
>>   IMAGE_ID="12",
>>   READONLY="NO",
>>   SAVE="NO",
>>   SOURCE="/var/lib/one/datastores/1/a4d9b6af3313f826d9113b4e3b0ac25b",
>>   TARGET="hda",
>>   TM_MAD="ssh",
>>   TYPE="FILE" ]
>> FEATURES=[
>>   ACPI="yes" ]
>> MEMORY="512"
>> NAME="ubuntu-cloud64-on-qcow2"
>> NIC=[
>>   BRIDGE="br0",
>>   IP="10.*.*.*",
>>   MAC="02:00:0a:00:00:5d",
>>   NETWORK="Server 10.0.0.x network with br0",
>>   NETWORK_ID="9",
>>   VLAN="NO" ]
>> OS=[
>>   ARCH="x86_64",
>>   BOOT="hd",
>>   INITRD="/initrd.img",
>>   KERNEL="/vmlinuz" ]
>> REQUIREMENTS="FALSE"
>> VMID="62"
>>
>> VIRTUAL MACHINE HISTORY
>>  SEQ HOST            REASON           START            TIME
>> PROLOG_TIME
>>    0 metrics-backend none    11/02 12:11:01    0d 00h28m06s    0d
>> 00h06m08s
>>
>>
>> Thank you.
>> --
>> Best regards,
>>      ~ Xasima ~
>>
>> _______________________________________________
>> Users mailing list
>> Users at lists.opennebula.org
>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>
>>
>


-- 
Best regards,
     ~ Xasima ~
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.opennebula.org/pipermail/users-opennebula.org/attachments/20121114/d3afab06/attachment-0002.htm>


More information about the Users mailing list