[one-users] retrieve info from OpenVZ host running libvirtd

knawnd at gmail.com knawnd at gmail.com
Mon Feb 21 04:52:00 PST 2011


Hi Hector,

Héctor Sanjuán wrote on 21/02/11 15:27:
> Hi Nikolay,
>
> I had a similar problem and it came down to im driver files not having
> the proper execution rights. Can you maybe check that?
Thanks for reply!
FN$ ll /usr/lib/one/remotes/im/ovz.d/ovz.rb
-rwxr-xr-x 1 root root 3305 Feb 21 09:23 
/usr/lib/one/remotes/im/ovz.d/ovz.rb

CN$ ll /tmp/one/im/run_probes
-rwxr-xr-x 1 oneadmin cloud 1773 Feb 18 17:34 /tmp/one/im/run_probes

execution right looks for me OK.

Nikolay.
>
> Héctor
>
> El 21/02/11 13:17, knawnd at gmail.com escribió:
>> Hi Carlos,
>>
>> Carlos Martín Sánchez wrote on 21/02/11 14:04:
>>> Hi Nikolay,
>>>
>>> Could you paste the output of
>>>       ssh<OVZ_cluster_node>  "/var/tmp/one/im/run_probes ovz"
>>> ?
>> [oneadmin at FN]$ ssh edugrid10 "/tmp/one/im/run_probes ovz"
>> Warning: Permanently added '<CN_hostname,CN_IP>' (RSA) to the list of
>> known hosts.
>>   [oneadmin at FN]$
>>
>> i.e. it looks like the command above returns space character.
>>
>> Nikolay.
>>>
>>> Change /var/tmp/one to your SCRIPTS_REMOTE_DIR if you changed it in your
>>> oned.conf file.
>>>
>>> Regards,
>>> Carlos.
>>>
>>> --
>>> Carlos Martín, MSc
>>> Project Major Contributor
>>> OpenNebula - The Open Source Toolkit for Cloud Computing
>>> www.OpenNebula.org<http://www.opennebula.org/>  | cmartin at opennebula.org
>>> <mailto:cmartin at opennebula.org>
>>>
>>>
>>> On 18 February 2011 17:27,<knawnd at gmail.com<mailto:knawnd at gmail.com>>
>>> wrote:
>>>
>>>      Dear all,
>>>
>>>      Trying to implement IM driver for OpenVZ cluster nodes I faced with
>>>      the following problem:
>>>
>>>      $ onehost list
>>>        ID NAME                           CLUSTER  RVM   TCPU   FCPU
>>>      ACPU   TMEM    FMEM STAT
>>>         2<cluster_node1>          xen_x64           0       200      200
>>>           200       3.1G     128M    on
>>>         3<cluster_node2>          ovz_x32           0            0  0
>>>         100           0K          0K     err
>>>
>>>      oned.log:
>>>      [...]
>>>      Fri Feb 18 18:54:47 2011 [InM][D]: Host 3 successfully monitored.
>>>      Fri Feb 18 18:54:47 2011 [ONE][E]: syntax error, unexpected $end,
>>>      expecting VARIABLE at line 2, columns 1:2
>>>      Fri Feb 18 18:54:47 2011 [InM][E]: Error parsing host information:
>>>      [...]
>>>
>>>      I added oneadmin user into /etc/sudoers file and commented out the
>>>      line "Defaults    requiretty":
>>>      $ egrep "oneadmin|requiretty" /etc/sudoers
>>>      #Defaults    requiretty
>>>      oneadmin        ALL=(ALL)       NOPASSWD: ALL
>>>
>>>      (I couldn't figure out more secure and elegant way to get info from
>>>      OpenVZ host having libvirt interface installed. Any suggestions on
>>>      that?)
>>>
>>>      and did the following changes in ovz.rb file:
>>>      $ diff /usr/lib/one/remotes/im/ovz.d/ovz.rb
>>>      /usr/lib/one/remotes/im/kvm.d/kvm.rb
>>>      29c29
>>>      <  nodeinfo_text = `sudo virsh -c openvz:///system nodeinfo`
>>>      ---
>>>       >  nodeinfo_text = `virsh -c qemu:///system nodeinfo`
>>>      90c90
>>>      <  print_info("HYPERVISOR","ovz")
>>>      ---
>>>       >  print_info("HYPERVISOR","kvm")
>>>
>>>      "sudo virsh -c openvz:///system nodeinfo" running under oneadmin
>>>      user from front-end node gives the following output:
>>>      [oneadmin@<ONE_front-end>$ ssh<OVZ_cluster_node>  sudo virsh -c
>>>      openvz:///system nodeinfo
>>>      CPU model:           i686
>>>      CPU(s):              1
>>>      CPU frequency:       2793 MHz
>>>      CPU socket(s):       1
>>>      Core(s) per socket:  1
>>>      Thread(s) per core:  1
>>>      NUMA cell(s):        1
>>>      Memory size:         2056216 kB
>>>
>>>      I was hoping to apply the same solution as written in "KVM Driver
>>>      Guide 2.0" (http://opennebula.org/documentation:rel2.0:kvmg) since
>>>      ONE uses the libvirt interface as well to interact with KVM hosts:
>>>      ...The following steps are required in the cluster nodes to get the
>>>      KVM driver running:
>>>      [...]
>>>          *      The user with access to these remotes hosts on behalf of
>>>      OpenNebula (typically<oneadmin>) has to pertain to the<libvirtd>
>>>      and<kvm>  groups in order to use the deaemon and be able to launch
>>> VMs.
>>>
>>>      but I can't following that way since there is no libvirt group and
>>>      libvirtd is run under root (I had to build libvirt from tarball
>>>      available at its home page since CentOS has a pretty old one in its
>>>      repos without openvz support).
>>>
>>>      Any help on that is highly appreciated!
>>>      Nikolay.
>>>      _______________________________________________
>>>      Users mailing list
>>>      Users at lists.opennebula.org<mailto:Users at lists.opennebula.org>
>>>      http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>>
>>>
>> _______________________________________________
>> Users mailing list
>> Users at lists.opennebula.org
>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>
>



More information about the Users mailing list