[one-users] Opennebula with LVM

Rhesa Mahendra rhesa at lintasmediadanawa.com
Thu Feb 27 02:05:38 PST 2014


Ruben,

Thanks for your answer, once again, why command ../lvm/monitor 
(vgdisplay) take to long to get info monitor LVM, so our frontend have 
many process, and make everything stuck, how to fix this? thanks,

Rhesa.

On 02/27/2014 05:02 AM, Ruben S. Montero wrote:
> Hi,
>
> Yes, given the use of clvm in OpenNebula I think we are safe without 
> fencing. I cannot think of a  split-brain condition where fencing 
> would be needed in our case.
>
> Cheers
>
> Ruben
>
>
> On Thu, Feb 27, 2014 at 1:23 AM, Rhesa Mahendra 
> <rhesa at lintasmediadanawa.com <mailto:rhesa at lintasmediadanawa.com>> wrote:
>
>     Ruben,
>
>     I get error in Fencing, fencing agent not working fine, so if one
>     node cannot connect fencing this cluster will be stuck, i read
>     from forum, this fence can connect to ipmi, i think opennebula
>     just need clvm, so i decide to use cluster without fence, i hope
>     everythink is fine, thanks.
>
>     Regards,
>     Rhesa Mahendra.
>
>     On 26 Feb 2014, at 23:09, "Ruben S. Montero"
>     <rsmontero at opennebula.org <mailto:rsmontero at opennebula.org>> wrote:
>
>>     Hi Rhesa
>>
>>     I agree that the problem is related to lvm, probably clvmd cannot
>>     acquire locking through DLM. I assume that as you are running the
>>     cluster during 3-4 days it is not mis-configured, I've seen this
>>     before related to networking problems (usually filtering
>>     multicast traffic), can you double check that iptables is
>>     allowing all the required cluster traffic?.
>>
>>     Also what is the output of clustat, during the failure?
>>
>>
>>     Cheers
>>
>>     Ruben
>>
>>
>>     On Wed, Feb 26, 2014 at 3:50 AM, Rhesa Mahendra
>>     <rhesa at lintasmediadanawa.com
>>     <mailto:rhesa at lintasmediadanawa.com>> wrote:
>>
>>         Guys,
>>
>>         I will create production use San Storage, so i think
>>         opennebula need LVM/CLVM for do, it's have been 3 month for
>>         do this, but after i create 50 VM use one template with 3
>>         node, this lvm/clvm not working fine, status VM still Prolog
>>         after two days, please see :
>>
>>
>>         0:00 bash -c if [ -x "/var/tmp/one/im/run_probes" ]; then
>>         /var/tmp/one/im/run_probes kvm /var/lib/one//datastores 4124
>>         20 0 idc-conode001; else
>>         14447 ?        S      0:00 /bin/bash
>>         /var/tmp/one/im/run_probes kvm /var/lib/one//datastores 4124
>>         20 0 idc-conode001
>>         14454 ?        S      0:00 /bin/bash
>>         /var/tmp/one/im/run_probes kvm /var/lib/one//datastores 4124
>>         20 0 idc-conode001
>>         14455 ?        S      0:00 /bin/bash
>>         /var/tmp/one/im/run_probes kvm /var/lib/one//datastores 4124
>>         20 0 idc-conode001
>>         14460 ?        S      0:00 /bin/bash
>>         ./collectd-client_control.sh kvm /var/lib/one//datastores
>>         4124 20 0 idc-conode001
>>         14467 ?        S      0:00 /bin/bash
>>         /var/tmp/one/im/kvm.d/../run_probes kvm-probes
>>         /var/lib/one//datastores 4124 20 0 idc-conode001
>>         14474 ?        S      0:00 /bin/bash
>>         /var/tmp/one/im/kvm.d/../run_probes kvm-probes
>>         /var/lib/one//datastores 4124 20 0 idc-conode001
>>         14475 ?        S      0:00 /bin/bash
>>         /var/tmp/one/im/kvm.d/../run_probes kvm-probes
>>         /var/lib/one//datastores 4124 20 0 idc-conode001
>>         14498 ?        S      0:00 /bin/bash ./monitor_ds.sh
>>         kvm-probes /var/lib/one//datastores 4124 20 0 idc-conode001
>>         14525 ?        S      0:00 /bin/bash ./monitor_ds.sh
>>         kvm-probes /var/lib/one//datastores 4124 20 0 idc-conode001
>>         14526 ?        S      0:00 sudo vgdisplay --separator :
>>         --units m -o vg_size,vg_free --nosuffix --noheadings -C vg-one-0
>>         14527 ?        S      0:00 vgdisplay --separator : --units m
>>         -o vg_size,vg_free --nosuffix --noheadings -C vg-one-0
>>         15417 ?        S      0:00 [kdmflush]
>>         15452 ?        Ss     0:00 sshd: oneadmin [priv]
>>         15454 ?        S      0:00 sshd: oneadmin at notty
>>         15455 ?        Ss     0:00 bash -s
>>         15510 ?        Ss     0:00 sshd: oneadmin [priv]
>>         15512 ?        S      0:00 sshd: oneadmin at notty
>>         15513 ?        Ss     0:00 sh -s
>>         15527 ?        S      0:00 sudo lvremove -f
>>         /dev/vg-one/lv-one-179-596-0
>>         15528 ?        S      0:00 lvremove -f
>>         /dev/vg-one/lv-one-179-596-0
>>
>>
>>         I use locking type 3, i have 3 node and 1 front end, i use
>>         cman and this is configuration cluster.conf
>>
>>         <?xml version="1.0"?>
>>         <cluster name="idccluster" config_version="9">
>>
>>           <clusternodes>
>>           <clusternode name="idc-vcoz01" votes="1"
>>         nodeid="1"><fence><method name="single"><device
>>         name="idc-vcoz01"/></method></fence></clusternode><clusternode name="idc-conode001"
>>         votes="1" nodeid="2"><fence><method name="single"><device
>>         name="idc-conode001"/></method></fence></clusternode><clusternode
>>         name="idc-conode002" votes="1" nodeid="3"><fence><method
>>         name="single"><device
>>         name="idc-conode002"/></method></fence></clusternode><clusternode
>>         name="idc-conode003" votes="1" nodeid="4"><fence><method
>>         name="single"><device
>>         name="idc-conode003"/></method></fence></clusternode></clusternodes>
>>
>>           <fencedevices>
>>           <fencedevice name="idc-vcoz01"
>>         agent="fence_ipmilan"/><fencedevice name="idc-conode001"
>>         agent="fence_ipmilan"/><fencedevice name="idc-conode002"
>>         agent="fence_ipmilan"/><fencedevice name="idc-conode003"
>>         agent="fence_ipmilan"/></fencedevices>
>>
>>           <rm>
>>             <failoverdomains/>
>>             <resources/>
>>           </rm>
>>         </cluster>
>>
>>         i shared /etc/cluster/cluster.conf use NFS,
>>         this command use cman_tools
>>
>>         Node  Sts   Inc   Joined               Name
>>            1   M    304   2014-02-20 16:08:37  idc-vcoz01
>>            2   M    288   2014-02-20 16:08:37  idc-conode001
>>            3   M    304   2014-02-20 16:08:37  idc-conode002
>>            4   M    312   2014-02-26 09:44:04  idc-conode003
>>
>>         i think, this vm cannot running because so take a long for
>>         waiting lvcreate or vgdisplay, see this:
>>
>>         30818 ?        S      0:00 sudo vgdisplay --separator :
>>         --units m -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30819 ?        S      0:00 sudo vgdisplay --separator :
>>         --units m -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30820 ?        S      0:00 sudo vgdisplay --separator :
>>         --units m -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30821 ?        S      0:00 sudo vgdisplay --separator :
>>         --units m -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30824 ?        S      0:00 sudo vgdisplay --separator :
>>         --units m -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30825 ?        S      0:00 sudo vgdisplay --separator :
>>         --units m -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30827 ?        S      0:00 sudo vgdisplay --separator :
>>         --units m -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30842 ?        S      0:00 vgdisplay --separator : --units m
>>         -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30843 ?        S      0:00 vgdisplay --separator : --units m
>>         -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30844 ?        S      0:00 vgdisplay --separator : --units m
>>         -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30845 ?        S      0:00 vgdisplay --separator : --units m
>>         -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30846 ?        S      0:00 sudo vgdisplay --separator :
>>         --units m -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30847 ?        S      0:00 vgdisplay --separator : --units m
>>         -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30852 ?        S      0:00 vgdisplay --separator : --units m
>>         -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30853 ?        S      0:00 vgdisplay --separator : --units m
>>         -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>         30857 ?        S      0:00 vgdisplay --separator : --units m
>>         -o vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>
>>
>>         or :
>>
>>
>>         30859 ?        S      0:00 sudo lvcreate -L20480.00M -n
>>         lv-one-179-610-0 vg-one
>>         30860 ?        S      0:00 lvcreate -L20480.00M -n
>>         lv-one-179-610-0 vg-one
>>
>>         If i try to restart all server, and all service everything is
>>         fine, but after 3 or 4 days, this problem come again.
>>         This Infrastructure will be production, and i think i must
>>         find out how to fix this, iam not ready if this configuration
>>         will be production, so please help me, and thanks.
>>
>>         Rhesa.
>>         _______________________________________________
>>         Users mailing list
>>         Users at lists.opennebula.org <mailto:Users at lists.opennebula.org>
>>         http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>
>>
>>
>>
>>     -- 
>>     -- 
>>     Ruben S. Montero, PhD
>>     Project co-Lead and Chief Architect
>>     OpenNebula - Flexible Enterprise Cloud Made Simple
>>     www.OpenNebula.org <http://www.OpenNebula.org> |
>>     rsmontero at opennebula.org <mailto:rsmontero at opennebula.org> |
>>     @OpenNebula
>
>
>
>
> -- 
> -- 
> Ruben S. Montero, PhD
> Project co-Lead and Chief Architect
> OpenNebula - Flexible Enterprise Cloud Made Simple
> www.OpenNebula.org <http://www.OpenNebula.org> | 
> rsmontero at opennebula.org <mailto:rsmontero at opennebula.org> | @OpenNebula

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.opennebula.org/pipermail/users-opennebula.org/attachments/20140227/4bd16855/attachment-0002.htm>


More information about the Users mailing list