<html>
<head>
<meta content="text/html; charset=ISO-8859-1"
http-equiv="Content-Type">
</head>
<body bgcolor="#FFFFFF" text="#000000">
Ruben,<br>
<br>
Thanks for your answer, once again, why command ../lvm/monitor
(vgdisplay) take to long to get info monitor LVM, so our frontend
have many process, and make everything stuck, how to fix this?
thanks,<br>
<br>
Rhesa.<br>
<br>
<div class="moz-cite-prefix">On 02/27/2014 05:02 AM, Ruben S.
Montero wrote:<br>
</div>
<blockquote
cite="mid:CAGi56tc+4Dj2Mv7q36whephjvt_1PkYTii3SN-1ErqndAE96ZQ@mail.gmail.com"
type="cite">
<div dir="ltr">Hi,
<div><br>
</div>
<div>Yes, given the use of clvm in OpenNebula I think we are
safe without fencing. I cannot think of a split-brain
condition where fencing would be needed in our case.</div>
<div><br>
</div>
<div>Cheers</div>
<div><br>
</div>
<div>Ruben</div>
</div>
<div class="gmail_extra"><br>
<br>
<div class="gmail_quote">On Thu, Feb 27, 2014 at 1:23 AM, Rhesa
Mahendra <span dir="ltr"><<a moz-do-not-send="true"
href="mailto:rhesa@lintasmediadanawa.com" target="_blank">rhesa@lintasmediadanawa.com</a>></span>
wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0
.8ex;border-left:1px #ccc solid;padding-left:1ex">
<div dir="auto">
<div>Ruben,</div>
<div><br>
</div>
<div>I get error in Fencing, fencing agent not working
fine, so if one node cannot connect fencing this cluster
will be stuck, i read from forum, this fence can connect
to ipmi, i think opennebula just need clvm, so i decide
to use cluster without fence, i hope everythink is fine,
thanks.<br>
<br>
Regards,
<div>Rhesa Mahendra.</div>
</div>
<div>
<div class="h5">
<div><br>
On 26 Feb 2014, at 23:09, "Ruben S. Montero" <<a
moz-do-not-send="true"
href="mailto:rsmontero@opennebula.org"
target="_blank">rsmontero@opennebula.org</a>>
wrote:<br>
<br>
</div>
<blockquote type="cite">
<div>
<div dir="ltr">
<div>Hi Rhesa</div>
<div><br>
</div>
<div>I agree that the problem is related to lvm,
probably clvmd cannot acquire locking through
DLM. I assume that as you are running the
cluster during 3-4 days it is not
mis-configured, I've seen this before related
to networking problems (usually filtering
multicast traffic), can you double check that
iptables is allowing all the required cluster
traffic?. </div>
<div><br>
</div>
<div>Also what is the output of clustat, during
the failure?</div>
<div><br>
</div>
<div><br>
</div>
<div>Cheers</div>
<div><br>
</div>
<div>Ruben</div>
</div>
<div class="gmail_extra"><br>
<br>
<div class="gmail_quote">
On Wed, Feb 26, 2014 at 3:50 AM, Rhesa
Mahendra <span dir="ltr"><<a
moz-do-not-send="true"
href="mailto:rhesa@lintasmediadanawa.com"
target="_blank">rhesa@lintasmediadanawa.com</a>></span>
wrote:<br>
<blockquote class="gmail_quote"
style="margin:0 0 0 .8ex;border-left:1px
#ccc solid;padding-left:1ex">
Guys,<br>
<br>
I will create production use San Storage, so
i think opennebula need LVM/CLVM for do,
it's have been 3 month for do this, but
after i create 50 VM use one template with 3
node, this lvm/clvm not working fine, status
VM still Prolog after two days, please see :<br>
<br>
<br>
0:00 bash -c if [ -x
"/var/tmp/one/im/run_probes" ]; then
/var/tmp/one/im/run_probes kvm
/var/lib/one//datastores 4124 20 0
idc-conode001; else<br>
14447 ? S 0:00 /bin/bash
/var/tmp/one/im/run_probes kvm
/var/lib/one//datastores 4124 20 0
idc-conode001<br>
14454 ? S 0:00 /bin/bash
/var/tmp/one/im/run_probes kvm
/var/lib/one//datastores 4124 20 0
idc-conode001<br>
14455 ? S 0:00 /bin/bash
/var/tmp/one/im/run_probes kvm
/var/lib/one//datastores 4124 20 0
idc-conode001<br>
14460 ? S 0:00 /bin/bash
./collectd-client_control.sh kvm
/var/lib/one//datastores 4124 20 0
idc-conode001<br>
14467 ? S 0:00 /bin/bash
/var/tmp/one/im/kvm.d/../run_probes
kvm-probes /var/lib/one//datastores 4124 20
0 idc-conode001<br>
14474 ? S 0:00 /bin/bash
/var/tmp/one/im/kvm.d/../run_probes
kvm-probes /var/lib/one//datastores 4124 20
0 idc-conode001<br>
14475 ? S 0:00 /bin/bash
/var/tmp/one/im/kvm.d/../run_probes
kvm-probes /var/lib/one//datastores 4124 20
0 idc-conode001<br>
14498 ? S 0:00 /bin/bash
./monitor_ds.sh kvm-probes
/var/lib/one//datastores 4124 20 0
idc-conode001<br>
14525 ? S 0:00 /bin/bash
./monitor_ds.sh kvm-probes
/var/lib/one//datastores 4124 20 0
idc-conode001<br>
14526 ? S 0:00 sudo vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-0<br>
14527 ? S 0:00 vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-0<br>
15417 ? S 0:00 [kdmflush]<br>
15452 ? Ss 0:00 sshd: oneadmin
[priv]<br>
15454 ? S 0:00 sshd:
oneadmin@notty<br>
15455 ? Ss 0:00 bash -s<br>
15510 ? Ss 0:00 sshd: oneadmin
[priv]<br>
15512 ? S 0:00 sshd:
oneadmin@notty<br>
15513 ? Ss 0:00 sh -s<br>
15527 ? S 0:00 sudo lvremove -f
/dev/vg-one/lv-one-179-596-0<br>
15528 ? S 0:00 lvremove -f
/dev/vg-one/lv-one-179-596-0<br>
<br>
<br>
I use locking type 3, i have 3 node and 1
front end, i use cman and this is
configuration cluster.conf<br>
<br>
<?xml version="1.0"?><br>
<cluster name="idccluster"
config_version="9"><br>
<br>
<clusternodes><br>
<clusternode name="idc-vcoz01"
votes="1"
nodeid="1"><fence><method
name="single"><device
name="idc-vcoz01"/></method></fence></clusternode><clusternode
name="idc-conode001" votes="1"
nodeid="2"><fence><method
name="single"><device
name="idc-conode001"/></method></fence></clusternode><clusternode
name="idc-conode002" votes="1"
nodeid="3"><fence><method
name="single"><device
name="idc-conode002"/></method></fence></clusternode><clusternode
name="idc-conode003" votes="1"
nodeid="4"><fence><method
name="single"><device
name="idc-conode003"/></method></fence></clusternode></clusternodes><br>
<br>
<fencedevices><br>
<fencedevice name="idc-vcoz01"
agent="fence_ipmilan"/><fencedevice
name="idc-conode001"
agent="fence_ipmilan"/><fencedevice
name="idc-conode002"
agent="fence_ipmilan"/><fencedevice
name="idc-conode003"
agent="fence_ipmilan"/></fencedevices><br>
<br>
<rm><br>
<failoverdomains/><br>
<resources/><br>
</rm><br>
</cluster><br>
<br>
i shared /etc/cluster/cluster.conf use NFS,<br>
this command use cman_tools<br>
<br>
Node Sts Inc Joined Name<br>
1 M 304 2014-02-20 16:08:37
idc-vcoz01<br>
2 M 288 2014-02-20 16:08:37
idc-conode001<br>
3 M 304 2014-02-20 16:08:37
idc-conode002<br>
4 M 312 2014-02-26 09:44:04
idc-conode003<br>
<br>
i think, this vm cannot running because so
take a long for waiting lvcreate or
vgdisplay, see this:<br>
<br>
30818 ? S 0:00 sudo vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30819 ? S 0:00 sudo vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30820 ? S 0:00 sudo vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30821 ? S 0:00 sudo vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30824 ? S 0:00 sudo vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30825 ? S 0:00 sudo vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30827 ? S 0:00 sudo vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30842 ? S 0:00 vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30843 ? S 0:00 vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30844 ? S 0:00 vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30845 ? S 0:00 vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30846 ? S 0:00 sudo vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30847 ? S 0:00 vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30852 ? S 0:00 vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30853 ? S 0:00 vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
30857 ? S 0:00 vgdisplay
--separator : --units m -o vg_size,vg_free
--nosuffix --noheadings -C vg-one-1<br>
<br>
<br>
or :<br>
<br>
<br>
30859 ? S 0:00 sudo lvcreate
-L20480.00M -n lv-one-179-610-0 vg-one<br>
30860 ? S 0:00 lvcreate
-L20480.00M -n lv-one-179-610-0 vg-one<br>
<br>
If i try to restart all server, and all
service everything is fine, but after 3 or 4
days, this problem come again.<br>
This Infrastructure will be production, and
i think i must find out how to fix this, iam
not ready if this configuration will be
production, so please help me, and thanks.<br>
<br>
Rhesa.<br>
_______________________________________________<br>
Users mailing list<br>
<a moz-do-not-send="true"
href="mailto:Users@lists.opennebula.org"
target="_blank">Users@lists.opennebula.org</a><br>
<a moz-do-not-send="true"
href="http://lists.opennebula.org/listinfo.cgi/users-opennebula.org"
target="_blank">http://lists.opennebula.org/listinfo.cgi/users-opennebula.org</a><br>
</blockquote>
</div>
<br>
<br clear="all">
<div><br>
</div>
-- <br>
<div dir="ltr">
<div>
<div>-- <br>
</div>
</div>
Ruben S. Montero, PhD<br>
Project co-Lead and Chief Architect
<div>OpenNebula - Flexible Enterprise Cloud
Made Simple<br>
<a moz-do-not-send="true"
href="http://www.OpenNebula.org"
target="_blank">www.OpenNebula.org</a> | <a
moz-do-not-send="true"
href="mailto:rsmontero@opennebula.org"
target="_blank">rsmontero@opennebula.org</a>
| @OpenNebula</div>
</div>
</div>
</div>
</blockquote>
</div>
</div>
</div>
</blockquote>
</div>
<br>
<br clear="all">
<div><br>
</div>
-- <br>
<div dir="ltr">
<div>
<div>-- <br>
</div>
</div>
Ruben S. Montero, PhD<br>
Project co-Lead and Chief Architect
<div>OpenNebula - Flexible Enterprise Cloud Made Simple<br>
<a moz-do-not-send="true" href="http://www.OpenNebula.org"
target="_blank">www.OpenNebula.org</a> | <a
moz-do-not-send="true"
href="mailto:rsmontero@opennebula.org" target="_blank">rsmontero@opennebula.org</a>
| @OpenNebula</div>
</div>
</div>
</blockquote>
<br>
</body>
</html>