[one-users] Opennebula with LVM
Rhesa Mahendra
rhesa at lintasmediadanawa.com
Thu Feb 27 19:16:57 PST 2014
Ruben,
This VM is running normal, i just create 7 VM but in process i see so
many process /lvm/monitor see this:
8645 ? SN 0:00 sh -c
/var/lib/one/remotes/datastore/lvm/monitor
PERTX0RSSVZFUl9BQ1RJT05fREFUQT48REFUQVNUT1JFPjxJRD4xMjU8L0lEPjxVSUQ+MzwvVUlEPjxHSUQ+MDwvR0lEPjxVTkFNRT5yaGVzYTwvVU5BTUU+PEdOQU1FPm9uZWFkbWluPC9HTkFNRT48TkFNRT5MVk0tU1RPUkU8L05BTUU+PFBFUk1JU1NJT05TPjxPV05FUl9VPjE8L09XTkVSX1U+PE9XTkVSX00+MTwvT1dORVJfTT48T1dORVJfQT4wPC9PV05FUl9BPjxHUk9VUF9VPjE8L0dST1VQX1U+PEdST1VQX00+MDwvR1JPVVBfTT48R1JPVVBfQT4wPC9HUk9VUF9BPjxPVEhFUl9VPjA8L09USEVSX1U+PE9USEVSX00+MDwvT1RIRVJfTT48T1RIRVJfQT4wPC9PVEhFUl9BPjwvUEVSTUlTU0lPTlM+PERTX01BRD5sdm08L0RTX01BRD48VE1fTUFEPmx2bTwvVE1fTUFEPjxCQVNFX1BBVEg+L3Zhci9saWIvb25lLy9kYXRhc3RvcmVzLzEyNTwvQkFTRV9QQVRIPjxUWVBFPjA8L1RZUEU+PERJU0tfVFlQRT4yPC9ESVNLX1RZUEU+PENMVVNURVJfSUQ+LTE8L0NMVVNURVJfSUQ+PENMVVNURVI+PC9DTFVTVEVSPjxUT1RBTF9NQj4xMDk5NTEwODwvVE9UQUxfTUI+PEZSRUVfTUI+MTA3MDUxOTM8L0ZSRUVfTUI+PFVTRURfTUI+Mjg5OTE0PC9VU0VEX01CPjxJTUFHRVM+PElEPjE4MTwvSUQ+PElEPjE4MjwvSUQ+PElEPjE4NDwvSUQ+PElEPjE4NTwvSUQ+PElEPjE4NjwvSUQ+PElEPjE4NzwvSUQ+PElEPjE4ODwvSUQ+PElEPjE4OTwvSUQ+PC9JTUFHRVM+PFRFTVBMQVRFPjxCUklER0VfTElTVD48IVtDREFUQVtsb2NhbGhvc3RdXT48L0JSSURHRV9MSVNUPjxDTE9ORV9UQVJHRVQ+PCFbQ0RBVEFbU0VMRl1dPjwvQ0xPTkVfVEFSR0VUPjxEQVRBU1RPUkVfQ0FQQUNJVFlfQ0hFQ0s+PCFbQ0RBVEFbbm9dXT48L0RBVEFTVE9SRV9DQVBBQ0lUWV9DSEVDSz48RElTS19UWVBFPjwhW0NEQVRBW0JMT0NLXV0+PC9ESVNLX1RZUEU+PERTX01BRD48IVtDREFUQVtsdm1dXT48L0RTX01BRD48TE5fVEFSR0VUPjwhW0NEQVRBW05PTkVdXT48L0xOX1RBUkdFVD48VE1fTUFEPjwhW0NEQVRBW2x2bV1dPjwvVE1fTUFEPjxUWVBFPjwhW0NEQVRBW0lNQUdFX0RTXV0+PC9UWVBFPjxWR19OQU1FPjwhW0NEQVRBW3ZnLW9uZV1dPjwvVkdfTkFNRT48L1RFTVBMQVRFPjwvREFUQVNUT1JFPjwvRFNfRFJJVkVSX0FDVElPTl9EQVRBPg==
125 ; echo ExitCode: $? 1>&2
8647 ? SN 0:00 /bin/bash
/var/lib/one/remotes/datastore/lvm/monitor
PERTX0RSSVZFUl9BQ1RJT05fREFUQT48REFUQVNUT1JFPjxJRD4xMjU8L0lEPjxVSUQ+MzwvVUlEPjxHSUQ+MDwvR0lEPjxVTkFNRT5yaGVzYTwvVU5BTUU+PEdOQU1FPm9uZWFkbWluPC9HTkFNRT48TkFNRT5MVk0tU1RPUkU8L05BTUU+PFBFUk1JU1NJT05TPjxPV05FUl9VPjE8L09XTkVSX1U+PE9XTkVSX00+MTwvT1dORVJfTT48T1dORVJfQT4wPC9PV05FUl9BPjxHUk9VUF9VPjE8L0dST1VQX1U+PEdST1VQX00+MDwvR1JPVVBfTT48R1JPVVBfQT4wPC9HUk9VUF9BPjxPVEhFUl9VPjA8L09USEVSX1U+PE9USEVSX00+MDwvT1RIRVJfTT48T1RIRVJfQT4wPC9PVEhFUl9BPjwvUEVSTUlTU0lPTlM+PERTX01BRD5sdm08L0RTX01BRD48VE1fTUFEPmx2bTwvVE1fTUFEPjxCQVNFX1BBVEg+L3Zhci9saWIvb25lLy9kYXRhc3RvcmVzLzEyNTwvQkFTRV9QQVRIPjxUWVBFPjA8L1RZUEU+PERJU0tfVFlQRT4yPC9ESVNLX1RZUEU+PENMVVNURVJfSUQ+LTE8L0NMVVNURVJfSUQ+PENMVVNURVI+PC9DTFVTVEVSPjxUT1RBTF9NQj4xMDk5NTEwODwvVE9UQUxfTUI+PEZSRUVfTUI+MTA3MDUxOTM8L0ZSRUVfTUI+PFVTRURfTUI+Mjg5OTE0PC9VU0VEX01CPjxJTUFHRVM+PElEPjE4MTwvSUQ+PElEPjE4MjwvSUQ+PElEPjE4NDwvSUQ+PElEPjE4NTwvSUQ+PElEPjE4NjwvSUQ+PElEPjE4NzwvSUQ+PElEPjE4ODwvSUQ+PElEPjE4OTwvSUQ+PC9JTUFHRVM+PFRFTVBMQVRFPjxCUklER0VfTElTVD48IVtDREFUQVtsb2NhbGhvc3RdXT48L0JSSURHRV9MSVNUPjxDTE9ORV9UQVJHRVQ+PCFbQ0RBVEFbU0VMRl1dPjwvQ0xPTkVfVEFSR0VUPjxEQVRBU1RPUkVfQ0FQQUNJVFlfQ0hFQ0s+PCFbQ0RBVEFbbm9dXT48L0RBVEFTVE9SRV9DQVBBQ0lUWV9DSEVDSz48RElTS19UWVBFPjwhW0NEQVRBW0JMT0NLXV0+PC9ESVNLX1RZUEU+PERTX01BRD48IVtDREFUQVtsdm1dXT48L0RTX01BRD48TE5fVEFSR0VUPjwhW0NEQVRBW05PTkVdXT48L0xOX1RBUkdFVD48VE1fTUFEPjwhW0NEQVRBW2x2bV1dPjwvVE1fTUFEPjxUWVBFPjwhW0NEQVRBW0lNQUdFX0RTXV0+PC9UWVBFPjxWR19OQU1FPjwhW0NEQVRBW3ZnLW9uZV1dPjwvVkdfTkFNRT48L1RFTVBMQVRFPjwvREFUQVNUT1JFPjwvRFNfRFJJVkVSX0FDVElPTl9EQVRBPg==
125
8699 ? SN 0:00 /bin/bash
/var/lib/one/remotes/datastore/lvm/monitor
PERTX0RSSVZFUl9BQ1RJT05fREFUQT48REFUQVNUT1JFPjxJRD4xMjU8L0lEPjxVSUQ+MzwvVUlEPjxHSUQ+MDwvR0lEPjxVTkFNRT5yaGVzYTwvVU5BTUU+PEdOQU1FPm9uZWFkbWluPC9HTkFNRT48TkFNRT5MVk0tU1RPUkU8L05BTUU+PFBFUk1JU1NJT05TPjxPV05FUl9VPjE8L09XTkVSX1U+PE9XTkVSX00+MTwvT1dORVJfTT48T1dORVJfQT4wPC9PV05FUl9BPjxHUk9VUF9VPjE8L0dST1VQX1U+PEdST1VQX00+MDwvR1JPVVBfTT48R1JPVVBfQT4wPC9HUk9VUF9BPjxPVEhFUl9VPjA8L09USEVSX1U+PE9USEVSX00+MDwvT1RIRVJfTT48T1RIRVJfQT4wPC9PVEhFUl9BPjwvUEVSTUlTU0lPTlM+PERTX01BRD5sdm08L0RTX01BRD48VE1fTUFEPmx2bTwvVE1fTUFEPjxCQVNFX1BBVEg+L3Zhci9saWIvb25lLy9kYXRhc3RvcmVzLzEyNTwvQkFTRV9QQVRIPjxUWVBFPjA8L1RZUEU+PERJU0tfVFlQRT4yPC9ESVNLX1RZUEU+PENMVVNURVJfSUQ+LTE8L0NMVVNURVJfSUQ+PENMVVNURVI+PC9DTFVTVEVSPjxUT1RBTF9NQj4xMDk5NTEwODwvVE9UQUxfTUI+PEZSRUVfTUI+MTA3MDUxOTM8L0ZSRUVfTUI+PFVTRURfTUI+Mjg5OTE0PC9VU0VEX01CPjxJTUFHRVM+PElEPjE4MTwvSUQ+PElEPjE4MjwvSUQ+PElEPjE4NDwvSUQ+PElEPjE4NTwvSUQ+PElEPjE4NjwvSUQ+PElEPjE4NzwvSUQ+PElEPjE4ODwvSUQ+PElEPjE4OTwvSUQ+PC9JTUFHRVM+PFRFTVBMQVRFPjxCUklER0VfTElTVD48IVtDREFUQVtsb2NhbGhvc3RdXT48L0JSSURHRV9MSVNUPjxDTE9ORV9UQVJHRVQ+PCFbQ0RBVEFbU0VMRl1dPjwvQ0xPTkVfVEFSR0VUPjxEQVRBU1RPUkVfQ0FQQUNJVFlfQ0hFQ0s+PCFbQ0RBVEFbbm9dXT48L0RBVEFTVE9SRV9DQVBBQ0lUWV9DSEVDSz48RElTS19UWVBFPjwhW0NEQVRBW0JMT0NLXV0+PC9ESVNLX1RZUEU+PERTX01BRD48IVtDREFUQVtsdm1dXT48L0RTX01BRD48TE5fVEFSR0VUPjwhW0NEQVRBW05PTkVdXT48L0xOX1RBUkdFVD48VE1fTUFEPjwhW0NEQVRBW2x2bV1dPjwvVE1fTUFEPjxUWVBFPjwhW0NEQVRBW0lNQUdFX0RTXV0+PC9UWVBFPjxWR19OQU1FPjwhW0NEQVRBW3ZnLW9uZV1dPjwvVkdfTkFNRT48L1RFTVBMQVRFPjwvREFUQVNUT1JFPjwvRFNfRFJJVkVSX0FDVElPTl9EQVRBPg==
125
And see this for vgdisplay :
8711 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
8713 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
8882 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
8884 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9014 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9016 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9179 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9181 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9351 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9353 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9532 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9534 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9667 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9669 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9833 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
9835 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10021 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10023 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10176 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10178 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10329 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10331 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10492 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10494 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10668 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10670 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10846 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10848 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10997 ? S 0:00 sudo vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
10999 ? S 0:00 vgdisplay -o vg_size --units M -C
--noheadings --nosuffix vg-one
This VM running normal, but sometimes lvm/monitor will be check, and so
take a long for get information monitor, this process queue, and all
process like (create VM, create image) stuck, until one night, how to
fix it? thanks.
Rhesa.
On 02/27/2014 05:10 AM, Ruben S. Montero wrote:
> Hi Rhesa,
>
> Maybe we are just trying to submit to many VMs at the same time for
> your system, so cLVM get stuck. Are you experience this when deploying
> multiple VMs? If so we can either reduce the number of threads of the
> transfer driver to serialize the operations or tweak the scheduler to
> be less aggressive.
>
> Cheers
>
> Ruben
>
>
> On Thu, Feb 27, 2014 at 11:05 AM, Rhesa Mahendra
> <rhesa at lintasmediadanawa.com <mailto:rhesa at lintasmediadanawa.com>> wrote:
>
> Ruben,
>
> Thanks for your answer, once again, why command ../lvm/monitor
> (vgdisplay) take to long to get info monitor LVM, so our frontend
> have many process, and make everything stuck, how to fix this? thanks,
>
> Rhesa.
>
>
> On 02/27/2014 05:02 AM, Ruben S. Montero wrote:
>> Hi,
>>
>> Yes, given the use of clvm in OpenNebula I think we are safe
>> without fencing. I cannot think of a split-brain condition where
>> fencing would be needed in our case.
>>
>> Cheers
>>
>> Ruben
>>
>>
>> On Thu, Feb 27, 2014 at 1:23 AM, Rhesa Mahendra
>> <rhesa at lintasmediadanawa.com
>> <mailto:rhesa at lintasmediadanawa.com>> wrote:
>>
>> Ruben,
>>
>> I get error in Fencing, fencing agent not working fine, so if
>> one node cannot connect fencing this cluster will be stuck, i
>> read from forum, this fence can connect to ipmi, i think
>> opennebula just need clvm, so i decide to use cluster without
>> fence, i hope everythink is fine, thanks.
>>
>> Regards,
>> Rhesa Mahendra.
>>
>> On 26 Feb 2014, at 23:09, "Ruben S. Montero"
>> <rsmontero at opennebula.org <mailto:rsmontero at opennebula.org>>
>> wrote:
>>
>>> Hi Rhesa
>>>
>>> I agree that the problem is related to lvm, probably clvmd
>>> cannot acquire locking through DLM. I assume that as you are
>>> running the cluster during 3-4 days it is not
>>> mis-configured, I've seen this before related to networking
>>> problems (usually filtering multicast traffic), can you
>>> double check that iptables is allowing all the required
>>> cluster traffic?.
>>>
>>> Also what is the output of clustat, during the failure?
>>>
>>>
>>> Cheers
>>>
>>> Ruben
>>>
>>>
>>> On Wed, Feb 26, 2014 at 3:50 AM, Rhesa Mahendra
>>> <rhesa at lintasmediadanawa.com
>>> <mailto:rhesa at lintasmediadanawa.com>> wrote:
>>>
>>> Guys,
>>>
>>> I will create production use San Storage, so i think
>>> opennebula need LVM/CLVM for do, it's have been 3 month
>>> for do this, but after i create 50 VM use one template
>>> with 3 node, this lvm/clvm not working fine, status VM
>>> still Prolog after two days, please see :
>>>
>>>
>>> 0:00 bash -c if [ -x "/var/tmp/one/im/run_probes" ];
>>> then /var/tmp/one/im/run_probes kvm
>>> /var/lib/one//datastores 4124 20 0 idc-conode001; else
>>> 14447 ? S 0:00 /bin/bash
>>> /var/tmp/one/im/run_probes kvm /var/lib/one//datastores
>>> 4124 20 0 idc-conode001
>>> 14454 ? S 0:00 /bin/bash
>>> /var/tmp/one/im/run_probes kvm /var/lib/one//datastores
>>> 4124 20 0 idc-conode001
>>> 14455 ? S 0:00 /bin/bash
>>> /var/tmp/one/im/run_probes kvm /var/lib/one//datastores
>>> 4124 20 0 idc-conode001
>>> 14460 ? S 0:00 /bin/bash
>>> ./collectd-client_control.sh kvm
>>> /var/lib/one//datastores 4124 20 0 idc-conode001
>>> 14467 ? S 0:00 /bin/bash
>>> /var/tmp/one/im/kvm.d/../run_probes kvm-probes
>>> /var/lib/one//datastores 4124 20 0 idc-conode001
>>> 14474 ? S 0:00 /bin/bash
>>> /var/tmp/one/im/kvm.d/../run_probes kvm-probes
>>> /var/lib/one//datastores 4124 20 0 idc-conode001
>>> 14475 ? S 0:00 /bin/bash
>>> /var/tmp/one/im/kvm.d/../run_probes kvm-probes
>>> /var/lib/one//datastores 4124 20 0 idc-conode001
>>> 14498 ? S 0:00 /bin/bash ./monitor_ds.sh
>>> kvm-probes /var/lib/one//datastores 4124 20 0 idc-conode001
>>> 14525 ? S 0:00 /bin/bash ./monitor_ds.sh
>>> kvm-probes /var/lib/one//datastores 4124 20 0 idc-conode001
>>> 14526 ? S 0:00 sudo vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-0
>>> 14527 ? S 0:00 vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-0
>>> 15417 ? S 0:00 [kdmflush]
>>> 15452 ? Ss 0:00 sshd: oneadmin [priv]
>>> 15454 ? S 0:00 sshd: oneadmin at notty
>>> 15455 ? Ss 0:00 bash -s
>>> 15510 ? Ss 0:00 sshd: oneadmin [priv]
>>> 15512 ? S 0:00 sshd: oneadmin at notty
>>> 15513 ? Ss 0:00 sh -s
>>> 15527 ? S 0:00 sudo lvremove -f
>>> /dev/vg-one/lv-one-179-596-0
>>> 15528 ? S 0:00 lvremove -f
>>> /dev/vg-one/lv-one-179-596-0
>>>
>>>
>>> I use locking type 3, i have 3 node and 1 front end, i
>>> use cman and this is configuration cluster.conf
>>>
>>> <?xml version="1.0"?>
>>> <cluster name="idccluster" config_version="9">
>>>
>>> <clusternodes>
>>> <clusternode name="idc-vcoz01" votes="1"
>>> nodeid="1"><fence><method name="single"><device
>>> name="idc-vcoz01"/></method></fence></clusternode><clusternode
>>> name="idc-conode001" votes="1" nodeid="2"><fence><method
>>> name="single"><device
>>> name="idc-conode001"/></method></fence></clusternode><clusternode
>>> name="idc-conode002" votes="1" nodeid="3"><fence><method
>>> name="single"><device
>>> name="idc-conode002"/></method></fence></clusternode><clusternode
>>> name="idc-conode003" votes="1" nodeid="4"><fence><method
>>> name="single"><device
>>> name="idc-conode003"/></method></fence></clusternode></clusternodes>
>>>
>>> <fencedevices>
>>> <fencedevice name="idc-vcoz01"
>>> agent="fence_ipmilan"/><fencedevice name="idc-conode001"
>>> agent="fence_ipmilan"/><fencedevice name="idc-conode002"
>>> agent="fence_ipmilan"/><fencedevice name="idc-conode003"
>>> agent="fence_ipmilan"/></fencedevices>
>>>
>>> <rm>
>>> <failoverdomains/>
>>> <resources/>
>>> </rm>
>>> </cluster>
>>>
>>> i shared /etc/cluster/cluster.conf use NFS,
>>> this command use cman_tools
>>>
>>> Node Sts Inc Joined Name
>>> 1 M 304 2014-02-20 16:08:37 idc-vcoz01
>>> 2 M 288 2014-02-20 16:08:37 idc-conode001
>>> 3 M 304 2014-02-20 16:08:37 idc-conode002
>>> 4 M 312 2014-02-26 09:44:04 idc-conode003
>>>
>>> i think, this vm cannot running because so take a long
>>> for waiting lvcreate or vgdisplay, see this:
>>>
>>> 30818 ? S 0:00 sudo vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30819 ? S 0:00 sudo vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30820 ? S 0:00 sudo vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30821 ? S 0:00 sudo vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30824 ? S 0:00 sudo vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30825 ? S 0:00 sudo vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30827 ? S 0:00 sudo vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30842 ? S 0:00 vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30843 ? S 0:00 vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30844 ? S 0:00 vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30845 ? S 0:00 vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30846 ? S 0:00 sudo vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30847 ? S 0:00 vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30852 ? S 0:00 vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30853 ? S 0:00 vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>> 30857 ? S 0:00 vgdisplay --separator :
>>> --units m -o vg_size,vg_free --nosuffix --noheadings -C
>>> vg-one-1
>>>
>>>
>>> or :
>>>
>>>
>>> 30859 ? S 0:00 sudo lvcreate -L20480.00M -n
>>> lv-one-179-610-0 vg-one
>>> 30860 ? S 0:00 lvcreate -L20480.00M -n
>>> lv-one-179-610-0 vg-one
>>>
>>> If i try to restart all server, and all service
>>> everything is fine, but after 3 or 4 days, this problem
>>> come again.
>>> This Infrastructure will be production, and i think i
>>> must find out how to fix this, iam not ready if this
>>> configuration will be production, so please help me, and
>>> thanks.
>>>
>>> Rhesa.
>>> _______________________________________________
>>> Users mailing list
>>> Users at lists.opennebula.org
>>> <mailto:Users at lists.opennebula.org>
>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>>
>>>
>>>
>>>
>>> --
>>> --
>>> Ruben S. Montero, PhD
>>> Project co-Lead and Chief Architect
>>> OpenNebula - Flexible Enterprise Cloud Made Simple
>>> www.OpenNebula.org <http://www.OpenNebula.org> |
>>> rsmontero at opennebula.org <mailto:rsmontero at opennebula.org> |
>>> @OpenNebula
>>
>>
>>
>>
>> --
>> --
>> Ruben S. Montero, PhD
>> Project co-Lead and Chief Architect
>> OpenNebula - Flexible Enterprise Cloud Made Simple
>> www.OpenNebula.org <http://www.OpenNebula.org> |
>> rsmontero at opennebula.org <mailto:rsmontero at opennebula.org> |
>> @OpenNebula
>
>
>
>
> --
> --
> Ruben S. Montero, PhD
> Project co-Lead and Chief Architect
> OpenNebula - Flexible Enterprise Cloud Made Simple
> www.OpenNebula.org <http://www.OpenNebula.org> |
> rsmontero at opennebula.org <mailto:rsmontero at opennebula.org> | @OpenNebula
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.opennebula.org/pipermail/users-opennebula.org/attachments/20140227/649d493f/attachment-0002.htm>
More information about the Users
mailing list