[one-users] Opennebula with LVM

Ruben S. Montero rsmontero at opennebula.org
Fri Feb 28 01:55:45 PST 2014


Ok I see, it is about the monitoring frequency of the datastores, maybe you
could increase the value of MONITOR_INTERVAL in oned.conf, if you are in
4.4 this should not interfere with the intervals for hosts and VMs.


Cheers


On Fri, Feb 28, 2014 at 4:16 AM, Rhesa Mahendra <rhesa at lintasmediadanawa.com
> wrote:

>  Ruben,
>
> This VM is running normal, i just create 7 VM but in process i see so many
> process /lvm/monitor see this:
>
>  8645 ?        SN     0:00 sh -c
> /var/lib/one/remotes/datastore/lvm/monitor
> PERTX0RSSVZFUl9BQ1RJT05fREFUQT48REFUQVNUT1JFPjxJRD4xMjU8L0lEPjxVSUQ+MzwvVUlEPjxHSUQ+MDwvR0lEPjxVTkFNRT5yaGVzYTwvVU5BTUU+PEdOQU1FPm9uZWFkbWluPC9HTkFNRT48TkFNRT5MVk0tU1RPUkU8L05BTUU+PFBFUk1JU1NJT05TPjxPV05FUl9VPjE8L09XTkVSX1U+PE9XTkVSX00+MTwvT1dORVJfTT48T1dORVJfQT4wPC9PV05FUl9BPjxHUk9VUF9VPjE8L0dST1VQX1U+PEdST1VQX00+MDwvR1JPVVBfTT48R1JPVVBfQT4wPC9HUk9VUF9BPjxPVEhFUl9VPjA8L09USEVSX1U+PE9USEVSX00+MDwvT1RIRVJfTT48T1RIRVJfQT4wPC9PVEhFUl9BPjwvUEVSTUlTU0lPTlM+PERTX01BRD5sdm08L0RTX01BRD48VE1fTUFEPmx2bTwvVE1fTUFEPjxCQVNFX1BBVEg+L3Zhci9saWIvb25lLy9kYXRhc3RvcmVzLzEyNTwvQkFTRV9QQVRIPjxUWVBFPjA8L1RZUEU+PERJU0tfVFlQRT4yPC9ESVNLX1RZUEU+PENMVVNURVJfSUQ+LTE8L0NMVVNURVJfSUQ+PENMVVNURVI+PC9DTFVTVEVSPjxUT1RBTF9NQj4xMDk5NTEwODwvVE9UQUxfTUI+PEZSRUVfTUI+MTA3MDUxOTM8L0ZSRUVfTUI+PFVTRURfTUI+Mjg5OTE0PC9VU0VEX01CPjxJTUFHRVM+PElEPjE4MTwvSUQ+PElEPjE4MjwvSUQ+PElEPjE4NDwvSUQ+PElEPjE4NTwvSUQ+PElEPjE4NjwvSUQ+PElEPjE4NzwvSUQ+PElEPjE4ODwvSUQ+PElEPjE4OTwvSUQ+PC9JTUFHRVM+PFRFTVBMQVRFPjxCUklER0VfTElTVD48IVtDRE
> FUQVtsb2N
> hbGhvc3RdXT48L0JSSURHRV9MSVNUPjxDTE9ORV9UQVJHRVQ+PCFbQ0RBVEFbU0VMRl1dPjwvQ0xPTkVfVEFSR0VUPjxEQVRBU1RPUkVfQ0FQQUNJVFlfQ0hFQ0s+PCFbQ0RBVEFbbm9dXT48L0RBVEFTVE9SRV9DQVBBQ0lUWV9DSEVDSz48RElTS19UWVBFPjwhW0NEQVRBW0JMT0NLXV0+PC9ESVNLX1RZUEU+PERTX01BRD48IVtDREFUQVtsdm1dXT48L0RTX01BRD48TE5fVEFSR0VUPjwhW0NEQVRBW05PTkVdXT48L0xOX1RBUkdFVD48VE1fTUFEPjwhW0NEQVRBW2x2bV1dPjwvVE1fTUFEPjxUWVBFPjwhW0NEQVRBW0lNQUdFX0RTXV0+PC9UWVBFPjxWR19OQU1FPjwhW0NEQVRBW3ZnLW9uZV1dPjwvVkdfTkFNRT48L1RFTVBMQVRFPjwvREFUQVNUT1JFPjwvRFNfRFJJVkVSX0FDVElPTl9EQVRBPg==
> 125 ; echo ExitCode: $? 1>&2
>  8647 ?        SN     0:00 /bin/bash
> /var/lib/one/remotes/datastore/lvm/monitor
> PERTX0RSSVZFUl9BQ1RJT05fREFUQT48REFUQVNUT1JFPjxJRD4xMjU8L0lEPjxVSUQ+MzwvVUlEPjxHSUQ+MDwvR0lEPjxVTkFNRT5yaGVzYTwvVU5BTUU+PEdOQU1FPm9uZWFkbWluPC9HTkFNRT48TkFNRT5MVk0tU1RPUkU8L05BTUU+PFBFUk1JU1NJT05TPjxPV05FUl9VPjE8L09XTkVSX1U+PE9XTkVSX00+MTwvT1dORVJfTT48T1dORVJfQT4wPC9PV05FUl9BPjxHUk9VUF9VPjE8L0dST1VQX1U+PEdST1VQX00+MDwvR1JPVVBfTT48R1JPVVBfQT4wPC9HUk9VUF9BPjxPVEhFUl9VPjA8L09USEVSX1U+PE9USEVSX00+MDwvT1RIRVJfTT48T1RIRVJfQT4wPC9PVEhFUl9BPjwvUEVSTUlTU0lPTlM+PERTX01BRD5sdm08L0RTX01BRD48VE1fTUFEPmx2bTwvVE1fTUFEPjxCQVNFX1BBVEg+L3Zhci9saWIvb25lLy9kYXRhc3RvcmVzLzEyNTwvQkFTRV9QQVRIPjxUWVBFPjA8L1RZUEU+PERJU0tfVFlQRT4yPC9ESVNLX1RZUEU+PENMVVNURVJfSUQ+LTE8L0NMVVNURVJfSUQ+PENMVVNURVI+PC9DTFVTVEVSPjxUT1RBTF9NQj4xMDk5NTEwODwvVE9UQUxfTUI+PEZSRUVfTUI+MTA3MDUxOTM8L0ZSRUVfTUI+PFVTRURfTUI+Mjg5OTE0PC9VU0VEX01CPjxJTUFHRVM+PElEPjE4MTwvSUQ+PElEPjE4MjwvSUQ+PElEPjE4NDwvSUQ+PElEPjE4NTwvSUQ+PElEPjE4NjwvSUQ+PElEPjE4NzwvSUQ+PElEPjE4ODwvSUQ+PElEPjE4OTwvSUQ+PC9JTUFHRVM+PFRFTVBMQVRFPjxCUklER0VfTElTVD48IVtDRE
> FUQVtsb2N
> hbGhvc3RdXT48L0JSSURHRV9MSVNUPjxDTE9ORV9UQVJHRVQ+PCFbQ0RBVEFbU0VMRl1dPjwvQ0xPTkVfVEFSR0VUPjxEQVRBU1RPUkVfQ0FQQUNJVFlfQ0hFQ0s+PCFbQ0RBVEFbbm9dXT48L0RBVEFTVE9SRV9DQVBBQ0lUWV9DSEVDSz48RElTS19UWVBFPjwhW0NEQVRBW0JMT0NLXV0+PC9ESVNLX1RZUEU+PERTX01BRD48IVtDREFUQVtsdm1dXT48L0RTX01BRD48TE5fVEFSR0VUPjwhW0NEQVRBW05PTkVdXT48L0xOX1RBUkdFVD48VE1fTUFEPjwhW0NEQVRBW2x2bV1dPjwvVE1fTUFEPjxUWVBFPjwhW0NEQVRBW0lNQUdFX0RTXV0+PC9UWVBFPjxWR19OQU1FPjwhW0NEQVRBW3ZnLW9uZV1dPjwvVkdfTkFNRT48L1RFTVBMQVRFPjwvREFUQVNUT1JFPjwvRFNfRFJJVkVSX0FDVElPTl9EQVRBPg==
> 125
>  8699 ?        SN     0:00 /bin/bash
> /var/lib/one/remotes/datastore/lvm/monitor
> PERTX0RSSVZFUl9BQ1RJT05fREFUQT48REFUQVNUT1JFPjxJRD4xMjU8L0lEPjxVSUQ+MzwvVUlEPjxHSUQ+MDwvR0lEPjxVTkFNRT5yaGVzYTwvVU5BTUU+PEdOQU1FPm9uZWFkbWluPC9HTkFNRT48TkFNRT5MVk0tU1RPUkU8L05BTUU+PFBFUk1JU1NJT05TPjxPV05FUl9VPjE8L09XTkVSX1U+PE9XTkVSX00+MTwvT1dORVJfTT48T1dORVJfQT4wPC9PV05FUl9BPjxHUk9VUF9VPjE8L0dST1VQX1U+PEdST1VQX00+MDwvR1JPVVBfTT48R1JPVVBfQT4wPC9HUk9VUF9BPjxPVEhFUl9VPjA8L09USEVSX1U+PE9USEVSX00+MDwvT1RIRVJfTT48T1RIRVJfQT4wPC9PVEhFUl9BPjwvUEVSTUlTU0lPTlM+PERTX01BRD5sdm08L0RTX01BRD48VE1fTUFEPmx2bTwvVE1fTUFEPjxCQVNFX1BBVEg+L3Zhci9saWIvb25lLy9kYXRhc3RvcmVzLzEyNTwvQkFTRV9QQVRIPjxUWVBFPjA8L1RZUEU+PERJU0tfVFlQRT4yPC9ESVNLX1RZUEU+PENMVVNURVJfSUQ+LTE8L0NMVVNURVJfSUQ+PENMVVNURVI+PC9DTFVTVEVSPjxUT1RBTF9NQj4xMDk5NTEwODwvVE9UQUxfTUI+PEZSRUVfTUI+MTA3MDUxOTM8L0ZSRUVfTUI+PFVTRURfTUI+Mjg5OTE0PC9VU0VEX01CPjxJTUFHRVM+PElEPjE4MTwvSUQ+PElEPjE4MjwvSUQ+PElEPjE4NDwvSUQ+PElEPjE4NTwvSUQ+PElEPjE4NjwvSUQ+PElEPjE4NzwvSUQ+PElEPjE4ODwvSUQ+PElEPjE4OTwvSUQ+PC9JTUFHRVM+PFRFTVBMQVRFPjxCUklER0VfTElTVD48IVtDRE
> FUQVtsb2N
> hbGhvc3RdXT48L0JSSURHRV9MSVNUPjxDTE9ORV9UQVJHRVQ+PCFbQ0RBVEFbU0VMRl1dPjwvQ0xPTkVfVEFSR0VUPjxEQVRBU1RPUkVfQ0FQQUNJVFlfQ0hFQ0s+PCFbQ0RBVEFbbm9dXT48L0RBVEFTVE9SRV9DQVBBQ0lUWV9DSEVDSz48RElTS19UWVBFPjwhW0NEQVRBW0JMT0NLXV0+PC9ESVNLX1RZUEU+PERTX01BRD48IVtDREFUQVtsdm1dXT48L0RTX01BRD48TE5fVEFSR0VUPjwhW0NEQVRBW05PTkVdXT48L0xOX1RBUkdFVD48VE1fTUFEPjwhW0NEQVRBW2x2bV1dPjwvVE1fTUFEPjxUWVBFPjwhW0NEQVRBW0lNQUdFX0RTXV0+PC9UWVBFPjxWR19OQU1FPjwhW0NEQVRBW3ZnLW9uZV1dPjwvVkdfTkFNRT48L1RFTVBMQVRFPjwvREFUQVNUT1JFPjwvRFNfRFJJVkVSX0FDVElPTl9EQVRBPg==
> 125
>
> And see this for vgdisplay :
>
>  8711 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
>  8713 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
>  8882 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
>  8884 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
>  9014 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
>  9016 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
>  9179 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
>  9181 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
>  9351 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
>  9353 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
>  9532 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
>  9534 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
>  9667 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
>  9669 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
>  9833 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
>  9835 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
> 10021 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
> 10023 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
> 10176 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
> 10178 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
> 10329 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
> 10331 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
> 10492 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
> 10494 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
> 10668 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
> 10670 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
> 10846 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
> 10848 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
> 10997 ?        S      0:00 sudo vgdisplay -o vg_size --units M -C
> --noheadings --nosuffix vg-one
> 10999 ?        S      0:00 vgdisplay -o vg_size --units M -C --noheadings
> --nosuffix vg-one
>
>
> This VM running normal, but sometimes lvm/monitor will be check, and so
> take a long for get information monitor, this process queue, and all
> process like (create VM, create image) stuck, until one night, how to fix
> it? thanks.
>
> Rhesa.
> On 02/27/2014 05:10 AM, Ruben S. Montero wrote:
>
> Hi Rhesa,
>
>  Maybe we are just trying to submit to many VMs at the same time for your
> system, so cLVM get stuck. Are you experience this when deploying multiple
> VMs? If so we can either reduce the number of threads of the transfer
> driver to serialize the operations or tweak the scheduler to be less
> aggressive.
>
>  Cheers
>
>  Ruben
>
>
> On Thu, Feb 27, 2014 at 11:05 AM, Rhesa Mahendra <
> rhesa at lintasmediadanawa.com> wrote:
>
>>  Ruben,
>>
>> Thanks for your answer, once again, why command ../lvm/monitor
>> (vgdisplay) take to long to get info monitor LVM, so our frontend have many
>> process, and make everything stuck, how to fix this? thanks,
>>
>> Rhesa.
>>
>>
>> On 02/27/2014 05:02 AM, Ruben S. Montero wrote:
>>
>> Hi,
>>
>>  Yes, given the use of clvm in OpenNebula I think we are safe without
>> fencing. I cannot think of a  split-brain condition where fencing would be
>> needed in our case.
>>
>>  Cheers
>>
>>  Ruben
>>
>>
>> On Thu, Feb 27, 2014 at 1:23 AM, Rhesa Mahendra <
>> rhesa at lintasmediadanawa.com> wrote:
>>
>>>  Ruben,
>>>
>>>  I get error in Fencing, fencing agent not working fine, so if one node
>>> cannot connect fencing this cluster will be stuck, i read from forum, this
>>> fence can connect to ipmi, i think opennebula just need clvm, so i decide
>>> to use cluster without fence, i hope everythink is fine, thanks.
>>>
>>> Regards,
>>> Rhesa Mahendra.
>>>
>>> On 26 Feb 2014, at 23:09, "Ruben S. Montero" <rsmontero at opennebula.org>
>>> wrote:
>>>
>>>   Hi Rhesa
>>>
>>>  I agree that the problem is related to lvm, probably clvmd cannot
>>> acquire locking through DLM. I assume that as you are running the cluster
>>> during 3-4 days it is not mis-configured, I've seen this before related to
>>> networking problems (usually filtering multicast traffic), can you double
>>> check that iptables is allowing all the required cluster traffic?.
>>>
>>>  Also what is the output of clustat, during the failure?
>>>
>>>
>>>  Cheers
>>>
>>>  Ruben
>>>
>>>
>>>  On Wed, Feb 26, 2014 at 3:50 AM, Rhesa Mahendra <
>>> rhesa at lintasmediadanawa.com> wrote:
>>>
>>>> Guys,
>>>>
>>>> I will create production use San Storage, so i think opennebula need
>>>> LVM/CLVM for do, it's have been 3 month for do this, but after i create 50
>>>> VM use one template with 3 node, this lvm/clvm not working fine, status VM
>>>> still Prolog after two days, please see :
>>>>
>>>>
>>>> 0:00 bash -c if [ -x "/var/tmp/one/im/run_probes" ]; then
>>>> /var/tmp/one/im/run_probes kvm /var/lib/one//datastores 4124 20 0
>>>> idc-conode001; else
>>>> 14447 ?        S      0:00 /bin/bash /var/tmp/one/im/run_probes kvm
>>>> /var/lib/one//datastores 4124 20 0 idc-conode001
>>>> 14454 ?        S      0:00 /bin/bash /var/tmp/one/im/run_probes kvm
>>>> /var/lib/one//datastores 4124 20 0 idc-conode001
>>>> 14455 ?        S      0:00 /bin/bash /var/tmp/one/im/run_probes kvm
>>>> /var/lib/one//datastores 4124 20 0 idc-conode001
>>>> 14460 ?        S      0:00 /bin/bash ./collectd-client_control.sh kvm
>>>> /var/lib/one//datastores 4124 20 0 idc-conode001
>>>> 14467 ?        S      0:00 /bin/bash
>>>> /var/tmp/one/im/kvm.d/../run_probes kvm-probes /var/lib/one//datastores
>>>> 4124 20 0 idc-conode001
>>>> 14474 ?        S      0:00 /bin/bash
>>>> /var/tmp/one/im/kvm.d/../run_probes kvm-probes /var/lib/one//datastores
>>>> 4124 20 0 idc-conode001
>>>> 14475 ?        S      0:00 /bin/bash
>>>> /var/tmp/one/im/kvm.d/../run_probes kvm-probes /var/lib/one//datastores
>>>> 4124 20 0 idc-conode001
>>>> 14498 ?        S      0:00 /bin/bash ./monitor_ds.sh kvm-probes
>>>> /var/lib/one//datastores 4124 20 0 idc-conode001
>>>> 14525 ?        S      0:00 /bin/bash ./monitor_ds.sh kvm-probes
>>>> /var/lib/one//datastores 4124 20 0 idc-conode001
>>>> 14526 ?        S      0:00 sudo vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-0
>>>> 14527 ?        S      0:00 vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-0
>>>> 15417 ?        S      0:00 [kdmflush]
>>>> 15452 ?        Ss     0:00 sshd: oneadmin [priv]
>>>> 15454 ?        S      0:00 sshd: oneadmin at notty
>>>> 15455 ?        Ss     0:00 bash -s
>>>> 15510 ?        Ss     0:00 sshd: oneadmin [priv]
>>>> 15512 ?        S      0:00 sshd: oneadmin at notty
>>>> 15513 ?        Ss     0:00 sh -s
>>>> 15527 ?        S      0:00 sudo lvremove -f /dev/vg-one/lv-one-179-596-0
>>>> 15528 ?        S      0:00 lvremove -f /dev/vg-one/lv-one-179-596-0
>>>>
>>>>
>>>> I use locking type 3, i have 3 node and 1 front end, i use cman and
>>>> this is configuration cluster.conf
>>>>
>>>> <?xml version="1.0"?>
>>>> <cluster name="idccluster" config_version="9">
>>>>
>>>>   <clusternodes>
>>>>   <clusternode name="idc-vcoz01" votes="1" nodeid="1"><fence><method
>>>> name="single"><device
>>>> name="idc-vcoz01"/></method></fence></clusternode><clusternode
>>>> name="idc-conode001" votes="1" nodeid="2"><fence><method
>>>> name="single"><device
>>>> name="idc-conode001"/></method></fence></clusternode><clusternode
>>>> name="idc-conode002" votes="1" nodeid="3"><fence><method
>>>> name="single"><device
>>>> name="idc-conode002"/></method></fence></clusternode><clusternode
>>>> name="idc-conode003" votes="1" nodeid="4"><fence><method
>>>> name="single"><device
>>>> name="idc-conode003"/></method></fence></clusternode></clusternodes>
>>>>
>>>>   <fencedevices>
>>>>   <fencedevice name="idc-vcoz01" agent="fence_ipmilan"/><fencedevice
>>>> name="idc-conode001" agent="fence_ipmilan"/><fencedevice
>>>> name="idc-conode002" agent="fence_ipmilan"/><fencedevice
>>>> name="idc-conode003" agent="fence_ipmilan"/></fencedevices>
>>>>
>>>>   <rm>
>>>>     <failoverdomains/>
>>>>     <resources/>
>>>>   </rm>
>>>> </cluster>
>>>>
>>>> i shared /etc/cluster/cluster.conf use NFS,
>>>> this command use cman_tools
>>>>
>>>> Node  Sts   Inc   Joined               Name
>>>>    1   M    304   2014-02-20 16:08:37  idc-vcoz01
>>>>    2   M    288   2014-02-20 16:08:37  idc-conode001
>>>>    3   M    304   2014-02-20 16:08:37  idc-conode002
>>>>    4   M    312   2014-02-26 09:44:04  idc-conode003
>>>>
>>>> i think, this vm cannot running because so take a long for waiting
>>>> lvcreate or vgdisplay, see this:
>>>>
>>>> 30818 ?        S      0:00 sudo vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30819 ?        S      0:00 sudo vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30820 ?        S      0:00 sudo vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30821 ?        S      0:00 sudo vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30824 ?        S      0:00 sudo vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30825 ?        S      0:00 sudo vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30827 ?        S      0:00 sudo vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30842 ?        S      0:00 vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30843 ?        S      0:00 vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30844 ?        S      0:00 vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30845 ?        S      0:00 vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30846 ?        S      0:00 sudo vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30847 ?        S      0:00 vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30852 ?        S      0:00 vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30853 ?        S      0:00 vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>> 30857 ?        S      0:00 vgdisplay --separator : --units m -o
>>>> vg_size,vg_free --nosuffix --noheadings -C vg-one-1
>>>>
>>>>
>>>> or :
>>>>
>>>>
>>>> 30859 ?        S      0:00 sudo lvcreate -L20480.00M -n
>>>> lv-one-179-610-0 vg-one
>>>> 30860 ?        S      0:00 lvcreate -L20480.00M -n lv-one-179-610-0
>>>> vg-one
>>>>
>>>> If i try to restart all server, and all service everything is fine, but
>>>> after 3 or 4 days, this problem come again.
>>>> This Infrastructure will be production, and i think i must find out how
>>>> to fix this, iam not ready if this configuration will be production, so
>>>> please help me, and thanks.
>>>>
>>>> Rhesa.
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users at lists.opennebula.org
>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>>>
>>>
>>>
>>>
>>>  --
>>>  --
>>>  Ruben S. Montero, PhD
>>> Project co-Lead and Chief Architect
>>> OpenNebula - Flexible Enterprise Cloud Made Simple
>>> www.OpenNebula.org | rsmontero at opennebula.org | @OpenNebula
>>>
>>>
>>
>>
>>  --
>>  --
>>  Ruben S. Montero, PhD
>> Project co-Lead and Chief Architect
>> OpenNebula - Flexible Enterprise Cloud Made Simple
>> www.OpenNebula.org | rsmontero at opennebula.org | @OpenNebula
>>
>>
>>
>
>
>  --
>  --
>  Ruben S. Montero, PhD
> Project co-Lead and Chief Architect
> OpenNebula - Flexible Enterprise Cloud Made Simple
> www.OpenNebula.org | rsmontero at opennebula.org | @OpenNebula
>
>
>


-- 
-- 
Ruben S. Montero, PhD
Project co-Lead and Chief Architect
OpenNebula - Flexible Enterprise Cloud Made Simple
www.OpenNebula.org | rsmontero at opennebula.org | @OpenNebula
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.opennebula.org/pipermail/users-opennebula.org/attachments/20140228/32b0aaf9/attachment-0002.htm>


More information about the Users mailing list