<html>
  <head>
    <meta content="text/html; charset=UTF-8" http-equiv="Content-Type">
  </head>
  <body text="#000000" bgcolor="#FFFFFF">
    Hi Jaime<br>
    <div class="moz-cite-prefix">On 11/08/14 10:28, Jaime Melis wrote:<br>
    </div>
    <blockquote
cite="mid:CA+HrgRrpz_6LT=JK6EkO0k_rS-8SD7RwA-25Yt0WpPd+s3qs9g@mail.gmail.com"
      type="cite">
      <div dir="ltr">By the way, unfortunately this won't make it 4.8 as
        it's already feature freeze (and about to be released). I know
        you reported this a while back and in the end we didn't include
        it because we weren't sure. Sorry for not revisiting this issue
        earlier :S
        <div>
          <div><br>
            Before providing the patch for the docs, let me discuss this
            with the team, see if we can include it for 4.8.1</div>
        </div>
      </div>
    </blockquote>
    <br>
    Ok don't worry, we will include everything into a single new pull
    request to simplify the patch.<br>
    <br>
    Cheers<br>
    Alvaro<br>
    <blockquote
cite="mid:CA+HrgRrpz_6LT=JK6EkO0k_rS-8SD7RwA-25Yt0WpPd+s3qs9g@mail.gmail.com"
      type="cite">
      <div dir="ltr">
        <div>
          <div> ...</div>
          <div><br>
          </div>
          <div>cheers,</div>
          <div>Jaime</div>
          <div class="gmail_extra"><br>
            <div class="gmail_quote">On Mon, Aug 11, 2014 at 10:25 AM,
              Jaime Melis <span dir="ltr"><<a moz-do-not-send="true"
                  href="mailto:jmelis@opennebula.org" target="_blank">jmelis@opennebula.org</a>></span>
              wrote:<br>
              <blockquote class="gmail_quote" style="margin:0px 0px 0px
0.8ex;border-left-width:1px;border-left-color:rgb(204,204,204);border-left-style:solid;padding-left:1ex">
                <div dir="ltr">Hi Alvaro,
                  <div><br>
                  </div>
                  <div>Could you also provide a patch for the
                    documentation?</div>
                </div>
                <div>
                  <div>
                    <div class="gmail_extra"><br>
                      <br>
                      <div class="gmail_quote">On Mon, Aug 11, 2014 at
                        10:22 AM, Alvaro Simon Garcia <span dir="ltr"><<a
                            moz-do-not-send="true"
                            href="mailto:Alvaro.SimonGarcia@ugent.be"
                            target="_blank">Alvaro.SimonGarcia@ugent.be</a>></span>
                        wrote:<br>
                        <blockquote class="gmail_quote"
                          style="margin:0px 0px 0px
0.8ex;border-left-width:1px;border-left-color:rgb(204,204,204);border-left-style:solid;padding-left:1ex">
                          <div text="#000000" bgcolor="#FFFFFF"> Hi
                            Jaime, Javier<br>
                            <br>
                            Do you think that it would be possible to
                            include these ceph patches into next 4.8
                            release? At this moment we are applying
                            theses patches by hand in our configuration
                            to use different Ceph pools.<br>
                            <br>
                            <a moz-do-not-send="true"
                              href="https://github.com/OpenNebula/one/pull/27"
                              target="_blank">https://github.com/OpenNebula/one/pull/27</a><br>
                            <a moz-do-not-send="true"
                              href="https://github.com/OpenNebula/one/pull/28"
                              target="_blank">https://github.com/OpenNebula/one/pull/28</a><br>
                            <a moz-do-not-send="true"
                              href="https://github.com/OpenNebula/one/pull/29"
                              target="_blank">https://github.com/OpenNebula/one/pull/29</a><br>
                            <a moz-do-not-send="true"
                              href="https://github.com/OpenNebula/one/pull/30"
                              target="_blank">https://github.com/OpenNebula/one/pull/30</a><br>
                            <a moz-do-not-send="true"
                              href="https://github.com/OpenNebula/one/pull/31"
                              target="_blank">https://github.com/OpenNebula/one/pull/31</a><br>
                            <br>
                            Cheers<span><font color="#888888"><br>
                                Alvaro</font></span>
                            <div>
                              <div><br>
                                <br>
                                <div>On 17/07/14 14:11, Alvaro Simon
                                  Garcia wrote:<br>
                                </div>
                                <blockquote type="cite"> Hi Jaime<br>
                                  <br>
                                  Sorry for the late reply, I didn't see
                                  your mail before because I was
                                  included in cc. With this patch you
                                  don't need to include a new user into
                                  your ceph keyring. If you want to use
                                  a ceph datastore you only need to
                                  include the user and pool into the
                                  datastore template that's all. In our
                                  case we use livbirt datastore and we
                                  have created two ceph datastore for
                                  testing porposes (each one use a
                                  different pool as well):<br>
                                  <br>
                                  <br>
                                  $ onedatastore show ceph<br>
                                  DATASTORE 103
                                  INFORMATION                                                      
                                  <br>
                                  ID             : 103                 <br>
                                  NAME           : ceph                <br>
                                  USER           : oneadmin            <br>
                                  GROUP          : oneadmin            <br>
                                  CLUSTER        : -                   <br>
                                  TYPE           : IMAGE               <br>
                                  DS_MAD         : ceph                <br>
                                  TM_MAD         : ceph                <br>
                                  BASE PATH      :
                                  /var/lib/one//datastores/103<br>
                                  DISK_TYPE      : RBD                 <br>
                                  <br>
                                  DATASTORE
                                  CAPACITY                                                             
                                  <br>
                                  TOTAL:         : 87.6T               <br>
                                  FREE:          : 59.2T               <br>
                                  USED:          : 28.4T               <br>
                                  LIMIT:         : -                   <br>
                                  <br>
                                  PERMISSIONS                                                                    


                                  <br>
                                  OWNER          : um-                 <br>
                                  GROUP          : u--                 <br>
                                  OTHER          : ---                 <br>
                                  <br>
                                  DATASTORE
                                  TEMPLATE                                                             
                                  <br>
                                  BASE_PATH="/var/lib/one//datastores/"<br>
                                  BRIDGE_LIST="hyp004.cubone.os"<br>
                                  CEPH_HOST="ceph001.cubone.os
                                  ceph002.cubone.os ceph003.cubone.os"<br>
CEPH_SECRET="xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"<br>
                                  CEPH_USER="libvirt"<br>
                                  CLONE_TARGET="SELF"<br>
                                  DATASTORE_CAPACITY_CHECK="yes"<br>
                                  DISK_TYPE="RBD"<br>
                                  DS_MAD="ceph"<br>
                                  LN_TARGET="NONE"<br>
                                  POOL_NAME="one"<br>
                                  TM_MAD="ceph"<br>
                                  <br>
                                  IMAGES         <br>
                                  7              <br>
                                  8              <br>
                                  31             <br>
                                  32             <br>
                                  34             <br>
                                  35           <br>
                                  <br>
                                  <br>
                                  and the second one:<br>
                                  <br>
                                  $ onedatastore show "ceph two"<br>
                                  DATASTORE 104
                                  INFORMATION                                                      
                                  <br>
                                  ID             : 104                 <br>
                                  NAME           : ceph two            <br>
                                  USER           : oneadmin            <br>
                                  GROUP          : oneadmin            <br>
                                  CLUSTER        : -                   <br>
                                  TYPE           : IMAGE               <br>
                                  DS_MAD         : ceph                <br>
                                  TM_MAD         : ceph                <br>
                                  BASE PATH      :
                                  /var/lib/one//datastores/104<br>
                                  DISK_TYPE      : RBD                 <br>
                                  <br>
                                  DATASTORE
                                  CAPACITY                                                             
                                  <br>
                                  TOTAL:         : 87.6T               <br>
                                  FREE:          : 59.2T               <br>
                                  USED:          : 28.4T               <br>
                                  LIMIT:         : -                   <br>
                                  <br>
                                  PERMISSIONS                                                                    


                                  <br>
                                  OWNER          : um-                 <br>
                                  GROUP          : u--                 <br>
                                  OTHER          : ---                 <br>
                                  <br>
                                  DATASTORE
                                  TEMPLATE                                                             
                                  <br>
                                  BASE_PATH="/var/lib/one//datastores/"<br>
                                  BRIDGE_LIST="hyp004.cubone.os"<br>
                                  CEPH_HOST="ceph001.cubone.os
                                  ceph002.cubone.os ceph003.cubone.os"<br>
CEPH_SECRET="xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"<br>
                                  CEPH_USER="libvirt"<br>
                                  CLONE_TARGET="SELF"<br>
                                  DATASTORE_CAPACITY_CHECK="yes"<br>
                                  DISK_TYPE="RBD"<br>
                                  DS_MAD="ceph"<br>
                                  LN_TARGET="NONE"<br>
                                  POOL_NAME="two"<br>
                                  TM_MAD="ceph"<br>
                                  <br>
                                  As you can see we are using different
                                  pools in each one (so we don't need to
                                  include the pool name into ceph.conf
                                  either, and we are able to use several
                                  ceph clusters) and this change
                                  simplifies the ONE conf and ceph
                                  cluster admin as well. <br>
                                  <br>
                                  Cheers<br>
                                  Alvaro<br>
                                  <br>
                                  <div>On 03/07/14 17:08, Jaime Melis
                                    wrote:<br>
                                  </div>
                                  <blockquote type="cite">
                                    <div dir="ltr">Hi Alvaro,
                                      <div><br>
                                      </div>
                                      <div>thanks a lot for the
                                        contribution. I haven't tested
                                        it yet, but the code seems to be
                                        perfect.</div>
                                      <div><br>
                                      </div>
                                      <div>My main issue here is that
                                        I'm not entirely sure if we need
                                        this. Doesn't it make sense for
                                        the oneadmin user to be in the
                                        keyring and be able to run
                                        rados?  You mentioned this, but
                                        I'm having some trouble
                                        understanding what do you mean:</div>
                                      <div><br>
                                      </div>
                                      <div><span
                                          style="font-family:arial,sans-serif;font-size:13px">>


                                        </span><span
                                          style="font-family:arial,sans-serif;font-size:13px">This

                                          feature could be useful as
                                          well if you want to monitor
                                          several datastores that are
                                          using different </span><span
                                          style="font-family:arial,sans-serif;font-size:13px">ceph</span><span
style="font-family:arial,sans-serif;font-size:13px"> pools and users
                                          ids.</span></div>
                                      <div><span
                                          style="font-family:arial,sans-serif;font-size:13px">>


                                          You only have to include the
                                          id and pool info into the ONE </span><span
style="font-family:arial,sans-serif;font-size:13px">datastore</span><span
style="font-family:arial,sans-serif;font-size:13px"> template and the </span><span
style="font-family:arial,sans-serif;font-size:13px">monitoring</span><span
style="font-family:arial,sans-serif;font-size:13px"> script will use one
                                          or another depending on the DS
                                          conf.</span><br
                                          style="font-family:arial,sans-serif;font-size:13px">
                                      </div>
                                      <div><span
                                          style="font-family:arial,sans-serif;font-size:13px"><br>
                                        </span></div>
                                      <div><span
                                          style="font-family:arial,sans-serif;font-size:13px">With


                                          the current system you can
                                          monitor multiple ceph pools as
                                          long as the oneadmin user has
                                          right, isn't that so?</span></div>
                                      <div><br>
                                      </div>
                                      <div>Joel, would you like to weigh
                                        in? would you find this useful?</div>
                                      <div><br>
                                      </div>
                                      <div>cheers,<br>
                                        Jaime</div>
                                    </div>
                                    <div class="gmail_extra"><br>
                                      <br>
                                      <div class="gmail_quote">On Fri,
                                        Jun 20, 2014 at 3:23 PM, Javier
                                        Fontan <span dir="ltr"><<a
                                            moz-do-not-send="true"
                                            href="mailto:jfontan@opennebula.org"
                                            target="_blank">jfontan@opennebula.org</a>></span>
                                        wrote:<br>
                                        <blockquote class="gmail_quote"
                                          style="margin:0px 0px 0px
0.8ex;border-left-width:1px;border-left-color:rgb(204,204,204);border-left-style:solid;padding-left:1ex">Great,

                                          thanks! Our ceph dev is
                                          traveling right now, he will
                                          review the<br>
                                          patches when he arrives.<br>
                                          <br>
                                          Cheers<br>
                                          <br>
                                          On Thu, Jun 19, 2014 at 3:21
                                          PM, Alvaro Simon Garcia<br>
                                          <div>
                                            <div><<a
                                                moz-do-not-send="true"
                                                href="mailto:Alvaro.SimonGarcia@ugent.be"
                                                target="_blank">Alvaro.SimonGarcia@ugent.be</a>>


                                              wrote:<br>
                                              > Hi Javier<br>
                                              ><br>
                                              > We have modified the
                                              ceph monitor to take into
                                              account the datastore ceph<br>
                                              > pool and also the
                                              ceph user. This is a
                                              generic solution that
                                              could be useful<br>
                                              > for other datacenters
                                              as well, we have created a
                                              pull request in github if<br>
                                              > you are agree about
                                              this change and you want
                                              to include it in the next<br>
                                              > release.<br>
                                              ><br>
                                              > <a
                                                moz-do-not-send="true"
                                                href="https://github.com/OpenNebula/one/pull/27"
                                                target="_blank">https://github.com/OpenNebula/one/pull/27</a><br>
                                              ><br>
                                              > We only have modified
                                              these lines in<br>
                                              >
                                              /var/lib/one/remotes/datastore/ceph/monitor:<br>
                                              ><br>
                                              >> ---
                                              monitor.orig.190614  
                                               2014-06-19
                                              14:35:24.022755989 +0200<br>
                                              >> +++ monitor  
                                               2014-06-19
                                              14:49:34.043187892 +0200<br>
                                              >> @@ -46,10 +46,12
                                              @@<br>
                                              >>  while IFS= read
                                              -r -d '' element; do<br>
                                              >>    
                                               XPATH_ELEMENTS[i++]="$element"<br>
                                              >>  done <
                                              <($XPATH
                                              /DS_DRIVER_ACTION_DATA/DATASTORE/TEMPLATE/BRIDGE_LIST
                                              \<br>
                                              >> -
                                              /DS_DRIVER_ACTION_DATA/DATASTORE/TEMPLATE/POOL_NAME)<br>
                                              >> +
                                              /DS_DRIVER_ACTION_DATA/DATASTORE/TEMPLATE/POOL_NAME
                                              \<br>
                                              >> +
                                              /DS_DRIVER_ACTION_DATA/DATASTORE/TEMPLATE/CEPH_USER)<br>
                                              >><br>
                                              >>
                                               BRIDGE_LIST="${XPATH_ELEMENTS[j++]}"<br>
                                              >>
                                               POOL_NAME="${XPATH_ELEMENTS[j++]:-$POOL_NAME}"<br>
                                              >>
                                              +CEPH_USER="${XPATH_ELEMENTS[j++]}"<br>
                                              >><br>
                                              >>
                                               HOST=`get_destination_host`<br>
                                              >><br>
                                              >> @@ -61,7 +63,7 @@<br>
                                              >>  # ------------
                                              Compute datastore usage
                                              -------------<br>
                                              >><br>
                                              >>
                                               MONITOR_SCRIPT=$(cat
                                              <<EOF<br>
                                              >> -$RADOS df | $AWK
                                              '{<br>
                                              >> +$RADOS df -p
                                              ${POOL_NAME} --id
                                              ${CEPH_USER}| $AWK '{<br>
                                              >>      if (\$1 ==
                                              "total") {<br>
                                              >><br>
                                              >>          space =
                                              int(\$3/1024)<br>
                                              ><br>
                                              ><br>
                                              > CEPH_USER and
                                              POOL_NAME should be
                                              mandatory to create the
                                              ceph datastore.<br>
                                              ><br>
                                              > Cheers<br>
                                              > Alvaro<br>
                                              ><br>
                                              ><br>
                                              ><br>
                                              >> Hola Javi<br>
                                              >><br>
                                              >> Thanks a lot for
                                              your feedback. Yes we will
                                              modify the current<br>
                                              >> monitoring
                                              scripts to take into
                                              account this. This feature
                                              could be<br>
                                              >> useful as well if
                                              you want to monitor
                                              several datastores that
                                              are using<br>
                                              >> different ceph
                                              pools and users ids. You
                                              only have to include the
                                              id and<br>
                                              >> pool info into
                                              the ONE datastore template
                                              and the monitoring script
                                              will<br>
                                              >> use one or
                                              another depending on the
                                              DS conf.<br>
                                              >><br>
                                              >><br>
                                              >> Cheers and
                                              thanks!<br>
                                              >> Alvaro<br>
                                              >> On 2014-06-17
                                              14:55, Javier Fontan
                                              wrote:<br>
                                              >>><br>
                                              >>> CEPH_USER is
                                              used when generating the
                                              libvirt/kvm deployment
                                              file but<br>
                                              >>> not for DS
                                              monitoring:<br>
                                              >>><br>
                                              >>> * Deployment
                                              file generation:<br>
                                              >>><br>
                                              >>> <a
                                                moz-do-not-send="true"
href="https://github.com/OpenNebula/one/blob/one-4.6/src/vmm/LibVirtDriverKVM.cc#L461"
                                                target="_blank">https://github.com/OpenNebula/one/blob/one-4.6/src/vmm/LibVirtDriverKVM.cc#L461</a><br>
                                              >>> * Monitoring:<br>
                                              >>> <a
                                                moz-do-not-send="true"
href="https://github.com/OpenNebula/one/blob/one-4.6/src/datastore_mad/remotes/ceph/monitor#L64"
                                                target="_blank">https://github.com/OpenNebula/one/blob/one-4.6/src/datastore_mad/remotes/ceph/monitor#L64</a><br>
                                              >>><br>
                                              >>> Ceph is not
                                              may area of expertise but
                                              you may need to add those<br>
                                              >>> parameters to
                                              the monitor script a maybe
                                              to other scripts that use
                                              the<br>
                                              >>> "rados"
                                              command. It may also be
                                              possible to modify the
                                              RADOS command<br>
                                              >>> to have those
                                              parameters instead of
                                              modifying all the scripts:<br>
                                              >>><br>
                                              >>><br>
                                              >>> <a
                                                moz-do-not-send="true"
href="https://github.com/OpenNebula/one/blob/one-4.6/src/mad/sh/scripts_common.sh#L40"
                                                target="_blank">https://github.com/OpenNebula/one/blob/one-4.6/src/mad/sh/scripts_common.sh#L40</a><br>
                                              >>><br>
                                              >>> As I said I
                                              don't know much about Ceph
                                              and it may be those
                                              credentials<br>
                                              >>> could be set
                                              in a config file or so.<br>
                                              >>><br>
                                              >>> On Tue, Jun
                                              17, 2014 at 11:19 AM,
                                              Alvaro Simon Garcia<br>
                                              >>> <<a
                                                moz-do-not-send="true"
                                                href="mailto:Alvaro.SimonGarcia@ugent.be"
                                                target="_blank">Alvaro.SimonGarcia@ugent.be</a>>


                                              wrote:<br>
                                              >>>><br>
                                              >>>> Hi<br>
                                              >>>><br>
                                              >>>> We have
                                              included the admin keyring
                                              instead of libvirt user
                                              and it<br>
                                              >>>> works...<br>
                                              >>>> that
                                              means that we can run rbd
                                              or qemu-img wihtout the
                                              libvirt id, but<br>
                                              >>>> is<br>
                                              >>>> not the
                                              best solution. We have
                                              included the user into
                                              datastore conf:<br>
                                              >>>><br>
                                              >>>>
                                              CEPH_USER="libvirt"<br>
                                              >>>><br>
                                              >>>> but it
                                              seems that is not used by
                                              opennebula at the end<br>
                                              >>>><br>
                                              >>>> Cheers<br>
                                              >>>> Alvaro<br>
                                              >>>><br>
                                              >>>><br>
                                              >>>> On
                                              2014-06-17 10:09, Alvaro
                                              Simon Garcia wrote:<br>
                                              >>>>><br>
                                              >>>>> Hi
                                              all<br>
                                              >>>>><br>
                                              >>>>><br>
                                              >>>>> We
                                              have included our ONE
                                              nodes into Ceph cluster,
                                              cephx auth is working<br>
                                              >>>>> but
                                              OpenNebula is not able to
                                              detect the free space:<br>
                                              >>>>><br>
                                              >>>>><br>
                                              >>>>><br>
                                              >>>>>> $
                                              onedatastore show 103<br>
                                              >>>>>>
                                              DATASTORE 103 INFORMATION<br>
                                              >>>>>>
                                              ID             : 103<br>
                                              >>>>>>
                                              NAME           : ceph<br>
                                              >>>>>>
                                              USER           : oneadmin<br>
                                              >>>>>>
                                              GROUP          : oneadmin<br>
                                              >>>>>>
                                              CLUSTER        : -<br>
                                              >>>>>>
                                              TYPE           : IMAGE<br>
                                              >>>>>>
                                              DS_MAD         : ceph<br>
                                              >>>>>>
                                              TM_MAD         : ceph<br>
                                              >>>>>>
                                              BASE PATH      :
                                              /var/lib/one//datastores/103<br>
                                              >>>>>>
                                              DISK_TYPE      : RBD<br>
                                              >>>>>><br>
                                              >>>>>>
                                              DATASTORE CAPACITY<br>
                                              >>>>>>
                                              TOTAL:         : 0M<br>
                                              >>>>>>
                                              FREE:          : 0M<br>
                                              >>>>>>
                                              USED:          : 0M<br>
                                              >>>>>>
                                              LIMIT:         : -<br>
                                              >>>>>><br>
                                              >>>>>>
                                              PERMISSIONS<br>
                                              >>>>>>
                                              OWNER          : um-<br>
                                              >>>>>>
                                              GROUP          : u--<br>
                                              >>>>>>
                                              OTHER          : ---<br>
                                              >>>>><br>
                                              >>>>><br>
                                              >>>>>> $
                                              onedatastore list<br>
                                              >>>>>>  
                                                 ID NAME              
                                               SIZE AVAIL CLUSTER    
                                               IMAGES TYPE DS<br>
                                              >>>>>>
                                              TM<br>
                                              >>>>>>  
                                                  0 system          
                                               114.8G 85%   -          
                                                    0 sys -<br>
                                              >>>>>>
                                              shared<br>
                                              >>>>>>  
                                                  1 default          
                                              114.9G 84%   -            
                                                  2 img fs<br>
                                              >>>>>>
                                              ssh<br>
                                              >>>>>>  
                                                  2 files            
                                              114.9G 84%   -            
                                                  0 fil fs<br>
                                              >>>>>>
                                              ssh<br>
                                              >>>>>>  
                                                103 ceph                
                                               0M -     -              
                                                0 img ceph<br>
                                              >>>>>>
                                              ceph<br>
                                              >>>>><br>
                                              >>>>><br>
                                              >>>>> but
                                              if we run rados as
                                              oneadmin user:<br>
                                              >>>>><br>
                                              >>>>>> $
                                               rados df -p one --id
                                              libvirt<br>
                                              >>>>>>
                                              pool name       category  
                                                            KB    
                                               objects clones<br>
                                              >>>>>>
                                              degraded      unfound    
                                                    rd        rd KB    
                                                    wr        wr<br>
                                              >>>>>>
                                              KB<br>
                                              >>>>>>
                                              one             -        
                                                               0 0      
                                                   0<br>
                                              >>>>>> 0
                                                        0            0 0
                                                         0            0<br>
                                              >>>>>>  
                                                 total used        
                                              1581852           37<br>
                                              >>>>>>  
                                                 total avail  
                                              140846865180<br>
                                              >>>>>>  
                                                 total space  
                                              140848447032<br>
                                              >>>>><br>
                                              >>>>><br>
                                              >>>>> It's
                                              working correctly (we are
                                              using one pool and libvirt
                                              ceph id)<br>
                                              >>>>><br>
                                              >>>>> the
                                              oned.log only shows this
                                              info:<br>
                                              >>>>> Tue
                                              Jun 17 10:06:37 2014
                                              [InM][D]: Monitoring
                                              datastore default (1)<br>
                                              >>>>> Tue
                                              Jun 17 10:06:37 2014
                                              [InM][D]: Monitoring
                                              datastore files (2)<br>
                                              >>>>> Tue
                                              Jun 17 10:06:37 2014
                                              [InM][D]: Monitoring
                                              datastore ceph (103)<br>
                                              >>>>> Tue
                                              Jun 17 10:06:37 2014
                                              [ImM][D]: Datastore
                                              default (1) successfully<br>
                                              >>>>>
                                              monitored.<br>
                                              >>>>> Tue
                                              Jun 17 10:06:37 2014
                                              [ImM][D]: Datastore files
                                              (2) successfully<br>
                                              >>>>>
                                              monitored.<br>
                                              >>>>> Tue
                                              Jun 17 10:06:37 2014
                                              [ImM][D]: Datastore ceph
                                              (103) successfully<br>
                                              >>>>>
                                              monitored.<br>
                                              >>>>><br>
                                              >>>>> Any
                                              clue about how to debug
                                              this issue?<br>
                                              >>>>><br>
                                              >>>>>
                                              Thanks in advance!<br>
                                              >>>>>
                                              Alvaro<br>
                                              >>>>><br>
                                              >>>>><br>
                                              >>>>><br>
                                              >>>>>
                                              _______________________________________________<br>
                                              >>>>> Users
                                              mailing list<br>
                                              >>>>> <a
                                                moz-do-not-send="true"
                                                href="mailto:Users@lists.opennebula.org"
                                                target="_blank">Users@lists.opennebula.org</a><br>
                                              >>>>> <a
                                                moz-do-not-send="true"
                                                href="http://lists.opennebula.org/listinfo.cgi/users-opennebula.org"
                                                target="_blank">http://lists.opennebula.org/listinfo.cgi/users-opennebula.org</a><br>
                                              >>>><br>
                                              >>>>
                                              _______________________________________________<br>
                                              >>>> Users
                                              mailing list<br>
                                              >>>> <a
                                                moz-do-not-send="true"
                                                href="mailto:Users@lists.opennebula.org"
                                                target="_blank">Users@lists.opennebula.org</a><br>
                                              >>>> <a
                                                moz-do-not-send="true"
                                                href="http://lists.opennebula.org/listinfo.cgi/users-opennebula.org"
                                                target="_blank">http://lists.opennebula.org/listinfo.cgi/users-opennebula.org</a><br>
                                              >>><br>
                                              >>><br>
                                              >>
                                              _______________________________________________<br>
                                              >> Users mailing
                                              list<br>
                                              >> <a
                                                moz-do-not-send="true"
                                                href="mailto:Users@lists.opennebula.org"
                                                target="_blank">Users@lists.opennebula.org</a><br>
                                              >> <a
                                                moz-do-not-send="true"
                                                href="http://lists.opennebula.org/listinfo.cgi/users-opennebula.org"
                                                target="_blank">http://lists.opennebula.org/listinfo.cgi/users-opennebula.org</a><br>
                                              ><br>
                                              ><br>
                                              <br>
                                              <br>
                                              <br>
                                            </div>
                                          </div>
                                          <div>--<br>
                                            Javier Fontán Muiños<br>
                                            Developer<br>
                                            OpenNebula - Flexible
                                            Enterprise Cloud Made Simple<br>
                                            <a moz-do-not-send="true"
                                              href="http://www.OpenNebula.org"
                                              target="_blank">www.OpenNebula.org</a>
                                            | @OpenNebula | <a
                                              moz-do-not-send="true"
                                              href="http://github.com/jfontan"
                                              target="_blank">github.com/jfontan</a><br>
                                          </div>
                                          <div>
                                            <div>_______________________________________________<br>
                                              Users mailing list<br>
                                              <a moz-do-not-send="true"
href="mailto:Users@lists.opennebula.org" target="_blank">Users@lists.opennebula.org</a><br>
                                              <a moz-do-not-send="true"
href="http://lists.opennebula.org/listinfo.cgi/users-opennebula.org"
                                                target="_blank">http://lists.opennebula.org/listinfo.cgi/users-opennebula.org</a><br>
                                            </div>
                                          </div>
                                        </blockquote>
                                      </div>
                                      <br>
                                    </div>
                                    <br clear="all">
                                    <div><br>
                                    </div>
                                    -- <br>
                                    <div dir="ltr">
                                      <div>Jaime Melis<br>
                                        Project Engineer<br>
                                        OpenNebula - Flexible Enterprise
                                        Cloud Made Simple<br>
                                        <a moz-do-not-send="true"
                                          href="http://www.OpenNebula.org"
                                          target="_blank">www.OpenNebula.org</a>
                                        | <a moz-do-not-send="true"
                                          href="mailto:jmelis@opennebula.org"
                                          target="_blank">jmelis@opennebula.org</a></div>
                                    </div>
                                  </blockquote>
                                  <br>
                                </blockquote>
                                <br>
                              </div>
                            </div>
                          </div>
                          <br>
_______________________________________________<br>
                          Users mailing list<br>
                          <a moz-do-not-send="true"
                            href="mailto:Users@lists.opennebula.org"
                            target="_blank">Users@lists.opennebula.org</a><br>
                          <a moz-do-not-send="true"
                            href="http://lists.opennebula.org/listinfo.cgi/users-opennebula.org"
                            target="_blank">http://lists.opennebula.org/listinfo.cgi/users-opennebula.org</a><br>
                          <br>
                        </blockquote>
                      </div>
                      <br>
                      <br clear="all">
                      <div><br>
                      </div>
                      -- <br>
                      <div dir="ltr">
                        <div>Jaime Melis<br>
                          Project Engineer<br>
                          OpenNebula - Flexible Enterprise Cloud Made
                          Simple<br>
                          <a moz-do-not-send="true"
                            href="http://www.OpenNebula.org"
                            target="_blank">www.OpenNebula.org</a> | <a
                            moz-do-not-send="true"
                            href="mailto:jmelis@opennebula.org"
                            target="_blank">jmelis@opennebula.org</a></div>
                      </div>
                    </div>
                  </div>
                </div>
              </blockquote>
            </div>
            <br>
            <br clear="all">
            <div><br>
            </div>
            -- <br>
            <div dir="ltr">
              <div>Jaime Melis<br>
                Project Engineer<br>
                OpenNebula - Flexible Enterprise Cloud Made Simple<br>
                <a moz-do-not-send="true"
                  href="http://www.OpenNebula.org" target="_blank">www.OpenNebula.org</a>
                | <a moz-do-not-send="true"
                  href="mailto:jmelis@opennebula.org" target="_blank">jmelis@opennebula.org</a></div>
            </div>
          </div>
        </div>
      </div>
    </blockquote>
    <br>
  </body>
</html>