[one-users] File system performance testing suite tailored to OpenNebula
João Pagaime
joao.pagaime at gmail.com
Wed Sep 11 06:20:19 PDT 2013
Hello all,
the topic is very interesting
I wonder if anyone could answer this:
what is the penalty of using a file-system on top of a file-system? that
is what happens when the VM disk is a regular file on the hypervisor's
filesystem. I mean: the VM has its own file-system and then the
hypervisor maps that vm-disk on a regular file on another filesystem
(the hypervisor filesystem). Thus the file-system on top of a
file-system issue
putting the question the other way around: what is the benefit of using
raw disk-device (local disk, LVM, iSCSI, ...) as an open-nebula datastore?
didn't test this but I feel the benefit should be substantial
anyway simple bonnie++ tests within a VM show heavy penalties, comparing
test running in the VM and outside (directly on the hipervisor). That
isn't of course an opennebula related performance issue, but a more
general technology challenge
best regards,
João
Em 11-09-2013 13:10, Gerry O'Brien escreveu:
> Hi Carlo,
>
> Thanks for the reply. I should really look at XFS for the
> replication and performance.
>
> Do you have any thoughts on my second questions about qcow2 copies
> form /datastores/1 to /datastores/0 in a single filesystem?
>
> Regards,
> Gerry
>
>
> On 11/09/2013 12:53, Carlo Daffara wrote:
>> It's difficult to provide an indication of what a typical workload
>> may be, as it depends greatly on the
>> I/O properties of the VM that run inside (we found that the
>> "internal" load of OpenNebula itself to be basically negligible).
>> For example, if you have lots of sequential I/O heavy VMs you may get
>> benefits from one kind, while transactional and random I/O VMs may be
>> more suitably served by other file systems.
>> We tend to use fio for benchmarks (http://freecode.com/projects/fio)
>> that is included in most linux distributions; it provides for
>> flexible selection of read-vs-write patterns, can select different
>> probability distributions and includes a few common presets (like
>> file server, mail server etc.)
>> Selecting the bottom file system for the store is thus extremely
>> depending on application, feature and load. For example, we use in
>> some configurations BTRFS with compression (slow rotative devices,
>> especially when there are several of them in parallel), in other we
>> use ext4 (good, all-around balanced) and in other XFS. For example
>> XFS supports filesystem replication in a way similar to that of zfs
>> (not as sofisticated, though), excellent performance for multiple
>> parallel I/O operations.
>> ZFS in our tests tend to be extremely slow outside of a few "sweet
>> spots"; a fact confirmed by external benchmarks like this one:
>> http://www.phoronix.com/scan.php?page=article&item=zfs_linux_062&num=3 We
>> tried it (and we continue to do so, both for the FUSE and native
>> kernel version) but for the moment the performance hit is excessive
>> despite the nice feature set. BTRFS continue to improve nicely, and a
>> set of patches to implement send/receive like ZFS are here:
>> https://btrfs.wiki.kernel.org/index.php/Design_notes_on_Send/Receive
>> but it is still marked as experimental.
>>
>> I personally *love* ZFS, and the feature set is unparalleled.
>> Unfortunately, the poor license choice means that it never got the
>> kind of hammering and tuning that other linux kernel filesystem can get.
>> regards,
>> carlo daffara
>> cloudweavers
>>
>> ----- Messaggio originale -----
>> Da: "Gerry O'Brien" <gerry at scss.tcd.ie>
>> A: "Users OpenNebula" <users at lists.opennebula.org>
>> Inviato: Mercoledì, 11 settembre 2013 13:16:52
>> Oggetto: [one-users] File system performance testing suite tailored
>> to OpenNebula
>>
>> Hi,
>>
>> Are there any recommendations for a file system performance
>> testing
>> suite tailored to OpenNebula typical workloads? I would like to compare
>> the performance of zfs v. ext4. One of the reasons for considering zfs
>> is that it allows replication to a remote site using snapshot streaming.
>> Normal nightly backups, using something like rsync, are not suitable for
>> virtual machine images where a single block change means the whole image
>> has to be copied. The amount of change is to great.
>>
>> On a related issue, does it make sense to have datastores 0 and 1
>> in a single files system so that the instantiations of non-persistent
>> images does not require a copy from one file system to another? I have
>> in mind the case where the original image is a qcow2 image.
>>
>> Regards,
>> Gerry
>>
>
>
More information about the Users
mailing list