Hmm try searching the libvirt git for josh as an author you should see the 
commit from Josh Durgan about white listing rbd migration.




On May 11, 2013, at 10:53 AM, w sun <[email protected]> wrote:

> The reference Mike provided is not valid to me.  Anyone else has the same 
> problem? --weiguo
> 
> From: [email protected]
> Date: Sat, 11 May 2013 08:45:41 -0400
> To: [email protected]
> CC: [email protected]
> Subject: Re: [ceph-users] RBD vs RADOS benchmark performance
> 
> I believe that this is fixed in the most recent versions of libvirt, sheepdog 
> and rbd were marked erroneously as unsafe.
> 
> http://libvirt.org/git/?p=libvirt.git;a=commit;h=78290b1641e95304c862062ee0aca95395c5926c
> 
> Sent from my iPad
> 
> On May 11, 2013, at 8:36 AM, Mike Kelly <[email protected]> wrote:
> 
> (Sorry for sending this twice... Forgot to reply to the list)
> 
> Is rbd caching safe to enable when you may need to do a live migration of the 
> guest later on? It was my understanding that it wasn't, and that libvirt 
> prevented you from doing the migration of it knew about the caching setting.
> 
> If it isn't, is there anything else that could help performance? Like, some 
> tuning of block size parameters for the rbd image or the qemu
> 
> On May 10, 2013 8:57 PM, "Mark Nelson" <[email protected]> wrote:
> On 05/10/2013 07:21 PM, Yun Mao wrote:
> Hi Mark,
> 
> Given the same hardware, optimal configuration (I have no idea what that
> means exactly but feel free to specify), which is supposed to perform
> better, kernel rbd or qemu/kvm? Thanks,
> 
> Yun
> 
> Hi Yun,
> 
> I'm in the process of actually running some tests right now.
> 
> In previous testing, it looked like kernel rbd and qemu/kvm performed about 
> the same with cache off.  With cache on (in cuttlefish), small sequential 
> write performance improved pretty dramatically vs without cache.  Large write 
> performance seemed to take more concurrency to reach peak performance, but 
> ultimately aggregate throughput was about the same.
> 
> Hopefully I should have some new results published in the near future.
> 
> Mark
> 
> 
> 
> On Fri, May 10, 2013 at 6:56 PM, Mark Nelson <[email protected]
> <mailto:[email protected]>> wrote:
> 
>     On 05/10/2013 12:16 PM, Greg wrote:
> 
>         Hello folks,
> 
>         I'm in the process of testing CEPH and RBD, I have set up a small
>         cluster of  hosts running each a MON and an OSD with both
>         journal and
>         data on the same SSD (ok this is stupid but this is simple to
>         verify the
>         disks are not the bottleneck for 1 client). All nodes are
>         connected on a
>         1Gb network (no dedicated network for OSDs, shame on me :).
> 
>         Summary : the RBD performance is poor compared to benchmark
> 
>         A 5 seconds seq read benchmark shows something like this :
> 
>                 sec Cur ops   started  finished  avg MB/s  cur MB/s
>               last lat   avg
>             lat
>                   0       0         0         0         0 0         -
>                    0
>                   1      16        39        23   91.9586        92
>             0.966117  0.431249
>                   2      16        64        48   95.9602       100
>             0.513435   0.53849
>                   3      16        90        74   98.6317       104
>             0.25631   0.55494
>                   4      11        95        84   83.9735        40
>             1.80038   0.58712
>               Total time run:        4.165747
>             Total reads made:     95
>             Read size:            4194304
>             Bandwidth (MB/sec):    91.220
> 
>             Average Latency:       0.678901
>             Max latency:           1.80038
>             Min latency:           0.104719
> 
> 
>         91MB read performance, quite good !
> 
>         Now the RBD performance :
> 
>             root@client:~# dd if=/dev/rbd1 of=/dev/null bs=4M count=100
>             100+0 records in
>             100+0 records out
>             419430400 bytes (419 MB) copied, 13.0568 s, 32.1 MB/s
> 
> 
>         There is a 3x performance factor (same for write: ~60M
>         benchmark, ~20M
>         dd on block device)
> 
>         The network is ok, the CPU is also ok on all OSDs.
>         CEPH is Bobtail 0.56.4, linux is 3.8.1 arm (vanilla release + some
>         patches for the SoC being used)
> 
>         Can you show me the starting point for digging into this ?
> 
> 
>     Hi Greg, First things first, are you doing kernel rbd or qemu/kvm?
>       If you are doing qemu/kvm, make sure you are using virtio disks.
>       This can have a pretty big performance impact.  Next, are you
>     using RBD cache? With 0.56.4 there are some performance issues with
>     large sequential writes if cache is on, but it does provide benefit
>     for small sequential writes.  In general RBD cache behaviour has
>     improved with Cuttlefish.
> 
>     Beyond that, are the pools being targeted by RBD and rados bench
>     setup the same way?  Same number of Pgs?  Same replication?
> 
> 
> 
>         Thanks!
>         _________________________________________________
>         ceph-users mailing list
>         [email protected] <mailto:[email protected]>
>         http://lists.ceph.com/__listinfo.cgi/ceph-users-ceph.__com
>         <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
> 
> 
>     _________________________________________________
>     ceph-users mailing list
>     [email protected] <mailto:[email protected]>
>     http://lists.ceph.com/__listinfo.cgi/ceph-users-ceph.__com
>     <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
> 
> 
> 
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> _______________________________________________ ceph-users mailing list 
> [email protected] 
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to