On Wed, Aug 19, 2015 at 1:36 PM, Somnath Roy <[email protected]> wrote:
> Mark,
> Thanks for verifying this. Nice report !
> Since there is a big difference in memory consumption with jemalloc, I would 
> say a recovery performance data or client performance data during recovery 
> would be helpful.
>

The RSS memory usage in the report is per OSD I guess(really?). It
can't be ignored since it's really a great improvement memory usage.

> Thanks & Regards
> Somnath
>
> -----Original Message-----
> From: [email protected] 
> [mailto:[email protected]] On Behalf Of Mark Nelson
> Sent: Tuesday, August 18, 2015 9:46 PM
> To: ceph-devel
> Subject: Ceph Hackathon: More Memory Allocator Testing
>
> Hi Everyone,
>
> One of the goals at the Ceph Hackathon last week was to examine how to 
> improve Ceph Small IO performance.  Jian Zhang presented findings showing a 
> dramatic improvement in small random IO performance when Ceph is used with 
> jemalloc.  His results build upon Sandisk's original findings that the 
> default thread cache values are a major bottleneck in TCMalloc 2.1.  To 
> further verify these results, we sat down at the Hackathon and configured the 
> new performance test cluster that Intel generously donated to the Ceph 
> community laboratory to run through a variety of tests with different memory 
> allocator configurations.  I've since written the results of those tests up 
> in pdf form for folks who are interested.
>
> The results are located here:
>
> http://nhm.ceph.com/hackathon/Ceph_Hackathon_Memory_Allocator_Testing.pdf
>
> I want to be clear that many other folks have done the heavy lifting here.  
> These results are simply a validation of the many tests that other folks have 
> already done.  Many thanks to Sandisk and others for figuring this out as 
> it's a pretty big deal!
>
> Side note:  Very little tuning other than swapping the memory allocator and a 
> couple of quick and dirty ceph tunables were set during these tests. It's 
> quite possible that higher IOPS will be achieved as we really start digging 
> into the cluster and learning what the bottlenecks are.
>
> Thanks,
> Mark
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the 
> body of a message to [email protected] More majordomo info at  
> http://vger.kernel.org/majordomo-info.html
>
> ________________________________
>
> PLEASE NOTE: The information contained in this electronic mail message is 
> intended only for the use of the designated recipient(s) named above. If the 
> reader of this message is not the intended recipient, you are hereby notified 
> that you have received this message in error and that any review, 
> dissemination, distribution, or copying of this message is strictly 
> prohibited. If you have received this communication in error, please notify 
> the sender by telephone or e-mail (as shown above) immediately and destroy 
> any and all copies of this message in your possession (whether hard copies or 
> electronically stored copies).
>



-- 
Best Regards,

Wheat
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to