Re: Solr hardware memory question

2013-12-12 Thread Michael Della Bitta
uction hardware now, > before profiling - I get to control much of the design and construction so > I don't argue with this!) > > Thanks for all the comments everyone, all very much appreciated :) > Gil > > > -Original Message- > From: Toke Eskildsen [mailto:t..

RE: Solr hardware memory question

2013-12-12 Thread Toke Eskildsen
On Thu, 2013-12-12 at 11:10 +0100, Hoggarth, Gil wrote: > Thanks for this - I haven't any previous experience with utilising SSDs > in the way you suggest, so I guess I need to start learning! There's a bit of divide in the Lucene/Solr-world on this. Everybody agrees that SSDs in themselves are gr

RE: Solr hardware memory question

2013-12-12 Thread Hoggarth, Gil
profiling - I get to control much of the design and construction so I don't argue with this!) Thanks for all the comments everyone, all very much appreciated :) Gil -Original Message- From: Toke Eskildsen [mailto:t...@statsbiblioteket.dk] Sent: 11 December 2013 12:02 To: solr-use

Re: Solr hardware memory question

2013-12-12 Thread Toke Eskildsen
On Thu, 2013-12-12 at 02:46 +0100, Joel Bernstein wrote: > Curious how many documents per shard you were planning? 350-500 million, optimized to a single segment as the data are not changing. > The number of documents per shard and field type will drive the amount > of a RAM needed to sort and fa

Re: Solr hardware memory question

2013-12-11 Thread Otis Gospodnetic
Hi Gil, I'd look at the number and type of fields you sort and facet on (this stuff likes memory). I'd keep in mind heaps over 32 GB use bigger pointers, so maybe more smaller heaps are better than one big one. You didn't mention the # of CPU cores, but keep that in mind when sharding. When a que

Re: Solr hardware memory question

2013-12-11 Thread Joel Bernstein
Curious how many documents per shard you were planning? The number of documents per shard and field type will drive the amount of a RAM needed to sort and facet. On Wed, Dec 11, 2013 at 7:02 AM, Toke Eskildsen wrote: > On Tue, 2013-12-10 at 17:51 +0100, Hoggarth, Gil wrote: > > We're probably go

Re: Solr hardware memory question

2013-12-11 Thread Toke Eskildsen
On Tue, 2013-12-10 at 17:51 +0100, Hoggarth, Gil wrote: > We're probably going to be building a Solr service to handle a dataset > of ~60TB, which for our data and schema typically gives a Solr index > size of 1/10th - i.e., 6TB. Given there's a general rule about the > amount of hardware memory re

Re: Solr hardware memory question

2013-12-10 Thread Ryan Cutter
Shawn's right that if you're going to scale this big you'd be very well served to spend time getting the index as small as possible. In my experience if your searches require real-time random access reads (that is, the entire index needs to be fast), you don't want to wait for HDD disk reads. Get

RE: Solr hardware memory question

2013-12-10 Thread Hoggarth, Gil
rag.org] Sent: 10 December 2013 17:37 To: solr-user@lucene.apache.org Subject: Re: Solr hardware memory question On 12/10/2013 9:51 AM, Hoggarth, Gil wrote: > We're probably going to be building a Solr service to handle a dataset > of ~60TB, which for our data and schema typically gives

Re: Solr hardware memory question

2013-12-10 Thread Shawn Heisey
On 12/10/2013 9:51 AM, Hoggarth, Gil wrote: > We're probably going to be building a Solr service to handle a dataset > of ~60TB, which for our data and schema typically gives a Solr index > size of 1/10th - i.e., 6TB. Given there's a general rule about the > amount of hardware memory required shoul