Hi Antonio, Besides thinking in terms of documents, you also need to think in terms of index size on the file system vs. the amount of RAM your search application/server can use. 50M documents may be doable on a single server if those documents are not too large and you have sufficient RAM. It gets even better if your index doesn't change very often and if you can get decent hit ratios on the various Solr caches.
If you are indexing largish documents, or even something as small as an average web page, 50M docs may be too much on a "commodity box" (say dual core 8 GB RAM box) Otis -- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch ----- Original Message ---- > From: Antonio Eggberg <antonio_eggb...@yahoo.se> > To: solr-user@lucene.apache.org > Sent: Monday, December 29, 2008 3:41:48 PM > Subject: how large can the index be? > > Hi, > > We are running successfully a solr index of 3 million docs. I have just been > informed that our index size will increase to 50 million. I been going > through > the doc > > http://wiki.apache.org/solr/DistributedSearch > > Seems like we will loose out on the date facet and some more other stuff that > we > use. which is important to us. So far we been using 1 index and 1 machine. > > Can I still stick with my 1 index but have many query servers? We don't > update > our index so often this are rather static data. Over the past year we have > updated the index data a total of 3 times and about 300 records :) > > Can someone provide some idea how/what should I do to deal with new datasets?. > > Thanks for your help. > > > __________________________________________________________ > Går det långsamt? Skaffa dig en snabbare bredbandsuppkoppling. > Sök och jämför priser hos Kelkoo. > http://www.kelkoo.se/c-100015813-bredband.html?partnerId=96914325