Hi Otis, Thanks for your fast answer.
I understand perfectly your points. I will explain my limitations ... --Multiple smaller indices you can split them across several servers, but you can't do that with a monolithic index. The index will be allocated in a SAN that is not under my election. I can decide to split the index or use a monolithic one but not the allocation --With multiple smaller indices you can choose to search only a subset of them, should that make sense for your app. --How much does it cost to have 1 server with a LOT of RAM that serving this index will need? Maybe it's cheaper to have multiple smaller machines. This index will be an index public and i will always need to search in the entire index. I understand the problem of the RAM, but if I use multiple index and then i search in all of them i will use less RAM? The index will have 10 fields, all of them excepting the content will be small and I will only sort be score. If someone have any experience of how much ram i will need or something about the response times with this kind of index it would be very usefull for me. --How long does it take you to rebuild one big index, should it get corrupted vs. rebuilding only a subset of your data? This is a very important aspect, but my primary objective must be the response time. I thought about using different index with different solr but the problem is the mixture of results and how to sort them...so i think (but not sure) that using only one index it will be faster knowing that i will always need to search in the entire index. Any help or suggestion will be very usefull. Thank you very much for your attention 2008/6/14 Otis Gospodnetic <[EMAIL PROTECTED]>: > Roberto, > > Here is some food for thought... > > Multiple smaller indices you can split them across several servers, but you > can't do that with a monolithic index. > > With multiple smaller indices you can choose to search only a subset of > them, should that make sense for your app. > How much does it cost to have 1 server with a LOT of RAM that serving this > index will need? Maybe it's cheaper to have multiple smaller machines. > > How long does it take you to rebuild one big index, should it get corrupted > vs. rebuilding only a subset of your data? > How long does it take you to copy the index around the network after you > optimize it vs. copying only a subset, or multiple subsets in parallel? > > etc. > > Otis -- > Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch > > > ----- Original Message ---- > > From: Roberto Nieto <[EMAIL PROTECTED]> > > To: solr-user@lucene.apache.org > > Sent: Saturday, June 14, 2008 7:31:28 AM > > Subject: doubt with an index of 300gb > > > > Hi users, > > > > I´m going to create a big index of 300gb in a SAN where i have 4TB. I > read > > many entries in the mail list talking about using multiple index with > > multicore. I would like to know what kind of benefit can i have > > using multiple index instead of one big index if i dont have problems > with > > the disk? I know that the optimizes and the commits would be faster with > > smaller indexs, but in search? The RAM use would be the same using 10 > > indexes of 30gb than using 1 index of 300gb? Any suggestion or experience > > will be very usefull for me. > > > > Thanks in advance. > > > > Rober. > >