Maybe also have a look at these links. http://www.hathitrust.org/blogs/large-scale-search/performance-5-million-volumes http://www.hathitrust.org/blogs/large-scale-search
On Fri, 13 Jan 2012 15:49:06 +0100, Daniel Brügge <dan...@bruegge.eu> wrote: > Hi, > > it's definitely a problem to store 5TB in Solr without using sharding. I > try to split data over solr instances, > so that the index will fit in my memory on the server. > > I ran into trouble with a Solr using 50G index. > > Daniel > > On Jan 13, 2012, at 1:08 PM, mustafozbek wrote: > >> I am an apache solr user about a year. I used solr for simple search >> tools >> but now I want to use solr with 5TB of data. I assume that 5TB data will >> be >> 7TB when solr index it according to filter that I use. And then I will >> add >> nearly 50MB of data per hour to the same index. >> 1- Are there any problem using single solr server with 5TB data. (without >> shards) >> a- Can solr server answers the queries in an acceptable time >> b- what is the expected time for commiting of 50MB data on 7TB index. >> c- Is there an upper limit for index size. >> 2- what are the suggestions that you offer >> a- How many shards should I use >> b- Should I use solr cores >> c- What is the committing frequency you offered. (is 1 hour OK) >> 3- are there any test results for this kind of large data >> >> There is no available 5TB data, I just want to estimate what will be the >> result. >> Note: You can assume that hardware resourses are not a problem. >> >> >> -- >> View this message in context: >> http://lucene.472066.n3.nabble.com/Can-Apache-Solr-Handle-TeraByte-Large-Data-tp3656484p3656484.html >> Sent from the Solr - User mailing list archive at Nabble.com.