Unfortunately the answer for this can vary quite a bit based on a number of factors:
1. Whether or not fields are stored, 2. Document size, 3. Total term count, 4. Solr version etc. We have two major indexes, one for servicing online queries, and one for batch processing. Our batch index is performance critical and therefore was optimized for throughput, was stored in RAM, and has less stored fields than the online query one. The batch index shards are 25Gb or less, and we're trending toward smaller and more numerous shards. This is with 1.4, and I'm just finishing up on our migration to 3.6.1. Michael Della Bitta P.S. Why'd you CC honeybadger? Honeybadger don't care... ------------------------------------------------ Appinions | 18 East 41st St., Suite 1806 | New York, NY 10017 www.appinions.com Where Influence Isn’t a Game On Wed, Aug 29, 2012 at 5:17 PM, Michael Brandt <michael.j.bra...@colorado.edu> wrote: > Hi all, > > I am looking for information on how many documents may be indexed by a > single instance of Solr (not using shards) before performance issues are > encountered. In searching the internet I've come across some varying > answers; one answer suggest 50GBs is > problematic<http://lucene.472066.n3.nabble.com/Can-Apache-Solr-Handle-TeraByte-Large-Data-tp3656484p3656848.html>; > this blog > post<http://harish11g.blogspot.com/2012/02/apache-solr-sharding-amazon-ec2.html>on > sharding Solr in AWS says sharding is not necessary until you have > "millions of records," but is no more specific. > > What experiences have you had with this? At what point did you find it > necessary to scale up Solr, in terms of both number of records and size of > index (whether MB, GB, etc.)? > > Thanks, > Michael Brandt