Hello Otis,

Hi Otis,

We are using Solr to provide indexing for the full text of 5 million books
(About 4-6 terrabytes of text.)  Our index is currently around 3 terrabytes
distributed over 10 shards with about 310 GB of index per shard.  We are
using very large Solr documents (about 750MB of text or about 100,000
words/doc), and using CommonGrams to deal with stopwords/common words in
multiple languages.

I would be interested in contributing a chapter if this sounds interesting. 
More details about the project are available at: 
http://www.hathitrust.org/large_scale_search 
http://www.hathitrust.org/large_scale_search  and our blog: 
http://www.hathitrust.org/blogs/large-scale-search 
http://www.hathitrust.org/blogs/large-scale-search  (I'll be updating the
blog with details of current hardware and performance tests in the next week
or so)

Tom

Tom Burton-West
Digital Library Production Service
University of Michigan Library
-- 
View this message in context: 
http://old.nabble.com/Contributors---Solr-in-Action-Case-Studies-tp27166564p27249616.html
Sent from the Solr - User mailing list archive at Nabble.com.

Reply via email to