You might try using docvalues with the on disk option and try and let the OS manage all the memory needed for all the faceting/sorting. This would require Solr 4.2.
- Mark On Mar 21, 2013, at 2:56 AM, kobe.free.wo...@gmail.com wrote: > Hello All, > > Scenario: > > My data model consist of approx. 450 fields with different types of data. We > want to include each field for indexing as a result it will create a single > SOLR document with *450 fields*. The total of number of records in the data > set is *755K*. We will be using the features like faceting and sorting on > approx. 50 fields. > > We are planning to use SOLR 4.1. Following is the hardware configuration of > the web server that we plan to install SOLR on:- > > CPU: 2 x Dual Core (4 cores) | RAM: 12GB | Storage: 212 GB > > Questions : > > 1)What's the best approach when dealing with documents with large number of > fields. What's the drawback of having a single document with a very large > number of fields. Does SOLR support documents with large number of fields as > in my case? > > 2)Will there be any performance issue if i define all of the 450 fields for > indexing? Also if faceting is done on 50 fields with document having large > number of fields and huge number of records? > > 3)The name of the fields in the data set are quiet lengthy around 60 > characters. Will it be a problem defining fields with such a huge name in > the schema file? Is there any best practice to be followed related to naming > convention? Will big field names create problem during querying? > > Thanks! > > > > -- > View this message in context: > http://lucene.472066.n3.nabble.com/SOLR-Documents-with-large-number-of-fields-450-tp4049633.html > Sent from the Solr - User mailing list archive at Nabble.com.