Great answer Robert.

On Fri, Mar 9, 2012 at 12:06 PM, Robert Stewart <bstewart...@gmail.com> wrote:
> Split up index into say 100 cores, and then route each search to a specific 
> core by some mod operator on the user id:
>
> core_number = userid % num_cores
>
> core_name = "core"+core_number
>
> That way each index core is relatively small (maybe 100 million docs or less).
>
>
> On Mar 9, 2012, at 2:02 PM, Glen Newton wrote:
>
>> millions of cores will not work...
>> ...yet.
>>
>> -glen
>>
>> On Fri, Mar 9, 2012 at 1:46 PM, Lan <dung....@gmail.com> wrote:
>>> Solr has no limitation on the number of cores. It's limited by your 
>>> hardware,
>>> inodes and how many files you could keep open.
>>>
>>> I think even if you went the Lucene route you would run into same hardware
>>> limits.
>>>
>>> --
>>> View this message in context: 
>>> http://lucene.472066.n3.nabble.com/Lucene-vs-Solr-design-decision-tp3813457p3813511.html
>>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
>>
>>
>> --
>> -
>> http://zzzoot.blogspot.com/
>> -
>



-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076

Reply via email to