Hi Shawn
    Thanks a lot. It is  greatly  helpful.

2014-04-23 0:43 GMT+08:00 Shawn Heisey <s...@elyograg.org>:

> On 4/22/2014 10:02 AM, yypvsxf19870706 wrote:
>
>>      I am curious of the influences when have more than 2G docs in a
>> core.And we plan to have  5g docs/core.
>>
>>     Please give me some suggestion about how to plan num of docs in a
>> core ?
>>
>
> One Solr core contains one Lucene index.  It can't be divided further than
> that without a significant redesign.  Quick note: Although SolrCloud can
> handle five billion documents with no problem, you can't have five billion
> documents in a single shard/core.
>
> The only hard limitation in the entire system is that you can't have more
> than approximately 2 billion documents in a single Lucene index.  This is
> because a Java integer (which is a signed 32-bit number) is what gets used
> for internal Lucene document identifiers. Deleted documents count against
> that limit.  It is theoretically possible to overcome this limitation, but
> it would be a MAJOR change to Lucene, requiring major changes in Solr as
> well.
>
> The other limitations you can run into with a large SolrCloud are mostly a
> matter of configuration, system resources, and scaling to multiple servers.
>  They are not hard limitations in the software.
>
> I would never put more than about 1 billion documents in a single core.
>  For performance reasons, it would be a good idea to never exceed a few
> hundred million.  When a high query rate is required, loading only one Solr
> core per server may be a requirement.
>
> Thanks,
> Shawn
>
>

Reply via email to