There is an API now for dynamically loading, unloading, creating and
deleting cores.
Restarting a Solr with thousands of cores will take, I don't know, hours.

On Thu, Oct 21, 2010 at 10:44 PM, Tharindu Mathew <mcclou...@gmail.com> wrote:
> Hi Mike,
>
> I've also considered using a separate cores in a multi tenant
> application, ie a separate core for each tenant/domain. But the cores
> do not suit that purpose.
>
> If you check out documentation no real API support exists for this so
> it can be done dynamically through SolrJ. And all use cases I found,
> only had users configuring it statically and then using it. That was
> maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks.
>
> So your better off using a single index and with a user id and use a
> query filter with the user id when fetching data.
>
> On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind <rochk...@jhu.edu> wrote:
>> No, it does not seem reasonable.  Why do you think you need a seperate core
>> for every user?
>> mike anderson wrote:
>>>
>>> I'm exploring the possibility of using cores as a solution to "bookmark
>>> folders" in my solr application. This would mean I'll need tens of
>>> thousands
>>> of cores... does this seem reasonable? I have plenty of CPUs available for
>>> scaling, but I wonder about the memory overhead of adding cores (aside
>>> from
>>> needing to fit the new index in memory).
>>>
>>> Thoughts?
>>>
>>> -mike
>>>
>>>
>>
>
>
>
> --
> Regards,
>
> Tharindu
>



-- 
Lance Norskog
goks...@gmail.com

Reply via email to