Hi, Let me first roughly describe the scenario :-) We're trying to index online stored data for some thousand users. The schema.xml has a custom identifier for the user, so FQ can be applied and further filtering is only done for the user (more important, the user doesn't get to see results from data not belonging to him)
Unfortunatelly, the Index might become quite big ( we're indexing more that 50 TB Data, all kind of files, full text (indexed only, not stored) where possible, elsewhere fileinfos (size, date) and meta if available) So Question the is: We're thinking of starting out with multiple Solr instances (either in their own containers or MultiCore, guess that's not the important point), on 1 to n machines. Lets just pretend: we do modulo 5 on the user number and assign it to one of the two machines. The index gets distributed on QuerySlaves ( 1-m dependend on the need). So now the Question: Is there a way to split a too big index into smaller ones? Do I have to create more instances at the beginning, so that I will not run out of power and space? (which will ad quite a bit of redundance of data) Lets say I miscalculated and used only 2 indices, but now I see I need at least 4. Any idea will be very welcome, Thanks, Nico