Hi,

SolrCloud is designed with an assumption that you should be able to upload your 
whole disk-based conf folder into ZK, and that you should be able to add an 
empty Solr node to a cluster and it would download all config from ZK. So 
immediately a splitting strategy automatically handled by ZkSolresourceLoader 
for large files could be one way forward, i.e. store synonyms.txt as e.g. 
__001_synonyms.txt __002_synonyms.txt....

Feel free to open a JIRA issue for this so we can get a proper resolution.

--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com

7. mai 2013 kl. 09:55 skrev Roman Chyla <roman.ch...@gmail.com>:

> We have synonym files bigger than 5MB so even with compression that would
> be probably failing (not using solr cloud yet)
> Roman
> On 6 May 2013 23:09, "David Parks" <davidpark...@yahoo.com> wrote:
> 
>> Wouldn't it make more sense to only store a pointer to a synonyms file in
>> zookeeper? Maybe just make the synonyms file accessible via http so other
>> boxes can copy it if needed? Zookeeper was never meant for storing
>> significant amounts of data.
>> 
>> 
>> -----Original Message-----
>> From: Jan Høydahl [mailto:jan....@cominvent.com]
>> Sent: Tuesday, May 07, 2013 4:35 AM
>> To: solr-user@lucene.apache.org
>> Subject: Re: Solr Cloud with large synonyms.txt
>> 
>> See discussion here
>> http://lucene.472066.n3.nabble.com/gt-1MB-file-to-Zookeeper-td3958614.html
>> 
>> One idea was compression. Perhaps if we add gzip support to SynonymFilter
>> it
>> can read synonyms.txt.gz which would then fit larger raw dicts?
>> 
>> --
>> Jan Høydahl, search solution architect
>> Cominvent AS - www.cominvent.com
>> 
>> 6. mai 2013 kl. 18:32 skrev Son Nguyen <s...@trancorp.com>:
>> 
>>> Hello,
>>> 
>>> I'm building a Solr Cloud (version 4.1.0) with 2 shards and a Zookeeper
>> (the Zookeeer is on different machine, version 3.4.5).
>>> I've tried to start with a 1.7MB synonyms.txt, but got a
>> "ConnectionLossException":
>>> Caused by: org.apache.zookeeper.KeeperException$ConnectionLossException:
>> KeeperErrorCode = ConnectionLoss for /configs/solr1/synonyms.txt
>>>       at
>> org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
>>>       at
>> org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
>>>       at org.apache.zookeeper.ZooKeeper.setData(ZooKeeper.java:1266)
>>>       at
>> org.apache.solr.common.cloud.SolrZkClient$8.execute(SolrZkClient.java:270)
>>>       at
>> org.apache.solr.common.cloud.SolrZkClient$8.execute(SolrZkClient.java:267)
>>>       at
>> 
>> org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java
>> :65)
>>>       at
>> org.apache.solr.common.cloud.SolrZkClient.setData(SolrZkClient.java:267)
>>>       at
>> org.apache.solr.common.cloud.SolrZkClient.makePath(SolrZkClient.java:436)
>>>       at
>> org.apache.solr.common.cloud.SolrZkClient.makePath(SolrZkClient.java:315)
>>>       at
>> org.apache.solr.cloud.ZkController.uploadToZK(ZkController.java:1135)
>>>       at
>> org.apache.solr.cloud.ZkController.uploadConfigDir(ZkController.java:955)
>>>       at
>> org.apache.solr.core.CoreContainer.initZooKeeper(CoreContainer.java:285)
>>>       ... 43 more
>>> 
>>> I did some researches on internet and found out that because Zookeeper
>> znode size limit is 1MB. I tried to increase the system property
>> "jute.maxbuffer" but it won't work.
>>> Does anyone have experience of dealing with it?
>>> 
>>> Thanks,
>>> Son
>> 
>> 

Reply via email to