to 1) yes, because -Djute.maxbuffer is going to JAVA as a start parameter.
to 2) I don't know because i never use internal zookeeper
to 3) the configs are located at solr/server/solr/configsets/
- choose one configset, make your changes and upload it to zookeeper
- when creating a new collection choose your uploaded config
- whenever you change something at your config you have to upload it to
zookeeper
I don't know which Solr version you are using, but a good starting point with
solr cloud is
http://lucene.apache.org/solr/guide/6_6/solrcloud.html
Regards
Bernd
Am 02.08.19 um 07:59 schrieb Salmaan Rashid Syed:
Hi Bernd,
Sorry for noob questions.
1) What do you mean by restart? Do you mean that I shoud issue ./bin/solr
stop -all?
And then issue these commands,
bin/solr restart -cloud -s example/cloud/node1/solr -p 8983
bin/solr restart -c -p 7574 -z localhost:9983 -s example/cloud/node2/solr
2) Where can I find solr internal Zookeeper folder for issuing this command
SERVER_JVMFLAGS="$SERVER_JVMFLAGS -Djute.maxbuffer=10000000"?
3) Where can I find schema.xml and config.xmo files for Solr Cloud Cores to
make changes in schema and configuration? Or do I have to make chages in
the directory that contains managed-schema and config.xml files with which
I initialized and created collections? And then the solr will pick them up
from there when it restarts?
Regards,
Salmaan
On Thu, Aug 1, 2019 at 5:40 PM Bernd Fehling <bernd.fehl...@uni-bielefeld.de>
wrote:
Am 01.08.19 um 13:57 schrieb Salmaan Rashid Syed:
After I make the -Djute.maxbuffer changes to Solr, deployed in
production,
Do I need to restart the solr to be able to add synonyms >1MB?
Yes, you have to restart Solr.
Or, Was this supposed to be done before putting Solr to production ever?
Can we make chages when the Solr is running in production?
It depends on your system. In my cloud with 5 shards and 3 replicas I can
take one by one offline, stop, modify and start again without problems.
Thanks.
Regards,
Salmaan
On Tue, Jul 30, 2019 at 4:53 PM Bernd Fehling <
bernd.fehl...@uni-bielefeld.de> wrote:
You have to increase the -Djute.maxbuffer for large configs.
In Solr bin/solr/solr.in.sh use e.g.
SOLR_OPTS="$SOLR_OPTS -Djute.maxbuffer=10000000"
This will increase maxbuffer for zookeeper on solr side to 10MB.
In Zookeeper zookeeper/conf/zookeeper-env.sh
SERVER_JVMFLAGS="$SERVER_JVMFLAGS -Djute.maxbuffer=10000000"
I have a >10MB Thesaurus and use 30MB for jute.maxbuffer, works perfect.
Regards
Am 30.07.19 um 13:09 schrieb Salmaan Rashid Syed:
Hi Solr Users,
I have a very big synonym file (>5MB). I am unable to start Solr in
cloud
mode as it throws an error message stating that the synonmys file is
too large. I figured out that the zookeeper doesn't take a file greater
than 1MB size.
I tried to break down my synonyms file to smaller chunks less than 1MB
each. But, I am not sure about how to include all the filenames into
the
Solr schema.
Should it be seperated by commas like synonyms = "__1_synonyms.txt,
__2_synonyms.txt, __3synonyms.txt"
Or is there a better way of doing that? Will the bigger file when
broken
down to smaller chunks will be uploaded to zookeeper as well.
Please help or please guide me to relevant documentation regarding
this.
Thank you.
Regards.
Salmaan.