Thanks Guys. Very very helpful.
I will probably look at consolidate 4 Solr servers into 2 bigger/better server
- it gives more memory, and it cut down the replica the Leader needs to manage.
Also, I may look into write a script to monitor the tomcat log and if there is
OOM, kill tomcat, then re
Hello,
I have two collections and need to join the results on uniqueIds.
I am able to do that with Streaming Expressions- LeftOuterJoin. Is there any
way to use facets along with this?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-6-Use-facet-with-Streaming-Express
Hello,
I am looking for similar use case. Will it be possible for you to share the
corrected syntax ?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Streaming-expressions-malfunctioning-tp4281016p4290528.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi Bastien,
Have you tried upgrading to 6.1? SOLR-8812, mentioned earlier in the thread,
was released with 6.1, and is directly aimed at fixing the problem you are
having in 6.0 (also a problem in 5.5): when mm is not explicitly provided and
the query contains explicit operators (except for AN
Hi Steve,
I read the thread you sent me (SOLR-8812) and it seems that the 6.1
includes this fix, as you said.
I will upgrade.
Thank you!
Kind regards,
Bast
On 05/08/2016 14:37, Steve Rowe wrote:
Hi Bastien,
Have you tried upgrading to 6.1? SOLR-8812, mentioned earlier in the thread,
was r
If you need to aggregate after the join then you'll need to use the
rollup() function.
The rollup function requires the tuples be sorted by the group by fields.
So it's easiest to accomplish this using the hashOuterJoin, which doesn't
require a sort on the join keys.
If you're doing a parallel jo
I don't see any field level data exposed in the SolrDocumentList I get from
shardResponse.getSolrResponse().getResponse().get("response"). I see the
unique ID field and value. Is that by design or am I being stupid?
Separate but related question: the mergIds() method in the merge strategy
class -
On 8/3/2016 12:49 AM, liubiaoxin1 wrote:
> set exery core solrconfig.xml: 4096
Are you absolutely certain that you have set maxBooleanClauses on
*EVERY* core in that Solr instance?
This value is global, across the entire JVM, and the last core that
starts will set the value for all cores.
It is
Another alternative is to use the TermsQueryParser which is intended to
deal with very large lists of values that should be ORed together. It may
be useful if your query pattern matches its intent. see:
https://cwiki.apache.org/confluence/display/solr/Other+Parsers#OtherParsers-TermsQueryParser
Be
You don't really have to worry that much about memory consumed during indexing.
The ramBufferSizeMB setting in solrconfig.xml pretty much limits the amount of
RAM consumed, when adding a doc if that limit is exceeded then the
buffer is flushed.
So you can reduce that number, but it's default is 10
Many thanks for the assistance Hoss! After a couple of bumps, it worked
great.
I followed the recommendations (and read the explanation - thanks!)
Although I swear it threw the error once again, just to be sure I rebooted
everything (Zookeeper included) then reloaded the configs into Zooke
CLASSIFICATION: UNCLASSIFIED
I am trying to index from nutch 1.12 to SOLR 6.1.0.
Got this error.
java.lang.Exception:
org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error
from server at http://localhost:8983/solr/ARLInside: Exception writing document
id
https://emcstage.
what that error is telling you is that you have an unanalyzed term
that is, well, huge (i..e > 32K). Is your "content" field by chance a
"string" type? It's very rare that a term > 32K is actually useful.
You can't search on it except with, say, wildcards,there's no stemming
etc. So the first quest
CLASSIFICATION: UNCLASSIFIED
What I did was force nutch to truncate content to 32765 max before indexing
into solr and it solved my problem.
Thanks,
Kris
~~
Kris T. Musshorn
FileMaker Developer - Contractor – Catapult Technology Inc.
US Army Research Lab
Aberdeen
You also need to find out _why_ you're trying to index such huge
tokens, they indicate that something you're ingesting isn't
reasonable
Just truncating the input will index things, true. But a 32K token is
unexpected, and indicates what's in your index may not be what you
expect and may not be
15 matches
Mail list logo