You have to check your log4j.properties, usually located
server/resources/log4j.properties
There is a line about infostream logging, change it from OFF to ON.
# set to INFO to enable infostream log messages
log4j.logger.org.apache.solr.update.LoggingInfoStream=OFF
Regards
Bernd
Am 17.04.2018 u
On 4/17/2018 8:54 PM, Aristedes Maniatis wrote:
Is there any difference between using the tools supplied with Solr to
write configuration to Zookeeper or just writing directly to our
Zookeeper cluster?
We have tooling that makes it much easier to write directly to ZK
rather than having to use
On 4/17/2018 8:44 PM, Erick Erickson wrote:
The other possibility is that you have LuceneMatchVersion set to
5-something in solrconfig.xml.
It's my understanding that luceneMatchVersion does NOT affect index
format in any way, that about the only things that pay attention to this
value are a
On 4/17/2018 12:17 PM, Jay Potharaju wrote:
> After digging into the error a bit more ..I see that the error messages
> contain a call to lucenecodec54. I am using version solr 6.6.3. Any ideas
> why is lucene54 being referred here??
The 6.6 version uses index file formats that were last updated i
Is there any difference between using the tools supplied with Solr to
write configuration to Zookeeper or just writing directly to our
Zookeeper cluster?
We have tooling that makes it much easier to write directly to ZK rather
than having to use yet another tool to do it.
Thanks
Ari
Those codecs only change their number when their behavior changes IIUC.
So lucenecodec54 may be there for Lucene50StoredFieldsFormat still
exists in master/8.0
IOW this is normal.
The other possibility is that you have LuceneMatchVersion set to
5-something in solrconfig.xml.
Best,
Erick
On Tue
Terry:
Tika has a horrible problem to deal with and it's approaching a
miracle that it does so well ;)
Let's take a PDF file. Which vendor's version? From what _decade_? Did
that vendor adhere
to the spec? Every spec has gray areas so even good-faith efforts can
result in some version/vendor
beha
Hi Alexey,
First, thanks for moving the conversation to the mailing list. Discussion of
usage problems should take place here rather than in JIRA.
I locally set up Solr 7.3 similarly to you and was able to get things to work.
Problems with your setup:
1. Your update chain is missing the Log a
Hi Timothy,
As I understand it, Tika is integrated with Solr. All my indexed
documents declare that they've been parsed by tika. For the eml files
it's: |org.apache.tika.parser.mail.RFC822Parser Word docs show they
were parsed by ||org.apache.tika.parser.microsoft.ooxml.OOXMLParser PDF
files
In distributed search response writer is used twice
https://lucene.apache.org/solr/guide/7_1/distributed-requests.html
once slave node that's where response writer yields "json" content and it
upset aggregator node which is expect only javabin.
I hardly can comment on rrw, it's probably used for re
Hi Amrit,
The cdcr?action=ERRORS is returning consecutiveErrors=1 on the shards which
are not forwarding updates. Any clue does that gives?
1
1
0
bad_request
On Tue, Apr 17, 2018 at 1:22 PM, Amrit Sarkar
wrote:
> Susheel,
>
> At the time of core reload, logs must be complaining or atle
Hi,
Current solr server is 5.2 and I want to enable infoStream and updated the
solrconfig.xml.
Reload the config. But it doesn’t create any logs. Do I need to configure
anything else?
Thanks.
true
Shawn,
I first deleted the collection from the admin interface. It didn´t work.
When I deleted direct on command line it worked:
/opt/solr-6.6.2/bin/solr delete -c
Thanks for the advice on using schemaless on production. I understand the
potential problems, so I will first create schema autom
Chris,
Try to index few dummy documents and analyse if the tlogs are getting
cleared or not. Ideally on the restart, it clears everything and keeps max
2 tlog per data folder.
Amrit Sarkar
Search Engineer
Lucidworks, Inc.
415-589-9269
www.lucidworks.com
Twitter http://twitter.com/lucidworks
Linke
Hi Amrit, thanks for the reply.
I shut down all of the nodes on the source cluster after the buffer was
disabled, and there was no change to the tlogs.
On Tue, Apr 17, 2018 at 12:20 PM, Amrit Sarkar
wrote:
> Chris,
>
> After disabling the buffer on source, kind shut down all the nodes of
> sour
After digging into the error a bit more ..I see that the error messages
contain a call to lucenecodec54. I am using version solr 6.6.3. Any ideas
why is lucene54 being referred here??
Thanks
at
org.apache.solr.request.SimpleFacets.lambda$getFacetFieldCounts$0(SimpleFacets.java:809)
Did you send a commit after you sent the document?
On Tue, Apr 17, 2018 at 8:23 AM, Alexey Ponomarenko
wrote:
> Hi once more I am trying to implement named entities extraction using this
> manual
> https://lucene.apache.org/solr/7_3_0//solr-analysis-
> extras/org/apache/solr/update/processor/Ope
James,
That was it! Many, many thanks!
Gene
On Tue, Apr 17, 2018 at 8:57 AM, Dyer, James
wrote:
> (moving to solr-user@lucene.apache.org)
>
> Gene,
>
> I can reproduce your problem if I misspell the "spellcheck.dictionary"
> parameter in my query. But I see your query has "direct" which m
Hi once more I am trying to implement named entities extraction using this
manual
https://lucene.apache.org/solr/7_3_0//solr-analysis-extras/org/apache/solr/update/processor/OpenNLPExtractNamedEntitiesUpdateProcessorFactory.html
I am modified solrconfig.xml like this:
opennlp/en-ner-pe
Susheel,
At the time of core reload, logs must be complaining or atleast pointing to
some direction. Each leader of shard is responsible to spawn a threadpool
for cdcr replicator to get the data over.
Amrit Sarkar
Search Engineer
Lucidworks, Inc.
415-589-9269
www.lucidworks.com
Twitter http://twi
Thanks for the response Shawn !
In relation to this :
"I feel fairly sure that most of them are unwilling to document their
skills.
If information like that is documented, it might saddle a committer with
an obligation to work on issues affecting those areas when they may not
have the free ti
Chris,
After disabling the buffer on source, kind shut down all the nodes of
source cluster first and then start them again. The tlogs will be removed
accordingly. BTW CDCR doesn't abide by 100 numRecordsToKeep or 10 numTlogs.
Amrit Sarkar
Search Engineer
Lucidworks, Inc.
415-589-9269
www.lucidwo
Ok. My expectation was the response writer would not be used until the
final serialization of the result. If my response writer breaks the
response writer contract, exactly the way rawResponseWriter does and just
out puts a filed value how does that work? Does rawResponseWriter support
cloud mode?
Hi,
Has anyone gone thru this issue where few shard leaders are forwarding
updates to their counterpart leaders in target cluster while some of the
shards leaders are not forwarding the updates.
on Solr 6.6, 4 of the shards logs I see below entries and their
counterpart in target are getting upd
DISABLEBUFFER on source cluster would solve this problem.
On Tue, Apr 17, 2018 at 9:29 AM, Chris Troullis
wrote:
> Hi,
>
> We are attempting to use CDCR with solr 7.2.1 and are experiencing odd
> behavior with transaction logs. My understanding is that by default, solr
> will keep a maximum of 1
Hi
Has anyone seen issues with group faceting on multivalued fields in solr 6x?
Can any of the committers comment?
Thanks
Jay
> On Apr 16, 2018, at 1:44 PM, Jay Potharaju wrote:
>
> I deleted my collection and rebuilt it to check if there are any issues with
> indexing. I didn't see any error
Please post the exact results. Many a times the high cpu utilisation may be
a boon as it improves query response times
On Tue, 17 Apr 2018, 13:55 mganeshs, wrote:
> Regarding query times, we couldn't see big improvements. Both are more or
> less same.
>
> Our main worry is that, why CPU usage is
That's what should happen.
Expected mime type application/octet-stream but got application/json.
Distributed search coordinator expect to merge slave responses in javabin
format. But slave's wt indicated json.
As far as I know only javabin might be used to distributed search
underneath. Coordinat
On 4/17/2018 8:15 AM, Kojo wrote:
> I am trying schemaless mode and it seems to works very nice, and there is
> no overhead to write a custom schema for each type of collection that we
> need to index.
> However we are facing a strange problem. Once we have created a collection
> and indexed data o
I have just deleted using command line and worked as expected!
2018-04-17 11:15 GMT-03:00 Kojo :
> Hi all,
>
> I am trying schemaless mode and it seems to works very nice, and there is
> no overhead to write a custom schema for each type of collection that we
> need to index.
> However we are f
Hi all,
I am trying schemaless mode and it seems to works very nice, and there is
no overhead to write a custom schema for each type of collection that we
need to index.
However we are facing a strange problem. Once we have created a collection
and indexed data on that collection, if we need to ma
(moving to solr-user@lucene.apache.org)
Gene,
I can reproduce your problem if I misspell the "spellcheck.dictionary"
parameter in my query. But I see your query has "direct" which matches the
"name" element of one of your spellcheckers. I think the actual problem in
your case might be that y
+1 to Charlie's guidance.
And...
>60,000 documents, mostly pdfs and emails.
> However, there's a premium on precision (and recall) in searches.
Please, oh, please, no matter what you're using for content/text extraction
and/or OCR, run tika-eval[1] on the output to ensure that that you are gett
Hi,
We are attempting to use CDCR with solr 7.2.1 and are experiencing odd
behavior with transaction logs. My understanding is that by default, solr
will keep a maximum of 10 tlog files or 100 records in the tlogs. I assume
that with CDCR, the records will not be removed from the tlogs until it ha
Sure
with 1 shard 1 replica this request works fine
1. Request URL:
http://localhost:8983/solr/images/image?q=id:1
2. Request Method:
GET
3. Status Code:
200 OK
logs are clean
with 2 shards 2 replicas the same request fails and in the logs
INFO - 2018-04-17 13:20:32.052; [c
On 4/16/2018 7:32 PM, gadelkareem wrote:
I cannot complain cuz it actually worked well for me so far but..
I still do not understand if Solr already paginates the results from the
full import, why not do the same for the delta. It is almost the same query:
`select id from t where t.lastmod > ${s
On 4/17/2018 5:35 AM, Alessandro Benedetti wrote:
Apache Lucene/Solr is a big project, is there anywhere in the official
Apache Lucene/Solr website where each committer list the modules of
interest/expertise ?
No, there is no repository like that. Each committer knows what their
own expertise
Hi Erick,
I have a curiosity/suggestion regarding how to speed up pending( or
forgotten ) Jiras,
is there a way to find out the most suitable committer(s) for the task and
tag them ?
Apache Lucene/Solr is a big project, is there anywhere in the official
Apache Lucene/Solr website where each commit
Hi Christopher,
if you model your documents with a nested document approach ( like the one
you mentioned) you should be able to achieve your requirement following
this interesting blog [1] :
*" ToParentBlockJoinQuery supports several score calculation modes. For
example, a score for a parent cou
Regarding query times, we couldn't see big improvements. Both are more or
less same.
Our main worry is that, why CPU usage is so high in 6.5.1 and above ? What's
going wrong ?
Is any one else facing this sort of issue ? If yes, how to bring down the
CPU usage? Is there any settings which we need
I have a multivalued field for session timings (where i store timestamps) of
groups document. e.g. session_timings: [1526882026, 1513882026, 1533882026
]. My sorting logic is the groups should be listed sorted based on their
upcoming session time.
For example, Group A has three session_timings = [
On 16/04/2018 19:48, Terry Steichen wrote:
I have from time-to-time posted questions to this list (and received
very prompt and helpful responses). But it seems that many of you are
operating in a very different space from me. The problems (and
lessons-learned) which I encounter are often very
43 matches
Mail list logo