Hi all!
Sometimes, in logs is this ERROR:
ERROR - 2015-09-10 11:52:19.940; org.apache.solr.common.SolrException;
org.apache.solr.common.SolrException:
org.apache.solr.client.solrj.SolrServerException: IOException occured when
talking to server at: http://x.x.x.x:8080/solr/corename_shard1_replica1
ink* two queries may return same set of documents, but
> I'm not sure about scoring/ranking.
>
> I suggest you ask to java-user@lucene mailing list as for Lucene API.
>
> Regards,
> Tomoko
>
>
>
> 2015-01-21 19:12 GMT+09:00 ku3ia <
> demesg@
> >:
>
Any ideas?
--
View this message in context:
http://lucene.472066.n3.nabble.com/MultiPhraseQuery-Rewrite-to-BooleanQuery-tp4180638p4180820.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi folks!
I have a multiphrase query, for example, from units:
Directory indexStore = newDirectory();
RandomIndexWriter writer = new RandomIndexWriter(random(), indexStore);
add("blueberry chocolate pie", writer);
add("blueberry chocolate tart", writer);
IndexReader r = writer.getReader();
Hi, folks!
We are using parent/child architecture in our project and sometimes when
using child transformer ([child]) there are an exception:
"Parent query yields document which is not matched by parents filter,
docID=..."
Examples of query are:
http://localhost/solr/core/select?fq=id:123456789&
Erick Erickson wrote
> What version of Solr/Lucene?
First of all, was Lucene\Solr v.4.6, but later it was changed to Lucene\Solr
4.8. More later to the schema was added _root_ field and child doc support.
Full data re-index on each change was not done. But not so long ago I had
made an optimize to
Hi, Erick. Thanks for you response.
I'd checked my index via check index utility, and what I'm got:
3 of 41: name=_1ouwn docCount=518333
codec=Lucene46
compound=false
numFiles=11
size (MB)=431.564
diagnostics = {timestamp=1412166850391, os=Linux,
os.version=3.2.0-68-generic, mergeFactor
Hi folks!
I'm interesting in, can delete operation destroy Solr index, if optimize
command never perform?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-index-corrupt-question-tp4166810.html
Sent from the Solr - User mailing list archive at Nabble.com.
ku3ia wrote
> I can't find an example to post document with child boosted documents
> using json update handler.
> ...
> How to set the "boost" of child documents??
No ideas? Is it possible at all?
--
View this message in context:
http://lucene.472066.n3.nabble.co
Hi, folks!
I can't find an example to post document with child boosted documents using
json update handler.
{
"add":{
"boost":2.3,
"doc":{
"id":"123",
"title":{"boost":4,"value":"simple text"},
"_childDocuments_":[
{
"id":"123_0",
"child":"te
Hi all!
The question is how many collections I can put to one alias, using SolrCloud
alias collection API
https://cwiki.apache.org/confluence/display/solr/Collections+API#CollectionsAPI-api4
to process distributed requests? Is it limited?
Thanks.
--
View this message in context:
http://lucene.
Hi all!
I'm using Solr 4.6.0. I'd created three collections and combine them to
alias via CREATEALIAS API.
I run delete request via curl, ex
curl
"http://127.0.0.1:8080/solr/all/update?stream.body=%3Cdelete%3E%3Cquery%3EText:dummy%3C/query%3E%3C/delete%3E";
where "all" is alias if three collection
Hi, Ahmet! Thanks for your reply. I understand, that it is ok. And one more
question, based on
https://issues.apache.org/jira/browse/LUCENE-4963
>>We have some TokenFilters which are only broken with specific options.
This includes:
>>StopFilter, ..., LengthFilter when enablePositionIncrements=fals
Hi all! My question is not new in maillist, but I can't found a solution. I'm
using Solr 4.6.0. My text field definition is
In SolrConfig – LUCENE_46
Via core/Documents I'm adding a document, ex, {"id":"1","title":"Need for
speed"}. Keyword "for" is in
Hi all. I'd reviewed
https://cwiki.apache.org/confluence/display/solr/Distributed+Requests
but have a question. Does SolrCloud support preference search requests, like
Elasticsearch engine?
http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/search-request-preference.html
I'm in
OK. I understand. Will review check it. Many thanks.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Flush-buffer-exceptions-tp4128593p4128801.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi all!
I'm using Solr 4.6.0 and Jetty 8. Sometimes in jetty's logs are these errors
and warnings:
ERROR - 2014-03-27 17:11:15.022; org.apache.solr.common.SolrException;
null:org.eclipse.jetty.io.EofException
at
org.eclipse.jetty.http.HttpGenerator.flushBuffer(HttpGenerator.java:914)
Hi all!
Now I have a default search field, defined as
...
In a future, I will need to search using my current field (with KStem
filte
Erick Erickson wrote
> The algorithm is only sensitive to the shard ID, you should be able to
> freely move the data to another node.
>
> BTW, perhaps the easiest way to do this would be to set up a replica
> for the shards you care about on the new hardware (assuming
> connectivity) and let Solr
Hi all!
A have a solrcloud (v. 4.6.0) cluster of 5 shards. Four of them I need to
move to another server (configs and indexes). So as I understand I need to
clear zookeeper data and after restart it will be updated. My question is,
for example, in a future I need to update a specific document, is a
autocommit and /update/?commit=true works fine. I tell about, for example, I
send 807632 docs to index to my 3 shard cluster - everything is fine, but
when I'm trying to remove them, using POST request with small number of ids,
lets say 100 per request - some docs are still on index, but seems must
Hi all!
Does SolrCloud correct delete documents? When I send many requests via POST
with small number of ids – there are some documents left in index, which not
deleted.
Thanks.
--
View this message in context:
http://lucene.472066.n3.nabble.com/SolrCloud-problems-with-delete-tp4116026.html
S
ns.
>
> You can have a look at the patch. Feel free commenting
> https://issues.apache.org/jira/browse/SOLR-5478
>
> Best,
> Manuel
>
>
> On Thu, Jan 9, 2014 at 7:33 PM, ku3ia <
> demesg@
> > wrote:
>
>> Today I setup a simple SolrCloud with tow shard
Today I setup a simple SolrCloud with tow shards. Seems the same. When I'm
debugging a distributed search I can't catch a break-point at lucene codec
file, but when I'm using faceted search everything looks fine - debugger
stops.
Can anyone help me with my question? Thanks.
--
View this message
Hi!
https://issues.apache.org/jira/browse/SOLR-3855
Description
It would be nice if Solr supported DocValues:
for ID fields (fewer disk seeks when running distributed search),
Does docValues completely done for distributed search? for ID fields?
P.S. I'm using Solr 4.6.0 not in Cloud.
--
View
Any ideas?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Poor-performance-on-distributed-search-tp3590028p4106968.html
Sent from the Solr - User mailing list archive at Nabble.com.
Yonik Seeley-2-2 wrote
> On Wed, Dec 28, 2011 at 5:47 AM, ku3ia <
> demesg@
> > wrote:
>> So, based on p.2) and on my previous researches, I conclude, that the
>> more
>> documents I want to retrieve, the slower is search and main problem is
>> the
>&g
Hi all!
My index is dynamically updated. This means, that every day I have new data,
and every day I remove unused documents from it. Approximately, I know
number of documents, which I'm indexing per day.
Today I had tested a situation. Simple imagine, there is an one collection
and two shards wi
Hi, I have near the same problems with cloud state
see
http://lucene.472066.n3.nabble.com/Replicated-zookeeper-td4018984.html
--
View this message in context:
http://lucene.472066.n3.nabble.com/Testing-Solr-Cloud-with-ZooKeeper-tp4018900p4019264.html
Sent from the Solr - User mailing list archi
Hi!
I'm trying to setup SolrCloud with replicated zookeeper, but have a problem.
I'm using Jetty 8 (not embedded), Zookeeper 3.3.6, SolrCloud 4.0 from
branch, Ubuntu 12.04 LTS.
My configs are:
Four Jetty instances running on ports 8080, 8081, 8082 and 8083
Jetty1.sh:
JAVA_OPTIONS="$JAVA_OPTIONS
Hi Tomás!!!
The first three questions are major for me. Many thanks for your response.
About number of shards and documents in it I'll try to test.
Thanks.
--
View this message in context:
http://lucene.472066.n3.nabble.com/SolrCloud-general-questions-tp4017769p4017836.html
Sent from the Solr
Hi all!
We plan to migrate from Solr 3.5 to SolrCloud 4.0. We pass some tests and I
want to conform results with you.
So, what I have on tests:
Ubuntu 12.04 LTS, Oracle JDK 7u7, Jetty 8, SolrCloud 4.0, 4 shards (4 JVM's
on the same machine on different ports [9080, 9081, 9082, 9083]), no
replicas
In some cases multi-shard architecture might significantly slow down the
search process at this index size...
By the way, how much RAM do you use?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Negative-value-in-numFound-tp3986398p3986438.html
Sent from the Solr - User maili
Hm... Have you any errors in logs? During search, during indexing?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Negative-value-in-numFound-tp3986398p3986426.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi!
Can you please show your hardware parameters, version of Solr, that you're
using and schema.xml file?
thanks.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Negative-value-in-numFound-tp3986398p3986408.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi, all!
It may be seems strange, but can you who read this post answer at some
questions. I want to understand, that maybe I want to much from my Solr, so:
1) Solr version;
2) Summary doc count;
3) Shards count (if exists);
4) rows count at query (from ... into);
5) Average queries per minute (QP
Hi. Thanks for your responses. Yesterday I tried openjdk-6-jre package from
ubuntu 10.04 LTS repos. I'll monitor the situation, but seems it works! (c)
--
View this message in context:
http://lucene.472066.n3.nabble.com/sun-java6-alternatives-for-Solr-3-5-tp3781792p3784278.html
Sent from the Solr
Hi all!
I had installed an Ubuntu 10.04 LTS. I had added a 'partner' repository to
my sources list and updated it, but I can't find a package sun-java6-*:
root@ubuntu:~# apt-cache search java6
default-jdk - Standard Java or Java compatible Development Kit
default-jre - Standard Java or Java compati
Hi, all
Thanks for your responses.
I'd tried
[NOW/DAY-30DAY+TO+NOW/DAY-1DAY-1SECOND]
and seems it works fine for me.
Thanks a lot!
--
View this message in context:
http://lucene.472066.n3.nabble.com/Date-filter-query-tp3764349p3766139.html
Sent from the Solr - User mailing list archive at Nabbl
Hi,
>>First: I am really surprised that the difference between explicit
>>Date-Values and the more friendly date-keywords is that large.
Maybe it is that I use shards. I have 11 shards, summary ~310M docs.
>>Did you made a server restart between both tests?
I tried to run these test one after a
Hi, Em, thanks for your response. But seems a have a problem.
I wrote a script, which sends a queries (curl based), with a certain delay.
I had made a dictionary of matched words. I run my script with 500ms delay
during 60 seconds. Take look at catalina logs:
INFO: [] webapp=/solr path=/select
par
Hi all!
Please advice me:
1) q=test&fq=date:[NOW-30DAY+TO+NOW]
2) q=test&fq=date:[2012-01-23T00:00:00Z+TO+2012-02-21T23:59:59Z]
3) q=test+AND+date:[NOW-30DAY+TO+NOW]
4) q=test+AND+date:[2012-01-23T00:00:00Z+TO+2012-02-21T23:59:59Z]
where date:
Which of these queries will be faster by QTime at
Hi!
I'm using Solr 3.5. I have two shards. Now I'm using default and my own
defined request handler to search by these shards:
192.168.1.1:8080/solr,192.168.1.2:8080/solr
So, urls I have:
http://192.168.1.1:8080/solr/select/?q=test&rows=0&qt=distributed:
{"responseHeader":{"
Hi all.
Due to my code review, I discovered next things:
1) as I wrote before, seems there is a low disk read speed;
2) at ~/solr-3.5/solr/core/src/java/org/apache/solr/response/XMLWriter.java
and in the same classes there is a writeDocList => writeDocs method, which
contains a cycle for of all doc
Hi!
Today I'd added loginfo to Solr here:
~/solr-3.5/solr/core/src/java/org/apache/solr/servlet/SolrDispatchFilter.java
to method
private void writeResponse(SolrQueryResponse solrRsp, ServletResponse
response,
QueryResponseWriter responseWriter,
SolrQueryRequest solr
tomas.zerolo wrote
>
> But then the results would be wrong? Suppose the documents are not evenly
> distributed (wrt the sort criterium) across all the shards. In an extreme
> case, just imagine all 2000 top-most documents are on shard 3. You would
> get
> the 500 top-most (from shard 3) and some
project2501 wrote
>
> I see what you are asking. This is an interesting question. It seems
> inefficient for Solr to apply the
> requested rows to all shards only to discard most of the results on merge.
> That would consume lots of resources not used in the final result set.
>
Yeah, like Erick
>>Uhm, either I misunderstand your question or you're doing
>>a lot of extra work for nothing
>>The whole point of sharding it exactly to collect the top N docs
>>from each shard and merge them into a single result. So if
>>you want 10 docs, just specify rows=10. Solr will query all
>>the
Hi, Erick. Thanks for your advice.
>>Here's another test. Add &debugQuery=on to your query and post the
results.
Here is for 2K rows:
0
53153
on
*,score
127.0.0.1:8080/solr/shard1,127.0.0.1:8080/solr/shard2,127.0.0.1:8080/solr/shard3,127.0.0.1:8080/solr/shard4
true
0
(mainstreaming)
2000
>>
Hi, Erick!
>>Right, are you falling afoul of the recursive shard thing? That is,
>>if you shards point back to itself. As far as I understand, your
>>shards parameter in your request handler shouldn't point back
>>to itself
No, my request handler don't points itself cause default is false.
>> OK, so your speed differences are pretty much dependent upon whether you
specify
>> rows=2000 or rows=10, right? Why do you need 2,000 rows?
Yes, big difference is 10 v. 2K records. Limit of 2K rows is setted by
manager and I can't decrease it. It is a minimum row count needed to process
data.
Hi, Erick, thanks for your reply
Yeah, you are right - document cache is default, but I tried to decrease and
increase values but I didn't get the desired result.
I tried the tests. Here are results:
>>1> try with "&rows=10"
successfully started at 19:48:34
Queries interval is: 10 queries per mi
Hi, all!
I have a problem with distributed search. I downloaded one shard from my
production. It has:
* ~29M docs
* 11 fields
* ~105M terms
* size of shard is: 13GB
On production there are near 30 the same shards. I split this shard to 4
more smaller shards, so now I have:
small shard1:
docs: 6.2M
53 matches
Mail list logo