I believe I figured out what the issue is. Even though we do not open a new
searcher on master during full import, the slave anyway replicates the
index after auto commits! (Is this desired behavior?) Since "clean=true"
this meant all the docs were deleted on slave and a partial index got
replicate
Found and here,
may be it's because my filter query is changing for each new user. Better i
keep it in main query
http://lucene.472066.n3.nabble.com/fq-vs-q-td495570.html
--
View this message in context:
http://lucene.472066.n3.nabble.com/Adding-filter-query-slows-down-avg-response-time-tp41205
On 1 March 2014 09:39, nutchsolruser wrote:
> You may think this is silly question but let me ask this because i am
> confused ,
> http://www.lucidworks.com/webinar-solr-4-the-nosql-search-server/ this says
> Solr is NoSQL but many other links dont have solr in their list as NoSQL
> database.
>
>
Well, I'd certainly try removing parts of the query to see
what was actually in the index.
I don't see anything obvious though...
Erick
On Fri, Feb 28, 2014 at 8:06 PM, Arun Rangarajan
wrote:
> Thx, Erick and Chris.
>
> This is indeed very strange. Other queries which do not restrict by the
>
I am finding users with same nick name in certain area. when i send both
queries in q parameter to solr it works really fast . but if i send location
query in fq then it slows down too much. why it is so? why adding fq to
query degrades my performance?
nickname:"nick name"
{!geofilt pt=20.2284,80.
You may think this is silly question but let me ask this because i am
confused ,
http://www.lucidworks.com/webinar-solr-4-the-nosql-search-server/ this says
Solr is NoSQL but many other links dont have solr in their list as NoSQL
database.
http://en.wikipedia.org/wiki/NoSQL
http://en.wikipedia.o
Thx, Erick and Chris.
This is indeed very strange. Other queries which do not restrict by the
date field are returning results, so the index is definitely not empty. Has
it got something to do with the date query part, with NOW/DAY or something
in here?
first_publish_date:[NOW/DAY-33DAYS TO NOW/DA
: This is odd. The full import, I think, deletes the
: docs in the index when it starts.
Yeah, if you are doing a full-import everyday, and you don't want it to
delete all docs when it starts, you need to specify "clearn=false"
https://cwiki.apache.org/confluence/display/solr/Uploading+Structur
Any pointer in this will be helpful, is there a way to avoid using group by
queries and achieve similar results
or way to enable caching for group by queries
--
View this message in context:
http://lucene.472066.n3.nabble.com/Group-query-not-cached-in-SOLR-tp4120159p4120547.html
Sent from the
This is odd. The full import, I think, deletes the
docs in the index when it starts.
If you check our index directory on the slave, is it empty
after the full import starts? If so, check your solr log
on the slave... does it show a replication?
Shooting in the dark...
Erick
On Fri, Feb 28, 201
That really shouldn't be happening IF indexing is shut off. Otherwise
the slave is taking a snapshot of the master index and synching.
bq: The slave has about 33 more documents and one fewer
segements (according to Overview in solr admin
Sounds like the master is still indexing and you've deleted
Yep. One alternative is something I just found out about;
ParseDateFieldUpdateProcessorFactory
Best,
Erick
On Thu, Feb 27, 2014 at 3:12 PM, solr2020 wrote:
> Hi,
>
> We are using 'solr.TrieDateField' type for indexing a date column in Solr.
> By default triedate will index date columns as U
Hi Furkan
I have read that before but I haven't added any new classes or changed
anything with my setup. I just created more collections in solr. How will
that increase perm gen space ? Doesn't solr intern strings at all ?
Interned strings also go to the perm gen space right?
- Nitin
On Fri, F
Thanks, Jack.
>
How is first_publish_date defined?
with "date" being
Yes, we need to fix the Boolean operators AND, OR and NOT as mentioned in
http://searchhub.org/2011/12/28/why-not-and-or-and-not/ but I believe that
is not an issue here, because the same query returns results few mins
bef
The Solr function query documentation (
https://wiki.apache.org/solr/FunctionQuery#exists) says:
exists(query({!v='year:2012'})) will return true for docs with year=2012
I have a document like:
{
id: 1,
user_type: ADMIN,
like_score: 1
}
id, user_type and like_score are all indexed and stor
Hi,
We have name searches on Solr for millions of documents. User may search like
"Morrison Dave" or other may search like "Morrison David". What's the best way
to handle that both brings similar results. Adding Synonym is the option we are
using right.
But we may need to add around such 50,0
Hi;
You can also check here:
http://stackoverflow.com/questions/3717937/cmspermgensweepingenabled-vs-cmsclassunloadingenabled
Thanks;
Furkan KAMACI
2014-02-26 22:35 GMT+02:00 Josh :
> Thanks Timothy,
>
> I gave these a try and -XX:+CMSPermGenSweepingEnabled seemed to cause the
> error to happe
Hi;
Jack has an answer for a PermGen usages:
"PermGen memory has to do with number of classes loaded, rather than
documents.
Here are a couple of pages that help explain Java PermGen issues. The
bottom
line is that you can increase the PermGen space, or enable unloading of
classes, or at least t
Yes my indexer runs as a service on a different box, it has 24 threads pushing
docs to solr atomically. No the solr master is not virtual, it has 64 GB main
memory and dual quad xeon cpus. The cpu utilization is not maxed out from what
I can see in 'top'. Right now it says 38%. The other thi
Is it indexing data from over the network? (high data throughput would
increase latency) Is it a virtual machine? (Other machines causing slow
downs) Another possible option is the network card is offloading processing
onto the CPU which is introducing latency when the CPU is under load.
On Fri,
Hi guys,
Got an odd thing going on right now. Indexing into my master server (solr
3.6.1) has slowed and it is because when solr runs ping shows latency. When I
stop solr though, ping returns to normal. This has been happening
occasionally, rebooting didn't help. This is the first time I no
Hello,
We have recently moved to Solr cloud in our application, but we still do have
single solr instance which we use for testing purposes.
We already had explain plan working in single instance, now after moving to
solr cloud it does not show any explanation field in its response.
Thsi is how
Hi Chris,
The enum option is working for us, with suitable minDf settings. We are
able to do faceting with decent speed using this.
Thanks a lot,
Dave
On Fri, Feb 28, 2014 at 9:09 AM, David Miller wrote:
> Hi Chris,
>
> Thanks for the info. I have looked into the "docValues" option earlier.
>
Hi
I am seeing the Perm Gen usage increase as i keep adding more collections.
What kind of strings get interned in solr? (Only schema , fields,
collection metadata or the data itself?)
Will Permgen space (atleast interned strings) increase proportional to the
size of the data in the collections
Hi,
I'm using Solr 4.5.0, I have a single master replicating to a single slave.
Only the master is being indexed to - never the slave. The master is committed
once each night. After the first commit and replication the numDoc counts are
identical. After the next nightly commit and after the
We have created some scripts that can do this for you - basically
reconstruct (by looking at information in ZK) solr.xml, core.properties
etc on the new machine as they where on the machine that crashed. Our
procedure when a machine crashes is
* Remove it from rack, replace it by a similar mach
Look at the parsed_query by setting the debugQuery=true parameter.
I think what is happening is that the query parser will generate a separate
dismax query for each term and each dismax query will require at least one
of its fields to contain the term. I suspect that some of your qf fields do
Thanks for taking time on this...
Here is my request handler definition:
edismax
explicit
10
all_text number party name all_code ent_name
all_text number^3 name^5 party^3 all_code^2
ent_name^7
id description
AND
Name which is indexed
Hi,
From the URLs you provided, it is not clear that you use edismax query parser
at all. Thats why I asked complete list of parameters. Can you paste request
handler definition from solrconfig.xml?
And what do you expect and what is not working for you.
On Friday, February 28, 2014 7:30
explicit
For all handlers I have the same setting.
Another observation I have is,
I'm getting results when I use, 'q.op=OR' the default operator set in
solrconfig.xml is 'AND'
the query working fine is:
http://localhost:8080/solr/collection1/select?q=bank+america&wt=json&indent=true&q.op=OR
How is first_publish_date defined?
After queries start failing, do an explicit query of some of the document
IDs that you think should be present and see what the first_publish_date
field contains.
Also, Solr and Lucene queries are not strict Boolean, so ANDing of a purely
negative term requ
All,
Just playing around with the SuggestComponent, trying to compare results
with the old-style spell-check-based suggester. Tried this config
against a string field:
json
true
true
10
default
suggest2
Solr server version 4.2.1
I am facing a strange issue with a date query like this:
q=first_publish_date:[NOW/DAY-33DAYS TO NOW/DAY-3DAYS] AND
-tag_id:268702&fq=(burial_score:[* TO 0.49] AND
-tag_id:286006)&rows=1&sort=random_906313237 asc&fl=id
The only process by which we add documents to the c
Can give parameters defined in defaults sections of request handler /
solrconfig.xml?
By the way echoParams=all will list all parameters.
On Friday, February 28, 2014 5:18 PM, sureshrk19 wrote:
Ahmet,
Thanks for the reply..
Here is the query:
http://localhost:8080/solr/collection1/select?q
Hi Jan,
There are a few ways to do that, but no, nothing is automatic.
1) If your node is alive, you can create new replicas on the new node,
let them replicate, verify they are ok, then delete the replicas on the
old node and shut it down.
2) If your node is dead, create new replicas on the
Hi Jan,
There are a few ways to do that, but no, nothing is automatic.
1) If your node is alive, you can create new replicas on the new node,
let them replicate, verify they are ok, then delete the replicas on the
old node and shut it down.
2) If your node is dead, create new replicas on the
I’m pretty sure the default config will unlock on startup.
- Mark
http://about.me/markrmiller
On Feb 28, 2014, at 3:50 AM, Chen Lion wrote:
> Dear all,
> I hava a problem i can't understand it.
>
> I use solr 4.6.1, and 2 nodes, one leader and one follower, both have the
> write.lock file.
>
Ahmet,
thanks for this, but I do not think this actually meets my requirements. My
intent is not that of harcoding the facet field and value I want to exclude,
but to be able to apply the exclusion variant regardless (I currently have 3
facet field and ~ 5 million of records). Before posting my que
Thanks.
We find little evidence that page/disk cache is causing this issue. We use sar
to collect statistics. Here is the statistics on a node where the query took
maximum time. (out of 5 shards, one with most data takes long time) However,
we're reducing heap size and testing in QA.
Ahmet,
Thanks for the reply..
Here is the query:
http://localhost:8080/solr/collection1/select?q=a+of+b&fq=type%3AEntity&wt=json&indent=true
And here is my stopwords_en.txt content
a
an
and
are
as
at
be
but
by
for
if
in
into
is
it
no
not
of
on
or
--
View this message in context:
http://l
Hi,
I am a bit confused about how solr cloud disaster recovery is supposed
to work exactly in the case of loosing a single node completely.
Say I have a solr cloud cluster with 3 nodes. My collection is created
with numShards=3&replicationFactor=3&maxShardsPerNode=3, so there is
no data loss when
Hi Chris,
Thanks for the info. I have looked into the "docValues" option earlier. But
docValues doesn't support textField and we require textField to enable
various tokenizer and analyzers (like shingle, pattern filter etc.) We
require the faceting to be on terms with in the text field, not as a w
Hi,
Let's say you have accomplished what you want. You have a .txt with the tokens
tomerge, like "European" and "Parliament". What is your use case then? What is
your high level goal?
MappingCharFilter approach is closer (to your .txt approach) than
PatternReplaceCharFilterFactory approach.
ups I sent it prematurely. Here is the correct one :
solrQuery.add(CommonParams.FQ, "-{!term f=author}Dickens, Janet");
On Friday, February 28, 2014 12:42 PM, Ahmet Arslan wrote:
Hi,
This should do the trick : solrQuery.add(CommonParams.FQ, "fq=-{!term
f=author}Dickens, Janet");
Ahmet
Hi,
This should do the trick : solrQuery.add(CommonParams.FQ, "fq=-{!term
f=author}Dickens, Janet");
Ahmet
On Friday, February 28, 2014 11:21 AM, idioma wrote:
Ahmet,
thank you for your reply, much appreciated. Let me answer to your
question(s):
1) "Your example response () looks like custo
Have you tried to just use a copyField? For example, I had a similar use
case where I needed to have particular field (f1) tokenized but also
needed to facet on the complete contents.
For that, I created a copyField
f1 used tokenizers and filters but f2 was just a plain string. You then
Hi Ahmet!!
I went ahead and did something I thought it was not a clean solution and
then when I read your post and I found we thought of the same solution,
including the European_Parliament with the _ :)
So I guess there would be no way to do this more cleanly, maybe only
implementing my own Tok
Ahmet,
thank you for your reply, much appreciated. Let me answer to your
question(s):
1) "Your example response () looks like customized."
It is not, but I have not included the code that generates it. In a
nutshell, I have two XSLT files, one that transforms the Solr query into
something that my
Dear all,
I hava a problem i can't understand it.
I use solr 4.6.1, and 2 nodes, one leader and one follower, both have the
write.lock file.
I did not think i could create index since the write.lock file exists,
right?
But I could, why?
Jiahui Chen
Hi Suresh,
Can you give us full set of parameters you use for edismax? qf, mm, etc.
And content of your stopwords.txt. Is a listed there too?
Ahmet
On Friday, February 28, 2014 8:54 AM, sureshrk19 wrote:
Hi All,
I'm having a problem while searching for some string with a word defined in
stop
50 matches
Mail list logo