Thanks Jack . Yes, this seems so !!
However I would like to fix this at code level by setting the noCFSRatio to
1.0 . But in solr 3.6.1 i am not able to find the build.xml file .
I suppose the build process has been changed since 1.3 ,can you throw some
light on how I can build source code post th
: Is it possible to let user's define their position in search when location
: is queried? Let's say that I am UserA and when you make a search with
: Moscow, my default ranking is 258. By clicking a button, something like
: "Boost Me!", I would like to see UserA as the first user when search is d
: We are trying to run solr on https, these are few of the issues or
: problems that are coming up. Just wanted to understand if anyone else is
: facing these problems,
There are currently some known issues using SolrCloud with https, Sami is
working on it...
https://issues.apache.org/jira/br
You may simply be encountering the situation where the merge size is greater
than 10% of the index size, as per this comment in the code:
/** If a merged segment will be more than this percentage
* of the total size of the index, leave the segment as
* non-compound file even if compound file i
Here is another way, without using synonyms: in data preparation, you can
create a new token 'bats_sports' for all common words in different categories.
You can do this in a separate field that you do not store, just index. Now, if
you search with a category you would send in 'bats bats_sports'
You don't have to worry about background merges when optimizing, it
won't error out.
Optimize is a little heavy handed though.
You might just use expunge deletes and/or try a low merge factor -
though with the latest tiered merge policy, I think you have to use a
different knob than merge factor
On Wed, Sep 19, 2012 at 5:27 PM, Yonik Seeley wrote:
> Folks,
>
> Some changes have been committed in the past few days related to
> SOLR-3815 as part of the groundwork
> for SOLR-3755 (shard splitting).
>
> The resulting clusterstate.json now looks like the following:
>
> {"collection1":{
> "
Also note that some of the Solr request parameters are lists of fields where
space is the delimiter and can NOT be escaped.
For example, the "fl" parameter uses both comma and space as delimiters, and
the e/dismax field list parameters use space as the field delimiter.
-- Jack Krupansky
Logs are dependent on the servlet contianer you use -- ie: with the
solr example, log messages are written to the console were you run
"start.jar" and can be configured to point elsewhere based on how you
configure jetty.
in your case it looks like you are using tomcat, so you'll want to check
Logs are dependent on the servlet contianer you use -- ie: with the
solr example, log messages are written to the console were you run
"start.jar" and can be configured to point elsewhere based on how you
configure jetty.
in your case it looks like you are using tomcat, so you'll want to check
Hi,
I have been making changes to my schema and unfortauntly I now get a error
when viewing my Solr tomcat admin page.
The error doesnt seem to explain the problem:
HTTP Status 500 - Severe errors in solr configuration. Check your log files
for more detailed information on what may be wrong. If y
Folks,
Some changes have been committed in the past few days related to
SOLR-3815 as part of the groundwork
for SOLR-3755 (shard splitting).
The resulting clusterstate.json now looks like the following:
{"collection1":{
"shard1":{
"range":"8000-",
"replicas":{"Rogue:8
the latter: the document (eg product) has a category, and the synonyms
would be applied at index time. sports-related "bat" synonyms to
baseball "bats", and halloween-themed "bat" synonyms to scary "bats",
for example.
On 09/19/2012 05:08 PM, Erick Erickson wrote:
Not that I know of, syno
well, I've certainly been wrong before, so it may not be so bad. Time
will tell...
Erick
On Wed, Sep 19, 2012 at 5:08 PM, Aleksey Vorona wrote:
> Thank you for that insight. I, myself, would've liked to remove the spaces,
> but it is not possible in that particular project.
>
> I see that I need
Not that I know of, synonyms are an all-or-nothing on a field.
But how would you indicate the context at index time as opposed to
query time? Especially at query time, there's very little in the way of
context to figure out what the category was.
Or were you thinking that the document had a categ
Thank you for that insight. I, myself, would've liked to remove the
spaces, but it is not possible in that particular project.
I see that I need to learn more about Lucene. Hopefully that will help
me avoid some of those headaches to come.
-- Aleksey
On 12-09-19 11:42 AM, Erick Erickson wrot
The mm parameter, as I understand it, doesn't really play with phrase.
So you're looking for the phrase
"this amazing sample query", "amazing" must be in the phrase.
and phrase slop reorders things, counting the reordering as "slop", so
the approach would not
do what you want anyway, i.e.
"this am
> already-optimized, single-segment index
That part is interesting... if true, then the type of "insanity" you
saw should be impossible, and either the insanity detection or
something else is broken.
-Yonik
http://lucidworks.com
Some function queries also use the field cache. I *think* those usually use
the segment level cache, but I'm not sure.
On Wed, Sep 19, 2012 at 4:36 PM, Yonik Seeley wrote:
> The other thing to realize is that it's only "insanity" if it's
> unexpected or not-by-design (so the term is rather mis-n
On Wed, Sep 19, 2012 at 4:25 PM, Mark Miller wrote:
> bq. I believe there were some changes made to the clusterstate.json
> recently that are not backwards compatible.
>
> Indeed - I think yonik committed something the other day - we prob
> should send an email out about this.
Yeah, I was just in
Hi Guys
The solr server which was running fine for last few months, is now failing
during startup wuth following error
Sep 19, 2012 12:53:25 PM org.apache.solr.common.SolrException log
SEVERE: java.lang.RuntimeException: java.io.FileNotFoundException:
/var/lib/solr/default/index/.cfs (No such fil
bq. I believe there were some changes made to the clusterstate.json
recently that are not backwards compatible.
Indeed - I think yonik committed something the other day - we prob
should send an email out about this. Not sure exactly how easy an
upgrade is or what steps to take - it may be somethin
LotsOfCores ( http://wiki.apache.org/solr/LotsOfCores ) is intended to
dynamically juggle loading (and unloading ) required cores where the total
number of cores is very large.
We're approaching that situation, but it looks like LotsOfCores isn't quite
ready for prime time yet. Are there any othe
The other thing to realize is that it's only "insanity" if it's
unexpected or not-by-design (so the term is rather mis-named).
It's more for core developers - if you are just using Solr without
custom plugins, don't worry about it.
-Yonik
http://lucidworks.com
On Wed, Sep 19, 2012 at 3:27 PM, To
Hi Tomás,
> This probably means that you are using the same field for faceting and for
> sorting (tf_normalizedTotalHotttnesss), sorting uses the segment level
> cache and faceting uses by default the global field cache. This can be a
> problem because the field is duplicated in cache, and then it
Hi Aaron, here there is some information about the "insanity count":
http://wiki.apache.org/solr/SolrCaching#The_Lucene_FieldCache
As for the SUBREADER type, the javadocs say:
"Indicates an overlap in cache usage on a given field in sub/super readers."
This probably means that you are using the s
Is there an existing TokenFilterFactory that can conditionally insert
synonyms based on a given document attribute, say category? Some
synonyms only make sense in context: "bats" in Sports is different from
"bats" in "Party and Novelty".
It seems the synonyms.txt file would need an additional
> SEVERE: Full Import failed:java.lang.RuntimeException:
> java.lang.RuntimeException:
> org.apache.solr.handler.dataimport.DataImportHandlerException:
> Could not
> load driver: com.mysql.jdbc.Driver Processing Document # 1
You need to put mysql-connector-java-5.1.*.jar to lib folder.
Hi,
i've set up and solr4 application and wanted to do an dataimport via DB. So
I'm using the DIH for the solr4 beta. If I go to the datahandler in the
admin/ panel, and click on import, I get the following messages in the log:
SEVERE: Full Import failed:java.lang.RuntimeException:
java.la
1) so how does this look like for example?
2) without curl? how does this look like? i am very confused because they use
curl in the example but say at the same time that we should not use curl. also
i have not installed curl
Original-Nachricht
> Datum: Wed, 19 Sep 2012 11:47:54
> But I wonder what "text_ws" means?
it means text whitespace you can find its definition in schema.xml. search for
'text_ws' It uses whitespacetokenizer.
Thanks, it worked after editing my schema line to:
But I wonder what "text_ws" means?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-Keep-Punctuation-in-Parsed-Phrase-Query-tp4008977p4008984.html
Sent from the Solr - User mailing list archive at Nabble.com.
: user:~/solr/example/exampledocs$ java -jar post.jar test.pdf doesnt work
1) you can use post.jar to send PDFs, but you have to use the option to
tell solr you are sending a PDF file - because by default it assumes you
are posting XML. you can see the problem by looking at the output from
po
BANG!
On Wed, Sep 19, 2012 at 2:38 PM, vempap wrote:
> Shoot me. Thanks, I did not notice that the doc has ".. e a .." in the
> content. Thanks again for immediate reply :)
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/SpanNearQuery-distance-issue-tp4008973p4008
I would _really_ recommend that you re-do your schema and
take spaces out of your field names. That may require that
you change your indexing program to not send spaces in dynamic
field names
This is the kind of thing that causes endless headaches as time
goes forward.
You don't _have_ to, bu
So I have to create a java file and compile it just for this purpose? like
http://wiki.apache.org/solr/ContentStreamUpdateRequestExample?
No way to do this via post.jar (and without curl? or an other already existing
implementation via command line ...) also there is no way mentioned how it can
> I am working with apache-solr-3.6.0 on windows machine. I
> would like to
> search for phrases which contain punctuation marks.
> Example:
>
> "He said: Hi"
> I tried to escape the punctuation marks using \ so my url
> was:
>
> http://localhost:8983/solr/select/?q="He%20said\:%20Hi"&version=2.2
Shoot me. Thanks, I did not notice that the doc has ".. e a .." in the
content. Thanks again for immediate reply :)
--
View this message in context:
http://lucene.472066.n3.nabble.com/SpanNearQuery-distance-issue-tp4008973p4008978.html
Sent from the Solr - User mailing list archive at Nabble.co
Hi;
I am working with apache-solr-3.6.0 on windows machine. I would like to
search for phrases which contain punctuation marks. Example:
"He said: Hi"
I tried to escape the punctuation marks using \ so my url was:
http://localhost:8983/solr/select/?q="He%20said\:%20Hi"&version=2.2&start=0&rows=1
> I've a issue with respect to the distance measure of
> SpanNearQuery in
> Lucene. Let's say I've following two documents:
>
> DocID: 6, cotent:"1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17
> 18 19 20 1001
> 1002 1003 1004 1005 1006 1007 1008 1009 1100",
> DocID: 7, content:"a b c d e a b c f g h
Hello All,
I've a issue with respect to the distance measure of SpanNearQuery in
Lucene. Let's say I've following two documents:
DocID: 6, cotent:"1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 1001
1002 1003 1004 1005 1006 1007 1008 1009 1100",
DocID: 7, content:"a b c d e a b c f g h i j k
> user:~/solr/example/exampledocs$ java
> -jar post.jar test.pdf doesnt work
>
> Index binary documents such as Word and PDF with Solr Cell
> (ExtractingRequestHandler).
> how do i do his?
>
> http://lucene.apache.org/solr/api-4_0_0-BETA/doc-files/tutorial.html
>
>
> http://wiki.apache.org/sol
Hi all,
In reviewing a solr instance with somewhat variable performance, I
noticed that its fieldCache stats show an insanity_count of 1 with the
insanity type SUBREADER:
---snip---
insanity_count : 1
insanity#0 : SUBREADER: Found caches for descendants of
ReadOnlyDirectoryReader(segments_k
_6h9(
On 12-09-19 11:04 AM, Ahmet Arslan wrote:
I have a field with space in its name (that is a dynamic
field). How can I execute search on it?
I tried "q=aattr_box%20%type_sc:super" and it did not work
The field name is "aattr_box type"
How about q=aattr_box\ type_sc:super
That works! Thank you!
> I have a field with space in its name (that is a dynamic
> field). How can I execute search on it?
>
> I tried "q=aattr_box%20%type_sc:super" and it did not work
>
> The field name is "aattr_box type"
How about q=aattr_box\ type_sc:super
Hi,
I have a field with space in its name (that is a dynamic field). How can
I execute search on it?
I tried "q=aattr_box%20%type_sc:super" and it did not work
The field name is "aattr_box type"
-- Aleksey
user:~/solr/example/exampledocs$ java -jar post.jar test.pdf doesnt work
Index binary documents such as Word and PDF with Solr Cell
(ExtractingRequestHandler).
how do i do his?
http://lucene.apache.org/solr/api-4_0_0-BETA/doc-files/tutorial.html
http://wiki.apache.org/solr/ExtractingRequestHa
> So I want to enable highlighting on my results. When I run
> the query like
> this:
>
> http://localhost:8080/solr/select?q=book&hl=true
Try explicitly setting the field(s) that you want to highlight.
To enable highlighting your field must me stored="true"
See : http://wiki.apache.org/solr/Fi
Hi,
So I want to enable highlighting on my results. When I run the query like
this:
http://localhost:8080/solr/select?q=book&hl=true
I don't get any highlighted results. I am assuming that more is needed to
actually enable highlighting. Commented out at the bottom of my
solrconfig.xml is this:
I have used this setting to reduce gc pauses with CMS - java 6 u23
XX:+ParallelRefProcEnabled
With this setting, jvm does gc of weakrefs with multiple threads and pauses are
low.
Please use this option only when you have multiple cores.
For me, CMS gives better results
Sent from my iPhone
On
Ooh, that is a nasty one. Is this JDK 7 only or also in 6?
It looks like the "-XX:ConcGCThreads=1" option is a workaround, is that right?
We've had some 1.6 JVMs behave in the same way that bug describes, but I
haven't verified it is because of finalizer problems.
wunder
On Sep 19, 2012, at 5:
also, did you re create the cluster after upgrading to a newer
version? I believe there were some changes made to the
clusterstate.json recently that are not backwards compatible.
--
Sami Siren
On Wed, Sep 19, 2012 at 6:21 PM, Sami Siren wrote:
> Hi,
>
> I am having troubles understanding the
The front-end code protection that I mentioned was outside of Solr. At that
time, requests with very large start values were slow, so we put code in the
front end to never request those. Even if the user wanted page 5000 of the
results, they would get page 100.
Now, those requests are fast, so
Via this bug: https://issues.apache.org/jira/browse/SOLR-2631
> ... Solr can infinite loop, use 100% CPU and stack overflow, if you
> execute the following HTTP request:
> - http://localhost:8983/solr/select?qt=/admin/ping
> - http://localhost:8983/solr/admin/ping?qt=/admin/ping
I am ru
Hi,
I am having troubles understanding the reason for that NPE.
First you could try removing the line #102 in HttpClientUtility so
that logging does not prevent creation of the http client in
SyncStrategy.
--
Sami Siren
On Wed, Sep 19, 2012 at 5:29 PM, Markus Jelsma
wrote:
> Hi,
>
> Since the
How are you triggering an infinite loop in your requests to Solr?
Erik
On Sep 19, 2012, at 11:12 , Mike Gagnon wrote:
> [ I am sorry for breaking the thread, but my inbox has neither received my
> original post to the mailing list, nor Otis's response (so I can't reply to
> his response)
[ I am sorry for breaking the thread, but my inbox has neither received my
original post to the mailing list, nor Otis's response (so I can't reply to
his response) ]
Thanks a bunch for your response Otis. Let me more thoroughly explain my
experimental workload and why I am surprised Solr works s
On 19 September 2012 20:10, wrote:
> I want to process a pdf file see "Indexing Data" from
> http://lucene.apache.org/solr/api-4_0_0-BETA/doc-files/tutorial.html
>
> the directory "update" doesnt even exist:
> SimplePostTool: POSTing files to http://localhost:8983/solr/update..
>
Sorry
There's nothing in that tutorial that mentions an update "directory". /update
is a URL endpoint that requires Solr be up and running.
Please post the entire set of steps that you're trying and the exact
(copy/pasted) error messages you're receiving.
And once you index a PDF file, you don't ret
I want to process a pdf file see "Indexing Data" from
http://lucene.apache.org/solr/api-4_0_0-BETA/doc-files/tutorial.html
the directory "update" doesnt even exist:
SimplePostTool: POSTing files to http://localhost:8983/solr/update..
fails because the /update directory is not there and also has
Hi,
Since the 2012-09-17 11:10:41 build shards start to have trouble coming back
online. When i restart one node the slices on the other nodes are throwing
exceptions and cannot be queried. I'm not sure how to remedy the problem but
stopping a node or restarting it a few times seems to help it.
Hello,
You can add this request parameter in the "defaults" section of your
request handler named "/select" in solrconfig.xml like this:
true
and as long as you use this request handler you won't need to
explicitly specify thi
Hi,
I was wondering if it is possible to set up highlighting so it is on by
default, and doesnt need to add to the URL. For example:
http://localhost:8080/solr/select?q=book&hl=true
I would like to have it so highlighting is on even if the URL is this:
http://localhost:8080/solr/select?q=boo
Take a look at admin/analysis on the text_general type You'll see that
StandardTokenizer is breaking the input strings up into individual tokens
on the colons and hyphens, so
2010-01-27T00:00:00Z
becomes the tokens
2010 01 27T00 00 00Z
admin/analysis should be your first reflex when you encounter
I _think_ the getReader calls are being triggered by the autoSoftCommit being
at one second. If so, this is probably OK. But bumping that up would nail
whether that's the case...
About RamBufferSizeMB. This has nothing to do with the size of the segments!
It's just how much memory is consumed befo
Two in one morning
The JVM bug I'm familiar with is here:
http://bugs.sun.com/bugdatabase/view_bug.do?bug_id=7112034
FWIW,
Erick
On Wed, Sep 19, 2012 at 8:20 AM, Shawn Heisey wrote:
> On 9/18/2012 9:29 PM, Lance Norskog wrote:
>>
>> There is a known JVM garbage collection bug that causes th
Would boosting (or sorting) by geodist work? See:
http://wiki.apache.org/solr/SpatialSearch#geodist_-_The_distance_function
Which you can use for a "boost query" as well as sorting.
Of course you need to get the lat/lon of your users to make this work,
but there are a number of services that can
I think this is the weak reference bug maybe?
http://bugs.sun.com/bugdatabase/view_bug.do?bug_id=7112034
Best
Erick
On Tue, Sep 18, 2012 at 11:29 PM, Lance Norskog wrote:
> The same answer as in another thread:
>
> There is a known JVM garbage collection bug that causes this. It has to do
> wi
Please review:
http://wiki.apache.org/solr/UsingMailingLists
There's nothing in your problem statement that's diagnosable. What did
you try? What
were the results? Details matter.
4.0 is in process of being prepped for release. 30 days was a
straw-man proposal.
Best
Erick
On Wed, Sep 19, 2012
On 9/18/2012 9:29 PM, Lance Norskog wrote:
There is a known JVM garbage collection bug that causes this. It has to do with
reclaiming Weak references, I think in WeakHashMap. Concurrent garbage
collection collides with this bug and the result is that old field cache data
is retained after clos
> We're having difficulty with some wildcard searches in Solr
> 4.0Beta. We're using a copyField to write a "tdate" to a
> "text_general" field. We are using the default definition
> for the "text_general" field type.
>
> indexed="true" stored="true" />
> type="text_general" indexed="tr
I'm still not sure I understand what it is you're trying to do. Index-time or
query-time boosts would probably be neater and more predictable than multiple
field instances, though.
http://wiki.apache.org/solr/UpdateXmlMessages#Optional_attributes_for_.22field.22
http://wiki.apache.org/solr/SolrR
Hi
Using SolrCloud I have added the following to solrconfig.xml (actually
the node in zookeeper)
512
After that I expected that my Lucene index segment files would be a bit
bigger than 1KB as I'm indexing very small documents
Enabling the infoStream I see a lot of "flush at getReader" (on
The "Sawtooth curve" is normal. It means that memory use slowly goes up, this
triggers a garbage collection pass, which frees the memory very quickly.
You can also turn off parallel garbage collection. This is slower, but will not
trigger the SUN bug. (If that really is the problem.)
- Orig
Hi Senthil,
try using the following:
CoreContainer coreContainer = new CoreContainer.Initializer().initialize();
SolrServer solrServer = new EmbeddedSolrServer(coreContainer,
"collection1");
Hope it helps,
Tommaso
2012/9/19 Senthil Kk Mani
>
> Hi,
>
> I am facing an issue while trying to us
Hi,
I am facing an issue while trying to use the solrj EmbeddedServer to load a
core. I am trying to load the example/solr packaged with solr-3.6.1.
It works perfectly fine through CommonHTTPSolrSever. I am able to query and
fetch the document.
I used the following jar files to compile and run.
seems the /update directory is missing? I use solr 4.0.0 beta
can not process pdf files because of it
also when will the final version be released? thought it it 30 days after beta?
how can we get the files which contain the searched queries / content?
77 matches
Mail list logo