Hi Dmitry,
On Mon, Aug 12, 2013 at 9:36 AM, Dmitry Kan wrote:
> Hi Roman,
>
> Good point. I managed to run the command with -C and double quotes:
>
> python solrjmeter.py -a -C "g1,cms" -c hour -x ./jmx/SolrQueryTest.jmx
>
> As a result got several files (html, css, js, csv) in the running dir
: "Collections that do not specify numShards at collection creation time use
: custom sharding and default to the "implicit" router. Document updates
: received by a shard will be indexed to that shard, unless a "*shard*"
: parameter or document field names a different shard."
the word "implicit"
Awesome, thanks - that fixed it and working perfectly now. The comma syntax
is a little cryptic but I can live with that.
And the new Intersects syntax is much easier to read so thanks for that too!
On Tue, Aug 13, 2013 at 5:52 AM, David Smiley (@MITRE.org) [via Lucene] <
ml-node+s472066n4084084
Hoss,
>From the Solr 4.1 release highlights, under the SolrCloud enhancements
section:
"Collections that do not specify numShards at collection creation time use
custom sharding and default to the "implicit" router. Document updates
received by a shard will be indexed to that shard, unless a "*sh
Hi Jack,
Thank you for the suggestions.
I'm using DIH and the field names mobile,home,office comes as database
fields. And 'phones' field is defined in schema.xml as multiValued.
in DIH I'm using sub-entity for 'phones' fields.
Let me give a try based on your suggestion :-).
Thank you
--
View
Hi Tamanjit,
Not really :-).
**phones** is an multiValued. Usually solr will produce list of array of
values without naming each row something like this
Micheal
Schumacher
090933434343
052323232323
073628362782
I wanted to be this way.
Micheal
Schumacher
090933434343
: If that is the case, I think that my settings are correct. I still do not
: explain why I have such growth on all the shards at the same time.
you are missunderstanding how SolrCLoud works.
Unless you go out of your way to override hte document routing, Solr will
compute a logical shard to
I am in stuck below problem a few days. If someone have experienced the same
as this problems like me. Please give a hint. Thank you.
Originally, there is not the lib folder under solr, so I created it for
adding several jar files. ( I already used the ext folder which located in
example/lib/ext.
Hi,
I'm getting the same error on 4.4.0 (just downloaded) reproducibly with
these steps:
# curl
'http://localhost:8983/solr/admin/cores?action=CREATE&collection=sockmonkey&name=sockmonkeycore1&numShards=1&shard=shard1&replicationFactor=1'
# curl
'http://localhost:8983/solr/admin/cores?action=CREA
How are you indexing documents?
You need to either URL encode things or XML encode, sounds like.
Erik
On Aug 12, 2013, at 14:49, John Randall wrote:
> I am using an HTTP GET to add docs to Solr. All the docs load fine as long as
> none contain an ampersand. I get a syntax error when a doc
To answer my own post about the subtle difference between the shard and
replicate examples, it looks like the difference is in the numShards
parameter.
If you define numShards to be = 2, and then creating more shards than 2
will give you replicates. Is that correct?
If that is the case, I think
Greg, what do you mean by 'manually setting the shard on each document'?
I explicitly push the documents to their respective shard/port numbers.
Something like
curl http://localhost:shardport/solr/update --data-binary file.csv -H
'Content-type:text/csv; charset=ISO-8859-1'
I guess that this is
Yes, my config xmls were correct and the first thing I did before posting
here was to check with the Solr sever. That worked. I just solved this
right now. Turns out my problem was in the following line:
SolrParams solrParams = SolrRequestParsers.parseQueryString("*:*");
That was not doing it
Thanks Shawn for the detailed instructions.
About the router: it is implicit.
About the replicas: I followed the example at
http://wiki.apache.org/solr/SolrCloud
I start the shards with the following (paths and ports simplified):
cd /.../solr/shard1/
/usr/bin/java -Djetty.port=1 -Dbootstrap_con
Michael,
We replaced Lucene jars but run into a problem with incompatible version of
Apache HttpComponents. Still figuring it out.
Dmitriy
--
View this message in context:
http://lucene.472066.n3.nabble.com/Problem-running-Solr-indexing-in-Amazon-EMR-tp4083636p4084121.html
Sent from the Solr
On 8/12/2013 4:50 PM, Thierry Thelliez wrote:
Hello, I am trying to set a four shard system for the first time. I do
not understand why all the shards data are growing at about the same rate
when I push the documents to only one shard.
The four shards represent four calendar years. And for no
Are you manually setting the shard on each document? If not, documents
will be hashed across all the shards.
-Greg
On Mon, Aug 12, 2013 at 3:50 PM, Thierry Thelliez <
thierry.thelliez.t...@gmail.com> wrote:
> Hello, I am trying to set a four shard system for the first time. I do
> not unders
Hello, I am trying to set a four shard system for the first time. I do
not understand why all the shards data are growing at about the same rate
when I push the documents to only one shard.
The four shards represent four calendar years. And for now, on a
development machine, these four shards r
hi Dmitriy,
Just out of curiosity, have you tried replacing the Lucene jars with a
bootstrap action?
Michael Della Bitta
Applications Developer
o: +1 646 532 3062 | c: +1 917 477 7906
appinions inc.
“The Science of Influence Marketing”
18 East 41st Street
New York, NY 10017
t: @appinions
Michael,
Amazon Hadoop distribution has Lucene 2.9.4 jars in /lib directory and they
conflict with Solr 4.4 we are using. Once we pass that problem we run into
conflict with Apache HttpComponents you describe. I think the best bet would
be for us to build our own AMI to avoid these dependencies.
I've simplified things from my previous email, and I'm still seeing errors.
Using solr 4.4.0 with two nodes, starting with a single shard. Collection
is named "marin", host names are dumbo and solrcloud1. I bring up an empty
cloud and index 50 documents. I can query them and everything looks fi
It sounds like something like
https://issues.apache.org/jira/browse/SOLR-5005
On Mon, Aug 12, 2013 at 3:40 PM, PoM wrote:
> Hello,
>
> I'm working with SOLR for a while now and i have the following problem.
> We are doing boundry searches on SOLR to find towns near a town someone
> searched on.
Are you able to use the regular Solr server to do comparable tasks?
Best to learn Solr using the standard setup and tutorial and then reproduce
problems there first, before getting too deep into "embedded".
-- Jack Krupansky
-Original Message-
From: Delip Rao
Sent: Monday, August 12
Hello,
I'm trying to teach myself how to use EmbeddedSolrServer for indexing and
querying. I wrote a simple program -- Index a bunch of documents and query
all documents. I'm using Solr 4.4.0 jars and I can't see any output from
the query, i.e. I get 0 hits.
13/08/12 12:41:17 INFO core.SolrCore:
Whoops; I copy'ed your error of using commas. I meant:
Less clear: grantRoundDates:"Intersects(0 2013224 2014231 300)"
More clear: grantRoundDates:["0 2013224" TO "2014231 300"]
On 8/12/13 3:13 PM, "David Smiley (@MITRE.org)" wrote:
> Less clear: grantRoundDates:"Intersects(0
One of my previous mails to the group helped me simulate short-circuiting
OR behavior using (thanks to yonik)
_val_:"def(query(cond1,cond2,..))"
where if cond1 is true the query returns without executing the subsequent
conditions.
While it works successfully for single attribute, I am trying to
Beautiful! It worked. Thanks so much.
From: Shawn Heisey
To: solr-user@lucene.apache.org
Sent: Monday, August 12, 2013 3:12 PM
Subject: Re: Problem escaping ampersands in HTTP GET
On 8/12/2013 12:49 PM, John Randall wrote:
> I am using an HTTP GET to add docs
Hi Dan,
FYI the main reference page on this technique is here:
http://wiki.apache.org/solr/SpatialForTimeDurations
(note the slight buffering needed for the query shape).
You got the exception you got because you separated the dimensions of your
query rectangle using a comma, which got Spatia
On 8/12/2013 12:49 PM, John Randall wrote:
I am using an HTTP GET to add docs to Solr. All the docs load fine as long as none contain an
ampersand. I get a syntax error when a doc contains a field, for example, with the phrase
"Jack & Jill".
How can I escape the ampersand so that the doc loads
I am using an HTTP GET to add docs to Solr. All the docs load fine as long as
none contain an ampersand. I get a syntax error when a doc contains a field,
for example, with the phrase "Jack & Jill".
How can I escape the ampersand so that the doc loads normally?
Thanks in advance.
Hi,
I would suggest boosting over sorting. Something along:
radius:[0 TO 10]^100 OR radius:[10 TO *]
Regards,
Aloke
On Mon, Aug 12, 2013 at 6:43 PM, Raymond Wiker wrote:
> It will probably have better performance than having a "plan b" query that
> executes if the first query fails...
>
>
>
Jack Krupansky-2 wrote
> Also, be aware that the spaces in your query need to be URL-encoded.
> Depending on how you are sending the command, you may have to do that
> encoding yourself.
>
> -- Jack Krupansky
It's a good possibility that that's the problem. I've been doing queries in
different
It's not even documents by ID... it's just IDs. Anything that does lookups
by ID could be used.
Michael Della Bitta
Applications Developer
o: +1 646 532 3062 | c: +1 917 477 7906
appinions inc.
“The Science of Influence Marketing”
18 East 41st Street
New York, NY 10017
t: @appinions
1) the "terM" parser doesn't use the "df" default field param -- it
requires an "f" local param.
2) because the "f" localparam is required, you can't really use "term" in
defType
3) what you can do is something like this...
q={!term f=id v=$qq}&qq=your_input_id
..where the "q" param can be
Clear, thanks for response.
So, if I have two fields
generateWordParts="0" generateNumberParts="0"
catenateWords="0" catenateNumbers="0" catenateAll="1"
splitOnCaseChange="0" />
mapping="mapping-FoldToASCII.txt"/>
Hi,
We are using Solr Search Engine.t
The application needs the following data:
1. Faceted output based on a field.
2. One Record for each facet.
As the application is an existing one, tweaking the application to perform 2
calls would be tedious and hence the above mentioned 2 tasks should be
per
On Mon, Aug 12, 2013 at 12:25 PM, Mathias Lux wrote:
>
> Another thing for not using the the SORTED_SET and SORTED
> implementations is, that Solr currently works with Strings on that and
> I want to have a small memory footprint for millions of images ...
> which does not go well with immutables.
Dmitriy,
I don't believe that EMR does include Solr or Lucene in their EMR AMIs. But
there was a recent AMI update that ruined some things for us. Have you
tried using an older AMI?
One headache for us has been that the EMR AMI uses an older version of
Apache HttpComponents than that of Solr 4.3,
Hi Robert,
I'm basically "mis-using" Solr for content based image search. So I
have indexed fields (hashes) for candidate selection, i.e. 1,500
candidate results retrieved with the IndexSearcher by hashes, which I
then have to re-rank based on numeric vectors I'm storing in byte[]
arrays. I had an
You haven't mentioned what language your update client code is written in
... so i'm not sure what classes "DateTime" and "Convert" are and what
their default behavior is regarding timezones, but w/o that info it's
still pretty clear where your two mistakes are...
: string strStartDate = "26/
That does seem odd. To store and fetch documents by ID, use a document store.
MongoDB might be a good match.
wunder
On Aug 12, 2013, at 8:47 AM, Jack Krupansky wrote:
> This sounds fishy... why are you using Solr if you are only indexing one
> field and it is the ID field?! That sounds like a
This sounds fishy... why are you using Solr if you are only indexing one
field and it is the ID field?! That sounds like a particularly lame use of
Solr.
Declare the field as a "string" field and then Solr won't break it up into
tokens (other than special characters like slash.)
Or just set
Hi,
We have SolrCloud (4.4.0) cluster (5 shards and 2 replicas) on 10 boxes with
about 450 mil documents (~90 mil per shard). We're loading 1000 or less
documents in CSV format every few minutes. In Solr3, with 300 mil documents, it
used to take 30 seconds to load 1000 documents while in Solr4,
Well, i couldnt get it work but maybe thats because im not a solr expert.
What im trying to do is:
I have an index with only one indexed field. This field is an id so I don't
want the standard queryparser to try to break it up in tokens. On the client
side I use solrj like this:
SolrQuery solr
There is nothing needed in either solrconfig or schema.xml to use the
TermQParserPlugin. It is built into Solr.
If you need an example of the syntax:
http://lucene.apache.org/solr/4_4_0/solr-core/org/apache/solr/search/TermQParserPlugin.html
or
http://wiki.apache.org/solr/QueryParser
-- Jack
Hi,
can anyone provide a working example (solrconfig.xml,schema.xml) using the
TermQParserPlugin? I always get a Nullpointer-Exception on startup:
8920 [searcherExecutor-4-thread-1] ERROR org.apache.solr.core.SolrCore û
java.lang.NullPointerException
at
org.apache.solr.search.TermQParser
Also, be aware that the spaces in your query need to be URL-encoded.
Depending on how you are sending the command, you may have to do that
encoding yourself.
-- Jack Krupansky
-Original Message-
From: S L
Sent: Monday, August 12, 2013 10:51 AM
To: solr-user@lucene.apache.org
Subject:
On Mon, Aug 12, 2013 at 8:38 AM, Mathias Lux wrote:
> Hi!
>
> I'm basically searching for a method to put byte[] data into Lucene
> DocValues of type BINARY (see [1]). Currently only primitives and
> Strings are supported according to [1].
>
> I know that this can be done with a custom update hand
Erick Erickson wrote
> What do you get when you add &debugQuery=true? That should show you the
> results of the query parsing, which often adds clues.
>
> FWIW,
> Erick
When I was trying to debug this last night I noticed that when I added
"&debugQuery=true" to queries I would only get debug outp
After two weeks it's still down. What could be the problem?
On 31-7-2013 16:40, Anshum Gupta wrote:
It perhaps is just replaying the transaction logs and coming up. Wait for
it is what I'd say.
The admin UI as of now doesn't show replaying of transaction log as
'recovering', it does so only duri
Hi,
Example:
I want that stringTweakedNoIDF would be a stringTweaked but with the extra
similarity.
Regards
Bruno
On Mon, Aug 12, 2013 at 3:07 PM, tamanjit.bin...@yahoo.co.in <
tamanjit.bin...@yahoo.co.in> wrote:
> You would need to p
Unfortunately, no. Multivalued fields are a great and powerful, but very
simple feature, with the emphasis on simplicity. There are all sorts of
hacks that you can use to get around that limitation, but no direct feature
for "nested documents".
1. You can form a composite string: "home: 052323
You would need to provide a Solr file that would be the basic field type and
do rest of analysis on it. Is this what you want?
eg. fieldType name="textSpellPhrase" class="solr.TextField"
positionIncrementGap="100" stored="false"
multiValued="true">
--
View this message in context:
http://lucen
What are trying to achieve specifically? Reuse chain definitions?
Regards,
Alex
On 12 Aug 2013 09:41, "Bruno René Santos" wrote:
> Hi,
>
> Is it possible to extend a fieldtype from another fieldtype on the
> schema.xml?
>
> Regards
> Bruno
>
> --
> Bruno René Santos
> Lisboa - Portugal
>
Quoted phrases will be passed to the analyzer as one string, so there a
white space tokenizer is needed.
-- Jack Krupansky
-Original Message-
From: Andrea Gazzarini
Sent: Monday, August 12, 2013 6:52 AM
To: solr-user@lucene.apache.org
Subject: Re: Tokenization at query time
Hi Tanguy
Hi,
Is it possible to extend a fieldtype from another fieldtype on the
schema.xml?
Regards
Bruno
--
Bruno René Santos
Lisboa - Portugal
Hi Stefan,
It is apparently a browser "feature": works fine in Chrome (Version
28.0.1500.95).
A side note: would "return false;" following the DOM instruction help here?
Dmitry
On Wed, Aug 7, 2013 at 6:59 PM, Stefan Matheis wrote:
> Hey Dmitry
>
> That sounds a bit odd .. those are more like
Hi!
That's what I'm doing currently, but it ends up in StoredField
implementations, which create an overhead on decompression I want to
avoid.
cheers,
Mathias
On Mon, Aug 12, 2013 at 3:11 PM, Raymond Wiker wrote:
> base64-encode the binary data? That will give you strings, at the expense
> of s
Hi Roman,
Good point. I managed to run the command with -C and double quotes:
python solrjmeter.py -a -C "g1,cms" -c hour -x ./jmx/SolrQueryTest.jmx
As a result got several files (html, css, js, csv) in the running directory
(any way to specify where the output should be stored in this case?)
W
2c worth,
We do lots of facet lookups to allow 'prettyprint' versions of facet names.
We do this on the client-side, though. The reason is then the lookups can
be different for different locations/users etc. - makes it easy for
localization.
It's also very easy to implement such a lookup, without h
I'm guessing that this will only be a partial solution, as you'll want to
use the "real" field name for filtering, and the encoding problems will
still have to be solved (since there will be a need to support the national
characters in the field values as well as the field names...)
Personally, I
Have you tried looking at source code itself? Between simple organizer like
keyword and complex language ones, you should be able to get an idea. Then
ask specific follow up questions.
Regards,
Alex
On 12 Aug 2013 09:29, "dhaivat dave" wrote:
> Hello All,
>
> I want to create custom tokenis
Hello All,
I want to create custom tokeniser in solr 4.4. it will be very helpful if
some one share any tutorials or information on this.
Many Thanks,
Dhaivat Dave
What do you get when you add &debugQuery=true? That should show you the
results of the query parsing, which often adds clues.
FWIW,
Erick
On Mon, Aug 12, 2013 at 1:35 AM, S L wrote:
> Jack Krupansky-2 wrote
> > What query parser and release of Solr are you using?
> >
> > There was a bug at one
Have you seen the "key" parameter here:
http://wiki.apache.org/solr/SimpleFacetParameters#key_:_Changing_the_output_key
it allows you to label the output key anything you want, and since these are
field names, this seems to-able.
Best,
Erick
On Mon, Aug 12, 2013 at 4:02 AM, Aleksander Akerø wro
Why are you using the core creation commands rather than the
collection commands? The latter are intended for SolrCloud...
Best
Erick
On Mon, Aug 12, 2013 at 4:51 AM, Prasi S wrote:
> Hi,
> I have setup solrcloud in solr 4.4, with 2 solr's in 2 tomcat servers and
> Zookeeper.
>
> I setup Zooke
It will probably have better performance than having a "plan b" query that
executes if the first query fails...
On Mon, Aug 12, 2013 at 2:27 PM, PoM wrote:
> That would actually be a decent solution, although it isn't the best i will
> try if it gives any performance issues
>
>
>
> --
> View th
base64-encode the binary data? That will give you strings, at the expense
of some storage overhead.
On Mon, Aug 12, 2013 at 2:38 PM, Mathias Lux wrote:
> Hi!
>
> I'm basically searching for a method to put byte[] data into Lucene
> DocValues of type BINARY (see [1]). Currently only primitives an
Hi
I manually edited core.properties file by setting /loadOnStartup=false/ on
all cores. But here i am facing a problem. After starting the solr cloud
servers, I couldnt index to any collection till particular interval. I am
getting exception like "No Live Solr Servers".
=> If we submit a index re
Hi!
I'm basically searching for a method to put byte[] data into Lucene
DocValues of type BINARY (see [1]). Currently only primitives and
Strings are supported according to [1].
I know that this can be done with a custom update handler, but I'd
like to avoid that.
cheers,
Mathias
[1] http://wik
That would actually be a decent solution, although it isn't the best i will
try if it gives any performance issues
--
View this message in context:
http://lucene.472066.n3.nabble.com/SOLR-OR-query-want-1-of-the-2-results-tp4083957p4083969.html
Sent from the Solr - User mailing list archive at N
Can you not get more number of results i.e. for a larger distance and sort by
distance?
--
View this message in context:
http://lucene.472066.n3.nabble.com/SOLR-OR-query-want-1-of-the-2-results-tp4083957p4083966.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hello,
I'm working with SOLR for a while now and i have the following problem.
We are doing boundry searches on SOLR to find towns near a town someone
searched on.
The problem is that when it does not find anything in a radius of 10 km for
example i'm required to do another query that searches in
Hi Tanguy,
thanks for fast response. What you are saying corresponds perfectly with
the behaviour I'm observing.
Now, other than having a big problem (I have several other fields both
in the pf and qf where spaces doesn't matter, field types like the
"text_en" field type in the example schema)
Chris,
Thanks for your reply. Please find the below process What we followed on Sample
Test Solr:
Step1: Indexing data to Solr ( Schema looks like this)
Step2: Indexing code:
string strStartDate = "26/5/2000 8:38:16 AM";
DateTime dt = new DateTime();
dt = Convert.ToDateTime(strSt
Hello Andrea,
I think you face a rather common issue involving keyword tokenization and query
parsing in Lucene:
The query parser splits the input query on white spaces, and then each token is
analysed according to your configuration.
So those queries with a whitespace won't behave as expected be
Would it not be convenient to replace these alphabets with "pretty" alphabets
in the application?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Facet-field-display-name-tp4083915p4083928.html
Sent from the Solr - User mailing list archive at Nabble.com.
So your doc would look something like this:
Micheal
Schumacher
090933434343
052323232323
073628362782
090933434343
052323232323
073628362782
--
View this message in context:
http://lucene.472066.n3.nabble.com/multiValued-How-do-I-specify-name-to-each-value-returned-by-multi
I think what you require here is to have different fields for mobile/landline
etc. Based on where and what data you fetch these fields will be populated.
You can also have a multivalued field that will contain all these fields
(copyField can come in handy) for search etc.
--
View this message in
Hi all,
I have a field (among others)in my schema defined like this:
Basically, both at index and query time the field value is normalized
like this.
Mag. 778 G 69 => mag778g69
Now, in my solrconfig I'm using a search handler like this:
...
Hi All,
I have a multiValued field, and it returns the result in the following
manner
Micheal
Schumacher
090933434343
052323232323
073628362782
In the above document phones is multiValued and it works fine :-). However I
would like to name each row in the 'phones' with meaning
Hi,
I have setup solrcloud in solr 4.4, with 2 solr's in 2 tomcat servers and
Zookeeper.
I setup Zookeeper with a collection "firstcollection" and then i give the
belwo command
http://localhost:8080/solr/admin/cores?action=CREATE&name=mycore_sh1&collection=firstcollection&shard=shard1
Similarly,
Awesome, thank you.
Would be great to see anything about this difference in wiki.
For anyone else suffering from this, originally I started with:
fulltext(query.gsub(/(\$|&|\+|,|\/|:|;|=|\?|@)/) { "\\#{$1}" }) do
minimum_match(1)
end
But be careful with the "+" sign, since it could be used in
Hi
I wondered if there was some way to configure a display name for facet
fields. Either that or some way to display nordic letters without it
messing up the faceting.
Say I wanted a facet field called "område" (norwegian, "area" in english).
Then I would have to create the field something like t
84 matches
Mail list logo