Hi list,
while I can't get solr 4.3 with run-jetty-run up and running under eclipse
for debugging I tried to switch back to slf4j and followed
the steps of http://wiki.apache.org/solr/SolrLogging
Unfortunately eclipse bothers me with an error:
The import org.apache.log4j.AppenderSkeleton cannot be
Thank you both for your answers.
I really like the idea of explaining the changes for luceneMatchVersion
in more detail. Maybe this could even go into the release notes?
Thanks,
Andreas
Shawn Heisey wrote on 10.05.2013 15:27:
On 5/10/2013 5:11 AM, Jan Høydahl wrote:
Hi,
The fastest way to
On 5/15/2013 12:31 AM, Shawn Heisey wrote:
> If we assume that you've taken every possible step to reduce Solr's Java
> heap requirements, you might be able to do a heap of 8 to 16GB per
> server, but the actual heap requirement could be significantly higher.
> Adding this up, you get a bare minimu
On 5/14/2013 11:00 PM, pankaj.pand...@wipro.com wrote:
> We have to setup a billion document index using Apache Solr(about 2 billion
> docs). I need some assistance on choosing the right configuration for the
> environment setup. I have been going through the Solr documentation, but
> couldn't f
Hi,
We have to setup a billion document index using Apache Solr(about 2 billion
docs). I need some assistance on choosing the right configuration for the
environment setup. I have been going through the Solr documentation, but
couldn't figure out what would be the best configuration for same.
Try https://issues.apache.org/jira/browse/SOLR-4685
It allows you to return put JSON from a string field.
Also to convert a XML field to JSON you can use a plugin for DIH
https://issues.apache.org/jira/browse/SOLR-4692
On Monday, May 13, 2013, Chris Hostetter wrote:
>
> : I don't want to use P
On 5/14/2013 5:37 AM, Benjamin Ryan wrote:
> I have seen posts etc on how to setup a core that has a merged schema and
> request handlers to enable search across two other cores.
> Can anybody give an example of how the request handler in the "merged" core
> is defined so that it will query over
Hi,
Currently 3.3 but I could move to 4.x if needed.
Regards,
Ben
--
Dr Ben Ryan
Jorum Technical Manager
5.12 Roscoe Building
The University of Manchester
Oxford Road
Manchester
M13 9PL
Tel: 0160 275 6039
E-mail: ben
Hi Rishi,
If you have your cluster up and running, just add the nodes and they will
get evenly assigned to the shards. As of now, the replication factor is not
persisted.
On Wed, May 15, 2013 at 1:07 AM, Rishi Easwaran wrote:
> Ok looks like...I have to go to every node, add a replica individua
Oh, yes, definitely... you can certainly do it yourself.
It would be nice to have a request handler that did the document
"characterization", returning the most significant terms. Then the app could
simply send a request for that to the first collection and then turn it into
a query for the ot
I know this can be done, because I did it in Ultraseek. It was the first
feature I wrote, probably around 1997.
Take the most significant terms from the document and turn them into a query.
You can use the tf-idf scores of the terms as weights. I tested my queries
(without weights) from Ultrase
Do you mean pointing to a document in one collection and retrieving
documents from another collection that are similar to the first document?
Not that I know of.
Sounds like a reasonable Jira request.
But... the too collections would need to be fairly similar overall
themselves, because the c
Hi!
Im new in Solr, I would like to do a MoreLikeThis search for an ID into
other collections, not only in the collection that owns that ID
Is that posible? do you have an example?
Thank you very much!
--
View this message in context:
http://lucene.472066.n3.nabble.com/Search-MoreLikeThis-ID-
Hi,
I updated TIKA to 1.3 in Solr 3.6.2 myself. Everything seemed to work fine, but
extracting exe files is now broken. I also tried Solr 4.3.0 where TIKA 1.3 is
already integrated (SOLR-4416) and both nightly builds
solr-4.4-2013-05-13_21-22-06 / solr-5.0-2013-05-14_16-36-39 but ended up with
Sounds like I may not have all of the conditions correctly identified. Or
rather, things in our environment that I did not think were a factor may very
well be. For one thing, I see I neglected to mention that this was a grouping
(field collapsing) query. Let me give you more complete informati
Try removing the 2nd level of nesting. You can either remove it entirely or
combine the middle-level and inner queries with a join. If this "solves" the
problem, then we have the basis of a failing unit test, a jira issue, and a
potential solution. I realize the workaround I propose here migh
Which version of Solr do you use?
2013/5/14 Benjamin Ryan
> Hi,
> I have seen posts etc on how to setup a core that has a
> merged schema and request handlers to enable search across two other cores.
> Can anybody give an example of how the request handler in
> t
On 5/14/2013 1:44 PM, Lee, Peter wrote:
> The cause:
> We had an unintentional case that occurred through one request handler such
> that after the query was processed through the RH and the parameters were
> set, there were fq variables set but no actual "q" (query) param.
> - Under solr 3.5
In some sense, if all you want to do is send over a URL, e.g.
http://localhost:8993/, it's not out of the question to
use the java url stuff as exemplified at
http://www.cafeaulait.org/course/week12/22.html
or
http://stackoverflow.com/questions/7500342/using-sockets-to-fetch-a-webpage-with-java
Bu
On 5/14/2013 3:13 AM, Luis Cappa Banda wrote:
> I know that, but I was wondering if it exists another way just to set the
> complete query (including q, fq, sort, etc.) embedded in a SolrQuery object
> as the same way that you query using some kind of RequestHandler. That way
> would be more flexib
We had the same though about the toString being the symptom.
We have performed heap analysis on four separate heap 4GB dumps and they all
point to an issue with a single SolrDocument containing all data returned by
every sub entity query. Stepping through the code on a live instance we can
s
To get a re-written query with the top suggestions, specify
"spellcheck.collate=true". Begin reading from here
(http://wiki.apache.org/solr/SpellCheckComponent#spellcheck.collate) to see all
the options you have related to "collate".
Solr cannot return results from a collation automatically.
The reason it is writing all the imput fields for that document is this
particular error message appends "doc" to the end, which is a subclass of
SolrInputDocument, which has a "toString" that shows all the fields. Not sure
if this in particular changed, but I suspect this is a symptom not a ca
Peter,
Thanks for taking the time to spell out what you were going
through. It's great to have details like to to mull over.
Jason
On
2013-05-14 12:44, Lee, Peter wrote:
> Thank you one and all for your
input.
>
> The problem we were tripping over turned out NOT to be
related to using s
Thank you one and all for your input.
The problem we were tripping over turned out NOT to be related to using solrj
3.5 client and solr 4.2.1 server. We are, as of this minute, feeling confident
we can put this into production. We are still doing some testing, but thus far
we have hit all of th
Ok looks like...I have to go to every node, add a replica individually,
create the cores and add them to the collection.
ex:
http://newNode1:port/solr/admin/cores?action=CREATE&name=testCloud1_shard1_replica3&collection=testCloud1&shard=shard1&collection.configName=myconf
http://newNode2:port/
Hi all,
I'm new to Solr, and I'm trying to set up authentication so each core will
have a Jetty user specifically able to use that core as well an
administrator with access to all cores and the administration dashboard.
I'm using start.jar from the Solr download to start Jetty.
I set up the real
Any chance you may have had a different analyzer or parameter values when
you indexed compared to now? Like, maybe the data wasn't originally indexed
as lower case?
Or, that maybe some of the term occurrences have adjacent punctuation
(comma, period, parentheses, etc.) that the word delimiter
Thanks for all the great work on Solr. We have used it for over a year and have
been very satisfied with it.
However ,we have noticed that some of the recent changes have affected import
caching in a not so good way. We are using Solr 4.2.0.
We use full and delta imports. We only use a delta
Normally whenever I see a 503, that means the Solr Cloud has one of its
shards down, i.e. there isn't a full collection available.
You can see them at index time if you have lost connection to zookeeper but
searches should be ok.
If you see them on searches, it (in my experience) means the collecti
Hello all,
I am currently trying to determine what is the cause of some odd behaviour
when performing fuzzy queries in Solr 4.2.1. I have a field that is
configured as follows:
Hi,
I am beginning to work on SOLR cloud implementation.
I created a collection using the collections API
http://myhost:port/solr/admin/collections?action=CREATE&name=testCloud1&numShards=6&replicationFactor=2&collection.configName=myconf&maxShardsPerNode=1
My cluster now has 6 shards and 2 rep
Is this a UI issue, where the UI isn't quite sure or aware or is confused
about traditional master/slave mode and SolrCloud mode?
-- Jack Krupansky
-Original Message-
From: Shawn Heisey
Sent: Tuesday, May 14, 2013 1:26 PM
To: solr-user@lucene.apache.org
Subject: Re: How Can Leader Lis
Forgot to mention that formdataUploadLimitInKB must be increased on all solr
shards.
-Original Message-
From: Sergiu Bivol [mailto:sbi...@blackberry.com]
Sent: Tuesday, May 14, 2013 1:55 PM
To: solr-user@lucene.apache.org
Subject: RE: Why Solr Returns Error for a Search that has rows para
I spent some time chasing this as well; in our case only sharded requests
requesting >500K rows were failing.
If you enable FINE logging on the main (coordinating) shard you will notice a
number of "400 Bad Request" errors, complaining about an upload limit being
exceeded. Increasing formdataUp
Venkata,
Solr is a neat webapp. It doesn't spins threads (almost). It's spin in
servlet container threads. You need to configure tomcat/jetty.
On Tue, May 14, 2013 at 4:17 PM, Dmitry Kan wrote:
> venkata,
>
> If you are after search scaling, then the webapp server (like tomcat, jetty
> etc) han
On 5/14/2013 8:41 AM, Furkan KAMACI wrote:
> I use Solr 4.2.1 at SolrCloud. When I look at graph at cloud link of admin
> page, I see that one of my node is listed as Leader. However when I
> click core selector and clik my core there writes: Replication (Slave). It
> is leader but a slave instead
Hi Mark;
I have same situation with my Solr 4.2.1 nodes. These are from my startup
parameters:
-XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=75
-XX:NewRatio=3 -XX:MaxTenuringThreshold=8 -XX:+CMSParallelRemarkEnabled
-XX:+ParallelRefProcEnabled -XX:+UseLargePages -XX:+AggressiveOpts
Hello,
We're working with Solr with the Spellchecker module. It works fine
and returns suggestions as expected, but we need the search results
using the first suggestion instead of the suggestions. It is that
possible?
I entered that URL from my browser, not used SolrJ.
2013/5/14 Upayavira
> That error looks like a SolrJ version incompatibility, not because of
> your rows param. However, your rows param is extremely high, and will
> likely cause Solr to blow up in other ways. Solr isn't designed to
> return al
That error looks like a SolrJ version incompatibility, not because of
your rows param. However, your rows param is extremely high, and will
likely cause Solr to blow up in other ways. Solr isn't designed to
return all documents like that, only the n best matches, where n is a
low number.
If you ne
When I make a query:
:8983/solr/select?q=*:*
I have numFound: 4264579
and then when I make that query:
:8983/solr/select?q=*:*&rows=4264579
or even that:
:8983/solr/select?q=*:*&rows=4264578
org.apache.solr.client.solrj.SolrServerException:
java.lang.RuntimeException: Invalid version (expe
Hi All,
I'm creating 2 distinct sets of facet results using a key local param, e.g.:
facet.field={!ex=dt key=my_facet_key}the_facet_field&facet.field={!ex=dt
key=some_other_facet_key}the_facet_field
I would then like to refer to these 'pseudo' field later in the request
string. I thought this wo
On 05/14/2013 03:44 PM, Andre Bois-Crettez wrote:
On 05/13/2013 03:12 PM, Achim Domma wrote:
I'm mainly interested in showing the terms which each result document has in
common with the reference document.
regards,
Achim
It seems a good job for highlighting ?
http://docs.lucidworks.com/displa
I'm using a combination of tika and custom code to extract text from files.
(with solrj)
I was looking at the amount of files I had in my index and noticed many of
them where missing.
Then I went to the solradmin panel and noticed this in the logfiles:
SEVERE
SolrCore
java.
I use Solr 4.2.1 at SolrCloud. When I look at graph at cloud link of admin
page, I see that one of my node is listed as Leader. However when I
click core selector and clik my core there writes: Replication (Slave). It
is leader but a slave instead of being master. Do I miss anything? Can we
use Mas
Thank you for the pointer, this helps me understand what is going on. I
took out the highlighting branch. It turns out there is truncation logic
somewhere down there, because when I did, I had the full document text.
Peace. Michael
On Mon, May 13, 2013 at 6:45 PM, Erik Hatcher wrote:
> #fiel
On 05/13/2013 03:12 PM, Achim Domma wrote:
I'm mainly interested in showing the terms which each result document has in
common with the reference document.
regards,
Achim
It seems a good job for highlighting ?
http://docs.lucidworks.com/display/solr/Highlighting
http://wiki.apache.org/solr/Sol
The actual state is a mix of the clusterstate.json and the ephemeral live nodes
- a node may be listed as active or whatever, and if it's live node is not up,
it doesn't matter - it's considered down.
- Mark
On May 14, 2013, at 8:08 AM, Furkan KAMACI wrote:
> Node is shown as down at admin pa
When I check my SolrCloud nodes' log files (Solr 4.2.1) I get see that line
at all of them:
INFO: [collection1] webapp=/solr path=/admin/file/
params={file=admin-extra.html&_=1368535805558} status=0 QTime=5
May 14, 2013 3:50:27 PM org.apache.solr.core.SolrCore execute
INFO: [collection1] webapp=/s
That's also part of the basic, traditional Lucene search syntax:
+Java +mysql +php TCL Perl Selenium -ethernet -switching -routing
-- Jack Krupansky
-Original Message-
From: Kamal Palei
Sent: Tuesday, May 14, 2013 8:33 AM
To: solr-user@lucene.apache.org
Subject: Re: Can we search some
Thanks Jack Krupansky
Your solution having key as* **+Java +mysql +php TCL Perl Selenium* worked
nicely.
If I have to extend it something like, I need to search all documents,
those have
1. Mandatory Keywords present are *Java, MySql*
2. Atleast one keyword out of* TCL Perl Selenium* should be pr
venkata,
If you are after search scaling, then the webapp server (like tomcat, jetty
etc) handles allocation of threads per client connection (maxThreads for
jetty for instance). Inside one client request SOLR uses threads for
various tasks, but I don't have any exact figures (not sure if wiki has
Node is shown as down at admin page. It says there is one replica for that
shard but leader is dead (no new leader is selected!) however when I check
zookeeper information from /clusterstate.json at admin page I see that:
"shard2":{
"range":"b333-e665",
"state":"active",
"replicas":{
"10.
I'm developing a custom SolrServer that executes raw queries. Those raw
queries are the ones I mentioned before: String ones. The idea is that that
custom SolrServer execute internally the raw String query via HttpClient
(GET or POST) and then returns a QueryResponse. I want to be respectful
with S
Hi,
I have seen posts etc on how to setup a core that has a merged
schema and request handlers to enable search across two other cores.
Can anybody give an example of how the request handler in the
"merged" core is defined so that it will query over the two other s
Why are you parsing te response to a QueryResponse object? I think it's
much easier to parse a json format response to you business object.
2013/5/14 Luis Cappa Banda
> Yeah, unfortunately that's what I'm doing right now, but parsing the
> resultant String from the HttpClient response into a Qu
I am using Solr4.2 , I have few queries on new fuzzy implementation in
Solr4+
1) I come to know that Solr4+ accepts maximum editing distance to 2 (2
insertion, deletion, replacements). Is there any way , i can configure this
maximum editing distance limit ??
2) although I set editing distance t
We have a number of queries that produce good results based on the textual
data, but are contextually wrong (for example, an "SSD hard drive" search
matches the music album "SSD hip hop drives us crazy".
Textually a fair match, but SSD is a term that strongly relates to technical
documents.
Yeah, unfortunately that's what I'm doing right now, but parsing the
resultant String from the HttpClient response into a QueryResponse seems to
be imposible... I think that this would be a great feature to include in
next Solr release. I'm sure that many people will find useful to execute
raw Stri
I don't see any method in SolrServer like what you want.But I think you
just need write some code with HttpClient ,make a http request use the
complete query string.
2013/5/14 Luis Cappa Banda
> I know that, but I was wondering if it exists another way just to set the
> complete query (including
Added
--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
14. mai 2013 kl. 06:19 skrev "Dao Xuan, Hoang" :
> Hi Admins,
>
> My name is Eric. I got an account at http://wiki.apache.org/solr/ with user
> name is Eric D. Please add me to the Contributor Group. We currently h
Hey, hello!
Which class or classes do you suggest to check out it's source code?
Thank you,
2013/5/14 Upayavira
> Look at the source code for SolrJ, you'll find it is just a glorified
> HashMap, and you might find a way to interact with it that suits you (I
> think the end URL is just the res
Hi,
Query pasted in my post, is returning 1 record with 0 highlights, if i just
remove hl.maxAnalyzedChars= -1 from my query, it return proper highlight...
same query with some different random id ,working fine and returning
highlights properly , while few records return 0 highlights with
hl.max
Look at the source code for SolrJ, you'll find it is just a glorified
HashMap, and you might find a way to interact with it that suits you (I
think the end URL is just the result of calling toString() on the
HashMap).
Upayavira
On Tue, May 14, 2013, at 10:13 AM, Luis Cappa Banda wrote:
> I know t
I know that, but I was wondering if it exists another way just to set the
complete query (including q, fq, sort, etc.) embedded in a SolrQuery object
as the same way that you query using some kind of RequestHandler. That way
would be more flexible because you don't need to parse the complete query
hi,
the solrQuery.setQuery() method just set the 'q' param in the query string.
You need call other method to set up all necessary param:
solrQuery.setFields(); --> this set up the 'fl' param
solrQuery.setStart(); --> this set up the 'start' param
solrQuery.setRows(); --> this set up the 'rows' par
Hello, guys!
I would like to do something like this. Let's suppose we have:
*
*
*(...) *
*
*
*String query = "q=*:*&start=0&rows=20&sort=date%20desc";*
*
*
*SolrQuery solrQuery = new SolrQuery();*
*solrQuery.setQuery(query);*
*
*
*server.query(solrQuery);*
*
*
*(...)*
I tried that and it fails
Is it possible instead to store in your solr index a bounding box of store
location + delivery radius, do a bounding box intersection between your
user's point + radius (as a bounding box) and the shop's delivery bounding
box. If you want further precision, the frange may work assuming it's a
post-
At first I thought you were referring to Filters in Lucene at query time
(i.e. bitset filters) but I think you are referring to token filters at
indexing/text analysis time?
I have had success writing my own Filter as the link presents. The key is
that you should write a custom class that extends
Hi, thanks for the links and for your help. The server is now running third
day in a row with no issues. What is done:
1. Applied these GC tuning options: -XX:+UseConcMarkSweepGC
-XX:CMSInitiatingOccupancyFraction=80
2. Optimized the schema and index size (decreased at least 8 times).
3. Updated th
71 matches
Mail list logo