Hi
Currently i am using solr-4.3 for my product. I will be creating a
collection for each user. So number of collections keeps on increasing. I
have hosted 3 solr servers and 3 zookeeper servers, each of size 400GB with
8 GB RAM.
There is possibility of memory(400GB) gets filled at sometime. Cu
Hi,
I just found SOLR-4663 beeing patched in the latest update I did.
Does anyone know any other solution to use ONE physical index for various
purposes?
Why? I would like to use different solconfig.xmls in terms of cache sizes,
result window size, etc. per business case for optimal performance,
Hi,
Have a look at the wiki page for multi-select faceting:
http://wiki.apache.org/solr/SimpleFacetParameters#Multi-Select_Faceting_and_LocalParams.
Alan Woodward
www.flax.co.uk
On 26 Jul 2013, at 07:23, Ifnu bima wrote:
> Hi,
>
> I'm currently looking at zappos solr implementation on their
Hi Erick,
First Of all sorry for the late reply.
The scenario is as follows
1. Create a solr set up on two machines say (ip1 and ip2) with shard=1 and
external zoo-keeper
2. Now if i create a core x on machine with ip1 only and use the query
http://ip1:port1/solr/x/select?q=*:*&distrib=false
htt
Hi James,
Thanks for you reply. I got it worked and below was my old query.
http://localhost:8080/solr_3.6.1_spellcheck/test_spellcheck/spellcheck?q=8956632541&spellcheck=true
now I changed the q to spellcheck.q and it started working. This is the response
0210108956632541589566325415
Regar
I'm in the process to create a service gateway from a SQL database bur
externally acting as a Solr Server. I have implemented the XML, JSON and PHPs
Response format but using sharing I'm receiving requests for a "javabin"
format. Looking into javadoc I have found a JavaBinCodec encoder receiving
I think I've confused myself here (not hard these days!), I have the
branch_4x code checked out, and that version definitely supports starting
Solr with no cores at all. I still get an Admin UI and I can then use that
to create cores/collections starting from a clean slate.
Does that work in 4.4.
Hi Alan,
It works flawlessly, thank you so much :D
Hi!
When SolrClound executes a query, it creates shard requests, which is sent
to one replica of each shard. Total QTime is determined by the slowest
shard response (plus some extra time). [For simplicity, let's assume that
no stored fields are requested.]
I suffer from a situation where in every
Thank you Jack, I've totally missed this type of plug-in.
Artem.
25.07.2013 23:38, Jack Krupansky пишет:
Yeah, those are the rules. They are more of a heuristic that manages to work
most of the time reasonably well, but like most heuristics, it is not
perfect.
In this particular case, your bes
Hi
This is a urgent call, I am grouping the solr documents by a field name and
want to get the Range(Min and Max) value for another field in that group.
StatsComponent works fine on all the document as whole rendering the max
and min of a field, is it possible to get the StatsComponent per group
Hey guys, i have a Solr 4.3 instance running in my server, but Im having some
troubles with encoding URL querystring.
Im currently encoding my query characters, so, when its searched for "Café",
its actually encoded to "caf%E9" and "cão" is encoded to "c%E3o".
My URLencoding in tomcat is "iso-8859
Here is an example from example solrconfig file:
content^0.5 anchor^1.0 title^1.2
content^0.5 anchor^1.5 title^1.2 site^1.5
What is the difference between qf and pf parameters, they both boost
fields both there should be a difference?
This was at example solrconfig file:
dismax
explicit
0.01
content^0.5 anchor^1.0 title^1.2
content^0.5 anchor^1.5 title^1.2 site^1.5
url
100
true
*:*
title url content
0
title
0
url
regex
On 26 July 2013 18:41, Furkan KAMACI wrote:
>
> Here is an example from example solrconfig file:
>
> content^0.5 anchor^1.0 title^1.2
> content^0.5 anchor^1.5 title^1.2 site^1.5
>
> What is the difference between qf and pf parameters, they both boost
> fields both there should be a dif
Ok, I've found that there was not a problem at config.
2013/7/26 Furkan KAMACI
> This was at example solrconfig file:
>
>
>
> dismax
> explicit
> 0.01
> content^0.5 anchor^1.0 title^1.2
> content^0.5 anchor^1.5 title^1.2 site^1.5
> url
> 100
>
Look at org.apache.solr.response.BinaryResponseWriter. This serializes from
a SolrQueryResponse. There is no direct XML to JavaBin serialization.
It uses org.apache.solr.common.util.JavaBinCodec.
This is the server side processing. For the client side, see the SolrJ
package.
-- Jack Krupans
The "p" is for "phrase" or "proximity" boosting. "pf" doesn't change what
documents match, but gives a boost if all of the terms occur next to or near
each other, based on "ps" (phrase/proximity slop.)
pf2 is to two-term bigram boosting. pf3 is for three-term trigram boosting.
The terms are ad
On 7/26/2013 4:30 AM, Daniel Collins wrote:
> I think I've confused myself here (not hard these days!), I have the
> branch_4x code checked out, and that version definitely supports starting
> Solr with no cores at all. I still get an Admin UI and I can then use that
> to create cores/collections
On 7/26/2013 7:05 AM, Gustav wrote:
> Hey guys, i have a Solr 4.3 instance running in my server, but Im having some
> troubles with encoding URL querystring.
> Im currently encoding my query characters, so, when its searched for "Café",
> its actually encoded to "caf%E9" and "cão" is encoded to "c
Hi,
I think no, and I think there is a JIRA issue open for that.
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring -- http://sematext.com/spm
On Fri, Jul 26, 2013 at 2:32 PM, Vineet Mishra wrote:
> Hi
>
> This is a urgent call, I am grouping the solr document
Or simply use Flume Solr Sink and skip writing to local disk.
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring -- http://sematext.com/spm
On Thu, Jul 25, 2013 at 11:02 PM, Jack Krupansky
wrote:
> Use LucidWorks Search, define a file system data source and set
Hi,
1. Keep monitoring your disk (e.g. with SPM:
http://sematext.com/spm/solr-performance-monitoring/ ).
2. Add new SolrCloud nodes when disk is N% full
3. Create new Collection and specify where its data should live - see
createNodeSet under
http://wiki.apache.org/solr/SolrCloud#Managing_collecti
When I run that query:
solr/select?q=url:"ftp://"&wt=xml&fl=url
I get results as like that:
http://forum.whmdestek.com/ftp-makaleleri/
http://www.netadi.com/ftp-kurulumu.php
Why it does not make an exact search find: *ftp://* ?
On 26 July 2013 20:01, Furkan KAMACI wrote:
> When I run that query:
>
> solr/select?q=url:"ftp://"&wt=xml&fl=url
>
> I get results as like that:
>
>
> http://forum.whmdestek.com/ftp-makaleleri/
> http://www.netadi.com/ftp-kurulumu.php
>
> Why it does not make an exact search find: *ftp://* ?
Hi,
use &debugQuery=true and you'll probably immediately see how that
url:... stuff was interpreted. Maybe that ":" throws it off...
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring -- http://sematext.com/spm
On Fri, Jul 26, 2013 at 4:31 PM, Furkan KAMACI w
Is "url" a "string" field or is it "text" (and hence analyzed, which
typically removes all punctuation)?
A URL typically needs to be stored as both string and text so that the user
can do both exact match and keyword match.
-- Jack Krupansky
-Original Message-
From: Furkan KAMACI
S
Try using the Solr's Analysis tool(http://localhost:8983/solr/analysis) to
check all the tokenizers/filters chain that modifies your query.
What's probably happening is that the field "url" uses a tokenizer or filter
that removes the *special* chars "://".
--
View this message in context:
http
I just realised that you can use the "appends" attribute value in the request
handler config (of solrconfig.xml). By setting this, any additional fields
you add via the solrj API are appended.
Thanks
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-dynamically-add-geo
Hi
Is there somewhere in the stats page (e.g.
http://localhost:8983/solr/admin/mbeans?stats=true) that has the stats for
all of the request handlers combined?
I have a lot of request handlers that have their individual stats, but for a
birds eye view of performance it would be good to get a combi
I have a synonyms file as like that:
cart; shopping cart; market trolley
When I analyse my query I see that when I search cart these becomes
synonyms:
cart, shopping, market, trolley
so cart is synonym with shopping. How should I define my synonyms.txt file
that it will understand that cart is
Hey Erick
Thank you very much for your help.
So I dived into the solr code and read the
http://wiki.apache.org/solr/HowToContribute section. Really informative :-)
I created a Jira issue about my problem and I attached a patch file with a
implementation off pivot faceting with ngroup and visi
Hi,
I don't think there is a global stat for all request handlers combined.
You can do that with SPM for Solr, though.
Here's a marked up screenshot:
http://www.flickr.com/photos/otis/9369102411/
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring -- http://semat
That makes sense because the Query Converter has a hard time with numbers, but
spellcheck.q bypasses the Query Converter.
James Dyer
Ingram Content Group
(615) 213-4311
-Original Message-
From: Poornima Jay [mailto:poornima...@rocketmail.com]
Sent: Friday, July 26, 2013 4:05 AM
To: sol
Query converter is a plugin if the app has special requirements.
-- Jack Krupansky
-Original Message-
From: Dyer, James
Sent: Friday, July 26, 2013 11:33 AM
To: solr-user@lucene.apache.org
Subject: RE: Spell check SOLR 3.6.1 not working for numbers
That makes sense because the Query C
Well, that's one of the areas where Solr synonym support breaks down. The
LucidWorks Search query parser has a proprietary solution for that problem,
but it won't help you with bare Solr. Some people have used shingles.
In short, for query-time synonym phrases your best bet is to parse the quer
I have not fully worked through this myself yet. So if it is a silly
question, feel free to ignore. But if you know the answer of the top of
your head
I am trying to figure out how SolrCloud works when a container has a port
forwarded, so internal port number is different from the external one
When I search wikipedia the home page of wikipedia is not at first result:
http://www.wikipedia.org/
first result is that:
http://en.wikipedia.org/wiki/Spain
How can I say that domains of web sites should be first at SolrCloud? (I
want something like grouping at domains and boosting at url leng
Why Solr does not split that terms by*;* I think that it both split
by *;* and white space character?
2013/7/26 Jack Krupansky
> Well, that's one of the areas where Solr synonym support breaks down. The
> LucidWorks Search query parser has a proprietary solution for that problem,
> b
Hmmm... Actually, I think there was also a solution where you could specify
an alternate tokenizer for the synonym file which would not tokenize on
space, so that the full phrase would be passed to the query parser/generator
as a single "term" so that it would generate a phrase (if you have the
Should I re write it as like that:
shopping cart => market trolley, cart
or somethinglike that?
2013/7/26 Furkan KAMACI
> Why Solr does not split that terms by*;* I think that it both
> split by *;* and white space character?
>
>
> 2013/7/26 Jack Krupansky
>
>> Well, that's one
Let's assume that I have that urls at my index:
www.abc.com
www.abc.com/a
www.abc.com/b
www.abc.com/c
...
www.abc.com/x
How can I exact search for www.abc.com ? url:"www.abc.com" doesn't works
because it returns both www.abc.com/a, www.abc.com/b etc?
Separate fields for URL as string and URL as keywords makes sense. You can
also use the URL classifier update processor or a regex filter to have a
third field to match solely the domain name, if that is needed.
-- Jack Krupansky
-Original Message-
From: Furkan KAMACI
Sent: Friday, J
May be this wont work, but just a thought...Cant you use
PathHierarchyTokenizerFactory and configure as below?
In this example however we see the oposite configuration, so that a query
for Books/NonFic/Science/Physics would match documents containing
Books/NonFic, Books/NonFic/Science, or Books/No
Hello,
Example:
Attribute:valueOR Attribute:valueOR Attribute:value
When I clear facets, it clears the entire &fq statement. If I wanted to keep
one of the attribute within the query string, would this be possible through
solr or would I have to hardcode this?
Jay
--
View this message in c
Which code is this happening in? I think this is in client (Javascript)
code rather than in Solr. For Solr, whatever you pass in request parameters
is what it does. This seem to be happening earlier than Solr.
Regards,
Alex.
Personal website: http://www.outerthoughts.com/
LinkedIn: http://www.
This is current a hard-coded limit from what I've understood. From what
I remember, Mark said Yonik said that there are reasons to make the
packets that size. But whether this is empirically a Good Thing I don't know.
SOLR-4816 will address this a different way by making SolrJ batch up
the docs an
By definition, the autocommit interval defines when you can
search on documents you've added. Items should be
searchable after 10 seconds. Take a look at the logs to see
the commit messages
Best
Erick
On Thu, Jul 25, 2013 at 8:29 AM, tskom wrote:
> My actual solconfig.xml is:
>
>
>
>
Thanks for the answer Shawn,
The problem here is that in my client's application, the query beign encoded
in iso-8859-1 its a *must*. So, this is kind of a trouble here.
I just dont get how this encoding could work on queries in version 3.5, but
it doesnt in 4.3.
--
View this message in context
We have SolrCloud cluster (5 shards and 2 replicas) on 10 boxes. While running
stress tests, we want to monitor JVM heap size across 10 nodes. Is there a
utility which would connect to all nodes' jmx port and display all bean details
for the cloud?
Thanks!
bq: I am also trying to figure out if I can place
extra dimensions to the solr score which takes other attributes into
consideration
Have you looked at function queries? The whole point of them is
to do something that influences score, which may be quite
complex. There are ways to, say, multiply t
Or just put separate ZK ensembles in the two data centers, and
send the data to both when you index. Not tightly coupled, but
possible.
Best
Erick
On Thu, Jul 25, 2013 at 2:58 PM, Joshi, Shital wrote:
> Thanks for all answers.
>
> It appears that we will not have a data-center failure tolerant d
You can use SPM (i think):
http://sematext.com/spm/solr-performance-monitoring/
On Fri, Jul 26, 2013 at 1:36 PM, Joshi, Shital wrote:
> We have SolrCloud cluster (5 shards and 2 replicas) on 10 boxes. While
> running stress tests, we want to monitor JVM heap size across 10 nodes. Is
> there a u
I have used JMX with SOLR before..
http://docs.lucidworks.com/display/solr/Using+JMX+with+Solr
--
View this message in context:
http://lucene.472066.n3.nabble.com/monitor-jvm-heap-size-for-solrcloud-tp4080713p4080725.html
Sent from the Solr - User mailing list archive at Nabble.com.
We have been using newrelic (they have a free plan too) and gives all
needed info like: jvm heap usage in eden space, survivor space and old gen.
Garbage collection info, detailed info about the solr requests and its
response times, error rates etc.
I highly recommend using newrelic to monitor you
I was using Linux. I used the Java Service Wrapper and found what I needed!
It provides a way to wrap the start.jar so that it can be started and
stopped using linux daemon, helpful for my case of connecting Solr to a
chef recipe. I may write an explanation of this soon.
Hello,
I am using the SolrCloud with a zookeeper ensemble like on example C from
the wiki except with total of 3 shards and no replicas (oops). After
indexing a whole bunch of documents, shard 2 went down and I'm not sure
why. I tried restarting it with the jar command and I tried deleting shard1
I am trying to upgrade solr to 4.4 version, and looks like solr cann't load
the ShingleFilterFactory class.
417 [coreLoadExecutor-4-thread-1] ERROR org.apache.solr.core.CoreContainer
– Unable to create core: collection1
org.apache.solr.common.SolrException: Plugin init failure for [schema.xml]
fi
On 7/26/2013 2:03 PM, Gustav wrote:
> Thanks for the answer Shawn,
>
> The problem here is that in my client's application, the query beign encoded
> in iso-8859-1 its a *must*. So, this is kind of a trouble here.
> I just dont get how this encoding could work on queries in version 3.5, but
> it d
Hi Katie,
1. First things first, I would strongly advice to manually update/remove zk
or any other info when you're running things in the SolrCloud mode unless
you are sure of what you're doing.
2. Also, your node could be currently recovering from the transaction
log(did you issue a hard commit
Can you also let me know what version of Solr are you on?
On Sat, Jul 27, 2013 at 8:26 AM, Anshum Gupta wrote:
> Hi Katie,
>
> 1. First things first, I would strongly advice to manually update/remove
> zk or any other info when you're running things in the SolrCloud mode
> unless you are sure of
Dear list:
I have an ever-growing solr repository, and I need to process every single
document to extract statistics. What would be a reasonable process that
satifies the following properties:
- Exhaustive: I have to traverse every single document
- Incremental: in other words, it has to allow me
On 7/26/2013 11:02 PM, Joe Zhang wrote:
> I have an ever-growing solr repository, and I need to process every single
> document to extract statistics. What would be a reasonable process that
> satifies the following properties:
>
> - Exhaustive: I have to traverse every single document
> - Increme
Thanks for your kind reply, Shawn.
On Fri, Jul 26, 2013 at 10:27 PM, Shawn Heisey wrote:
> On 7/26/2013 11:02 PM, Joe Zhang wrote:
> > I have an ever-growing solr repository, and I need to process every
> single
> > document to extract statistics. What would be a reasonable process that
> > sati
On a related, inspired by what you said, Shawn, an auto increment id seems
perfect here. Yet I found there is no such support in solr. The UUID only
guarantees uniqueness.
On Fri, Jul 26, 2013 at 10:50 PM, Joe Zhang wrote:
> Thanks for your kind reply, Shawn.
>
> On Fri, Jul 26, 2013 at 10:27 P
On 7/26/2013 11:50 PM, Joe Zhang wrote:
> ==> Essentially we are doing paigination here, right? If performance is not
> the concern, given that the index is dynamic, does the order of
> entries remain stable over time?
Yes, it's pagination. Just like the other method that I've described in
detail
On Fri, Jul 26, 2013 at 11:18 PM, Shawn Heisey wrote:
> On 7/26/2013 11:50 PM, Joe Zhang wrote:
> > ==> Essentially we are doing paigination here, right? If performance is
> not
> > the concern, given that the index is dynamic, does the order of
> > entries remain stable over time?
>
> Yes, it's
On 7/27/2013 12:30 AM, Joe Zhang wrote:
> ==> so a "url" field would work fine?
As long as it's guaranteed unique on every document (especially if it is
your uniqueKey) and goes into the index as a single token, that should
work just fine for the range queries I've described.
Thanks,
Shawn
Thanks.
On Fri, Jul 26, 2013 at 11:34 PM, Shawn Heisey wrote:
> On 7/27/2013 12:30 AM, Joe Zhang wrote:
> > ==> so a "url" field would work fine?
>
> As long as it's guaranteed unique on every document (especially if it is
> your uniqueKey) and goes into the index as a single token, that should
69 matches
Mail list logo