And if I use dynamic fields to split multivalued field on different fields,
can I use this dynamic field in *updateRequestProcessorChain* ? I've tried
this, but seems dynamic values doesn't supported in langid.map.individual.fl
--
View this message in context:
http://lucene.472066.n3.nabble.c
It is a class writen by me which extends the UpdateRequestProcessorFactory
.I compiled it ,created the jar and added it to the solr lib folder.It is
the process suggested in the links i used to verify.
Thanks.
On Wed, Oct 30, 2013 at 4:09 PM, Koji Sekiguchi wrote:
> (13/10/30 22:09), lovely ka
Bumping up this thread as I'm facing similar issue . Any solution ?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Grouping-performance-problem-tp3995245p4098566.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
I've recently upgraded to SolrCloud (4.4) from Master-Slave mode. One of
the changes I did the in queries is to add group functionality to remove
duplicate results. The grouping is done on a specific field. But the change
seemed to have a huge effect on the query performance. The "group" op
Hi Susheel,
Have a look at this: http://wiki.apache.org/lucene-java/ScoresAsPercentages
You may really want to reconsider doing that.
On Thu, Oct 31, 2013 at 9:41 AM, sushil sharma wrote:
> Hi,
>
> We have a requirement where user would like to see a score (between 0 to
> 1) which can tell h
Hi,
We have a requirement where user would like to see a score (between 0 to 1)
which can tell how close the input search string is with result string. So if
input was very close but not exact matach, score could be .90 etc.
I do understand that we can get score from solr & divide by highest
So here is my use case with a little more detail. I'm working with
recurring events. Each event has an expression associated with it that
defines its recurrence pattern. For example, monthly, daily, yearly...
The event has metadata associated with it that is searchable. When a user
performs a s
A flat distribution of queries is a poor test. Real queries have a zipf
distribution. The flat distribution will get almost no benefit from caching, so
it will give too low a number and stress disk IO too much. The 99th percentile
is probably the same for both distributions, because that is domi
Hi Siva,
(13/10/30 18:12), sivaprasad wrote:
Hi,
We have a requirement where we need to send the matched synonyms as part of
Solr response.
I don't think that Solr has such function.
Do we need to customize the Solr response handler to do this?
So the answer is yes.
koji
--
http://soleami
I tried using synonyms but it doesn't actually change the stored text rather
just the indexed value.
I need a way to change the raw value stored in SOLR. May be I should use a
custom update processor to standardize the data.
--
View this message in context:
http://lucene.472066.n3.nabble.com/
(13/10/30 22:09), lovely kasi wrote:
Hi,
I made few changes to the solrconfig.xml, created a jar file,added it to
the lib folder of the solr and tried to start it.
THe changes in the solrconfig.xml are
LEAD_NOTES
category
Others
naiveBayesModel
Also note that function queries only return numbers (given their origin
in scoring). They cannot be used to create virtual string or text
fields.
Upayavira
On Wed, Oct 30, 2013, at 05:19 PM, Jack Krupansky wrote:
> A function query is simply returning a calculated result based on
> existing
> da
Thanks for the link
Sent from my iPhone
On Oct 30, 2013, at 4:06 PM, "Rajani Maski" wrote:
> Hi Eric,
>
> I have also developed mini-applications replacing GSA for some of our
> clients using Apache Nutch + Solr to crawl multi lingual sites and enable
> multi-lingual search. Nutch+Solr is ver
I too have come across this same exact problem. One thing that I have
found is that with autoGeneratePhraseQueries=true, you can find the case
where your index has 'z score' and your query is z-score, but with false it
will not find it. As to your specific problem with the single token zscore
in
As Shawn pointed out, seems like your client is actually sending out "*:*"
queries all of the times.
You perhaps have the wrong id for the search box or something that results
in your ajax library to never actually receive the actual input value, but
I'm just guessing.
On Thu, Oct 31, 2013 at 1:
Hi Eric,
I have also developed mini-applications replacing GSA for some of our
clients using Apache Nutch + Solr to crawl multi lingual sites and enable
multi-lingual search. Nutch+Solr is very stable and Nutch mailing list
provides a good support.
Reference link to start:
https://sites.google.
On 10/30/2013 1:55 PM, Reyes, Mark wrote:
solr.log file per Solr 4.5
http://pastebin.com/zSpERJZA
Your queries all look like the following, with different numbers for the
parameters json.wrf and _ (underscore) that I've never seen before, and
I assume Solr just ignores.
{json.wrf=jQuery171
On 10/30/2013 1:49 PM, Shalom Ben-Zvi Kazaz wrote:
we are continuously getting this exception during replication from
master to slave. our index size is 9.27 G and we are trying to replicate
a slave from scratch.
Its a different file each time , sometimes we get to 60% replication
before it fails
solr.log file per Solr 4.5
http://pastebin.com/zSpERJZA
Thanks Shawn,
Mark
On 10/30/13, 12:44 PM, "Shawn Heisey" wrote:
>On 10/30/2013 1:26 PM, Reyes, Mark wrote:
>> I am currently integrating JavaScript framework AJAX Solr to my domain.
>>I am trying to query words such as 'doctorate' or '
we are continuously getting this exception during replication from
master to slave. our index size is 9.27 G and we are trying to replicate
a slave from scratch.
Its a different file each time , sometimes we get to 60% replication
before it fails and sometimes only 10%, we never managed a successfu
On 10/30/2013 1:26 PM, Reyes, Mark wrote:
I am currently integrating JavaScript framework AJAX Solr to my domain. I am
trying to query words such as 'doctorate' or 'programs' but the console is
reporting '*:*' only the default wildcard.
Just curious if anyone has any helpful hints? The problem
... AND appologies to everyone for erroneously posting irrelevant stuff on the
list.
Michael
Fra: Michael Preminger [michael.premin...@hioa.no]
Sendt: 30. oktober 2013 20:34
To: solr-user@lucene.apache.org
Emne: SV: Evaluating a SOLR index with trec_eval
Hi, Tom!
Thanks alot. Ill check Ian's stuff and anticpate yours ...
As you know, the ProveIt is now terminated as an INEX track, but we still hope
to write a paper to a journal, summarizing what was done, and it would be nice
to have you on.
AND, youll be happy (or shocked) to know that this we
Hi Michael,
I know you are asking about Solr, but in case you haven't seen it, Ian
Soboroff has a nice little demo for Lucene:
https://github.com/isoboroff/trec-demo.
There is also the lucene benchmark code:
http://lucene.apache.org/core/4_5_1/benchmark/org/apache/lucene/benchmark/quality/packag
I am currently integrating JavaScript framework AJAX Solr to my domain. I am
trying to query words such as 'doctorate' or 'programs' but the console is
reporting '*:*' only the default wildcard.
Just curious if anyone has any helpful hints? The problem can be seen in detail
on Stackoverflow,
ht
Hi Alessandro,
What version of Solr are you running and what's the version of SolrJ? I am
guessing they are different.
On Wed, Oct 30, 2013 at 8:32 PM, Alessandro Benedetti <
benedetti.ale...@gmail.com> wrote:
> I have a zookeeper ensemble hotes in one amazon server.
> Using the CloudSolrServ
Markus and Jason
thanks for the info.
I will start to research Nutch. Writing a crawler, agree it is a rabbit
hole.
--
Eric Palmer
Web Services
U of Richmond
To report technical issues, obtain technical support or make requests for
enhancements please visit
http://web.richmond.edu/contact/t
Hi Michael,
Here's a good post by Erick Erickson about understanding commits and
transaction logs in Solr.
http://searchhub.org/2013/08/23/understanding-transaction-logs-softcommit-and-commit-in-sorlcloud/
About the replication, as soon as you post an update, here's what happens:
1. The update ge
Nutch is an excellent option. It should feel very comfortable for people
migrating away from the Google appliances.
Apache Droids is another possible way to approach, and I’ve found people using
Heretrix or Manifold for various use cases (and usually in combination with
other use cases where t
Hi Eric,
We have also helped some government institution to replave their expensive GSA
with open source software. In our case we use Apache Nutch 1.7 to crawl the
websites and index to Apache Solr. It is very effective, robust and scales
easily with Hadoop if you have to. Nutch may not be the
Hello all,
Been lurking on the list for awhile.
We are at the end of life for replacing two google mini search appliances used
to index our public web sites. Google is no longer selling the mini appliances
and buying the big appliance is not cost beneficial.
http://search.richmond.edu/
We wou
A function query is simply returning a calculated result based on existing
data - no new fields required.
Did you actually want to precompute a value, store it in the index, and then
query on it? If so, you could do that indexing with a custom or scripted
update processor.
Flesh out an examp
On 10/30/2013 10:44 AM, Elena Camossi wrote:
Actually, I had tried the instanceDir="ckan" but it didn't work either (with
the same error, just reporting a wrong path to solrconf.xml).
I used this configuration taking suggestion from here
http://stackoverflow.com/questions/16230493/apache-solr-una
Hi,
Hm, sorry for not helping with this particular issue directly, but it
looks like you are *uploading* your logs and indexing that way?
Wouldn't pushing them be a better fit when it comes to log indexing?
We recently contributed a Logstash output that can index logs to Solr,
which may be of inte
Sounds really close to what I'm looking for, but this sounds like it would
result in a new field on a document (or a new value for a field defined to
hold the result of a function). Would it be possible for a function query
to produce a new document so that I can associate the computed value with
Dear Shawn,
thanks a lot for your quick answer.
> -Original Message-
> From: Shawn Heisey [mailto:s...@elyograg.org]
> Sent: mercoledì 30 ottobre 2013 17:12
> To: solr-user@lucene.apache.org
> Subject: Re: solr 4.5.0 configuration Error:
> org.apache.solr.common.SolrException:org.apache.
I have set at multipartUploadLimitInKB parameter to 10240 (which was 2048
earlier)
multipartUploadLimitInKB="10240". Now it gives following error for same
files at place.
http://localhost:8983/solr/logsIndexing returned non ok status:500,
message:the request was rejected because its size (1048604
On 10/30/2013 9:24 AM, Elena Camossi wrote:
Hi everyone,
I'm trying to configure Solr 4.5.0 on Linux red Hat to work with CKAN and
Tomcat, but Solr cannot initialize the core (I'm configuring just one core,
but this is likely to change in the next future. I'm using contexts for this
set up). Tom
You could create a custom "value source" and then use it in a function query
embedded in your return fields list (fl).
So, the function query could use a function (value source) that takes a
field, fetches its value, performs some arbitrary calculation, and then
returns that value.
fl=id,nam
Hi,
The query z-score doesn't match a doc with zscore in the index. The analysis
tool shows that this query would match this data in the index, but it's the
edismax query parser step that seems to screw things up. Is there some
combination of autoGeneratePhraseQueries, WordDelimiterFilterFactor
Hi,
The query z-score doesn't match a doc with zscore in the index. The analysis
tool shows that this query would match this data in the index, but it's the
edismax query parser step that seems to screw things up. Is there some
combination of autoGeneratePhraseQueries, WordDelimiterFilterF
I'm currently using a SolrCoud setup and I index my data using a couple of
in-house indexing clients.
The clients process some files and post json messages containing added
documents in batches.
Initially my batch size was 100k docs and the post request took about 20-30
secs.
I switched to 10k batc
I have a zookeeper ensemble hotes in one amazon server.
Using the CloudSolrServer and trying to connect , I obtain this nreally
unusual error :
969 [main] INFO org.apache.solr.common.cloud.ConnectionManager - Client is
connected to ZooKeeper
1043 [main] INFO org.apache.solr.common.cloud.ZkStateRea
Hello!
Is there a simple way to evaluate a SOLR index with TREC_EVAL?
I mean:
* preparing a query file in some format Solr will understand, but where each
query has an ID
* getting results out in trec format, with these query IDs attached
Thanks
Michael
On Wed, 2013-10-30 at 14:24 +0100, Shawn Heisey wrote:
> On 10/30/2013 4:00 AM, Toke Eskildsen wrote:
> > Why would TRIM have any influence on whether or not a driver failure
> > also means server failure?
>
> I left out a step in my description.
>
> Lack of TRIM support in RAID means that I woul
I'd like to throw out a design question and see if its possible to solve
this with Solr.
I have a set of data that is computed that I'd like to make searchable.
Ideally, I'd like to have all documents indexed and call it the day, but
the nature of the data is such that it needs to be computed give
Hello,
As suggested by Chris, now I am accessing the files using java program and
creating SolrInputDocument, but i ran into this exception while doing
server.add(document). When i tried to increase "ramBufferSizeMB", it doesn't
let me make it more than 2 gig.
org.apache.solr.client.solrj.SolrSer
Escape any special characters with a backslash, or put the full term in
quotes.
-- Jack Krupansky
-Original Message-
From: jihyun suh
Sent: Wednesday, October 30, 2013 6:28 AM
To: solr-user@lucene.apache.org
Subject: query with colon in bq
I have a question about query with colon in
On 10/30/2013 4:00 AM, Toke Eskildsen wrote:
> On Tue, 2013-10-29 at 16:41 +0100, Shawn Heisey wrote:
>> If you put the index on SSD, you could get by with less RAM, but a RAID
>> solution that works properly with SSD (TRIM support) is hard to find, so
>> SSD failure in most situations effectively
Hi,
First, the feature will only detect ONE language per field, even if it is a
multi-valued field. In your case there is VERY little text for the detector, so
do not expect great detection quality. But I believe the detector chose ES as
language and mapped the whole field as tag_es. The reason
Wow again!
Thank you all very much for your insights.
We will certainly take all of this under consideration.
Erik: I want to upgrade but unfortunately, it's not up to me. You're right,
we definitely need to do it.
And SolrJ sounds interesting, thanks for the suggestions.
By the way, is ther
Hi,
I made few changes to the solrconfig.xml, created a jar file,added it to
the lib folder of the solr and tried to start it.
THe changes in the solrconfig.xml are
LEAD_NOTES
category
Others
naiveBayesModel
ap
Oops... need to note that the parameters have changed since Solr 4.4 - I
gave the link for 4.5.1, but for 4.4 and earlier, use:
http://lucene.eu.apache.org/solr/4_4_0/solr-core/org/apache/solr/update/processor/UniqFieldsUpdateProcessorFactory.html
(My book is for 4.4, but hasn't been updated fo
Unfortunately, atomic "add" is add to a "list" (append) rather than add to a
"set" (only unique values). But, you can use the unique fields update
processor (solr.UniqFieldsUpdateProcessorFactory) to de-dupe specified
multivalued fields.
See:
http://lucene.apache.org/solr/4_5_1/solr-core/org/a
that is what i'd call a compound key? :) using multiple attribute to generate a
unique key across multiple tables ..
On Wednesday, October 30, 2013 at 2:10 AM, dtphat wrote:
> yes, I've just used concat(id, '_', tableName) instead using compound key. I
> think this is an easy way.
> Thanks.
>
Hi All,
What should be the field type if I have to save solr's open bit set value
within solr document object and retrieve it later for search?
OpenBitSet bits = new OpenBitSet();
bits.set(0);
bits.set(1000);
doc.addField("SolrBitSets", bits);
What should be the field type of SolrBit
Robert:
Thanks. I'm on my way out the door, so I'll have to put up a JIRA with your
patch later if it hasn't been done already
Erick
On Tue, Oct 29, 2013 at 10:14 PM, Robert Muir wrote:
> I think its a bug, but thats just my opinion. i sent a patch to dev@
> for thoughts.
>
> On Tue, Oct 29,
Think it'll be a good thing to have.
I just created a JIRA for that.
https://issues.apache.org/jira/browse/SOLR-5403
Will try and get to it soon.
On Wed, Oct 30, 2013 at 4:28 PM, Shalin Shekhar Mangar <
shalinman...@gmail.com> wrote:
> Ah I misread your email. You are actually sending the updat
Ah I misread your email. You are actually sending the update twice and
asking about how to dedup the multi-valued field values.
No I don't think we have an update processor which can do that.
On Wed, Oct 30, 2013 at 4:18 PM, Shalin Shekhar Mangar <
shalinman...@gmail.com> wrote:
> Perhaps you a
I am not sure if optimistic concurrency would help in deduplicating but
yes, as Shalin points out, you'll be able to spot issues with your client
code.
On Wed, Oct 30, 2013 at 4:18 PM, Shalin Shekhar Mangar <
shalinman...@gmail.com> wrote:
> Perhaps you are running the update request more than
Perhaps you are running the update request more than once accidentally?
Can you try using optimistic update with _version_ while sending the
update? This way, if some part of your code is making a duplicate request
then Solr would throw an error.
See
https://cwiki.apache.org/confluence/display/so
I have a question about query with colon in bq.
Actually I use edismax and I set the q and bq just like this,
.../select?defType=edismax&q="1:100"^100 "1 100"^30&qf=Title^2.0
Body&bq=Title:(1:100)^6.0 Body:(1:100)^6.0
in this query phrase, I got the error in bq, undefined field 1.
How do I use qu
I am working on a offline tagging capability to tag records with a
thesaurus dictionary of key concepts. I am able to use the update="add"
option using xml and json update calls for a field to update specific
document field information. Although if I run the same atomic update query
twice then the
On Tue, 2013-10-29 at 16:41 +0100, Shawn Heisey wrote:
> If you put the index on SSD, you could get by with less RAM, but a RAID
> solution that works properly with SSD (TRIM support) is hard to find, so
> SSD failure in most situations effectively means a server failure. Solr
> and Lucene have a
On Wed, Oct 30, 2013 at 4:50 PM, Vineet Mishra wrote:
>
> I am making web server call to a website for Shortening the links, that is
> bit.ly but recieving a 403 Request Forbidden.
> Although if I use their webpage to short the web link its working good.
> Can any body tell me what might be the re
Hi All,
I am making web server call to a website for Shortening the links, that is
bit.ly but recieving a 403 Request Forbidden.
Although if I use their webpage to short the web link its working good.
Can any body tell me what might be the reason for such a vague behavior.
Here is the code includ
Hi,
We have a requirement where we need to send the matched synonyms as part of
Solr response.
Do we need to customize the Solr response handler to do this?
Regards,
Siva
--
View this message in context:
http://lucene.472066.n3.nabble.com/Return-the-synonyms-as-part-of-Solr-response-tp40983
Hi,
I experience the same problem, using version 4.4.0.
In my case:
2 Solr nodes -> 4 collections, each 1 shard and 2 replicas.
3 Zookeepers
Replicas can get state=down when a connection to Zookeeper is lost.
However, there are 2 more Zookeeper servers, so this shouldn't be a
problem right?
Hi,
I have a problem with SolrCloud in an specific test case and I wanted to
know if it is the way it should work or if is there any way to avoid this...
I have the next scenario:
- Three machines
- Each one with one zookeeper and one solr 4.1.0
- Each Solr stores 7 Million documents and the ind
On Tue, 2013-10-29 at 14:24 +0100, eShard wrote:
> I have a 1 TB repository with approximately 500,000 documents (that will
> probably grow from there) that needs to be indexed.
As Shawn point out, that isn't telling us much. If you describe the
documents, how and how often you index and how you
Hi Bayu ,
I did that but for solr 4.2, the catalaina.out has no exceptions at all.
Thanks
--
View this message in context:
http://lucene.472066.n3.nabble.com/Can-not-find-solr-core-on-admin-page-after-setup-tp4098236p4098385.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi Engy,
Have you copy solr's war (e.g. solr-4.5.1.war, for latest Solr
distribution) from Solr source distribution to Tomcat's webapps directory
(rename to solr.war on webapps dir.)?
After put that file and restarted the Tomcat, it will create 'solr' folder
under webapps.
Or, if you still found
72 matches
Mail list logo