Hi,
I want to compute my own stats in addition to solr default stats. How can i
enhance statistics in solr? How this thing can be achieved.. Solr compute
stats as cumulative, is there is any way to get per instant stats...??
Thanks... waiting for good replies..
--
View this message in con
anyone knows?
On Thu, Jul 19, 2012 at 5:48 PM, Roy Liu wrote:
> Hi,
>
> When I use Transformer to handle files, I always get NULL with
> row.get(columnName).
> anyone knows?
>
> --
> The following file is *data-config.xml*
>
>
> name="ds"
> driver="oracle.jdbc
Thanks. Are you saying that if we run low on memory, the MMapDirectory will
stop using it? The least used memory will be removed from the OS automatically?
Isee some paging. Wouldn't paging slow down the querying?
My index is 10gb and every 8 hours we get most of it in shared memory. The
memory
Hi,
I have requirements to place a document to a pre-determined position for
special filter query values, for instance when filter query is
fq=(field1:"xyz") place document abc as first result (the rest of the
result set will be ordered by sort=field2). I guess I have to plug in my
Java code as a
: When I run dismax queries I see there are no lookups in the
: queryResultCache. If I remove the field collapsing - lookups happen. I
: can't find any mention of this anywhere or think of reason why this should
I'm not very familiar with the grouping code, but i think the
crux of what you are
What is the best way to redirect a SolrQueryRequest to another core from
within a handler (custom SearchHandler)?
I've tried to find the SolrCore of the core I want to redirect to and
called the execute() method with the same params but it looks like the
SolrQueryRequest object already has the ol
Hi Bruno,
It's usually the maxThreads attribute in the tag in
$CATALINA_HOME/conf/server.xml. But I kind of doubt you're running out
of threads... maybe you could post some more details about the system
you're running Solr on.
Michael Della Bitta
Dear Solr User,
I don't know if it's here that my question must be posted but I'm sure
some users have already had my problem.
Actually, I do 1556 requests with 4 Http components with my program. If
I do these requests without delay (500ms)
before sending each requests I have around 10% of re
Thanks Mark!
On Thu, Jul 19, 2012 at 4:07 PM, Mark Miller wrote:
> https://issues.apache.org/jira/browse/SOLR-3649
>
> On Thu, Jul 19, 2012 at 3:34 PM, Briggs Thompson <
> w.briggs.thomp...@gmail.com> wrote:
>
> > This is unrelated for the most part, but the javabin update request
> handler
> >
: >
facet.field=testfield&facet.field=%7B!key=mylabel%7Dtestfield&f.mylabel.limit=1
: >
: > but the limit on the alias didn't seem to work. Is this expected?
:
: Per-field params don't currently look under the alias. I believe
: there's a JIRA open for this.
https://issues.apache.org/jira/bro
I did a search via both admin UI and /search
What I searched for was *:* as that was default in the search box in the admin
ui (so expected something that was not an 303 error).
Will post url and server logs tomorrow when I am back in the office.
But i think the admin url was not anything odd.
: So from StatsComponent the count for 'electronics' cat is 3, while
: FacetComponent report 14 'electronics'. Is this a bug?
:
: Following is the field definition for 'cat'.
:
FYI...
https://issues.apache.org/jira/browse/SOLR-3642
(The underlying problem is that the stats.facet feature doesn
: > try to do a search - throws 303 error
Can you be specific about how exactly you did the search?
Was this from the admin UI? what URL was in your browser location bar?
what values did you put in the form? what buttons did you click? what URL
was in your browser location bar when the error
Okay - I'll do the same in a bit and report back.
On Jul 19, 2012, at 5:23 PM, John-Paul Drawneek wrote:
> This is just out of the box.
>
> All I did was download solr 4 Alpha from the site.
> unpack
> follow instructions from wiki.
>
> admin console worked - great
>
> try to do a search - thr
This is just out of the box.
All I did was download solr 4 Alpha from the site.
unpack
follow instructions from wiki.
admin console worked - great
try to do a search - throws 303 error
Downloaded nightly build, same issue.
Also got errors from the other shard with error connecting due to maste
That's really odd - never seen or heard anything like it. A 303 is what a
server will respond with if you should GET a different URI...
This won't happen out of the box that I've ever seen...can you tells us
about any customization's you have made?
On Thu, Jul 19, 2012 at 1:08 PM, John-Paul Drawn
https://issues.apache.org/jira/browse/SOLR-3649
On Thu, Jul 19, 2012 at 3:34 PM, Briggs Thompson <
w.briggs.thomp...@gmail.com> wrote:
> This is unrelated for the most part, but the javabin update request handler
> does not seem to be working properly when calling solrj
> method*HttpSolrServer.de
Your password has an & in it. Since this is an XML file, you need to
turn it into an XML entity, so your password should be entered as:
8ty&2ty=6
Michael Della Bitta
Appinions, Inc. -- Where Influence Isn’t a Game.
http://www.appinions.com
On Th
This is unrelated for the most part, but the javabin update request handler
does not seem to be working properly when calling solrj
method*HttpSolrServer.deleteById(List ids)
*. A single Id gets deleted from the index as opposed to the full list. It
appears properly in the logs - shows delete of al
Robert,
So this is lossy: basically you can think of there being only 256
> possible values. So when you increased the number of terms only
> slightly by changing your analysis, this happened to bump you over the
> edge rounding you up to the next value.
>
> more information:
> http://lucene.apach
Hi Brandan,
I am not sure if get whats being suggested. Our delete worked fine, but now
no new data is going into the system.
Could you please throw some more light.
Regards,
Rohit
-Original Message-
From: Brendan Grainger [mailto:brendan.grain...@gmail.com]
Sent: 19 July 2012 17:33
T
>
> and get the following errors:
> ---
>
> [javac] warning: [options] bootstrap class path not set in conjunction
> with -source 1.6
> [javac]
> /home/swu/newproject/lucene_4x/lucene/analysis/opennlp/src/java/org/apache/lucene/analysis/opennlp/OpenNLPTokenizer.java:170:
> erro
I am following instruction
http://wiki.apache.org/solr/OpenNLP to test OpenNLP, Solr integration
1. pull 4.0 branch from trunk
2. apply patch LUCENE-2899 patch
(there are several LUCENE-2899 patch files, I took the one, 385KB,
02/Jul/12 08:05, I should only apply this one, correct ?)
3. ant compil
You might be running into the same issue someone else had the other day:
https://issues.apache.org/jira/browse/SOLR-3432
On Jul 19, 2012, at 1:23 PM, Rohit wrote:
> We delete some data from solr, post which solr is not accepting any
> commit's. What could be wrong?
>
>
>
> We don't see any
We delete some data from solr, post which solr is not accepting any
commit's. What could be wrong?
We don't see any error in logs or anywhere else.
Regards,
Rohit
Hello,
I have a similar problem, anything new about this issue?
My problem is that info logs go to stderr and not stdout, do you have an
explanation?
For the log level I use the file "logging.properties" with in it only one
line setting the level.
.l
First, turn off all your soft commit stuff, that won't help in your situation.
If you do leave autocommit on, make it a really high number
(let's say 1,000,000 to start).
You won't have to make 300M calls, you can batch, say, 1,000 docs
into each request.
DIH supports a bunch of different data so
Hi Jonatan,
Ideally you'd use a Solr API client that allowed batched updates, so
you'd be sending documents 100 at a time, say. Alternatively, if
you're good with Java, you could build an index by using the
EmbeddedSolrServer class in the same process as the code you use to
parse the documents. Bu
Hello,
I was wondering if there's other ways to import data in Solr than
posting xml/json/csv to the server URL (e.g. locally building the
index). Is the DataImporter only for database?
My data is in an enormous text file that is parsed in python, I get
clean json/xml out of it if I want, but the
On Thu, Jul 19, 2012 at 11:11 AM, Aaron Daubman wrote:
> Apologies if I didn't clearly state my goal/concern: I am not looking for
> the exact same scoring - I am looking to explain scoring differences.
> Deprecated components will eventually go away, time moves on, etc...
> etc... I would like
Read this, then you will see that MMapDirectory will use 0% of your Java Heap
space or free system RAM:
http://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html
Uwe
-
Uwe Schindler
H.-H.-Meier-Allee 63, D-28213 Bremen
http://www.thetaphi.de
eMail: u...@thetaphi.de
> -Or
Robert,
> I have a solr 1.4.1 instance and a solr 3.6.0 instance, both configured as
> > identically as possible (given deprecations) and indexing the same
> document.
>
> Why did you do this? If you want the exact same scoring, use the exact
> same analysis.
> This means specifying luceneMatchVer
Maybe I'm not understanding the problem, but I accomplish this by
having two fields. One for sorting, like so:
And then a string type field for faceting. Use a copyField directive
to get the same data in both, and then sort on the sort field, and
facet on the string field. The Mapping
Thanks for the reply.
To clarify, the idea is to search for authors with certain specialties (eg.
political, horror, etc.) and if they have any published titles relevant to
the user's query, then display those titles next to the author's name.
At first, I thought it would be great to have all t
On Thu, 2012-07-19 at 16:00 +0200, Spadez wrote:
> This seems to suggest you have to reindex Solr in its entirety and cant add a
> single document at a time, is this right?
>
> http://stackoverflow.com/questions/11247625/apache-solr-adding-editing-deleting-records-frequently
No. What is says is t
You can definitely do a single document at a time, but unless you're
using NRT, your changes won't be visible until you do a commit. Doing
a commit involves closing Searchers and reopening them, which is semi
expensive... depending on how you're doing caching, you wouldn't want
to do it too frequen
This seems to suggest you have to reindex Solr in its entirety and cant add a
single document at a time, is this right?
http://stackoverflow.com/questions/11247625/apache-solr-adding-editing-deleting-records-frequently
--
View this message in context:
http://lucene.472066.n3.nabble.com/Importing
we really need to resolve that issue soon...
On Jul 19, 2012, at 12:08 AM, Briggs Thompson wrote:
> Yury,
>
> Thank you so much! That was it. Man, I spent a good long while trouble
> shooting this. Probably would have spent quite a bit more time. I
> appreciate your help!!
>
> -Briggs
>
> On W
I'm not sure your point <3> makes sense. If you're searching by
author, how do you define "the four most relevant titles"? Relevant
to what?
If you are searching text of the publications, then displaying authors with
no publications seems unhelpful.
If you're searching the bios, how do you define
Can you file two JIRA issues for these?
bq. but does return reasonable results when distrib is turned off like so
It should default to distrib=false - I don't think /terms is distrib
aware/compatible.
bq. /browse returns this stack trace to the browser HTTP ERROR 500
We may be able to fix this
On Thu, Jul 19, 2012 at 12:10 AM, Aaron Daubman wrote:
> Greetings,
>
> I've been digging in to this for two days now and have come up short -
> hopefully there is some simple answer I am just not seeing:
>
> I have a solr 1.4.1 instance and a solr 3.6.0 instance, both configured as
> identically
Check your csv file for extraneous data? The other thing to do is look at
your logs to see if more informative information is there.
THere's really very little info to go on here, you might review:
http://wiki.apache.org/solr/UsingMailingListshttp://wiki.apache.org/solr/UsingMailingLists
Best
Er
A multiValued really doesn't make any sense. But your
log file should have something in it like this:
SEVERE: uniqueKey should not be multivalued
although it _is_ a bit hard to see on startup unless you've suppressed
the INFO level output.
See: https://issues.apache.org/jira/browse/SOLR-1570
Bes
While porting from 3.6.1 to 4.x I noticed the doubling content of some fields
in my index.
Didn't have this with 3.6.1.
This can also be seen with luke. I could trace it down to DIH so far.
Anyone seen this?
I'm using XPathEntityProcessor with RegexTransformer.
Will look into this closer tomorro
On Thu, 2012-07-19 at 13:49 +0200, Spadez wrote:
> It does seem really poor design to reimport 10,000 documents, when only one
> needs to be added. I dont like that, can you not insert a specific entry
> into Solr rather than reimporting everything?
Isn't that what you outlined in your option #1?
Hi,
We have used *dismax* in our SOLR config with /defaultOperator="OR"/ and
some *mm * settings. Recently, we have started using *defType=edismax * in
query params. With this change, we have observed significant drop in results
count. We doubt that SOLR is using default operator="AND" and hence r
On 19 July 2012 10:15, 晋鹏(Tomsdinary) wrote:
>
> Hi
> I wait to join this mailing list.
Please see the very first entry under
http://lucene.apache.org/solr/discussion.html
Regards,
Gora
Hi
I wait to join this mailing list.
This email (including any attachments) is confidential and may be legally
privileged. If you received this email in error, please delete it immediately
and do not copy it or use it for any purpose or disclose its con
Thank you for the reply. Ok, well that brings another question. I dont like
pre-optimisation, but I also dont like inefficiency, so lets see if I can
strike a balance.
It does seem really poor design to reimport 10,000 documents, when only one
needs to be added. I dont like that, can you not inser
I have set some of my fields to be NGram Indexed. Have also set analyzer both
at query as well as index level.
Most of the stuff works fine except for use cases where I simply interchange
couple of characters.
For an example: "springfield" retrieves correct matches, "springfi" retrieves
correc
On Thu, 2012-07-19 at 12:54 +0200, Spadez wrote:
> I want to import any new SQL results onto the server as quickly as possible
> so they are searchable but I dont want to overload the server. These are my
> new options:
>
> 1. Devise a script to run when a new SQL item is posted, to immediatly
> i
Hi,
Why is maxScore always returned with distributed search? It used to return only
if score was part of fl. Bug? Feature?
Thanks
Markus
Hi,
Lets say I am running an auction site. There are 20,000 entries. 100 entries
come from an on-site SQL database, the rest come from a generated txt file
from scrapped content.
I want to import any new SQL results onto the server as quickly as possible
so they are searchable but I dont want to
I have seen that issue several times, in my case it was always with an id
field, mysql db and linux. Same config but on windows did not show that
issue.
Never got to the bottom of it...as it was an id it was just working as it
was unique.
--
View this message in context:
http://lucene.472066.n
On Wed, 2012-07-18 at 20:30 +0200, Christopher Gross wrote:
> When I do a query, the results that come through retain their original
> case for this field, like:
> doc 1
> keyword: Blah Blah Blah
> doc 2
> keyword: Yadda Yadda Yadda
>
> But when I pull back facets, i get:
>
> blah blah blah (1)
>
55 matches
Mail list logo