> We construct our query by Lucene API
> before, as BooleanQuery, TermQuery
> those kind of things.
Okey, it seems that your field are not analyzed and you don't do any analysis
while construction of your query by Lucene API. Correct?
Then you can use your existing Java code directly inside a s
Where do you get your Lucene/Solr downloads from?
[] ASF Mirrors (linked in our release announcements or via the Lucene website)
[X] Maven repository (whether you use Maven, Ant+Ivy, Buildr, etc.)
[X] I/we build them from source via an SVN/Git checkout.
[] Other (someone in your company mirrors
Thanks a lot for your reply.That was very helpful.
We construct our lucene query after certain analysis(ex : words
segmentation, category identification). Do you mean we plugin those analysis
logic and query construction part onto solr, and solr takes the very
beginning input.
Kun
2011/1/20 Ahme
Thanks Jack!
I will give it a try, even though I finally have a Nutch configuration
that does exactly what I want it to do (except keeping an eye on updated
and deleted documents).
Erlend
On 19.01.11 16.52, Jack Krupansky wrote:
Take a look at Apache ManifoldCF (incubating, close to 0.1 re
Hasnain,
there is no need for any _additional_ looping? Of course, you have to loop
over initially, to get the results - but this should be enough.
use result/doc/[@name=id] to check if lst[@name=highlighting]/lst[@name=$id]
exists, and if so .. replace the original content w/ the highlighted one
Did i write wt? Oh dear. The q and w are too close =)
> Markus,
>
> Its not wt its qt, wt for response type,
> Also qt is not for Query Parser its for Request Handler ,In solrconfig.xml
> there are many Request Handlers can be Defined using "dismax" Query Parser
> Or Using "lucene" Query Parser.
>
Ok,
I tryed to use nested queries this way:
wt=json&indent=true&fl=qFR&q=sarkozy
_query_:"{!tag=test}chirac"&facet=true&facet.field={!ex=test}studyDescriptionId
It resulted in this error:
"facet_counts":{
"facet_queries":{},
"exception":"java.lang.NullPointerException\n\tat
org.apache.solr.re
Hi list,
after sending full-import=true&clean=true&commit=true
Solr 4.x (apache-solr-4.0-2010-11-24_09-25-17) responds with:
- DataImporter doFullImport
- DirectUpdateHandler2 deleteAll
...
- DocBuilder finish
- SolrDeletionPolicy.onCommit: commits:num=2
- SolrDeletionPolicy updateCommits
- SolrIn
Hello list,
I got a theoretical question about a Multicore-Situation:
I got two cores: active, inactive
The active core serves all the queries.
The inactive core is the tricky thing:
I create an optimized index outside the environment and want to insert that
optimized index 1 to 1 into the ina
Hi David,
if your transaction id will be handled at document level, you can just add a
field in your schema named "transaction_id" - that's it. All you have to do
is to insert that transaction_id everytime you do an update (Solr does not
generate an transaction_id by default and I don't know of a
Hi all,
we have implemented a Solr based search in our web application. We have one
master server that maintains the index which is replicated to the slaves using
the built-in Solr replication.
This has been working fine so far, but suddenly the replication does not send
the modified files to
Hallo..
I have create multicore search and will search in more then one Core!
Now i have done:
http://192.168.105.59:8080/solr/mail/select?wt=phps&q=*:*&shards=192.168.105.59:8080/solr/mail,192.168.105.59:8080/solr/mail11
But Error...
HTTP Status 500 - Map size must not be negative
java.lang.I
That looks like this issue:
https://issues.apache.org/jira/browse/SOLR-2278
On Thursday 20 January 2011 13:02:41 Jörg Agatz wrote:
> Hallo..
>
> I have create multicore search and will search in more then one Core!
>
> Now i have done:
>
> http://192.168.105.59:8080/solr/mail/select?wt=phps&q=*
Since it seems to be no voodoo available I did it on the client side.
I send a first request to get the facets and a second to get the documents and
their highlighting.
It works well but requires more processing.
>
> From: Xavier SCHEPLER
> Sent: Thu Jan
On which events did you configure master to perform replication? replicateAfter
Regards,
Stevo.
On Thu, Jan 20, 2011 at 12:53 PM, Thomas Kellerer wrote:
> Hi all,
>
> we have implemented a Solr based search in our web application. We have one
> master server that maintains the index which is rep
I will be looking into JConsole.
One more question regarding caching. When we talk about warm-up queries does
that mean that some of the complex queries (esp those which require high I/O
e.g. phrase queries) will really be very slow (on lets say an index of
200GB) if they are not cached? I am talk
Here is our configuration:
true
commit
startup
stopwords.txt,stopwords_de.txt,stopwords_en.txt,synonyms.txt
Stevo Slavić, 20.01.2011 13:26:
On which events did you configure master to perform replication? replicateAfter
Regards,
Stevo.
On Thu, Jan 20, 2011 at 12:53 PM, Thomas Ke
Thanks guys. I read (and actually digested this time) how multivalued fields
work and now realize my question came from a 'structured language/dbms'
background. The multivalued field is stored basically as a single value with
extra spacing between terms (the positionIncrementGap previously mention
Is that it? Of all the strange, esoteric, little understood configuration
settings available in solrconfig.xml, the only thing that affects Index
Speed vs Query Speed is turning on/off the Query Cache and RamBufferSize?
And for the latter, why wouldn't RamBufferSize be the same for both...that
is,
Thomas Kellerer, 20.01.2011 12:53:
Hi all,
we have implemented a Solr based search in our web application. We
have one master server that maintains the index which is replicated
to the slaves using the built-in Solr replication.
This has been working fine so far, but suddenly the replication do
I think that the actual name of the property is "solr.data.dir". You can
check this in the element in solrconfig.xml.
Best,
*Juan*
On Wed, Jan 19, 2011 at 4:04 PM, Fred Gilmore wrote:
> I've checked the archive, and plenty of people have suggested an
> arrangement where you can have two core
Thanks for the responses.
Right now each document has a fairly small amount of indexed data such
as title, author, language, subjects, and various media
characteristics. Indexing or reindexing a document is very fast,
updating a batch of 100 documents takes less than 1/10th of a second.
What impac
So if on startup index gets replicated, then commit probably isn't
being called anywhere on master.
Is that index configured to autocommit on master, or do you commit
from application code? If you commit from application code, check if
commit actually gets issued to the slave.
Regards,
Stevo.
On
Stevo Slavić, 20.01.2011 15:42:
So if on startup index gets replicated, then commit probably isn't
being called anywhere on master.
No, the index is not replicated on startup (same behaviour: "no files to
download")
Is that index configured to autocommit on master, or do you commit
from appl
Not too elegant but valid check would be to bring slave down, delete
it's index data directory, then to commit a change to index to master,
then bring slave up, and wait for pollInterval to pass and replication
to occur.
Regards,
Stevo.
On Thu, Jan 20, 2011 at 4:16 PM, Thomas Kellerer wrote:
> S
Thanks to both you and Markus.
Fred
On 1/20/2011 7:47 AM, Juan Grande wrote:
I think that the actual name of the property is "solr.data.dir". You
can check this in the element in solrconfig.xml.
Best,
*Juan*
On Wed, Jan 19, 2011 at 4:04 PM, Fred Gilmore
mailto:fgilm...@mail.utexas.edu>>
On 1/20/2011 1:42 AM, kun xiong wrote:
Thar example string means our query is BooleanQuery containing
BooleanQuerys.
I am wondering how to write a complicated BooleanQuery for dismax, like (A
or B or C) and (D or E)
Or I have to use Lucene query parser.
You can't do it with dismax. You might
On Thu, Jan 20, 2011 at 4:59 AM, Xavier SCHEPLER
wrote:
> Ok,
> I tryed to use nested queries this way:
> wt=json&indent=true&fl=qFR&q=sarkozy
> _query_:"{!tag=test}chirac"&facet=true&facet.field={!ex=test}studyDescriptionId
> It resulted in this error:
> "facet_counts":{
> "facet_queries":{},
>
Hi,
I am building a social networking site.For searching profiles, i am trying
to implement solr.
But here i am facing a problem.As a social networking site, the data base is
going to get more updates/inserts frequently.That means,the search is going
to be in real time.How can we achieve this usin
http://wiki.apache.org/solr/NearRealtimeSearchTuning
http://lucene.472066.n3.nabble.com/Tuning-Solr-caches-with-high-commit-rates-NRT-td1461275.html
http://lucene.472066.n3.nabble.com/NRT-td2276967.html#a2278477
-
--- System
I want to provide a list of facets to my visitors order alphabetically, for
example, for the 'features' facet I have:
data-config.xml:
schema.xml:
But this doesnt give me the facets in an alphabetical order.
Besides the features facet, I also have some
Hi,
I am building tag cloud for products by using facets.I made tag names as
facets and i am taking facets count as reference to display tag cloud.Each
product has tags with their own weightage.Let us say,
For example
prod1 has tag called “Light Weight” with weightage 20,
prod2 has tag called “
> We construct our lucene query after certain analysis(ex :
> words
> segmentation, category identification).
By analysis, I referring charfilter(s)+tokenizer+tokenfilter(s) combination.
> Do you mean we
> plugin those analysis
> logic and query construction part onto solr, and solr takes
> the
Salman,
Yeah, that first cache is too small. Double it and evictions may go away.
Re warm-up queries - yes, they'll be slower than when executed later when they
get pulled out of the cache. Plus if these are warm-up queries, the relevant
parts of the index may not be in the OS buffer cache.
> But this doesnt give me the facets in an alphabetical
> order.
>
> Besides the features facet, I also have some other facets
> that ALSO need to
> be shown in alphabetical order. How to approach this?
&facet.sort=false
http://wiki.apache.org/solr/SimpleFacetParameters#facet.sort
On Tue, Jan 18, 2011 at 6:04 PM, Grant Ingersoll wrote:
> As devs of Lucene/Solr, due to the way ASF mirrors, etc. works, we really
> don't have a good sense of how people get Lucene and Solr for use in their
> application. Because of this, there has been some talk of dropping Maven
> support for
> On Tue, Jan 18, 2011 at 6:04 PM, Grant Ingersoll wrote:
>
>> As devs of Lucene/Solr, due to the way ASF mirrors, etc. works, we really
>> don't have a good sense of how people get Lucene and Solr for use in their
>> application. Because of this, there has been some talk of dropping Maven
>> su
If you are using Lucene's trunk (to be 4.0) builds, read on...
I just committed LUCENE-2872, which is a hard break on the index file format.
If you are living on Lucene's trunk then you have to remove any
previously created indices and re-index, after updating.
The change cuts over to a faster o
Hi, I have a little problem regarding indexing, that i don't know
how to solve, i need to index the same data in different ways
into the same field. The problem is a normalization problem, and
here is an example:
I have a special character \uA732, which i need to normalize in
two different ways fo
Are you showing the facets with &facet parameters in your request?
Then you can ask for the facets to be returned sorted by byte-order with
facet.sort=index.
Got nothing to do with your schema, let alone your DIH import
configuration that you showed us. Just a matter of how you ask Solr for
Maybe?: Just keep the 'weightages' in an external store of some kind
(rdbms, nosql like mongodb, just a straight text config file that your
app loads into a hash internally, whatever), rather than Solr, and have
your app look them up for each facet value to be displayed, after your
app fetches
Hi all,
I have an index containing a couple of million documents.
Documents are grouped into "groups", each group contains from 1000-2
documents.
The problem:
Each group has defined permission settings. It can be viewed by public,
viewed by registred users, or viewed by a list of users (each
Hey there,
I'm looking for an configuration for Solr 1.4 that
accomplishes the following:
Given the input "abc xyz foo" I would like to add at least the following
token combinations to the index:
abc
abc xyz
abc xyz foo
abc foo
xyz
xyz foo
Why do you want to do this, what is it meant to accomplish? There might
be a better way to accomplish what it is you are trying to do; I can't
think of anything (which doesn't mean it doesn't exist) that what you're
actually trying to do would be required in order to do. What sorts of
queries
Does Solr support the Opensearch format? If so could someone point me
to the correct documentation?
Thanks - Tod
On 20.01.11 22:19, Jonathan Rochkind wrote:
> On 1/20/2011 4:03 PM, Martin Jansen wrote:
>> I'm looking for an configuration for Solr 1.4 that
>> accomplishes the following:
>>
>> Given the input "abc xyz foo" I would like to add at least the following
>> token combinations to the index:
>>
>>
Aha, I have no idea if there actually is a better way of achieving that,
auto-completion with Solr is always tricky and I personally have not
been happy with any of the designs I've seen suggested for it. But I'm
also not entirely sure your design will actually work, but neither am I
sure it w
No, not exactly. In general, people don't expose their Solr API direct
to the world -- they front Solr with some software that is exposed to
the world. (If you do expose your Solr API directly to the world, you
will need to think carefully about security, and make sure you aren't
letting anyone
Hi Martin,
The co-occurrence filter I'm working on at
https://issues.apache.org/jira/browse/LUCENE-2749 would do what you want (among
other things). Still vaporware at this point, as I've only put a couple of
hours into it, so don't hold your breath :)
Steve
> -Original Message-
> Fro
You might also see https://issues.apache.org/jira/browse/SOLR-2143
On Jan 20, 2011, at 4:50 PM, Jonathan Rochkind wrote:
> No, not exactly. In general, people don't expose their Solr API direct to the
> world -- they front Solr with some software that is exposed to the world. (If
> you do expos
Hi,
One of the things about Document Security is that it never involves
just one thing. There are a lot of things to consider, and
unfortunately, they're generally non-trivial.
Deciding how to store/hold/retrieve permissions is certainly one of
those things, and you're right, you should avoid att
Okey, thanks very much.
2011/1/21 Ahmet Arslan
> > We construct our lucene query after certain analysis(ex :
> > words
> > segmentation, category identification).
>
> By analysis, I referring charfilter(s)+tokenizer+tokenfilter(s)
> combination.
>
> > Do you mean we
> > plugin those analysis
> >
I'm not sure how you COULD do searching without having the permissions in the
documents. I mentally use the model of unix filesystems, as a starter. Simple,
but powerful. If I needed a separate table for permissions, or index, I'd have
to do queries, with GINORMOUS amounts of OR statements.
I s
I'm thinking of using something like this:
http://www.xaprb.com/blog/2006/08/16/how-to-build-role-based-access-control-in-sql/
http://www.xaprb.com/blog/2006/08/18/role-based-access-control-in-sql-part-2/
- Original Message
From: Dennis Gearon
To: solr-user@lucene.apache.org
Sent: T
On Thu, Jan 20, 2011 at 4:08 PM, shm wrote:
> Hi, I have a little problem regarding indexing, that i don't know
> how to solve, i need to index the same data in different ways
> into the same field. The problem is a normalization problem, and
> here is an example:
>
> I have a special character \u
(apologies for the cross posting)
*** PLEASE NOTE - the deadline for submitting papers has been extended by 1
week to 1/28/2011! ***
Please consider submitting a paper to the ICSE 2011 Software Engineering for
Cloud Computing (SECLOUD) Workshop to be held Sunday, May 22, 2011, at the
Hilton Ha
Does anyone know if there is a function in solr that allows us to log
optimize times? ie - the length of time optimization takes. I can find alot
of questions of how long optimization should take, but thus far nothing on
how to access how long a particular run actually took.
Thanks!!
--
View thi
Looks like this is a bug and I should write a jira issue for it?
Regards
Bernd
Am 20.01.2011 11:30, schrieb Bernd Fehling:
> Hi list,
>
> after sending full-import=true&clean=true&commit=true
> Solr 4.x (apache-solr-4.0-2010-11-24_09-25-17) responds with:
> - DataImporter doFullImport
> - Direc
Hi All,
Is there is any way in SOLR or any plug-in through which the folders and
documents in FTP location can be indexed.
/ Pankaj Bhatt.
Hi Rok,
I have used about 25 ids with OR Operator and its working fine for
me.Just Have to Increase the MaxBoolClouse parameter and also have to
configure max header size on Servlet container to enable for big query
requests.
-
Thanx:
Grijesh
--
View this message in context:
http://luc
Hi All
I want to integrate Surround Query Parser with solr, To do this i have
downloaded jar file from the internet and and then pasting that jar file in
web-inf/lib
and configured query parser in solrconfig.xml as
now when i load solr admin page following exception comes
org.apache.solr.co
We have tried that as well, but the slave still claims to have a higher index
version, even when the index files were deleted completely
Regards
Thomas
Stevo Slavić, 20.01.2011 16:52:
Not too elegant but valid check would be to bring slave down, delete
it's index data directory, then to commit
On Fri, Jan 21, 2011 at 12:21 PM, pankaj bhatt wrote:
> Hi All,
> Is there is any way in SOLR or any plug-in through which the folders and
> documents in FTP location can be indexed.
[...]
What format are these documents in? Which parts of the documents
do you want to index?
In general, this ca
On Tue, 2011-01-11 at 12:12 +0100, Julien Piquot wrote:
> I would like to be able to prune my search result by removing the less
> relevant documents. I'm thinking about using the search score : I use
> the search scores of the document set (I assume there are sorted by
> descending order), norm
Would you do that with 1000's of users? How expensive in processor time is it?
Have you ever benchmarked it?
Dennis Gearon
Signature Warning
It is always a good idea to learn from your own mistakes. It is usually a
better
idea to learn from others’ mistakes, so you do not ha
that's a pretty good idea, using 'delta score'
Dennis Gearon
Signature Warning
It is always a good idea to learn from your own mistakes. It is usually a
better
idea to learn from others’ mistakes, so you do not have to make them yourself.
from 'http://blogs.techrepublic.com.
I haven't tried myself, but you could look at solandra :
https://github.com/tjake/Lucandra
- Espen
On Thu, Jan 20, 2011 at 6:30 PM, stockii wrote:
>
> http://wiki.apache.org/solr/NearRealtimeSearchTuning
>
> http://lucene.472066.n3.nabble.com/Tuning-Solr-caches-with-high-commit-rates-NRT-td14612
Sounds to me like you either have to find a way NOT to use a parser that is a
child class of:
org.apache.solr.search.QParserPlugin
(not sure if that's possible), or you have to find out what's wrong with the
file. Where did you get it, have you talked to the author?
Dennis Gearon
Signatu
68 matches
Mail list logo