Hi,
I am going to evaluate some Lucene/Solr capabilities on handling faceted
queries, in particular, with a single facet field that contains large number
(say up to 1 million) of distinct values. Does anyone have some experience
on how lucene performs in this scenario?
e.g.
Doc1 has tags A B C D
Works perfectly. Thank you. I didn't know this tokenizer does nothing before
:)
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-define-a-lowercase-fieldtype-without-tokenizer-tp4040500p4040507.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
I don't want the field to be tokenized because Solr doesn't support sorting
on a tokenized field. In order to do case insensitive sorting I need to copy
a field to a lowercase but not tokenized field. How to define this?
I did below but it says I need to specify a tokenizer or a class for
ana
Thanks for the quick reply. Seems like you are suggesting to add explicitly
AND operator. I don't think this solves my problem.
I found it somewhere, and this
works.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Search-match-all-tokens-in-Query-Text-tp4037758p4037762.h
Hello,
I have a field text with type text_general here.
Hi,
While using ContentStreamUpdateRequest up = new
ContentStreamUpdateRequest("/update/extract");
The two ways of adding a file are
up.addFile(File)
up.addContentStream(ContentStream)
However my raw files are stored on some remote storage devices. I am able to
get an InputStream object for the
Thanks Mr.Yagami. I'll look into that.
Jack, for the latter two options, they both require reading the entire text
file into memory, right?
Bing
--
View this message in context:
http://lucene.472066.n3.nabble.com/Send-plain-text-file-to-solr-for-indexing-tp4004515p4004772.html
Sent from the S
So in order to use solrcell I'll have to add a number of dependent libraries,
which is one of what I'm trying to avoid. The second thing is, solrcell
still parses the plain text files and I don't want it to make any change to
those of my exported files.
Any ideas?
Bing
--
View this message in c
Hello,
I used to use solrcell, which has built-in tika support to handle both
extraction and indexing of raw documents. Now I got another text extraction
provider to convert raw document to a plain text txt file so I want to let
solr bypass that extraction phase. Is there a way I can send the plai
Most of existing comparisons were done on Solr3.x or earlier against ES.
After Solr4 added those cloud concepts similar to ES's, there are really
less differences. Solr is more heavier loaded and was not designed for
maximize elasticity In my opinion. It's not hard to decide which way to go
as long
You may write a client using solrj and loop through all files in that folder.
Something like,
ContentStreamUpdateRequest up = new
ContentStreamUpdateRequest("/update/extract");
up.addFile(new File(fileLocation), null);
ModifiableSolrParams p = new ModifiableSolrParams();
p.add("literal.id", str);
Great comments. Thanks to you all.
Bing
--
View this message in context:
http://lucene.472066.n3.nabble.com/Getting-Suggestions-without-Search-Results-tp4000968p4001192.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
I'm having a spell check component that does auto-complete suggestions. It
is part of "last-components" of my /select search handler. So apart from
normal search results I also get a list of suggestions.
Now I want to split things up. Is there a way that I can only get
suggestions of a query
Got it at
https://svn.apache.org/repos/asf/lucene/dev/trunk/solr/solrj/src/test/org/apache/solr/client/solrj/SolrExampleTests.java
Problem solved.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr4-0-Partially-update-document-tp4000875p4000878.html
Sent from the Solr -
Hi,
Several days ago I came across some solrj test code on partially updating
document field values. Sadly I forgot where that was. In Solr 4.0, "/update"
is able to take in document id and fields as hashmaps like
"id": "doc1"
"field1": {"set":"new_value"}
Just trying to figure out what's the so
I remember I did set the 15sec autocommit and still saw the Tlogs growing
unboundedly. But sounds like theoretically it should not if I index in a
constant rate. I'll probably try it again sometime.
For the peersync, I think solr cloud now uses push-replication over pull.
Hmm, it makes sense to ke
Thanks for the information. It definitely helps a lot. There're
numDeletesToKeep = 1000; numRecordsToKeep = 100; in UpdateLog so this should
probably be what you're referring to.
However when I was doing indexing the total size of TLogs kept on
increasing. It doesn't sound like the case where the
Hello,
I'm a bit confused with the purpose of Transaction Logs (Update Logs) in
Solr.
My understanding is, update request comes in, first the new item is put in
RAM buffer as well as T-Log. After a soft commit happens, the new item
becomes searchable but not hard committed in stable storage. Conf
Hello,
>From spell check component I'm able to get the collation query and its # of
hits. Is it possible to have solr execute the collated query automatically
and return doc search results without resending it on client side?
Thanks,
Bing
--
View this message in context:
http://lucene.472066.
Hello,
Background is that I want to use both Suggest and SpellCheck features in a
single query to have alternatives returned at one time. Right now I can only
specify one of them using spellcheck.dictionary at query time.
default
..
suggest
I agree. We chose embedded to minimize the maintenance cost of http solr
servers.
One more concern. Even if I have only one node doing indexing, other nodes
need to reopen index reader periodically to catch up with new changes,
right? Is there a solr request that does this?
Thanks,
Bing
--
Vie
Thanks Kuli and Mikhail,
Using either termcomponent or suggester I could get some suggested terms but
it's still confusing me how to get the respective field names. In order to
get that, Use TermComponent I'll need to do a term query to every possible
field. Similar things as using SpellCheckCompo
Makes sense. Thank you.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Multiple-Embedded-Servers-Pointing-to-single-solrhome-index-tp3999451p4000180.html
Sent from the Solr - User mailing list archive at Nabble.com.
Not quite understand but I'd explain the problem I had. The response would
contain only fields and a list of field values that match the query.
Essentially it's querying for field values rather than documents. The
underlying use case would be, when typing in a quick search box, the drill
down menu
Thanks for the response but wait... Is it related to my question searching
for field values? I was not asking how to use wildcards though.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Does-Solr-support-Value-Search-tp3999654p3999817.html
Sent from the Solr - User mailing
Hi folks,
With StandardDirectoryFactory, index is stored under data/index in forms of
frq, tim, tip and a few other files. While index grows larger, more files
are generated and sometimes it merges a few of them. It's like there're some
kinds of separation and merging strategies there.
My questio
Hi folks,
Just wondering if there is a query handler that simply takes a query string
and search on all/part of fields for field values?
e.g.
q=*admin*
Response may look like
author: [admin, system_admin, sub_admin]
last_modifier: [admin, system_admin, sub_admin]
doctitle: [AdminGuide, AdminMan
Thanks Lance. The use case is to have a cluster of nodes which runs the same
application with EmbeddedSolrServer on each of them, and they all point to
the same index on NFS. Every application is designed equal, meaning that
everyone may index and/or search.
In such way, after every commit the wr
Hi,
I'm trying to use two embedded solr servers pointing to a same solrhome /
index. So that's something like
System.setProperty("solr.solr.home", "SomeSolrDir");
CoreContainer.Initializer initializer = new
CoreContainer.Initializer();
CoreContainer coreContainer = initial
29 matches
Mail list logo