Hi All,
I think Solr 5.1+ supports streaming API that can be used for my need.
Though it is not working for me right now. I will send another email for
that.
On Thu, Aug 6, 2015 at 3:08 PM, Selvam wrote:
> Dear Toke,
>
> Thanks for your input. Infact my scenario is much more complex, let me
Hi All,
I am trying to use Streaming API in Solr 5.2.
For eg as per
https://cwiki.apache.org/confluence/display/solr/Streaming+Expressions
I tried this from my linux terminal,
1) curl --data-urlencode
'stream=search(gettingstarted,q="*:*",fl="id",sort="id")'
http://localhost:8983/solr/gettings
On 8/7/2015 1:37 AM, Selvam wrote:
> https://cwiki.apache.org/confluence/display/solr/Streaming+Expressions
>
> I tried this from my linux terminal,
> 1) curl --data-urlencode
> 'stream=search(gettingstarted,q="*:*",fl="id",sort="id")'
> http://localhost:8983/solr/gettingstarted/stream
>
> Thre
Hey,
I'm playing around with the suggester component, and it works perfectly
as described: Suggestions for 'logitech mouse' include 'logitech mouse
g500' and 'logitech mouse gaming'.
However, when the words in the record supplying the suggester do not
follow each other as in the search terms, no
Hi,
Thanks for your update, yes, I was missing the cloud mode, I am new to the
world of Solr cloud. Now I have enabled a single node (with two shards &
replicas) that runs on 8983 port along with zookeeper running on 9983 port.
When I run,
curl --data-urlencode
'stream=search(gettingstarted,q="*
Hi Mikhail, you are my hero!
I added your patch to 4.10.4, recompiled, tested and pushed from testing stage
to online system. What a difference!
Right after restart the performance increase for faceting is times 10.
Qtime for MatchAllDocsQuery(*:*) and docValues and faceting went down from
around
Hi,
Sorry, it is working now.
curl --data-urlencode
'stream=search(gettingstarted,q="*:*",fl="id",sort="id asc")'
http://localhost:8983/solr/gettingstarted/stream
I missed *'asc'* in sort :)
Thanks for the help Shawn Heisey.
On Fri, Aug 7, 2015 at 3:46 PM, Selvam wrote:
> Hi,
>
> Thanks for
Took me a while but I tried it and its works perfectly. Thanks a lot !!
Ritesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Search-for-All-CAPS-words-tp4219893p4221597.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi Upayavira,
Yes, I have create boost which is come from outside and I can't index them
into the doc because it need to be realtime sort.
How can i do?
other solution for the sort in this case?
Thank you very much,
Chun.
--
View this message in context:
http://lucene.472066.n3.nabble.com/H
Shawn:
thanks, we found an intermediate solution by serializing our data structure
using string representation, perhaps less optimal than using binary format
directly.
In the original router with JavaBinCodec we found, that
BinaryResponseWriter should also be extended. But the following method is
Hi,
There is a new error handling framework in trunk (SOLR-7441) for the
Streaming API, Streaming Expressions.
So if you're purely in testing mode, it will be much easier to work in
trunk then Solr 5.2.
If you run into errors in trunk that are still confusing please continue to
report them so we
Hi all,
We need to boost a field in a document if field matches certain criteria.
For example:
if title contains "Secrete" , then we want to boost the field to 100 .
For this we have the below code in solrj api while indexing the document:
Collection docs = new ArrayList();
SolrInput
Hello Everyone,
I have indexed 16 million documents in Solr
Cloud. Created 4 nodes and 8 shards with single replica.
I am trying to make concurrent indexing and searching on those indexed
documents. Trying to make 100 concurrent indexing calls along with 100
concurrent sea
Looks like my issue is that my nextDoc call is consuming the first
position, and then on the call to nextPosition it's moving past where I
want it to be. I believe that I have this working properly now by checking
if the current position should or shouldn't be incremented.
On Thu, Aug 6, 2015 at
The thing is that you are trying to introduce custom xml tags which
require changing the response writers. Instead, if you just used
nested maps/lists or SimpleOrderedMap/NamedList then every response
writer should be able to just directly write the output. Nesting is
not a problem.
On Fri, Aug 7,
I have an application that knows enough to tell me that a document has been
updated, but not which document has been updated.There aren't that many
documents in this core/collection - just a couple of 1000. So far I've just
been pumping them all to the update handler every week, but the bu
Use the DedupUpdateProcessor, which can compute a signature based upon
the specified fields.
Upayavira
On Fri, Aug 7, 2015, at 03:56 PM, Davis, Daniel (NIH/NLM) [C] wrote:
> I have an application that knows enough to tell me that a document has
> been updated, but not which document has been upda
How many CPUs do you have? 100 concurrent indexing calls seems like
rather a lot. You're gonna end up doing a lot of context switching,
hence degraded performance. Dunno what others would say, but I'd aim for
approx one indexing thread per CPU.
Upayavira
On Fri, Aug 7, 2015, at 02:58 PM, Nitin So
Thanks - key is that signature field will not be id, and overwriteDupes will be
false:
false
sig
-Original Message-
From: Upayavira [mailto:u...@odoko.co.uk]
Sent: Friday, August 07, 2015 11:22 AM
To: solr-user@lucene.apache.org
Subject: Re: Only indexing changed documents
On 8/7/2015 8:56 AM, Davis, Daniel (NIH/NLM) [C] wrote:
> I have an application that knows enough to tell me that a document has been
> updated, but not which document has been updated.There aren't that many
> documents in this core/collection - just a couple of 1000. So far I've just
> be
Your soft commit time of 3 seconds is quite aggressive,
I'd lengthen it to as long as possible.
Ugh, looked at your query more closely. Adding commit=true to every update
request is horrible performance wise. Let your autocommit process
handle the commits is the first thing I'd do. Second, I'd try
Add &debug=all&debug.explain.structured=true to your query.
What you're looking for is evidence that your boost is being used.
Likely your problem is that the score is indeed taking into account
your boost, but other scoring factors are keeping your specific
document from coming out at the top.
I
On 8/7/2015 11:48 AM, Shawn Heisey wrote:
> On 8/7/2015 8:56 AM, Davis, Daniel (NIH/NLM) [C] wrote:
> > ... snip...
> > Each document has an id I wish to use as the unique ID, but I also want to
> > compute a signature. Is there some way I can use an
> > updateRequestProcessorChain to throw a
You might consider different implementations. The FST-based suggesters
only work in order by design.
AnalyzingInfixSuggester or perhaps FreeTextSuggester are possibilities.
Here's something you might find useful:
http://lucidworks.com/blog/solr-suggester/
Best,
Erick
On Fri, Aug 7, 2015 at 5:1
i have added docValues="true" to my existing schema and I have seen
exponential increase in the size of the index.
The reason in going with docValues is to improve the faceting query time.
schema:
Is this ok? am i doing anything wrong in the schema?
My crude approximation is that docValues increases the disk size,
but that's mostly due to serializing data that would be built in-memory,
basically serializing these structures. So while they increase
disk size, the memory requirements aren't increased. AFAIK, the
in-memory size is a bit smaller a
On 8/7/2015 10:25 AM, naga sharathrayapati wrote:
> i have added docValues="true" to my existing schema and I have seen
> exponential increase in the size of the index.
>
> The reason in going with docValues is to improve the faceting query time.
>
> schema:
> multiValued="false" docValues="true"/
JVM-Memory has gone up from 3% to 17.1%
On Fri, Aug 7, 2015 at 12:10 PM, Shawn Heisey wrote:
> On 8/7/2015 10:25 AM, naga sharathrayapati wrote:
> > i have added docValues="true" to my existing schema and I have seen
> > exponential increase in the size of the index.
> >
> > The reason in going
Hi, Upayavira
RAM = 28GB
CPU = 4 processes..
On Fri, Aug 7, 2015 at 8:53 PM Upayavira wrote:
> How many CPUs do you have? 100 concurrent indexing calls seems like
> rather a lot. You're gonna end up doing a lot of context switching,
> hence degraded performance. Dunno what othe
Hi Erick,
You said that soft commit should be more than 3000 ms.
Actually, I need Real time searching and that's why I need soft commit fast.
commit=true => I made commit=true because , It reduces by indexed data size
from 1.5GB to 500MB on* each shard*. When I did commit=false then,
bq: How much limitations does Solr has related to indexing and searching
simultaneously? It means that how many simultaneously calls, I made for
searching and indexing once?
None a-priori. It all depends on the hardware you're throwing at it. Obviously
indexing with 100 threads is going to eat up
Hi Erick,
posting files to Solr via curl =>
Rather than posting files via curl. Which is better SolrJ or post.jar... I
don't use both things. I wrote a python script for indexing and using
urllib and urllib2 for indexing data via http.. I don't have any option to
use SolrJ Right no
On 8/7/2015 11:47 AM, naga sharathrayapati wrote:
> JVM-Memory has gone up from 3% to 17.1%
In my experience, a healthy Java application (after the heap size has
stabilized) will have a heap utilization graph where the low points are
between 50 and 75 percent. If the low points in heap utilizatio
bq: So, How much minimum concurrent threads should I run?
I really can't answer that in the abstract, you'll simply have to
test.
I'd prefer SolrJ to post.jar. If you're not going to SolrJ, I'd imagine that
moving from Python to post.jar isn't all that useful.
But before you do anything, see wha
I upgraded to Java 1.8.51 no change,
I downgraded solr to 5.1.0 no change?
It must have something to do with the OS, this server is CentOS 6
Scott
--
Even if you're on the right track, you'll get run over if you just sit there.
Will Rogers
There's not nearly enough information here to help.
What was the error? What was the command? What did the logs show?
You might review:
http://wiki.apache.org/solr/UsingMailingLists
Best,
Erick
On Fri, Aug 7, 2015 at 5:07 PM, Scott Derrick wrote:
> I upgraded to Java 1.8.51 no change,
>
> I do
Thanks Erick for your suggestion. I will remove commit = true and use solr
5.2 and then get back to you again. For further help. Thanks.
On Sat, Aug 8, 2015 at 4:07 AM Erick Erickson
wrote:
> bq: So, How much minimum concurrent threads should I run?
>
> I really can't answer that in the abstract
37 matches
Mail list logo