my solrcloud,3 shards,and 2replica,
and one shard docs is duplicate,the document router is compositeId
who can help me?
--
View this message in context:
http://lucene.472066.n3.nabble.com/solrcloud-How-to-delete-a-doc-at-a-specific-shard-tp4249354.html
Sent from the Solr - User mailing list arc
Sorry for taking so long. I can confirm that SOLR-8418 is fixed for me
in a self-built 5.5.0 snapshot. Now the next obvious question is, any
ETA for a release?
Regards,
Ere
31.12.2015, 19.15, Erick Erickson kirjoitti:
Ere:
Can you help with testing the patch if it's important to you? Ramkuma
Hi,
An input will be helpful.
Thanks,
Modassar
On Wed, Jan 6, 2016 at 12:39 PM, Modassar Ather
wrote:
> Hi,
>
>
> *q=fl1:net*&facet.field=fl&facet.limit=50&stats=true&stats.field={!cardinality=1.0}fl*
> is returning cardinality around 15 million. It is taking around 4 minutes.
> Similar respon
Hi,
I have two date fields, d_a and d_b, both of type solr.TrieDateField,
that represent different events associated with a particular document.
The interval between these dates is relevant for corner-case statistics.
The interval is calculated as the difference: sub(d_b,d_a) and I've been
a
On Fri, 2016-01-08 at 10:55 +0500, Zap Org wrote:
> i wanted to ask that i need to index after evey 15 min with hard commit
> (real time records) and currently have 5 zookeeper instances and 2 solr
> instances in one machine serving 200 users with 32GB RAM. whereas i wanted
> to serve more than 10,
On Wed, 2016-01-06 at 12:39 +0530, Modassar Ather wrote:
> *q=fl1:net*&facet.field=fl&facet.limit=50&stats=true&stats.field={!cardinality=1.0}fl*
> is returning cardinality around 15 million. It is taking around 4 minutes.
Is this a single shard or multiple?
Anyway, you might have better luck try
Hi Toke,
Is this a single shard or multiple?
It is 12 shard cluster without replicas and has around 90+ GB on each shard.
Thanks for sharing the link. I will look into that.
Regards,
Modassar
On Fri, Jan 8, 2016 at 4:28 PM, Toke Eskildsen
wrote:
> On Wed, 2016-01-06 at 12:39 +0530, Modassar A
Hi,
do you mean you want to use a PEAR to provide the Annotator for the Solr
UIMA UpdateProcessor ?
Can you please detail a bit more your needs?
Regards,
Tommaso
2016-01-08 1:57 GMT+01:00 techqnq :
> implemented custom annotator and generated the PEAR file.
> Windos has the PEAR installer util
Thanks for the replies. The problem I'm trying to solve is to automate
whatever steps I can in configuring Solr for our customer. Rather than an
admin have to edit schema.xml, I thought it would be easier and less
error-prone to do it programmatically. But I'm a novice, so if there is a
better,
On 1/8/2016 6:30 AM, Bob Lawson wrote:
> Thanks for the replies. The problem I'm trying to solve is to automate
> whatever steps I can in configuring Solr for our customer. Rather than an
> admin have to edit schema.xml, I thought it would be easier and less
> error-prone to do it programmaticall
Hi Erick,
I used solr5.3.1 and I sincerely expected response times with replica
configuration near to response times without replica configuration.
Do you agree with me?
I read here
http://lucene.472066.n3.nabble.com/Solr-Cloud-Query-Scaling-td4110516.html
that "Queries do not need to be routed
Hi Erick,
I used solr5.3.1 and I sincerely expected response times with replica
configuration near to response times without replica configuration.
Do you agree with me?
I read here
http://lucene.472066.n3.nabble.com/Solr-Cloud-Query-Scaling-td4110516.html that
"Queries do not need to be routed
Bob,
XY problem means that you are presenting the imagined solution without
presenting the problem to solve. In other words, you are presenting X (solve
for X), without a full statement of the equation to be solved for X.
My guess at your problem is the same as my problem - editing Solr confi
Hi Matteo,
the questions are two:
- "Why are response times on a solr cloud collecton with 1 replica
higher than on solr cloud without replica"
Configuration1: solrCloud with two 8 cores VMs each with 8
shards of 17M docs
Configuration2: solrClous with two 8 cores VMs
Bob,
Not sure why you would want to do this. You can set up Solr to guess the
schema. It creates a file called manage_schema.xml for an override. This is
the case with 5.3 I came across it by accident setting it up the first time
and I was a little annoyed but it made for a quick setup. Your progr
On 1/8/2016 7:55 AM, Luca Quarello wrote:
> I used solr5.3.1 and I sincerely expected response times with replica
> configuration near to response times without replica configuration.
>
> Do you agree with me?
>
> I read here
> http://lucene.472066.n3.nabble.com/Solr-Cloud-Query-Scaling-td4110516
Hello,
I have Solr Cloud configured to stores its index files on a Kerberized HDFS
(I followed documentation at
https://cwiki.apache.org/confluence/display/solr/Running+Solr+on+HDFS), and
have been able to index some documents with the files being written to the
HDFS as expected. However, it appea
First, Daniel nailed the XY problem, but this isn't that...
You're correct that hand-editing the schema file is error-prone.
The managed schema API is your friend here. There are
several commercial front-ends that already do this.
The managed schema API is all just HTTP, so there's nothing
preclu
Here's a longer form of Toke's answer:
https://lucidworks.com/blog/2012/07/23/sizing-hardware-in-the-abstract-why-we-dont-have-a-definitive-answer/
BTW, on the surface, having 5 ZK nodes isn't doing you any real good.
Zookeeper isn't really involved in serving queries or handling
updates, it's pur
I'm going to side-step your primary question and say that it's nearly
always best to do your calculations up-front during indexing to make
queries more efficient and thus serve more requests on the same
hardware. This assumes that the stat you're interested in is
predictable of course...
Best,
Eri
This simply shouldn't be the case if by "duplicate" you mean it has
the same id (i.e. the field defined as the uniqueKey in schema.xml).
If you do have docs in different shards with the same ID, then
something is very strange about your setup.
What version of Solr BTW?
Assuming you mean "same con
Yes, I want to use PEAR file to provide my custom annotator for the solr UIMA
UpdateProcessor.
Basically I have written a custom annotator to capture the certain type of
data from "content" and copies over to other solr field. Generated the PEAR
file using eclipse UIMA plugins. All well till now.
Hi Luca,
It looks like your queries are complex wildcard queries. My theory is that
you are CPU-bounded, for a single query one CPU core for each shard will be
at 100% for the duration of the sub-query. Smaller shards make these
sub-queries faster which is why 16 shards is better than 8 in your cas
For anyone wanting to know an answer, I used
facet.query={!frange l=0 u=3110400}ms(d_b,d_a)
facet.query={!frange l=3110401 u=6220800}ms(d_b,d_a)
facet.query={!frange l=6220801 u=15552000}ms(d_b,d_a)
etc ...
Not the prettiest nor most efficient but accomplishes what I need
Hi Shawn,
I expect that indexing is a little bit slower with replication but in my
case is 3 times worst. I don't explain this.
The monitored consumption of resources is:
All the test have point out an I/O utilization of 100MB/s during
loading data on disk cache, disk cache utilizatio
Hi,
I'm having trouble identifying a replica to delete...
I've created a 3-shard cluster, all 3 created on a single host, then
added a replica for shard2 onto another host, no problem so far.
Now I want to delete the original shard, but got this error when trying
a *replica* param value I th
I’m pretty sure you could change the name when you ADDREPLICA using a core.name
property. I don’t know if you can when you initially create the collection
though.
The CLUSTERSTATUS command will tell you the core names:
https://cwiki.apache.org/confluence/display/solr/Collections+API#Collectio
Thanks for the pointer Jeff,
For SolrCloud it turned out to be...
&property.coreNodeName=xxx
btw, for your app, isn't "slice" old notation?
On 08/01/16 22:05, Jeff Wartes wrote:
I’m pretty sure you could change the name when you ADDREPLICA using a core.name
property. I don’t know if you
I read the client was happy, so I am only curious to know more :)
Apart the readibility, shouldn't be more efficient to put the filters
directly in the main query if you don't cache ?
( checking into the code when not caching is adding a lucene boolean query,
with specifically 0 score, maybe this i
Hi Tomas,
I give you other details.
- The fragment field contains 3KB xml messages.
- The queries that I used for the test are (I only change the word to
search inside the fragment field between requests): curl "
http://localhost:8983/solr/sepa/select?q=+fragment%3A*A*+&fq=marked%3AT
Honestly, I have no idea which is "old". The solr source itself uses slice
pretty consistently, so I stuck with that when I started the project last year.
And logically, a shard being an instance of a slice makes sense to me. But one
significant place where they word shard is exposed is the def
solr version is 5.2.0,
this problem is different shards with the same ID,
the document router is compositeId ,
and if i do this
../collection/update?commit=true&stream.body=idhere,
then this id is missing in whole solrcloud.
--
View this message in context:
http://lucene.472066.n3.nabble.com/so
Hi,
How do I specify a different directory for transaction logs? I tried using
the updatelog entry in solrconfig.xml and reloaded the collection but that
does not seem to work.
Is there another setting I need to change?
Thanks
Nitin
33 matches
Mail list logo