On Tue, May 2, 2017 at 10:22 AM, Damien Kamerman wrote:
> Hi all,
>
> I have a Solr v6.4.2 collection with 12 shards and 2 replicas. Each replica
> uses about 14GB disk usage. I'm using Solaris 11 and I see the 'Page cache'
> grow by about 7GB for each suggester replica I build. The suggester inde
Hello Team,
We have Solr setup with below configuration.
1) 1 collection with one shard
2) 4 Solr Nodes
2) and replication factor 4 with one replication to each Solr Node.
as of now, it's working fine.But going forward it Size may reach high and
we would need to add new Node.
Could you guys p
On 5/2/2017 4:24 AM, Venkateswarlu Bommineni wrote:
> We have Solr setup with below configuration.
>
> 1) 1 collection with one shard
> 2) 4 Solr Nodes
> 2) and replication factor 4 with one replication to each Solr Node.
>
> as of now, it's working fine.But going forward it Size may reach high
Hi ecos,
Is it giving solr documents when running mapreduce indexer tool with debug
mode.
if not can you run it with debug mode and send out any error.
--
View this message in context:
http://lucene.472066.n3.nabble.com/IndexFormatTooNewException-MapReduceIndexerTool-for-PDF-files-tp4332881p
Thanks Shawn for your clarifications,
I think showing a confirmation message saying that "The whole index will be
cleaned" when the clean option is checked will be good.
I always remove the check from the file
/opt/solr/server/solr-webapp/webapp/tpl/dataimport.html after installing
solr but when
On 5/1/2017 10:52 PM, Damien Kamerman wrote:
> I have a Solr v6.4.2 collection with 12 shards and 2 replicas. Each
> replica uses about 14GB disk usage. I'm using Solaris 11 and I see the
> 'Page cache' grow by about 7GB for each suggester replica I build. The
> suggester index itself is very small
On 5/2/2017 6:53 AM, Mahmoud Almokadem wrote:
> And for the dataimport I always use the old UI cause the new UI
> doesn't show the live update and sometimes doesn't show the
> configuration. I think there are many bugs on the new UI.
Do you know if these problems have been reported in the Jira is
On 5/1/2017 10:48 PM, ecos wrote:
> The cause of the error is:
> org.apache.lucene.index.IndexFormatTooNewException: Format version is not
> supported (resource: BufferedChecksumIndexInput (segments_1)): 4 (needs to
> be between 0 and 3).
>
> Reading out there I found the exception is thrown when L
Along with Shawn's comments, if you create a new collection,
consider "oversharding". Say you calculate (more later) you
can fit your collection in N shards, but you expect, over time,
for your collection to triple. _start out_ with 3N shards, many of
them will be co-located. As you get more docs m
thanks Mikhail, that sounds like it would help me as it allows you to set
createNodeSet on RESTORE calls
On Tue, May 2, 2017 at 2:50 PM, Mikhail Khludnev wrote:
> This sounds relevant, but different to https://issues.apache.org/
> jira/browse/SOLR-9527
> You may want to follow this ticket.
>
> O
Mahmoud,
Would it help to have field validation? If the DIH fields are still default
when you press execute, then field validation puts out a message and blocks any
clearing. Just an idea, please excuse if I am off track. -- Rick
On May 2, 2017 8:53:12 AM EDT, Mahmoud Almokadem wrote:
>Thanks S
Hi all,
I want to unload a core from Solr without deleting data-dir or instance-dir.
I’m performing some operations on the data-dir after this and then I would like
to reload the core from the same data-dir. These are the things I tried:
1. Reload api – throws an exception saying no such cor
: I specify a timeout on all queries,
Ah -- ok, yeah -- you mean using "timeAllowed" correct?
If the root issue you were seeing is in fact clocksource related,
then using timeAllowed would probably be a significant compounding
factor there since it would involve a lot of time checks in a s
Thanks Shawn,
We already use the admin UI for testing and bulk uploads. We are using curl
scripts for automation process.
I'll report the issues regarding the new UI on JIRA.
Thanks,
Mahmoud
On Tuesday, May 2, 2017, Shawn Heisey wrote:
> On 5/2/2017 6:53 AM, Mahmoud Almokadem wrote:
> > And
Hmm, has anyone measured the overhead of timeAllowed? We use it all the time.
If nobody has, I’ll run a benchmark with and without it.
wunder
Walter Underwood
wun...@wunderwood.org
http://observer.wunderwood.org/ (my blog)
> On May 2, 2017, at 9:52 AM, Chris Hostetter wrote:
>
>
> : I speci
I ran into the exact same situation recently. I unloaded from the browser
GUI which does not delete the data or instance dirs, but does delete
core.properties. I couldn't find any API either so I eventually manually
recreated core.properties and restarted Solr.
Would be nice if the core.propert
Yes you join more then one collection with Streaming Expressions. Here are
a few things to keep in mind.
* You'll likely want to use the parallel function around the largest join.
You'll need to use the join keys as the partitionKeys.
* innerJoin: requires that the streams be sorted on the join ke
Also take a look at the documentation for the "fetch" streaming expression.
Joel Bernstein
http://joelsolr.blogspot.com/
On Tue, May 2, 2017 at 2:03 PM, Joel Bernstein wrote:
> Yes you join more then one collection with Streaming Expressions. Here are
> a few things to keep in mind.
>
> * You'l
Hi,
I am getting the below exception all of a sudden with solr-6.3.0.
"null:org.apache.solr.common.SolrException: A previous ephemeral live node
still exists. Solr cannot continue. Please ensure that no other Solr
process using the same port is running already."
We are using external zookeeper an
Any ideas? "null:org.apache.solr.common.SolrException: A previous
ephemeral live node still exists. Solr cannot continue. Please ensure that
no other Solr process using the same port is running already."
Not sure, if JMX enablement has caused this.
Thanks,
Satya
On Tue, May 2, 2017 at 3:10 PM S
Well, if an ephemeral node exists, restarting your Zookeeper ensemble
will delete it. Not sure what the precursor here is.
Are you absolutely and totally sure you don't have a solr process
still running on the node you try and start the shows this error? 'ps
aux | grep solr' will show you all of t
IIRC, the core.properties file _is_ renamed to
core.properties.unloaded or something like that.
Yeah, this is something of a pain. The inverse of "unload" is "create"
but you have to know exactly how to create a core, and in SolrCloud
mode that's...interesting. It's much safer to bring the Solr no
I remember seeing some performance impact (even when not using it) and it
was attributed to the calls to System.nanoTime. See SOLR-7875 and SOLR-7876
(fixed for 5.3 and 5.4). Those two Jiras fix the impact when timeAllowed is
not used, but I don't know if there were more changes to improve the
perf
the core properties definitely disappears if you use a configset, as in
#
#Written by CorePropertiesLocator
#Tue May 02 20:19:40 UTC 2017
name=testcore
dataDir=/indexes/solrindexes/testcore
configSet=myconf
Using a conf directory, as in
#Written by CorePropertiesLocator
#Tue May 02 20:30:44
Satya
Say netstat --inet -lP
You might need to add -ipv4 to that command. The P might be lower case (I am on
the bus!). And the output might show misleading service names, see
/etc/services.
Cheers-- Rick
On May 2, 2017 3:10:30 PM EDT, Satya Marivada wrote:
>Hi,
>
>I am getting the below except
On 5/2/2017 10:53 AM, Shashank Pedamallu wrote:
> I want to unload a core from Solr without deleting data-dir or instance-dir.
> I’m performing some operations on the data-dir after this and then I would
> like to reload the core from the same data-dir. These are the things I tried:
>
> 1. Rel
Thank you Simon, Erick and Shawn for your replies. Unfortunately, restarting
Solr is not a option for me. So, I’ll try to follow the steps given by Shawn to
see where I’m standing. Btw, I’m using Solr 6.4.2.
Shawn, once again thank you very much for the detailed reply.
Thanks,
Shashank Pedamall
I have similar needs but for a slightly different use-case.
In my case, I am breaking up cores / indexes based on the month and year
so that I can add an alias that always points to the last few months,
but beyond that I want to simply unload the other indexes once they get
past a few months o
Shalin, yes I think it's a case of the Suggester build hitting the index
all at once. I'm thinking it's hitting all docs, even the ones without
fields relevant to the suggester.
Shawn, I am using ZFS, though I think it's comparable to other setups.
mmap() should still be faster, while the ZFS ARC
On 5/2/2017 6:46 PM, Damien Kamerman wrote:
> Shalin, yes I think it's a case of the Suggester build hitting the index
> all at once. I'm thinking it's hitting all docs, even the ones without
> fields relevant to the suggester.
>
> Shawn, I am using ZFS, though I think it's comparable to other setu
Thanks Shawn, I'll have to look closer into this.
On 3 May 2017 at 12:10, Shawn Heisey wrote:
> On 5/2/2017 6:46 PM, Damien Kamerman wrote:
> > Shalin, yes I think it's a case of the Suggester build hitting the index
> > all at once. I'm thinking it's hitting all docs, even the ones without
> >
Hi Joel,
Thanks for the info.
Regards,
Edwin
On 3 May 2017 at 02:04, Joel Bernstein wrote:
> Also take a look at the documentation for the "fetch" streaming expression.
>
> Joel Bernstein
> http://joelsolr.blogspot.com/
>
> On Tue, May 2, 2017 at 2:03 PM, Joel Bernstein wrote:
>
> > Yes you
Thanks for your response, Rick.
The search is based on fields with our custom type "text_en":
I tested it using Solr admin Analysis. It acts as expected. Both huntman's and
+huntman's are divided to huntman's and huntman in index and search:
Index:
SThuntman's
GFhuntman's
I'm running a three-node SolrCloud (tested with versions 6.4.2 and
6.5.0) with 1 shard and 3 replicas, and I'm having trouble getting the
collection backup API to actually do the job. This is the request I use
to initiate the backup:
http://localhost:8983/solr/admin/collections?action=BACKUP&n
Hi Joel,
I am getting this error when I used the innerHashJoin.
"EXCEPTION":"Invalid stream expression innerHashJoin(parallel(innerJoin
I also can't find the documentation on innerHashJoin for the Streaming
Expressions.
Are you referring to hashJoin?
Regards,
Edwin
On 3 May 2017 at 13:20, Z
35 matches
Mail list logo