Re: Is that solr supports multi version operations?

2018-09-19 Thread zhenyuan wei
Thanks for your explaination. @Alexandre Rafalovitch  @Walter Underwood

My case is use SOLR as  an Index Service of  some NoSQL systems,it is
 a  common Requirement   to
guarantee the consistency of index&source data .
There maybe  TWO ways to write source data/index:
 1. write index to solr first, then write source data to NoSQL system.
if write NoSQL failed,I want to rollback solr update。due to solr
not support rollback,I have ever
   think to use multi-version to implement this  feature, but ,feel
disappointed。

 2. write source data first, then wirte index to solr.  this is my
current implementation。and I found it fit for me。










Alexandre Rafalovitch  于2018年9月19日周三 下午1:41写道:

> I think if you try hard enough, it is possible to get Solr to keep
> multiple documents that would normally keep only the latest version.
> They will just have different internal lucene id.
>
> This may of course break a lot of other things like SolrCloud and
> possibly facet counts.
>
> So, I would ask the actual business case first. It is entirely
> possible that there are other ways to achieve the desired objectives.
>
> Regards,
>Alex.
>
> On 19 September 2018 at 00:17, Walter Underwood 
> wrote:
> > No. Solr only has one version of a document. It is not a multi-version
> database.
> >
> > Each replica will return the newest version it has.
> >
> > wunder
> > Walter Underwood
> > wun...@wunderwood.org
> > http://observer.wunderwood.org/  (my blog)
> >
> >> On Sep 18, 2018, at 7:11 PM, zhenyuan wei  wrote:
> >>
> >> Hi all,
> >>add solr document with overwrite=false will keepping multi version
> >> documents,
> >> My question is :
> >>1.  How to search newest documents?with what options?
> >>2.  How to delete  old version < newest version  documents?
> >>
> >> for example:
> >> {
> >>"id":"1002",
> >>"name":["james"],
> >>"_version_":1611998319085617152,
> >>"name_str":["james"]},
> >>  {
> >>"id":"1002",
> >>"name":["lily"],
> >>"_version_":1611998307815522304,
> >>"name_str":["lily"]},
> >>  {
> >>"id":"1002",
> >>"name":["lucy"],
> >>"_version_":1611998248265842688,
> >>"name_str":["lucy"]}]
> >>
> >> 1. curl  http://localhost:8983/solr/collection001/query?q=*:*   return
> all
> >> ,
> >>how to search to make response return the newest one?
> >> 2. how to delete  document of version
> >> [1611998307815522304,1611998248265842688] ,
> >> which is older then 1611998319085617152.
> >
>


Re: Implementing NeuralNetworkModel RankNet in Solr LTR

2018-09-19 Thread Zheng Lin Edwin Yeo
Hi Koji,

Thanks for your reply and provide the information.
Just to check, is this supported in Solr 7.4.0?

Regards,
Edwin

On Wed, 19 Sep 2018 at 11:02, Koji Sekiguchi 
wrote:

> Hi,
>
>  > https://github.com/airalcorn2/Solr-LTR#RankNet
>  >
>  > Has anyone tried on this before? And what is the format of the training
>  > data that this model requires?
>
> I haven't tried it, but I'd like to inform you that there is another
> project of LTR we've been
> developed:
>
> https://github.com/LTR4L/ltr4l
>
> It has many LTR algorithms based on neural network, SVM and boosting.
>
> Koji
>
> On 2018/09/12 11:44, Zheng Lin Edwin Yeo wrote:
> > Hi,
> >
> > I am working on to implementing Solr LTR in Solr 7.4.0 by using the
> > NeuralNetworkModel for the feature selection and model training, and I
> have
> > found this site which uses RankNet:
> > https://github.com/airalcorn2/Solr-LTR#RankNet
> >
> > Has anyone tried on this before? And what is the format of the training
> > data that this model requires?
> >
> > Regards,
> > Edwin
> >
>


CACHE -> fieldValueCache usage

2018-09-19 Thread Vincenzo D'Amore
Hi all,

sorry if I bothered you all but in these days I'm just struggling what's
going on with my production servers...

Looking at Solr Admin Panel I've found the CACHE -> fieldValueCache tab
where all the values are 0.

class:org.apache.solr.search.FastLRUCache
description:Concurrent LRU Cache(maxSize=1, initialSize=10,
minSize=9000, acceptableSize=9500, cleanupThread=false)
stats:
CACHE.searcher.fieldValueCache.cumulative_evictions:0
CACHE.searcher.fieldValueCache.cumulative_hitratio:0
CACHE.searcher.fieldValueCache.cumulative_hits:0
CACHE.searcher.fieldValueCache.cumulative_inserts:0
CACHE.searcher.fieldValueCache.cumulative_lookups:0
CACHE.searcher.fieldValueCache.evictions:0
CACHE.searcher.fieldValueCache.hitratio:0
CACHE.searcher.fieldValueCache.hits:0
CACHE.searcher.fieldValueCache.inserts:0
CACHE.searcher.fieldValueCache.lookups:0
CACHE.searcher.fieldValueCache.size:0
CACHE.searcher.fieldValueCache.warmupTime:0

what do you thing, is that normal? Given that this stats comes from a
production server I suppose to have some number here and looking at
solrconfig.xml I don't see any configuration regarding the fieldValueCache.
Don't should I see something here?

Cheers,
Vincenzo

-- 
Vincenzo D'Amore


Re: Is that solr supports multi version operations?

2018-09-19 Thread Walter Underwood
You are doing the right thing. Always write to the repository first, then
write to Solr. The repository is the single source of truth.

We write to the repository, then have a process that copies new items
to Solr.

wunder
Walter Underwood
wun...@wunderwood.org
http://observer.wunderwood.org/  (my blog)

> On Sep 19, 2018, at 3:03 AM, zhenyuan wei  wrote:
> 
> Thanks for your explaination. @Alexandre Rafalovitch  @Walter Underwood
> 
>My case is use SOLR as  an Index Service of  some NoSQL systems,it is
> a  common Requirement   to
> guarantee the consistency of index&source data .
>There maybe  TWO ways to write source data/index:
> 1. write index to solr first, then write source data to NoSQL system.
>if write NoSQL failed,I want to rollback solr update。due to solr
> not support rollback,I have ever
>   think to use multi-version to implement this  feature, but ,feel
> disappointed。
> 
> 2. write source data first, then wirte index to solr.  this is my
> current implementation。and I found it fit for me。
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> Alexandre Rafalovitch  于2018年9月19日周三 下午1:41写道:
> 
>> I think if you try hard enough, it is possible to get Solr to keep
>> multiple documents that would normally keep only the latest version.
>> They will just have different internal lucene id.
>> 
>> This may of course break a lot of other things like SolrCloud and
>> possibly facet counts.
>> 
>> So, I would ask the actual business case first. It is entirely
>> possible that there are other ways to achieve the desired objectives.
>> 
>> Regards,
>>   Alex.
>> 
>> On 19 September 2018 at 00:17, Walter Underwood 
>> wrote:
>>> No. Solr only has one version of a document. It is not a multi-version
>> database.
>>> 
>>> Each replica will return the newest version it has.
>>> 
>>> wunder
>>> Walter Underwood
>>> wun...@wunderwood.org
>>> http://observer.wunderwood.org/  (my blog)
>>> 
 On Sep 18, 2018, at 7:11 PM, zhenyuan wei  wrote:
 
 Hi all,
   add solr document with overwrite=false will keepping multi version
 documents,
 My question is :
   1.  How to search newest documents?with what options?
   2.  How to delete  old version < newest version  documents?
 
 for example:
{
   "id":"1002",
   "name":["james"],
   "_version_":1611998319085617152,
   "name_str":["james"]},
 {
   "id":"1002",
   "name":["lily"],
   "_version_":1611998307815522304,
   "name_str":["lily"]},
 {
   "id":"1002",
   "name":["lucy"],
   "_version_":1611998248265842688,
   "name_str":["lucy"]}]
 
 1. curl  http://localhost:8983/solr/collection001/query?q=*:*   return
>> all
 ,
   how to search to make response return the newest one?
 2. how to delete  document of version
 [1611998307815522304,1611998248265842688] ,
 which is older then 1611998319085617152.
>>> 
>> 



Re: Is that solr supports multi version operations?

2018-09-19 Thread Shawn Heisey

On 9/18/2018 8:11 PM, zhenyuan wei wrote:

Hi all,
 add solr document with overwrite=false will keepping multi version
documents,
My question is :
 1.  How to search newest documents?with what options?
 2.  How to delete  old version < newest version  documents?


When Solr is compiling results, it will only return one copy of a 
particular document (based on uniqueKey value).All other copies will be 
removed.


I suspect (but do not know for sure) that which document will be 
returned is not defined.  On a multi-shard index, if different copies 
are in different shards, which one is returned will be decided by which 
shard answers the query first, or maybe which one answers last.  If 
multiple copies exist in the same core, that's probably more 
deterministic, but it might not be the copy you wanted.


Solr isn't designed to have multiple versions of the same uniqueKey in 
the index.  Lucene itself doesn't care -- it's going to return all of 
them -- but if you want to be sure which one is returned, you'd need to 
write the Lucene-based software yourself instead of using Solr.


As you mentioned in the last message, writing to your true data store 
and then writing to Solr if that succeeds is a better option.  Or you 
could simply write to your data store and then have your indexing 
software detect and read the new records from there.


Thanks,
Shawn



RE: weird error for accessing solr

2018-09-19 Thread Gu, Steve (CDC/OD/OADS) (CTR)
A little update.

For the client machine where solr admin page behaves differently, it turns out 
that the requests on the page like */admin/* were never served.  I think it is 
related to the server setting that might prevent these urls with "/admin/" from 
being sent.

In essence, it is not a solr issue.

Thanks a lot
Steve

-Original Message-
From: Erick Erickson  
Sent: Tuesday, September 18, 2018 4:18 PM
To: solr-user 
Subject: Re: weird error for accessing solr

bq. can you share *ALL* of...

from both machines!
On Tue, Sep 18, 2018 at 12:40 PM Shawn Heisey  wrote:
>
> On 9/18/2018 12:23 PM, Gu, Steve (CDC/OD/OADS) (CTR) wrote:
> > I have set up my solr as a standalone service and the its url is 
> > http://solr.server:8983/solr.  I opened 8983 on  solr.server to anyone, and 
> > solr can be accessed from laptops/desktops.  But when I tried to access the 
> > solr from some servers, I got the error of SolrCore Initialization 
> > Failures.  The left nav on the page is shown but indicates that the solr is 
> > set up as SolrCloud, which is not.
>
> On the dashboard when you see the Cloud tab, can you share *ALL* of 
> what's under JVM in the Args section?
>
> Thanks,
> Shawn
>


Re: weird error for accessing solr

2018-09-19 Thread Erick Erickson
Thanks for bringing closure to this, Whew!
On Wed, Sep 19, 2018 at 8:04 AM Gu, Steve (CDC/OD/OADS) (CTR)
 wrote:
>
> A little update.
>
> For the client machine where solr admin page behaves differently, it turns 
> out that the requests on the page like */admin/* were never served.  I think 
> it is related to the server setting that might prevent these urls with 
> "/admin/" from being sent.
>
> In essence, it is not a solr issue.
>
> Thanks a lot
> Steve
>
> -Original Message-
> From: Erick Erickson 
> Sent: Tuesday, September 18, 2018 4:18 PM
> To: solr-user 
> Subject: Re: weird error for accessing solr
>
> bq. can you share *ALL* of...
>
> from both machines!
> On Tue, Sep 18, 2018 at 12:40 PM Shawn Heisey  wrote:
> >
> > On 9/18/2018 12:23 PM, Gu, Steve (CDC/OD/OADS) (CTR) wrote:
> > > I have set up my solr as a standalone service and the its url is 
> > > http://solr.server:8983/solr.  I opened 8983 on  solr.server to anyone, 
> > > and solr can be accessed from laptops/desktops.  But when I tried to 
> > > access the solr from some servers, I got the error of SolrCore 
> > > Initialization Failures.  The left nav on the page is shown but indicates 
> > > that the solr is set up as SolrCloud, which is not.
> >
> > On the dashboard when you see the Cloud tab, can you share *ALL* of
> > what's under JVM in the Args section?
> >
> > Thanks,
> > Shawn
> >


Re: SOLR 7.0 DIH out of memory issue with sqlserver

2018-09-19 Thread Erick Erickson
Has this ever worked? IOW, is this something that's changed or has
just never worked?

The obvious first step is to start Solr with more than 1G of memory.
Solr _likes_ memory and a 1G heap is quite small. But you say:
"Increasing the heap size further doesnt start SOLR instance itself.".
How much RAM do you have on your machine? What other programs are
running? You should be able to increase the heap and start Solr if you
have the RAM on your machine so I'd figure out what's behind that
issue first. I regularly start Solr with 16 or 32G of memory on my
local machines, I know of installations running Solr with 60G heaps so
this points to something really odd about your environment.

When you "increase it further", exactly _how_ does Solr fail to start?
What appears in the Solr logs? etc. Really, troubleshoot that issue
first I'd recommend.

If DIH still needs a ridiculous amount of memory, it's usually the
JDBC driver trying to read all the rows into memory at once and you'll
have to explore the jdbc driver settings in detail.

Best,
Erick
On Tue, Sep 18, 2018 at 11:16 PM Tanya Bompi  wrote:
>
> Hi,
>   I am using the Microsoft Jdbc driver 6.4 version in Solr 7.4.0 . I have
> tried removing the selectMethod=Cursor and still it runs out of heap space.
> Do we have anyone who has faced similar issue.
>
> Thanks
> Tanya
>
>
> On Tue, Sep 18, 2018 at 6:38 PM Shawn Heisey  wrote:
>
> > On 9/18/2018 4:48 PM, Tanya Bompi wrote:
> > >I have the SOLR 7.0 setup with the DataImportHandler connecting to the
> > > sql server db. I keep getting OutOfMemory: Java Heap Space when doing a
> > > full import. The size of the records is around 3 million so not very
> > huge.
> > > I tried the following steps and nothing helped thus far.
> >
> > See this wiki page:
> >
> > https://wiki.apache.org/solr/DataImportHandlerFaq
> >
> > You already have the suggested fix -- setting responseBuffering to
> > adaptive.  You might try upgrading the driver.  If that doesn't work,
> > you're probably going to need to talk to Microsoft about what you need
> > to do differently on the JDBC url.
> >
> > I did find this page:
> >
> >
> > https://docs.microsoft.com/en-us/sql/connect/jdbc/using-adaptive-buffering?view=sql-server-2017
> >
> > This says that when using adaptive buffering, you should avoid using
> > selectMethod=cursor.  So you should try removing that parameter.
> >
> > Thanks,
> > Shawn
> >
> >


Seeking a simple way to test my index.

2018-09-19 Thread Chip Calhoun
I've got a Solr instance which crawls roughly 3,500 seed pages, depth of 1, at 
240 institutions, all but 1 of which I don't control. I recrawl once a month or 
so. Naturally if one of the sites I crawl changes, then I need to know to 
update my seed URLs. I've been checking this by hand, which was tenable when my 
site was smaller, but is now completely unreasonable.


Is there a way to test my index without actually having to run a lot of manual 
searches? Perhaps an output I could skim? Any suggestions would be helpful.


Thanks,

Chip


Re: Seeking a simple way to test my index.

2018-09-19 Thread Alexandre Rafalovitch
Have you looked at Apache Nutch? Seems like the direct match for your
- growing - requirements and it does integrate with Solr. Or one of
the other solutions, like http://stormcrawler.net/
http://www.norconex.com/collectors/

Otherwise, this does not really feel like a Solr question.

Regards,
   Alex.

On 19 September 2018 at 14:01, Chip Calhoun  wrote:
> I've got a Solr instance which crawls roughly 3,500 seed pages, depth of 1, 
> at 240 institutions, all but 1 of which I don't control. I recrawl once a 
> month or so. Naturally if one of the sites I crawl changes, then I need to 
> know to update my seed URLs. I've been checking this by hand, which was 
> tenable when my site was smaller, but is now completely unreasonable.
>
>
> Is there a way to test my index without actually having to run a lot of 
> manual searches? Perhaps an output I could skim? Any suggestions would be 
> helpful.
>
>
> Thanks,
>
> Chip


Re: Seeking a simple way to test my index.

2018-09-19 Thread Chip Calhoun
I do use Nutch as my crawler, but just as my crawler, so I hadn't thought to 
look for an answer there. I will do so. thank you.


Chip


From: Alexandre Rafalovitch 
Sent: Wednesday, September 19, 2018 2:05:41 PM
To: solr-user
Subject: Re: Seeking a simple way to test my index.

Have you looked at Apache Nutch? Seems like the direct match for your
- growing - requirements and it does integrate with Solr. Or one of
the other solutions, like http://stormcrawler.net/
http://www.norconex.com/collectors/

Otherwise, this does not really feel like a Solr question.

Regards,
   Alex.

On 19 September 2018 at 14:01, Chip Calhoun  wrote:
> I've got a Solr instance which crawls roughly 3,500 seed pages, depth of 1, 
> at 240 institutions, all but 1 of which I don't control. I recrawl once a 
> month or so. Naturally if one of the sites I crawl changes, then I need to 
> know to update my seed URLs. I've been checking this by hand, which was 
> tenable when my site was smaller, but is now completely unreasonable.
>
>
> Is there a way to test my index without actually having to run a lot of 
> manual searches? Perhaps an output I could skim? Any suggestions would be 
> helpful.
>
>
> Thanks,
>
> Chip


UI fields signification

2018-09-19 Thread Jean-Marc Spaggiari
Hi dear SOLR community.

On this page of the documentation:
https://lucene.apache.org/solr/guide/6_6/core-specific-tools.html what are
the fields "current" and "gen" referring to? I have not been able to find
that anywhere :(

Thanks,

JMS


Re: UI fields signification

2018-09-19 Thread Alexandre Rafalovitch
I would say this is the relevant page for the "current" and
"generation" https://lucene.apache.org/solr/guide/6_6/index-replication.html

And I think generation refers to the actual Lucene index, so is
explained further here:
https://lucene.apache.org/core/6_6_0/core/org/apache/lucene/codecs/lucene62/package-summary.html#package.description

Regards,
   Alex.

On 19 September 2018 at 14:33, Jean-Marc Spaggiari
 wrote:
> Hi dear SOLR community.
>
> On this page of the documentation:
> https://lucene.apache.org/solr/guide/6_6/core-specific-tools.html what are
> the fields "current" and "gen" referring to? I have not been able to find
> that anywhere :(
>
> Thanks,
>
> JMS


Unnecessary Components

2018-09-19 Thread oddtyme
I am helping implement solr for a "downloadable library" of sorts. The
objective is that communities without internet access will be able to access
a library's worth of information on a small, portable device. As such, I am
working within strict space constraints. What are some non-essential
components of solr that can be cut to conserve space for more information?



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Re: Seeking a simple way to test my index.

2018-09-19 Thread Erick Erickson
Chip:

Another thing that might work for you are the streaming/export
capabilities. It has the capacity to efficiently return some data
(docValues only) for very large result sets. You'd have to have some
automated way to feed it what to look for.

But that's a fallback, I'd first look at Nutch as I bet someone's had
a similar problem before in Nutch-land ;)
On Wed, Sep 19, 2018 at 11:18 AM Chip Calhoun  wrote:
>
> I do use Nutch as my crawler, but just as my crawler, so I hadn't thought to 
> look for an answer there. I will do so. thank you.
>
>
> Chip
>
> 
> From: Alexandre Rafalovitch 
> Sent: Wednesday, September 19, 2018 2:05:41 PM
> To: solr-user
> Subject: Re: Seeking a simple way to test my index.
>
> Have you looked at Apache Nutch? Seems like the direct match for your
> - growing - requirements and it does integrate with Solr. Or one of
> the other solutions, like http://stormcrawler.net/
> http://www.norconex.com/collectors/
>
> Otherwise, this does not really feel like a Solr question.
>
> Regards,
>Alex.
>
> On 19 September 2018 at 14:01, Chip Calhoun  wrote:
> > I've got a Solr instance which crawls roughly 3,500 seed pages, depth of 1, 
> > at 240 institutions, all but 1 of which I don't control. I recrawl once a 
> > month or so. Naturally if one of the sites I crawl changes, then I need to 
> > know to update my seed URLs. I've been checking this by hand, which was 
> > tenable when my site was smaller, but is now completely unreasonable.
> >
> >
> > Is there a way to test my index without actually having to run a lot of 
> > manual searches? Perhaps an output I could skim? Any suggestions would be 
> > helpful.
> >
> >
> > Thanks,
> >
> > Chip


Re: SOLR 7.0 DIH out of memory issue with sqlserver

2018-09-19 Thread Tanya Bompi
Hi Erick,
  Thank you for the follow-up. I have resolved the issue with the increase
in heapSize and I am able to set the SOLR VM to initialize with a  3G heap
size and the subset of 1 mil records was fetched successfully. Although it
fails with the entire 3 mil records. So something is off with the adaptive
buffering setting as I see  it is not helping. And I also set the
autosoftcommit param. I might have to increase the heap size further to see
if it helps. I will keep posted if my issue doesnt resolve.

Thanks,
Tanya

On Wed, Sep 19, 2018 at 8:22 AM Erick Erickson 
wrote:

> Has this ever worked? IOW, is this something that's changed or has
> just never worked?
>
> The obvious first step is to start Solr with more than 1G of memory.
> Solr _likes_ memory and a 1G heap is quite small. But you say:
> "Increasing the heap size further doesnt start SOLR instance itself.".
> How much RAM do you have on your machine? What other programs are
> running? You should be able to increase the heap and start Solr if you
> have the RAM on your machine so I'd figure out what's behind that
> issue first. I regularly start Solr with 16 or 32G of memory on my
> local machines, I know of installations running Solr with 60G heaps so
> this points to something really odd about your environment.
>
> When you "increase it further", exactly _how_ does Solr fail to start?
> What appears in the Solr logs? etc. Really, troubleshoot that issue
> first I'd recommend.
>
> If DIH still needs a ridiculous amount of memory, it's usually the
> JDBC driver trying to read all the rows into memory at once and you'll
> have to explore the jdbc driver settings in detail.
>
> Best,
> Erick
> On Tue, Sep 18, 2018 at 11:16 PM Tanya Bompi 
> wrote:
> >
> > Hi,
> >   I am using the Microsoft Jdbc driver 6.4 version in Solr 7.4.0 . I have
> > tried removing the selectMethod=Cursor and still it runs out of heap
> space.
> > Do we have anyone who has faced similar issue.
> >
> > Thanks
> > Tanya
> >
> >
> > On Tue, Sep 18, 2018 at 6:38 PM Shawn Heisey 
> wrote:
> >
> > > On 9/18/2018 4:48 PM, Tanya Bompi wrote:
> > > >I have the SOLR 7.0 setup with the DataImportHandler connecting
> to the
> > > > sql server db. I keep getting OutOfMemory: Java Heap Space when
> doing a
> > > > full import. The size of the records is around 3 million so not very
> > > huge.
> > > > I tried the following steps and nothing helped thus far.
> > >
> > > See this wiki page:
> > >
> > > https://wiki.apache.org/solr/DataImportHandlerFaq
> > >
> > > You already have the suggested fix -- setting responseBuffering to
> > > adaptive.  You might try upgrading the driver.  If that doesn't work,
> > > you're probably going to need to talk to Microsoft about what you need
> > > to do differently on the JDBC url.
> > >
> > > I did find this page:
> > >
> > >
> > >
> https://docs.microsoft.com/en-us/sql/connect/jdbc/using-adaptive-buffering?view=sql-server-2017
> > >
> > > This says that when using adaptive buffering, you should avoid using
> > > selectMethod=cursor.  So you should try removing that parameter.
> > >
> > > Thanks,
> > > Shawn
> > >
> > >
>


Re: Unnecessary Components

2018-09-19 Thread Shawn Heisey

On 9/19/2018 1:48 PM, oddtyme wrote:

I am helping implement solr for a "downloadable library" of sorts. The
objective is that communities without internet access will be able to access
a library's worth of information on a small, portable device. As such, I am
working within strict space constraints. What are some non-essential
components of solr that can be cut to conserve space for more information?


For basic functionality, the entire contrib directory could probably be 
removed.  That's more than half of the download right there.


Some of the jars in solr-webapp/webapp/WEB-INF/lib can likely be 
removed.  Chances are that you won't need the jars starting with 
"hadoop" - those are for HDFS support.  That's another 11 MB.  If you 
don't need either HDFS or SolrCloud, you can remove the zookeeper jar, 
and I think you can also remove the curator jars.  If you're not 
accessing Solr with a JDBC driver, you won't need the calcite jars.    
If you're not dealing with oriental characters (and sometimes even if 
you ARE), you can probably do without lucene-analyzers-kuromoji.


With careful code analysis, you can probably find other jars that aren't 
needed, but there's not a huge amount of space saving to be gained with 
most of the others.


Thanks,
Shawn



Re: UI fields signification

2018-09-19 Thread Jean-Marc Spaggiari
Hi Alex,

Thanks for replying.

I also found this:
http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201404.mbox/%3c53483062.2000...@elyograg.org%3E
where it says "s basically means that Lucene has detected an index state
where something has made changes to the index, but those changes are not
yet visible.  To make them visible and return this status to 'true', do a
 commit or soft commit with openSearcher enabled."

That is interesting because my "current" flag almost never turns green, and
indeed my openSearcher is disabled... So I'm a bit lost about what it
should be ad if having "current" always red is really an issue or not... :-/

JMS

Le mer. 19 sept. 2018 à 14:44, Alexandre Rafalovitch  a
écrit :

> I would say this is the relevant page for the "current" and
> "generation"
> https://lucene.apache.org/solr/guide/6_6/index-replication.html
>
> And I think generation refers to the actual Lucene index, so is
> explained further here:
>
> https://lucene.apache.org/core/6_6_0/core/org/apache/lucene/codecs/lucene62/package-summary.html#package.description
>
> Regards,
>Alex.
>
> On 19 September 2018 at 14:33, Jean-Marc Spaggiari
>  wrote:
> > Hi dear SOLR community.
> >
> > On this page of the documentation:
> > https://lucene.apache.org/solr/guide/6_6/core-specific-tools.html what
> are
> > the fields "current" and "gen" referring to? I have not been able to find
> > that anywhere :(
> >
> > Thanks,
> >
> > JMS
>


RE: sharding and placement of replicas

2018-09-19 Thread Oakley, Craig (NIH/NLM/NCBI) [C]
I am still wondering whether anyone has ever seen any examples of this actually 
working (has anyone ever seen any example of SPLITSHARD on a two-node SolrCloud 
placing replicas of the each shard on different hosts than other replicas of 
the same shards)?


Anyone?

-Original Message-
From: Oakley, Craig (NIH/NLM/NCBI) [C]  
Sent: Friday, August 10, 2018 12:54 PM
To: solr-user@lucene.apache.org
Subject: RE: sharding and placement of replicas

Note that I usually create collections with commands which contain (for example)

solr/admin/collections?action=CREATE&name=collectest&collection.configName=collectest&numShards=1&replicationFactor=1&createNodeSet=

I give one node in the createNodeSet and then ADDREPLICA to the other node.

In case this were related, I now tried it a different way, using a command 
which contains

solr/admin/collections?action=CREATE&name=collectest5&collection.configName=collectest&numShards=1&replicationFactor=2&createNodeSet=

I gave both nodes in the createNodeSet in this case. It created one replica on 
each node (each node being on a different host at the same port). This is what 
I would consider the expected behavior (refraining from putting two replicas of 
the same one shard on the same node)

After this I ran a command including

solr/admin/collections?action=SPLITSHARD&collection=collectest5&shard=shard1&indent=on&async=test20180810h

The result was still the same: one of the four new shards was on one node and 
the other three were all together on the node from which I issued this command 
(including putting two replicas of the same shard on the same node).





I am wondering whether there are any examples of this actually working (any 
examples of SPLITSHARD occasionally placing replicas of the each shard on 
different hosts than other replicas of the same shards)


-Original Message-
From: Oakley, Craig (NIH/NLM/NCBI) [C] [mailto:craig.oak...@nih.gov] 
Sent: Thursday, August 09, 2018 5:08 PM
To: solr-user@lucene.apache.org
Subject: RE: sharding and placement of replicas

Okay, I've tried again with two nodes running Solr7.4 on different hosts.

Before SPLITSHARD, collectest2_shard1_replica_n1 was on the host nosqltest22, 
and collectest2_shard1_replica_n3 was on the host nosqltest11

After running SPLITSHARD (on the nosqltest22 node), only 
collectest2_shard1_0_replica0 was added to nosqltest11; nosqltest22 became the 
location for collectest2_shard1_0_replica_n5 and 
collectest2_shard1_1_replica_n6 and collectest2_shard1_1_replica0 (and so if 
nosqltest22 were to be down, shard1_1 would not be available).


-Original Message-
From: Erick Erickson [mailto:erickerick...@gmail.com] 
Sent: Tuesday, July 31, 2018 5:16 PM
To: solr-user 
Subject: Re: sharding and placement of replicas

Right, two JVMs on the same physical host with different ports are
"different Solrs" by default. If you had two replicas per shard and
both were on either Solr instance (same port) that would be
unexpected.

Problem is that this would have been a bug clear back in the Solr 4x
days so the fact that you say you saw it on 6.6 would be unexpected.

Of course if you have three replicas and two instances, I'd absolutely
expect that two replicas would be on one of them for each shard.

Best,
Erick

On Tue, Jul 31, 2018 at 12:24 PM, Oakley, Craig (NIH/NLM/NCBI) [C]
 wrote:
> In my case, when trying on Solr7.4 (in response to Shawn Heisey's 6/19/18 
> comment "If this is a provable and reproducible bug, and it's still a problem 
> in the current stable branch"), I had only installed Solr7.4 on one host, and 
> so I was testing with two nodes on the same host (different port numbers). I 
> had previously had the same symptom when the two nodes were on different 
> hosts, but that was with Solr6.6 -- I can try it again with Solr7.4 with two 
> hosts and report back.
>
> -Original Message-
> From: Shawn Heisey [mailto:apa...@elyograg.org]
> Sent: Tuesday, July 31, 2018 2:26 PM
> To: solr-user@lucene.apache.org
> Subject: Re: sharding and placement of replicas
>
> On 7/27/2018 8:26 PM, Erick Erickson wrote:
>> Yes with some fiddling as far as "placement rules", start here:
>> https://lucene.apache.org/solr/guide/6_6/rule-based-replica-placement.html
>>
>> The idea (IIUC) is that you provide a snitch" that identifies what
>> "rack" the Solr instance is on and can define placement rules that
>> define "don't put more than one thingy on the same rack". "Thingy"
>> here is replica, shard, whatever as defined by other placement rules.
>
> I'd like to see an improvement in Solr's behavior when nothing has been
> configured in auto-scaling or rule-based replica placement.  Configuring
> those things is certainly an option, but I think we can do better even
> without that config.
>
> I believe that Solr already has some default intelligence that keeps
> multiple replicas from ending up on the same *node* when possible ... I
> would like this to also be aware of *hosts*.
>
> Craig

Re: local "q.op=AND" ignored for edismax query

2018-09-19 Thread dshih
Thanks, Shawn.

We made a change to add q.op=AND as a separate param and found a few issues. 
For example, we have a query that filters out guest users in our product. 
It boils down to:

select?q=myname*&q.op=AND&fq=(-(site_role:"Guest"))

debugQuery shows this is parsed as the following, which does not work:
+(-(+site_role:Guest))

Previously, because we were setting q.op=AND as a local param, i.e.
"q={!q.op=AND}myname*", my understanding is the effective operator for the
fq param was OR.  So the fq param parsed as:
-site_role:Guest

I feel like we can fix the issue by just removing the outer parenthesis. 
But I haven't been able to figure out why the operator change (OR to AND)
has this effect.

Any insight appreciated!



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Re: Implementing NeuralNetworkModel RankNet in Solr LTR

2018-09-19 Thread Koji Sekiguchi

Hi Edwin,

> Just to check, is this supported in Solr 7.4.0?

Yes, it is.

https://github.com/LTR4L/ltr4l/blob/master/ltr4l-solr/ivy-jars.properties#L17

Koji

On 2018/09/19 19:40, Zheng Lin Edwin Yeo wrote:

Hi Koji,

Thanks for your reply and provide the information.
Just to check, is this supported in Solr 7.4.0?

Regards,
Edwin

On Wed, 19 Sep 2018 at 11:02, Koji Sekiguchi 
wrote:


Hi,

  > https://github.com/airalcorn2/Solr-LTR#RankNet
  >
  > Has anyone tried on this before? And what is the format of the training
  > data that this model requires?

I haven't tried it, but I'd like to inform you that there is another
project of LTR we've been
developed:

https://github.com/LTR4L/ltr4l

It has many LTR algorithms based on neural network, SVM and boosting.

Koji

On 2018/09/12 11:44, Zheng Lin Edwin Yeo wrote:

Hi,

I am working on to implementing Solr LTR in Solr 7.4.0 by using the
NeuralNetworkModel for the feature selection and model training, and I

have

found this site which uses RankNet:
https://github.com/airalcorn2/Solr-LTR#RankNet

Has anyone tried on this before? And what is the format of the training
data that this model requires?

Regards,
Edwin







Re: Implementing NeuralNetworkModel RankNet in Solr LTR

2018-09-19 Thread Zheng Lin Edwin Yeo
Ok, thank you.

Regards,
Edwin

On Thu, 20 Sep 2018 at 08:39, Koji Sekiguchi 
wrote:

> Hi Edwin,
>
>  > Just to check, is this supported in Solr 7.4.0?
>
> Yes, it is.
>
>
> https://github.com/LTR4L/ltr4l/blob/master/ltr4l-solr/ivy-jars.properties#L17
>
> Koji
>
> On 2018/09/19 19:40, Zheng Lin Edwin Yeo wrote:
> > Hi Koji,
> >
> > Thanks for your reply and provide the information.
> > Just to check, is this supported in Solr 7.4.0?
> >
> > Regards,
> > Edwin
> >
> > On Wed, 19 Sep 2018 at 11:02, Koji Sekiguchi <
> koji.sekigu...@rondhuit.com>
> > wrote:
> >
> >> Hi,
> >>
> >>   > https://github.com/airalcorn2/Solr-LTR#RankNet
> >>   >
> >>   > Has anyone tried on this before? And what is the format of the
> training
> >>   > data that this model requires?
> >>
> >> I haven't tried it, but I'd like to inform you that there is another
> >> project of LTR we've been
> >> developed:
> >>
> >> https://github.com/LTR4L/ltr4l
> >>
> >> It has many LTR algorithms based on neural network, SVM and boosting.
> >>
> >> Koji
> >>
> >> On 2018/09/12 11:44, Zheng Lin Edwin Yeo wrote:
> >>> Hi,
> >>>
> >>> I am working on to implementing Solr LTR in Solr 7.4.0 by using the
> >>> NeuralNetworkModel for the feature selection and model training, and I
> >> have
> >>> found this site which uses RankNet:
> >>> https://github.com/airalcorn2/Solr-LTR#RankNet
> >>>
> >>> Has anyone tried on this before? And what is the format of the training
> >>> data that this model requires?
> >>>
> >>> Regards,
> >>> Edwin
> >>>
> >>
> >
>


Re: Is that solr supports multi version operations?

2018-09-19 Thread zhenyuan wei
Yeah~, writing to true data store first, then write to solr.  I found it is
simple to guarantee the finally consistency,
only handling two main exception bellow:
1. If  failed to write to true data store,then client samply  retry its
request。
2. If write true data store success,and failed to write to solr, it will
retry to solr infinitely。
If write to solr is failed,and server was kill,I can use   the
transaction log of the true data store to replay and write to solr again。




Shawn Heisey  于2018年9月19日周三 下午10:38写道:

> On 9/18/2018 8:11 PM, zhenyuan wei wrote:
> > Hi all,
> >  add solr document with overwrite=false will keepping multi version
> > documents,
> > My question is :
> >  1.  How to search newest documents?with what options?
> >  2.  How to delete  old version < newest version  documents?
>
> When Solr is compiling results, it will only return one copy of a
> particular document (based on uniqueKey value).All other copies will be
> removed.
>
> I suspect (but do not know for sure) that which document will be
> returned is not defined.  On a multi-shard index, if different copies
> are in different shards, which one is returned will be decided by which
> shard answers the query first, or maybe which one answers last.  If
> multiple copies exist in the same core, that's probably more
> deterministic, but it might not be the copy you wanted.
>
> Solr isn't designed to have multiple versions of the same uniqueKey in
> the index.  Lucene itself doesn't care -- it's going to return all of
> them -- but if you want to be sure which one is returned, you'd need to
> write the Lucene-based software yourself instead of using Solr.
>
> As you mentioned in the last message, writing to your true data store
> and then writing to Solr if that succeeds is a better option.  Or you
> could simply write to your data store and then have your indexing
> software detect and read the new records from there.
>
> Thanks,
> Shawn
>
>


Re: SOLR 7.0 DIH out of memory issue with sqlserver

2018-09-19 Thread Erick Erickson
Tanya:

Good to hear. You probably want to configure hard  as
well, and in your case perhaps with openSearcher=true

Indexing is only half the problem. It's quite possible that what's
happening is your index is just growing and that's pushing the
boundaries of Java heap. What I'm thinking is that DIH may be
importing small batches, but as more docs are added the underlying
index grows by enough to hit an OOM with batch N+1.

You could just move the indexing off to a client which would give you
a way to see which it is (index growth or jdbc config). Here's a
sample:
https://lucidworks.com/2012/02/14/indexing-with-solrj/

Best,
Erick
On Wed, Sep 19, 2018 at 1:22 PM Tanya Bompi  wrote:
>
> Hi Erick,
>   Thank you for the follow-up. I have resolved the issue with the increase
> in heapSize and I am able to set the SOLR VM to initialize with a  3G heap
> size and the subset of 1 mil records was fetched successfully. Although it
> fails with the entire 3 mil records. So something is off with the adaptive
> buffering setting as I see  it is not helping. And I also set the
> autosoftcommit param. I might have to increase the heap size further to see
> if it helps. I will keep posted if my issue doesnt resolve.
>
> Thanks,
> Tanya
>
> On Wed, Sep 19, 2018 at 8:22 AM Erick Erickson 
> wrote:
>
> > Has this ever worked? IOW, is this something that's changed or has
> > just never worked?
> >
> > The obvious first step is to start Solr with more than 1G of memory.
> > Solr _likes_ memory and a 1G heap is quite small. But you say:
> > "Increasing the heap size further doesnt start SOLR instance itself.".
> > How much RAM do you have on your machine? What other programs are
> > running? You should be able to increase the heap and start Solr if you
> > have the RAM on your machine so I'd figure out what's behind that
> > issue first. I regularly start Solr with 16 or 32G of memory on my
> > local machines, I know of installations running Solr with 60G heaps so
> > this points to something really odd about your environment.
> >
> > When you "increase it further", exactly _how_ does Solr fail to start?
> > What appears in the Solr logs? etc. Really, troubleshoot that issue
> > first I'd recommend.
> >
> > If DIH still needs a ridiculous amount of memory, it's usually the
> > JDBC driver trying to read all the rows into memory at once and you'll
> > have to explore the jdbc driver settings in detail.
> >
> > Best,
> > Erick
> > On Tue, Sep 18, 2018 at 11:16 PM Tanya Bompi 
> > wrote:
> > >
> > > Hi,
> > >   I am using the Microsoft Jdbc driver 6.4 version in Solr 7.4.0 . I have
> > > tried removing the selectMethod=Cursor and still it runs out of heap
> > space.
> > > Do we have anyone who has faced similar issue.
> > >
> > > Thanks
> > > Tanya
> > >
> > >
> > > On Tue, Sep 18, 2018 at 6:38 PM Shawn Heisey 
> > wrote:
> > >
> > > > On 9/18/2018 4:48 PM, Tanya Bompi wrote:
> > > > >I have the SOLR 7.0 setup with the DataImportHandler connecting
> > to the
> > > > > sql server db. I keep getting OutOfMemory: Java Heap Space when
> > doing a
> > > > > full import. The size of the records is around 3 million so not very
> > > > huge.
> > > > > I tried the following steps and nothing helped thus far.
> > > >
> > > > See this wiki page:
> > > >
> > > > https://wiki.apache.org/solr/DataImportHandlerFaq
> > > >
> > > > You already have the suggested fix -- setting responseBuffering to
> > > > adaptive.  You might try upgrading the driver.  If that doesn't work,
> > > > you're probably going to need to talk to Microsoft about what you need
> > > > to do differently on the JDBC url.
> > > >
> > > > I did find this page:
> > > >
> > > >
> > > >
> > https://docs.microsoft.com/en-us/sql/connect/jdbc/using-adaptive-buffering?view=sql-server-2017
> > > >
> > > > This says that when using adaptive buffering, you should avoid using
> > > > selectMethod=cursor.  So you should try removing that parameter.
> > > >
> > > > Thanks,
> > > > Shawn
> > > >
> > > >
> >


SolrCoreInitializationException after restart of one solr node

2018-09-19 Thread Schaum Mallik
Hi Guys

I have a 3 zookeeper ensemble and 3 solr nodes running version 7.4.0.
Recently I had to restart one node and after I did that it started throwing
this exception.

{

  "error":{

"metadata":[

  "error-class","org.apache.solr.core.SolrCoreInitializationException",

  "root-error-class","org.apache.solr.common.SolrException"],

"msg":"SolrCore 'articles' is not available due to init failure: ",

"trace":"org.apache.solr.core.SolrCoreInitializationException: SolrCore
'articles' is not available due to init failure: \n\tat
org.apache.solr.core.CoreContainer.getCore(CoreContainer.java:1590)\n\tat
org.apache.solr.servlet.HttpSolrCall.init(HttpSolrCall.java:249)\n\tat
org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:469)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:377)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:323)\n\tat
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1634)\n\tat
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:533)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:146)\n\tat
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548)\n\tat
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:257)\n\tat
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:1595)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:255)\n\tat
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1253)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:203)\n\tat
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:473)\n\tat
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:1564)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:201)\n\tat
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1155)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:144)\n\tat
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:219)\n\tat
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:126)\n\tat
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132)\n\tat
org.eclipse.jetty.rewrite.handler.RewriteHandler.handle(RewriteHandler.java:335)\n\tat
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132)\n\tat
org.eclipse.jetty.server.Server.handle(Server.java:531)\n\tat
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:352)\n\tat
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:260)\n\tat
org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:281)\n\tat
org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:102)\n\tat
org.eclipse.jetty.io.ChannelEndPoint$2.run(ChannelEndPoint.java:118)\n\tat
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:333)\n\tat
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:310)\n\tat
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:168)\n\tat
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:126)\n\tat
org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:366)\n\tat
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:760)\n\tat
org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:678)\n\tat
java.lang.Thread.run(Thread.java:748)\nCaused by:
org.apache.solr.common.SolrException: \n\tat
org.apache.solr.cloud.ZkController.preRegister(ZkController.java:1685)\n\tat
org.apache.solr.core.CoreContainer.createFromDescriptor(CoreContainer.java:1128)\n\tat
org.apache.solr.core.CoreContainer.lambda$load$13(CoreContainer.java:681)\n\tat
com.codahale.metrics.InstrumentedExecutorService$InstrumentedCallable.call(InstrumentedExecutorService.java:197)\n\tat
java.util.concurrent.FutureTask.run(FutureTask.java:266)\n\tat
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)\n\tat
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)\n\tat
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)\n\t...
1 more\nCaused by: org.apache.solr.common.SolrException: No coreNodeName
for
CoreDescriptor[name=articles;instanceDir=/opt/solr/server/solr/configsets/articles]\n\tat
org.apache.solr.cloud.ZkController.checkStateInZk(ZkController.java:1716)\n\tat
org.apache.solr.cloud.ZkController.preRegister(ZkController.java:1654)\n\t...
8 more\n",

"code":500}}


I tried dropping all

Re: SolrCoreInitializationException after restart of one solr node

2018-09-19 Thread Shawn Heisey

On 9/19/2018 8:22 PM, Schaum Mallik wrote:

I have a 3 zookeeper ensemble and 3 solr nodes running version 7.4.0.
Recently I had to restart one node and after I did that it started throwing
this exception.



Caused by: org.apache.solr.common.SolrException: No coreNodeName
for
CoreDescriptor[name=articles;instanceDir=/opt/solr/server/solr/configsets/articles]


It is VERY weird for instanceDir to be under the configsets directory.  
Especially for SolrCloud.  I know you're in cloud mode because the 
ZK-related classes are heavily mentioned in the stacktrace.


Can you share the entire solr.log file after a restart?  If that 
directory really is the instanceDir, there should be a core.properties 
file.  What are its contents?


Did you use the service installer script?  If you did, having an 
instanceDir underneath configsets is even MORE strange.


Thanks,
Shawn



Re: SolrCoreInitializationException after restart of one solr node

2018-09-19 Thread Schaum Mallik
The data and index get stored under
/opt/solr/server/solr/articles_shard1_replica_n1.
The config directory when the collection was created, that time the path to
the config was given as '/opt/solr/server/solr/configsets/articles'. I
didn't use the service installer script. The other two solr nodes are
working without any issue. Any ideas how I can resolve this.

Below is the solr.log after restart

2018-09-20 03:01:32.853 INFO  (main) [   ] o.e.j.u.log Logging initialized
@950ms to org.eclipse.jetty.util.log.Slf4jLog

2018-09-20 03:01:33.015 INFO  (main) [   ] o.e.j.s.Server
jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git:
daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_171-b11

2018-09-20 03:01:33.044 INFO  (main) [   ] o.e.j.d.p.ScanningAppProvider
Deployment monitor [file:///opt/solr/server/contexts/] at interval 0

2018-09-20 03:01:33.334 INFO  (main) [   ]
o.e.j.w.StandardDescriptorProcessor NO JSP Support for /solr, did not find
org.apache.jasper.servlet.JspServlet

2018-09-20 03:01:33.343 INFO  (main) [   ] o.e.j.s.session
DefaultSessionIdManager workerName=node0

2018-09-20 03:01:33.343 INFO  (main) [   ] o.e.j.s.session No
SessionScavenger set, using defaults

2018-09-20 03:01:33.345 INFO  (main) [   ] o.e.j.s.session node0 Scavenging
every 66ms

2018-09-20 03:01:33.400 INFO  (main) [   ]
o.a.s.u.c.SSLCredentialProviderFactory Processing SSL Credential Provider
chain: env;sysprop

2018-09-20 03:01:33.436 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter Using
logger factory org.apache.logging.slf4j.Log4jLoggerFactory

2018-09-20 03:01:33.443 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter  ___
_   Welcome to Apache Solr™ version 7.4.0

2018-09-20 03:01:33.447 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter / __|
___| |_ _   Starting in cloud mode on port 8983

2018-09-20 03:01:33.447 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter \__
\/ _ \ | '_|  Install dir: /opt/solr

2018-09-20 03:01:33.448 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
|___/\___/_|_|Start time: 2018-09-20T03:01:33.448Z

2018-09-20 03:01:33.476 INFO  (main) [   ] o.a.s.c.SolrResourceLoader Using
system property solr.solr.home: /opt/solr/server/solr

2018-09-20 03:01:33.551 INFO  (zkConnectionManagerCallback-2-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected

2018-09-20 03:01:33.558 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
Loading solr.xml from SolrHome (not found in ZooKeeper)

2018-09-20 03:01:33.561 INFO  (main) [   ] o.a.s.c.SolrXmlConfig Loading
container configuration from /opt/solr/server/solr/solr.xml

2018-09-20 03:01:33.648 INFO  (main) [   ] o.a.s.c.SolrXmlConfig MBean
server found: com.sun.jmx.mbeanserver.JmxMBeanServer@15d9bc04, but no JMX
reporters were configured - adding default JMX reporter.

2018-09-20 03:01:34.479 INFO  (main) [   ] o.a.s.c.ZkContainer Zookeeper
client=zook1.local:2181,zook2.local:2182,zook3.local:2183

2018-09-20 03:01:34.503 INFO  (zkConnectionManagerCallback-6-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected

2018-09-20 03:01:34.603 WARN  (main) [   ] o.a.s.c.ZkController The
_default configset could not be uploaded. Please provide
'solr.default.confdir' parameter that points to a configset intended to be
the default. Current 'solr.default.confdir' value:
/opt/solr/server/solr/configsets/_default/conf

2018-09-20 03:01:34.608 INFO  (main) [   ] o.a.s.c.c.ZkStateReader Updated
live nodes from ZooKeeper... (0) -> (2)

2018-09-20 03:01:34.636 INFO  (main) [   ] o.a.s.c.Overseer Overseer
(id=null) closing

2018-09-20 03:01:34.652 INFO  (main) [   ]
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for
2147483647 transient cores

2018-09-20 03:01:34.652 INFO  (main) [   ] o.a.s.c.ZkController Register
node as live in ZooKeeper:/live_nodes/solr1:8983_solr

2018-09-20 03:01:34.663 INFO  (zkCallback-5-thread-1) [   ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)

2018-09-20 03:01:34.666 INFO  (main) [   ] o.a.s.c.CoreContainer
Initializing authorization plugin: solr.RuleBasedAuthorizationPlugin

2018-09-20 03:01:34.675 INFO  (main) [   ] o.a.s.c.CoreContainer
Initializing authentication plugin: solr.BasicAuthPlugin

2018-09-20 03:01:34.680 INFO  (main) [   ] o.a.s.c.CoreContainer
PKIAuthenticationPlugin is managing internode requests

2018-09-20 03:01:34.844 INFO  (zkConnectionManagerCallback-13-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected

2018-09-20 03:01:34.846 INFO  (main) [   ] o.a.s.c.c.ZkStateReader Updated
live nodes from ZooKeeper... (0) -> (3)

2018-09-20 03:01:34.849 INFO  (main) [   ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at
zook1.local:2181,zook2.local:2182,zook3.local:2183 ready

2018-09-20 03:01:34.870 INFO  (main) [   ] o.a.s.h.a.MetricsHistoryHandler
No .system collection, keeping metrics history in memory.

2018-09-20 03:01:34.940 INFO  (main) [   ] o.a.s.m.r.SolrJmxReporter JMX
monitoring for 'solr.node' (registry 'solr.node') enabled at server:
com

Re: SolrCoreInitializationException after restart of one solr node

2018-09-19 Thread Schaum Mallik
I also want to add one other things. I had moved from a single core solr
instance on solr 6.6 to the solr cloud few months back. I had ran the
indexupgrader tool on the indexes before I moved them to the solr cloud.

On Wed, Sep 19, 2018 at 7:29 PM Shawn Heisey  wrote:

> On 9/19/2018 8:22 PM, Schaum Mallik wrote:
> > I have a 3 zookeeper ensemble and 3 solr nodes running version 7.4.0.
> > Recently I had to restart one node and after I did that it started
> throwing
> > this exception.
> 
> > Caused by: org.apache.solr.common.SolrException: No coreNodeName
> > for
> >
> CoreDescriptor[name=articles;instanceDir=/opt/solr/server/solr/configsets/articles]
>
> It is VERY weird for instanceDir to be under the configsets directory.
> Especially for SolrCloud.  I know you're in cloud mode because the
> ZK-related classes are heavily mentioned in the stacktrace.
>
> Can you share the entire solr.log file after a restart?  If that
> directory really is the instanceDir, there should be a core.properties
> file.  What are its contents?
>
> Did you use the service installer script?  If you did, having an
> instanceDir underneath configsets is even MORE strange.
>
> Thanks,
> Shawn
>
>