Hello,
I use solr and dataimport to index xml files with a dtd.
The dtd is referenced like this
Previously we were using solr4 in a tomcat container.
During the import process, solr tries to validate the xml file with the
dtd.
To find it we were defining -Duser.dir=pathToDtD and solr could find
Thanks a lot Steve!
On Wed, Jul 11, 2018 at 10:24 AM Steve Rowe wrote:
> Hi Jerome,
>
> I was able to setup a configset to perform OpenNLP NER, loading the model
> files from local storage.
>
> There is a trick though[1]: the model files must be located *in a jar* or
> *in
Hi Steve,
Put models under " ${solr.solr.home}/lib/ " is not working.
I check the "ZkSolrResourceLoader" seems it will first try to find modes in
config set.
If not find, then it uses class loader to load from resources.
Regards,
Jerome
On Tue, Jul 10, 2018 at 9:58 A
Thanks Steve!
On Tue, Jul 10, 2018 at 5:20 AM Steve Rowe wrote:
> Hi Jerome,
>
> See the ref guide[1] for a writeup of how to enable uploading files larger
> than 1MB into ZooKeeper.
>
> Local storage should also work - have you tried placing OpenNLP model
> files in
Hi guys,
In Solrcloud mode, where to put the OpenNLP models?
Upload to zookeeper?
As I test on solr 7.3.1, seems absolute path on local host is not working.
And can not upload into zookeeper if the model size exceed 1M.
Regards,
Jerome
On Wed, Apr 18, 2018 at 9:54 AM Steve Rowe wrote:
>
e schema
> - pushing it to Zookeeper (managed schema API does this for you)
> - RELOAD the collection (do NOT delete it first).
>
> Since you can just RELOAD, I doubt this will be a high priority though.
>
> Thanks for reporting!
> Erick
>
>
> On Wed, Nov 23,
quest
to.
a> setDefaultCollection
b> sendRequest(SolrRequest request, String collection)
Please check
https://lucene.apache.org/solr/6_1_0/solr-solrj/org/apache/solr/client/solrj/impl/CloudSolrClient.html
Regards,
Jerome
On Wed, Nov 23, 2016 at 6:43 PM, Udit Tyagi wrote:
> Hi,
>
> I am a so
upload an configs as the same name as before, and the
managed-schema is the not modified version.
Then recreate the collection, it will throw an error, "core already
exists". But actually it's not.
After restart the whole cluster, recreate collection will success.
Regards,
Jerome
Solr 5.5 to make sure it consists only of one
up-to-date index segment.
Regards,
Jerome
On Tue, Nov 22, 2016 at 10:48 PM, Yonik Seeley wrote:
> It depends highly on what your requests look like, and which ones are
> slower.
> If you're request mix is heterogeneous, find the typ
exists". But actually not.
4. The restart the whole cluster, do the create again, everything works
fine.
I think when doing the delete collection, there's something still hold in
somewhere not deleted.
Please have a look
Regards,
Jerome
On Wed, Nov 23, 2016 at 10:16 AM, Jerome Yan
Hi all,
Here's my situation:
In cloud mode.
1. I created a collection called "test" and then modified the
managed-schemaI got an error as shown in picture 2.
2. To get enough error message, I checked solr logs and get message
shown in picture 3.
3. If I corrected the managed-sche
@Erick Please help😂
On Wed, Oct 12, 2016 at 10:21 AM, Jerome Yang wrote:
> Hi Shawn,
>
> I just check the clusterstate.json
> <http://192.168.33.10:18983/solr/admin/zookeeper?detail=true&path=%2Fclusterstate.json>
> which
> is restored for "restore_test_col
, it's a very serious bug I think.
Should this bug go into jira?
Please help!
Regards,
Jerome
On Tue, Oct 11, 2016 at 8:34 PM, Shawn Heisey wrote:
> On 10/11/2016 3:27 AM, Jerome Yang wrote:
> > Then, I index some new documents, and commit. I find that the
> > documents a
@Mark Miller Please help~
On Tue, Oct 11, 2016 at 5:32 PM, Jerome Yang wrote:
> Using curl do some tests.
>
> curl 'http://localhost:8983/solr/restore_test_collection/update?
> *commit=true*&wt=json' --data-binary @test.json -H
> 'Content-type:application/j
:8983/solr/restore_test_collection/update?
*commitWithin**=1000*&wt=json' --data-binary @test.json -H
'Content-type:application/json'
All replicas in shard1 have new documents include leader, and all new
documents route to shard1.
On Tue, Oct 11, 2016 at 5:27 PM, Jerome Yan
in shard1 and the leader of
shard1 don't have these new documents but other replicas do have these new
documents.
Anyone have this issue?
Really need your help.
Regards,
Jerome
That shard did receive update request, because it shows in the log.
And also commit request.
But no documents indexed.
On Tue, Sep 20, 2016 at 2:26 PM, Jerome Yang wrote:
> I'm sure I send documents to that shard. And execute commit.
>
> I also use curl to index, but not error o
of them are somewhere in the collection,
> that's a problem, assuming you are routing them properly when using
> the implicit router.
>
> Best,
> Erick
>
> On Sun, Sep 18, 2016 at 8:04 PM, Jerome Yang wrote:
> > Hi all,
> >
> > The situation is:
> > Thr
you confirm it's a bug or not?
Regards,
Jerome
collection is already exist.
The only way is to make things right is to clean it manually from zookeeper
and data directory.
How to prevent this happen?
Regards,
Jerome
Thanks a lot everyone!
By setting onlyIfDown=false, it did remove the replica. But still return a
failure message.
That confuse me.
Anyway, thanks Erick and Chris.
Regards,
Jerome
On Thu, Jul 21, 2016 at 5:47 AM, Chris Hostetter
wrote:
>
> Maybe the problem here is some confusion/amb
wn", it still show a failure:
Delete replica failed: Attempted to remove replica :
demo.public.tbl/shard0/core_node4 with onlyIfDown='true', but state is
'active'.
Is this the right behavior? If a hosts gone, I can not delete replicas in
this host?
Regards,
Jerome
On We
solr-lucene/org/apache/solr/common/cloud/ClusterState.html#liveNodesContain-java.lang.String->
).
So, is this a bug?
Regards,
Jerome
Dear Erick,
Thank you, I fond it's the problem of my text segmentation setting.
Anyway, thanks.
Regards,
Jerome
2015-06-21 0:43 GMT+08:00 Erick Erickson :
> Just that this _shouldn't_ be going on at all. Either
> 1> you've done something when setting up this collection
uot;:0.43969032,"docs":[
{},{},...
]
}
}
But, when I re-run the same query, it return :
{
"responseHeader":{
"status":0,
"QTime":14,
"params":{
"indent":"true",
"q":"a",
"wt":"json"}},
"response":{"numFound":0,"start":0,"maxScore":0.0,"docs":[]
},
"highlighting":{}}
Just some short word will show this kind of problem.
Do anyone know what's going on?
Thanks
Regards,
Jerome
tp://tokee.github.io/lucene-solr/
Right now we use solr 4.6, and we soon deliver our relsease, and I'm
afraid I won't have time to try this time, but I can try for next release
(next month I think).
Thanks very much again
Jerome
Dupont
jerome.dupont_at#bnf.fr
Participez à l'acquisition d'un Trésor national - Le manuscrit royal de
François I er Avant d'imprimer, pensez à l'environnement.
Hello,
I have a solr index (12 M docs, 45Go) with facets, and I'm trying to
improve facet queries performances.
1/ I tried to use docvalue on facet fields, it didn't work well
2/ I tried facet.threads=-1 in my querie, and worked perfectely (from more
15s to 2s for longest queries)
3/ I'm tryi
Hello
I'm indexing xml files with xpathEntityProcessor, and for some hundreads
documents on 12 millions are not processed.
When I tried to index only one of the KO documents it doesn't either index.
So it's not a matter of big number of documents.
We tried to do the xslt transformation external
Hello,
I have a question, I index documents and a small part them are skipped, (I
am in onError="skip" mode)
I'm trying to get a list of them, in order to analyse what's worng with
these documents
Is there a mean to get the list of skipped documents, and some more
information (my onError="skip" i
It works
I've done what you said:
_ In my request to get list of documents, I add a where clause filtering on
the select getting the documents to index:
where noticebib.numnoticebib LIKE '%${dataimporter.request.suffixeNotice}'"
_ And I called my dih on each shard with the parameter suffixeNotice
on both instances (even if it's not mandatory?
...
Regards
Jerome
Fermeture annuelle des sites François-Mitterrand et Richelieu du 2 au 15
septembre 2013 Avant d'imprimer, pensez à l'environnement.
some documents are not indexed (in particular
the firsts of the list) for some reason (due to our configuration), So when
I was trying on the ten first documents, it couldn't owrk.
Now I will try with 2 shards...
Jerome
Fermeture annuelle des sites François-Mitterrand et Richelieu du 2
Hello,
I'm trying to index documents with Data import handler and solrcloud at the
same time. (huge collection, need to make parallel indexing)
First I had a dih configuration whichs works with solr standalone.
(Indexing for two month every week)
I've transformed my configuration to "cloudify"
Thanks very much, it works, with dataSource (capital S) !!!
Finally, I didn't have to define a "CHEMINRELATIF" field in the
configuration, it's working without it.
This is the definive working configuration:
Hi,
Thanks for your anwser, it made me go ahead.
The name of the entity was not good, not consistent with schema
Now the first entity works fine: the query is done to the database and
returns the good result.
The problem is that the second entity, which is a XPathEntityProcessor
entity, doesn't r
Hello,
I want to use a index a huge list of xml file.
_ Using FileListEntityProcessor causes an OutOfMemoryException (too many
files...)
_ I can do it using a LineEntityProcessor reading a list of files,
generated externally, but I would prefer to generate the list in SOLR
_ So to avoid to mantai
The configuraiton works with LineEntityProcessor, with few documents (havn
(t test with many documents yet.
For information this the config
... fields de
Hello,
We are trying to use data import handler and particularly on a collection
which contains many file (one xml per document)
Our configuration works for a small amount of files, but dataimport fails
with OutofMemory Error when running it on 10M files (in several
directories...)
This is it
Thanks, From all the material i have looked at and searched I am inclined to
believe that those are indeed my options, any others are still welcome...
--
View this message in context:
http://lucene.472066.n3.nabble.com/solr-nested-multivalued-fields-tp3989114p3989260.html
Sent from the Solr - Use
I would like to produce the following result in a Solr search result but not
sure it is possible to do? (Using Solr 3.6)
John
Darby
Sue
Berger
However, i cant seem to manage getting this Tree like structre in my
results.
Hi Erick
On Tue, Jun 7, 2011 at 11:42 PM, Erick Erickson wrote:
> Well, this is odd. Several questions
>
> 1> what do your logs show? I'm wondering if somehow some data is getting
> Â Â rejected. I have no idea why that would be, but if you're seeing indexing
> Â Â exceptions that would explain i
Hi all,
I have a problem with my index. Even though I always index the same
data over and over again, whenever I try
a couple of searches (they are always the same as they are issued by a
unit test suite) I do not get the same
results, sometimes I get 3 successes and 2 failures and sometimes it
is
Hi Erick,
On Tue, Jan 25, 2011 at 1:38 PM, Erick Erickson wrote:
> Frankly, this puzzles me. It *looks* like it should be OK. One warning, the
> analysis page sometimes is a bit misleading, so beware of that.
>
> But the output of your queries make it look like the query is parsing as
> you
> exp
Erick,
On Mon, Jan 24, 2011 at 9:57 PM, Erick Erickson wrote:
> Hmmm, I don't see any screen shots. Several things:
> 1> If your stopword file has comments, I'm not sure what the effect would
> be.
>
Ha, I thought comments were supported in stopwords.txt
> 2> Something's not right here, or I'm
Hi Em, Erick
thanks for your feedback.
Em : yes Here is the stopwords.txt I use :
-
http://svn.apache.org/repos/asf/lucene/dev/trunk/modules/analysis/common/src/resources/org/apache/lucene/analysis/snowball/french_stop.txt
On Mon, Jan 24, 2011 at 6:58 PM, Erick Erickson wrote:
> Try submitting
Hi,
I have a problem with phrase queries, from times to times I do not get any
result
where as I know I should get returned something.
The search is run against a field of type "text" which definition is
available at the following URL :
- http://pastebin.com/Ncem7M8z
This field is defined with t
Shalin Shekhar Mangar wrote on 02/25/2010 07:38:39
AM:
> On Thu, Feb 25, 2010 at 5:34 PM, gunjan_versata
wrote:
>
> >
> > We are using SolrJ to handle commits to our solr server.. All runs
fine..
> > But whenever the commit happens, the server becomes slow and stops
> > responding.. therby result
Otis Gospodnetic wrote on 01/22/2010 12:20:45
AM:
> I'm missing the bigger context of this thread here, but from the
> snippet below - sure, commits cause in-memory index to get written
> to disk, that causes some IO, and that *could* affect search *if*
> queries are running on the same box. Wh
ysee...@gmail.com wrote on 01/20/2010 02:24:04 PM:
> On Wed, Jan 20, 2010 at 2:18 PM, Jerome L Quinn
wrote:
> > This is essentially the same problem I'm fighting with. Â Once in a
while,
> > commit
> > causes everything to freeze, causing add commands to timeout.
>
&g
ysee...@gmail.com wrote on 01/20/2010 02:24:04 PM:
> On Wed, Jan 20, 2010 at 2:18 PM, Jerome L Quinn
wrote:
> > This is essentially the same problem I'm fighting with. Â Once in a
while,
> > commit
> > causes everything to freeze, causing add commands to timeout.
>
&g
ysee...@gmail.com wrote on 01/19/2010 06:05:45 PM:
> On Tue, Jan 19, 2010 at 5:57 PM, Steve Conover
wrote:
> > I'm using latest solr 1.4 with java 1.6 on linux. Â I have a 3M
> > document index that's 10+GB. Â We currently give solr 12GB of ram to
> > play in and our machine has 32GB total.
> >
> >
Lance Norskog wrote on 01/16/2010 12:43:09 AM:
> If your indexing software does not have the ability to retry after a
> failure, you might with to change the timeout from 20 seconds to, say,
> 5 minutes.
I can make it retry, but I have somewhat real-time processes doing these
updates. Does an
Otis Gospodnetic wrote on 01/14/2010 10:07:15
PM:
> See those "waitFlush=true,waitSearcher=true" ? Do things improve if
> you make them false? (not sure how with autocommit without looking
> at the config and not sure if this makes a difference when
> autocommit triggers commits)
Looking at Dir
Hi, folks,
I am using Solr 1.3 pretty successfully, but am running into an issue that
hits once in a long while. I'm still using 1.3 since I have some custom
code I will have to port forward to 1.4.
My basic setup is that I have data sources continually pushing data into
Solr, around 20K adds
Otis Gospodnetic wrote on 11/13/2009 11:15:43
PM:
> Let's take a step back. Why do you need to optimize? You said: "As
> long as I'm not optimizing, search and indexing times are
satisfactory." :)
>
> You don't need to optimize just because you are continuously adding
> and deleting documents
Lance Norskog wrote on 11/13/2009 11:18:42 PM:
> The 'maxSegments' feature is new with 1.4. I'm not sure that it will
> cause any less disk I/O during optimize.
It could still be useful to manage the "too many open files" problem that
rears its ugly head on occasion.
> The 'mergeFactor=2' id
ysee...@gmail.com wrote on 11/13/2009 09:06:29 AM:
> On Fri, Nov 13, 2009 at 6:27 AM, Michael McCandless
> wrote:
> > I think we sorely need a Directory impl that down-prioritizes IO
> > performed by merging.
>
> It's unclear if this case is caused by IO contention, or the OS cache
> of the hot p
ysee...@gmail.com wrote on 11/13/2009 09:06:29 AM:
>
> On Fri, Nov 13, 2009 at 6:27 AM, Michael McCandless
> wrote:
> > I think we sorely need a Directory impl that down-prioritizes IO
> > performed by merging.
>
> It's unclear if this case is caused by IO contention, or the OS cache
> of the hot
Mark Miller wrote on 11/12/2009 07:18:03 PM:
> Ah, the pains of optimization. Its kind of just how it is. One solution
> is to use two boxes and replication - optimize on the master, and then
> queries only hit the slave. Out of reach for some though, and adds many
> complications.
Yes, in my us
Hi, everyone, this is a problem I've had for quite a while,
and have basically avoided optimizing because of it. However,
eventually we will get to the point where we must delete as
well as add docs continuously.
I have a Solr 1.3 index with ~4M docs at around 90G. This is a single
instance run
Mark Miller wrote on 01/26/2009 04:30:00 PM:
> Just a point or I missed: with such a large index (not doc size large,
> but content wise), I imagine a lot of your 16GB of RAM is being used by
> the system disk cache - which is good. Another reason you don't want to
> give too much RAM to the JV
"Lance Norskog" wrote on 01/20/2009 02:16:47 AM:
> "Lance Norskog"
> 01/20/2009 02:16 AM
> Java 1.5 has thread-locking bugs. Switching to Java 1.6 may cure this
> problem.
Thanks for taking time to look at the problem. Unfortunately, this is
happening on Java 1.6,
so I can't put the blame t
uspect I'll add a watchdog, no matter what's causing the problem here.
> However, you should figure out why you are running out of memory. You
> don't want to use more resources than you have available if you can help
it.
Definitely. That's on the agenda :-)
Thanks,
Julian Davchev wrote on 01/20/2009 10:07:48 AM:
> Julian Davchev
> 01/20/2009 10:07 AM
>
> I get SEVERE: Lock obtain timed out
>
> Hi,
> Any documents or something I can read on how locks work and how I can
> controll it. When do they occur etc.
> Cause only way I got out of this mess was rest
Hi, all.
I'm running solr 1.3 inside Tomcat 6.0.18. I'm running a modified query
parser, tokenizer, highlighter, and have a CustomScoreQuery for dates.
After some amount of time, I see solr stop responding to update requests.
When crawling through the logs, I see the following pattern:
Jan 12,
Hi, all. Are there any plans for putting together a bugfix release? I'm
not looking for particular bugs, but would like to know if bug fixes are
only going to be done mixed in with new features.
Thanks,
Jerry Quinn
66 matches
Mail list logo