Thanks Jan for your inputs.
I am keen to know about the way people keep running live sites while there
is a breaking change which calls for complete re-indexing.
we want to build a new index , with new schema (it may take couple of
hours) without impacting live e-commerce site.
any thoughts are w
Hi,
I am using SOLR1.4 version. I have a requirement to display the data as below.
Here we store 3 types of polygons MSA, AREA and HOOD.
One MSA can have more than one AREA types as children and one AREA can have
more than one HOOD types as children.
How to store this kind of d
Hi Erick,
I was able to resolve the issue with 'write.lock' files.
Using container.remove("core1") or using container.shutdown() is helping to
remove the 'write.lock' files.
-Shyam
Hi, all,
Thanks for the comment. Then I will abandon post.jar, and try to learn SolrJ
instead.
Best
Bing
--
View this message in context:
http://lucene.472066.n3.nabble.com/Indexing-content-in-XML-files-tp3702795p3705563.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi, Erick,
Thanks for your comment. Though I have some experience in Solr, I am
completely a newbie in SolrJ, and haven't tried using SolrJ to access Solr.
For now, I have a src package of solr3.5.0, and a SolrJ sc downloaded from
web that I want to incorporate into Solr and have a try. How woul
Here is how I got SolrJ to delete the write.lock file. I switched to the
CoreContainer's remove() method. So the new code is:
...
SolrCore curCore = container.remove("core1");
curCore.close();
Now my understanding for why it is working. Based on Solr source code, the
issue had to do with the core
Hello,
What options do I have to hide "ugly" data in the search results? For
example, I am crawling HTML pages and some documents have loose tags
or a long string such as "32lkj31U682860678Stock "
I could scrub the data before getting ingested into the index. (html
parsing, removing strings longe
On Wed, Feb 1, 2012 at 4:13 AM, Ramo Karahasan
wrote:
> This would mean, that i call somewhere in my application the url that is
> described in:
> http://wiki.apache.org/solr/DataImportHandlerDeltaQueryViaFullImport from my
> application?
[...]
Yes, if by application you mean the script used b
I'm not really following your specific example, but a worked through
example of the "index full breadcrumb" type approach darren was suggesting
for doing drill down i na hierarchy is described in slides 32-35 of
this presentation (which was recorded as a webcast)...
http://people.apache.org/%7
: References:
:
: Message-ID: <1327357980.36539.yahoomail...@web160302.mail.bf1.yahoo.com>
: Subject: Hierarchical faceting in UI
https://people.apache.org/~hossman/#threadhijack
Thread Hijacking on Mailing Lists
When starting a new discussion on a mailing list, please do not reply to
an e
: References: <1327606185216-3691370.p...@n3.nabble.com>
:
: <1327704368796-3694787.p...@n3.nabble.com>
: Message-ID: <1327708135.30472.yahoomail...@web160304.mail.bf1.yahoo.com>
: Subject: Strange things happen when I query with many facet.prefixes and fq
: filters
https://people.apache.org/
: Now in my case the indices are being built outside of Solr. So basically I
: create three sets of indices through Lucene API's. And at this point, I
: change the schema.xml and define the fields I have in these new indices. I
do you define a uniqueKey field in your schema.xml? does that field
Hi, iorixxx, Thanks.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Closed-Source-code-of-post-in-example-package-of-Solr-tp3702100p3705333.html
Sent from the Solr - User mailing list archive at Nabble.com.
(12/02/01 4:28), Vadim Kisselmann wrote:
Hmm, i don´t know, but i can test it tomorrow at work.
i´m not sure about the right syntax with hl.q. (?)
but i report :)
hl.q can accept same syntax of q, including local params.
koji
--
http://www.rondhuit.com/en/
This would mean, that i call somewhere in my application the url that is
described in:
http://wiki.apache.org/solr/DataImportHandlerDeltaQueryViaFullImport from my
application?
Ramo
-Ursprüngliche Nachricht-
Von: Igor MILOVANOVIC [mailto:pleti...@gmail.com]
Gesendet: Dienstag, 31. Jan
To have it updated on-demand you could just implement it inside your
application, in form of event trigger or hook (depending how is your
application's architecture).
For batch udpates it is just as simple as cron job script running as
fast as every minute. Limits (50k documents) are imposed on yo
I've been looking into this a bit further and am trying to figure out why
the FQ isn't getting applied.
Can anyone point me to a good spot in the code to start looking at how FQ
parameters are applied to query results in Solr4?
Thanks,
Mike
On Thu, Jan 26, 2012 at 10:06 PM, Mike Hugo wrote:
>
There's nothing built into Solr or DIH that automatically looks
to see if your DB has changed. People sometimes use
cron jobs or similar to fire off the delta-import query
on a regular schedule.
Best
Erick
On Tue, Jan 31, 2012 at 3:06 PM, Ramo Karahasan
wrote:
> Hi,
>
>
>
> i'm using DHI for ind
It may have been a permissions problem, or it stared working after the master
had done another fresh scheduled full-import and jumped an index version.
Timestamp issue?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Problem-with-replication-tp2294313p3704559.html
Sent from th
Hi Chris, Gora
thanks for the help. I am indeed writing a client conversing with solr with
http get calls.
Using your suggestions (in particular the echoParams tip) I managed to find
the problem.
Curiously it turns out that on sending messages I should not convert to
Utf8 or otherwise and
Hi,
i'm using DHI for indexing my data. Currently I always to a delta import
manually as described here:
http://wiki.apache.org/solr/DataImportHandlerDeltaQueryViaFullImport
The data is fetched from a database. How is it possible to update the index
if new data is inserted into the database
Actually, I get:
No files to download for index generation:
this is after deleting the data directory on the slave.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Problem-with-replication-tp2294313p3704457.html
Sent from the Solr - User mailing list archive at Nabble.com.
: Yes, refactoring the various example schema.xml's is what i have been
: doing up to now. The end results is usually quite verbose with a lot of
: redundancy. What is the most compact possible schema.xml?
>From my Solr OOTB Talk...
http://people.apache.org/~hossman/apachecon2011/
Nothing that I know of wil give you what you want OOB, but
there are two possibilities:
Query Elevantion Component is a broad-brush way to go,
but it's not very flexible so if 10 minutes looking at how it
works doesn't excite you, don't spend too much time on it.
If, at query time, you know your
Hmm, i don´t know, but i can test it tomorrow at work.
i´m not sure about the right syntax with hl.q. (?)
but i report :)
2012/1/31 Ahmet Arslan :
>> > Try the &fq option maybe?
>>
>> I thought so, unfortunately.
>> &fq will be the only option. I should rebuild my
>> application :)
>
> Could hl
I really don't understand this. It seems that not doing a full import
and just re-submitting the changed documents is something
you should consider.
Or just don't commit. Or consider using two cores, the idea here
is that you have your "live" core that serves requests, and
indexing to your new cor
: Programming in C++ (Using Qt's QString) I am wondering what conversion to
: apply before compiling words with such letters into the solrquery.
if you are implementing your own client and talking to Solr via HTTP then
how you escape/encode characters in URLs is largely dependent on how you
hav
On Tue, Jan 31, 2012 at 11:51 PM, RT wrote:
> Hi,
>
> there is a post going on encoding international characters. In the mean time
> based on this section:
>
> http://wiki.apache.org/solr/FAQ
>
> where it states that there may be a problem with the Container Servelet, I
> am using the jetty setup
Hi,
there is a post going on encoding international characters. In the mean
time based on this section:
http://wiki.apache.org/solr/FAQ
where it states that there may be a problem with the Container Servelet, I
am using the jetty setup from the example directory. And wondering whether
there
Hey Dipti -
Can you give the exact startup cmds you are using for each of the instances? I
have got Example C going, so I'll have to try and dig into whatever you are
seeing.
- mark
On Jan 27, 2012, at 12:53 PM, Dipti Srivastava wrote:
> Hi Mark,
> Did you get a chance to look into the issues
On Jan 31, 2012, at 1:03 PM, Phil Hoy wrote:
> Hi Mark,
>
> I am using the embedded zookeeper server, how would you recommend I connect
> to it so that I can remove the missing core or is it only possible when using
> a stand-alone zookeeper instance?
Nope, both cases are the same - you just
Hi,
Both Latin1 and Utf8 conversion yield the same negative results.
I get compagnieën back from SOLR as:
compagnieën
I post with: toLatin1() and retrieve from SOLR into QString with
QString::fromLatin1()
Rather dissapointing. Any ideas as to what I may be doing wrong are very
welcome at
Hi Mark,
I am using the embedded zookeeper server, how would you recommend I connect to
it so that I can remove the missing core or is it only possible when using a
stand-alone zookeeper instance?
You are of course correct the reload command as well a few others should cause
a resync with the
Works now ! thanks a lot
... i guess until a document with more than 2.147.483.647 chars
'happy night
+ Pierre
On Tue, Jan 31, 2012 at 5:23 PM, Ahmet Arslan wrote:
>> I'm trying to index word-ngrams using
>> the solr.ShingleFilterFactory,
>> (storing their positions + offset)
>> ...
>> > clas
Hi Gora,
thanks a lot for the below feedback. I use toLatin1() frequently and will
opt for that to see what it does for me.
Thanks again.
Kind regards,
Roland
Gora Mohanty wrote:
On Tue, Jan 31, 2012 at 1:50 PM, RT wrote:
Hi,
I am having a bit of trouble getting words with characters su
: it worked (I'm using Solr-3.4.0, not that it matters)!!
:
: I'll try to figure out what went wrong ...with my limited skills.
skimming the thread, i'm going to guess that even though you were adding
omitNorms=true and restarting solr you weren't re-indexing until Jan
suggested starting clean
: What does "/no_coord" mean in the dismax scoring output? I've looked
: through the wiki mail archives, lucidfind, and can't find any reference.
it's part of the BooleanQuery toString output if the BQ was constructed
with disableCoord=true
-Hoss
Did you rebuild the index? That would help since the index analyzer has
been changed.
On Tue, Jan 31, 2012 at 9:53 AM, Arkadi Colson wrote:
> The text field in the schema configuration looks like this. I changed
> catenateNumbers to 0 but it still doesn't work as aspected.
>
>
>
>
>
>
>
>
I need to index our data 10 times a day due to frequent data changes.
We have placed a mechanism where the data entered by the user in the front
end is submitted into the solr index directly through solrj server. So, if
the solrj commit occurs during the middle of indexing, I lose all the
records
Not sure if this helps but full-import deletes whole index using *:* query in
the beginning of import. Can you can disable this behavior by using
&clean=false&command=full-import
--- On Tue, 1/31/12, Andre Bois-Crettez wrote:
> From: Andre Bois-Crettez
> Subject: Re: Solrj commit affecting do
Why do you commit in the middle of a full import then, if you don't have
to ?
dprasadx wrote:
Hi, I am using solrj server to commit few changes in the data into the master
index through a java program. It works OK unless we do not do a full-import.
But when I do a full-import (say for 800 record
> > Try the &fq option maybe?
>
> I thought so, unfortunately.
> &fq will be the only option. I should rebuild my
> application :)
Could hl.q help? http://wiki.apache.org/solr/HighlightingParameters#hl.q
> I'm trying to index word-ngrams using
> the solr.ShingleFilterFactory,
> (storing their positions + offset)
> ...
> class="solr.TextField"
> positionIncrementGap="1">
>
> class="solr.HTMLStripCharFilterFactory"/>
> class="solr.WhitespaceTokenizerFactory" />
>
On Jan 31, 2012, at 4:49 AM, Phil Hoy wrote:
> Hi,
>
> I am running solrcloud and i am able to add cores
> http://wiki.apache.org/solr/SolrCloud#Creating_cores_via_CoreAdmin but how
> does one remove cores. If i use the core admin unload command, distributed
> queries then error as they still
Hi Erick,
> I didn't read your first post carefully enough, I was keying
> on the words "filter query". Your query does not have
> any filter queries! I thought you were talking
> about &fq=language:de type clauses, which is what
> I was responding to.
no problem, i understand:)
> Solr/Lucene ha
Thanks,
this helps a lot
greetings
Jens
Am 31.01.2012 13:53, schrieb Erik Hatcher-4 [via Lucene]:
>
> On Jan 31, 2012, at 04:42 , Jens Ellenberg wrote:
>
> > Hello,
> >
> > Is there a reference to this status-codes?
>
> Just the source code. SolrCore#setResponseHeaderValues, which
> predominat
I didn't read your first post carefully enough, I was keying
on the words "filter query". Your query does not have
any filter queries! I thought you were talking
about &fq=language:de type clauses, which is what
I was responding to. Solr/Lucene have no way of
interpreting an extended "q" clause and
Hi, I am using solrj server to commit few changes in the data into the master
index through a java program. It works OK unless we do not do a full-import.
But when I do a full-import (say for 800 records), and if I perform a solrj
commit in between the full-import indexing, I see a commit happens t
i m using 4.0 from trunk.
--
View this message in context:
http://lucene.472066.n3.nabble.com/product-popularity-score-gives-error-undefined-field-score-tp3701734p3703647.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi Erick,
thanks for your response:)
Here its my query:
(roomba OR irobot) AND language:de AND
url:"http://www.blog-gedanken.de/produkte/erste-erfahrung-mit-unserem-roomba-roboter-staubsauger/";
Url and language are fields in my schema.xml
With &hl=true&hl.fl=text,url i see this, but i want only
Hey Jonathan,
Any update?
We are experiencing the same thing you describe. As days go on these index
directories continue to collect. We have deleted timestamped indices that
are not currently in-use, but I've been nervous to remove the one simply
called 'index'. Did you end up doing that succ
I'm trying to index word-ngrams using the solr.ShingleFilterFactory,
(storing their positions + offset)
...
...
...
i'm testing it with a (big?) html document, [1.300.000 chars], with lots of tags
Looking at the index (using Schema browser web
The text field in the schema configuration looks like this. I changed
catenateNumbers to 0 but it still doesn't work as aspected.
generateNumberParts="1" catenateWords="1" catenateNumbers="1"
catenateAll="0" splitOnCaseChange="1"/>
protected="protwords.txt"/>
ignoreCase="true" e
This should be OK. There's no real issue with Solr docs having
up to 200 fields, and theres no real limitation on what
portion of those fields each doc has. In other words, only
having 20 out of 200 possible fields in a doc isn't a problem.
There's no overhead for "unused" fields.
Depending upon t
After changing the below suggested lines and compiling the branch_3x runs fine
now.
SolrException is gone.
Regards,
Bernd
Am 31.01.2012 14:21, schrieb Bernd Fehling:
On January 11th I downloaded branch_3x with svn into eclipse (indigo).
Compiled and tested it without problems.
Today I updated
Hi,
I'm running into some issues with solr scoring affecting ordering of query
results. Is it possible to run a Solr boolean query to check if
a document contains any search terms and get the results back without any
scoring mechanism besides presence or absence of any of the search
terms? Basical
Seeing the results with &debugQuery=on would help.
No, fq does NOT get translated into q params, it's a
completely separate mechanism so I'm not quite sure
what you're seeing.
Best
Erick
On Tue, Jan 31, 2012 at 8:40 AM, Vadim Kisselmann
wrote:
> Hi Ahmet,
>
> thanks for quick response :)
> I've
Also, be aware that Solr does NOT index arbitrary XML,
the XML used by the simple post tool is strictly formatted
in a way Solr understands.
A third possibility for arbitrary XML is to write a SolrJ
program that parses your XML and populates
SolrInputDocuments and sends those to Solr.
Best
Erick
Look at the Sort class. You just specify the field
you want to sort on, direction, and pass the class
to your IndexSearcher.search method.
Best
Erick
On Tue, Jan 31, 2012 at 1:24 AM, Peter Markey wrote:
> Hi Eric,
>
> I tried looking for a sample code to sort on Date but was unable to find
> one
Take a look at solrconfig.xml, the directives there. Either add
a path (relative) there or just plop the jar into one of the dirs
already specified.
Best
Erick
On Mon, Jan 30, 2012 at 10:38 PM, wrote:
>
> Hi Chantal,
>
> In the readme file at solr/contrib/analysis-extras/README.txt it says t
See below:
On Mon, Jan 30, 2012 at 10:16 PM, bing wrote:
> Hi, Erick Erickson,
>
> Your suggestions are sound.
>
> For (1), if I use SolrJ as the client to access Solr, then java coding
> becomes the most challenging part. Technically, I want to achieve the same
> effect with highlighting, faceti
Hi Erick,
Below is the sample flow.
String solrHome = "/opt/solr/home";
File solrXml = new File( solrHome, "solr.xml" );
container = new CoreContainer();
container.load(solrHome, solrXml);
SolrServer solr = new EmbeddedSolrServer(container, "core1");
solr.deleteByQuery("*:*");
SolrIn
OK, thanks Erick. Then I won't touch it. I was just wondering, if it would
make sense. But on the other hand the schema.xml is also replicated in my
setup, so maybe it's really confusing.
Thanks
Daniel
On Tue, Jan 31, 2012 at 3:07 PM, Erick Erickson wrote:
> I think it would be easy to get conf
We need more information on your setup. What version of Solr?
Best
Erick
On Mon, Jan 30, 2012 at 7:10 PM, abhayd wrote:
> hi
>
> I m trying to add some weight for popularity in the score returned by solr
> query.
> http://localhost:10101/solr/syx/select?q={!boost%20b=product(popularity,score)}SI
Ah, thanks for bringing closure. Should have occurred to me
when I saw your query
On Mon, Jan 30, 2012 at 2:55 PM, Yuhao wrote:
> Good question. I checked the output sent to Jetty. In the case where it
> returns a blank page, nothing at all is sent to Jetty. This raised my
> suspicion th
Right. Mutlivalued fields use fieldCache for
faceting (as I remember) whereas single valued
fields don't under some circumstances. See:
http://wiki.apache.org/solr/SolrCaching#The_Lucene_FieldCache
Before your change, you were probably using the
filterCache for what faceting you were doing.
So ye
I think it would be easy to get confused about what
was where, resulting in hard-to-track bugs because
the config file wasn't what you were expecting. I also
don't understand why you think this is desirable.
There might be an infinitesimal savings in memory,
due to not instantiating one analysis ch
After looking at the Carrot2 introduction, it seems this can be solved with
clustering but with pre-defined categories.
Does that make sense?
Remi
On Sun, Jan 29, 2012 at 8:42 PM, remi tassing wrote:
> Hi,
>
> Let's say Solr is setup and can return relevant urls. What if I wanted to
> get the
Unless you provide your schema configuration, there's
not much to go on here. Two things though:
1> look at the admin/analysis page to see how your
data is broken up into tokens.
2> at a guess you have WordDelimiterFilterFactory
in your chain and perhaps catenateNumbers="1"
Best
Erick
Oops, fat fingers... Anyway, this is surprising. Can you provide
more details on how you do your indexing?
Best
Erick
On Tue, Jan 31, 2012 at 8:59 AM, Erick Erickson wrote:
> On Mon, Jan 30, 2012 at 2:42 AM, Shyam Bhaskaran
> wrote:
>> Hi,
>>
>> We are using Solr 4.0 and after indexing every ti
On Mon, Jan 30, 2012 at 2:42 AM, Shyam Bhaskaran
wrote:
> Hi,
>
> We are using Solr 4.0 and after indexing every time it is observed that the
> write.lock remains without getting cleared and for the next indexing we have
> to delete the file to get the indexing process running.
>
> We use SolrSe
Hi Ahmet,
thanks for quick response :)
I've also discovered this failure.
I wonder that the query themselves works.
For example: query = language:de
I get results which only have language:de.
Also works the fq and i get only the "de"-result in my field "language".
I can't understand the behavior.
On January 11th I downloaded branch_3x with svn into eclipse (indigo).
Compiled and tested it without problems.
Today I updated my branch_3x from repository.
Compiled fine but get now SolrException when starting.
Jan 31, 2012 1:50:15 PM org.apache.solr.core.SolrCore initListeners
INFO: [] Added S
> in calalina.out i can see the following query:
> path=/select/
> params={hl=true&version=2.2&indent=on&rows=10&start=0&q=(roomba+OR+irobot)+AND+language:de}
> hits=1 status=0 QTime=65
>
> language:de is a filter, and shouldn't be highlighted.
> Do i have a thinking error, or is my query wrong? O
Hi,
i have problems with edismax, filter queries and highlighting.
First of all: can edismax deal with filter queries?
My case:
Edismax is my default requestHandler.
My query in SolrAdminGUI: (roomba OR irobot) AND language:de
You can see, that my q is "roomba OR irobot" and my fq is
"language:
On Jan 31, 2012, at 04:42 , Jens Ellenberg wrote:
> Hello,
>
> Is there a reference to this status-codes?
Just the source code. SolrCore#setResponseHeaderValues, which predominately
uses the codes specified in SolrException:
BAD_REQUEST( 400 ),
UNAUTHORIZED( 401 ), // not currently
We are planning to import data from various tables of ERP DB into a single
Solr/Lucene index.
Since these tables have overlapping columns we are planing to name the
corresponding document field as _. I have
following questions on this.
1) Does having long field name (_) affect
performance ?
2) W
> 2. Use the example package coming with Solr. This only
> requires to define
> the fields in the schema, and no additional configuration
> file needed.
> \apache-solr-3.5.0\example\exampledocs>java -jar post.jar
> *.xml
Bing, please see Hoss' explanation about intended usage of post.jar
http:/
Hello,
Is there a reference to this status-codes?
Erik Hatcher wrote
>
> It means the request was successful. If the status is non-zero (err,
> 1) then there was an error of some sort.
>
> Erik
>
> On Dec 4, 2008, at 9:32 AM, Robert Young wrote:
>
>> In the standard response format,
Hi, all,
I am investigating the indexing in XML files. Currently, I have two
findings:
1. Use DataImportHanlder. This requires to create one more configuration
file for DIH, data-config.xml, which defines the fields specifically for my
XML files.
2. Use the example package coming with Solr. Th
Hi,
I am running solrcloud and i am able to add cores
http://wiki.apache.org/solr/SolrCloud#Creating_cores_via_CoreAdmin but how does
one remove cores. If i use the core admin unload command, distributed queries
then error as they still query the removed core. Do I need to update zookeeper
som
On Tue, Jan 31, 2012 at 1:50 PM, RT wrote:
> Hi,
>
> I am having a bit of trouble getting words with characters such as:
>
> ė, į, ų etc into SOLR.
>
> Programming in C++ (Using Qt's QString) I am wondering what conversion to
> apply before compiling words with such letters into the solrquery.
>
>
Hello bing,
Le 31 janv. 2012 à 04:27, bing a écrit :
> I understand your point of missing "text_en" in the document. It is. Not
> "text_en" but "text" exists.
Unless you use copyField or upload the field as another element, it will not
get fed.
> But then it arises the question: isn't it dynami
Hi,
I am having a bit of trouble getting words with characters such as:
ë, á, ø etc into SOLR.
Programming in C++ (Using Qt's QString) I am wondering what conversion to
apply before compiling words with such letters into the solrquery.
Is UTF8 the correct encoding?
Thanks in advance.
Kind
84 matches
Mail list logo