Erick,
On Mon, Jan 24, 2011 at 9:57 PM, Erick Erickson wrote:
> Hmmm, I don't see any screen shots. Several things:
> 1> If your stopword file has comments, I'm not sure what the effect would
> be.
>
Ha, I thought comments were supported in stopwords.txt
> 2> Something's not right here, or I'm
Interestingly that worked. I deleted the slave index and restarted.
After the first replication I shut down the server, deleted the lock
file and started it again. It seems to be behaving itself now even
though a lock file seems to be recreated. Thanks a lot for the help.
This still seems like a bu
Hi,
just wanted to push this topic again.
Thank you!
Em wrote:
>
> By the way: although I am asking for SolrCloud explicitly again, I will
> take your advice and try distributed search first to understand the
> concept better.
>
> Regards
>
>
> Em wrote:
>>
>> Hi Lance,
>>
>> thanks for
i don't even know whether the regex expression that i'm using for my log is
correct or no.. i very much worried i couldn't proceed in my project already
1/3 rd of the timing is over.. please help.. this is just the first stage..
after this i have ti setup up all the log to be redirected to SYSLOG
Hi,
I am using the default example in the latest stable build
(apache-solr-4.0-2011-01-23_11-24-01).
I read the wiki on http://wiki.apache.org/solr/Suggester and my expectation
is that suggester would correct terms in addition to completing terms.
The handler for suggest is configured with spel
http://pastebin.com/tjCs5dHm
this is the log produced by the solr server
-
DINESHKUMAR . M
I am neither especially clever nor especially gifted. I am only very, very
curious.
--
View this message in context:
http://lucene.472066.n3.nabble.com/please-help-Problem-with-dataImportHandler-tp23
http://pastebin.com/CkxrEh6h
this is my sample log
-
DINESHKUMAR . M
I am neither especially clever nor especially gifted. I am only very, very
curious.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Getting-started-with-writing-parser-tp2278092p2326646.html
Sent from
: prod1 has tag called “Light Weight” with weightage 20,
: prod2 has tag called “Light Weight” with weightage 100,
:
: If i get facet for “Light Weight” , i will get Light Weight (2) ,
: here i need to consider the weightage in to account, and the result will be
: Light Weight (120)
:
: How can
: I have a group of subindex, each of which is a core in my solr now. I want
: to make one query for some of them, how can I do that? And classify response
: doc by index, using facet search?
some background:
"multi core" is when you have multiple solr "cores" on one solr instance;
each "core" c
Hello,
I have been looking at the solr synonym file that was an example, I
did not understand some notation:
aaa =>
bbb => 1 2
ccc => 1,2
a\=>a => b\=>b
a\,a => b\,b
fooaaa,baraaa,bazaaa
The first one says search for when query is aaa. am I correct?
the second one
: I tried following in my schema.xml, but I got
: org.apache.solr.common.SolrException: Error loading class
: 'solr.FinnishLightStemFilterFactory'
FinnishLightStemFilterFactory is a class that exists in SVN on the 3x and
trunk branches, but does not exist in the Solr 1.4.1 release (it was added
Hi,
You haven't defined the field in Solr's schema.xml configuration so it needs to
be added first. Perhaps following the tutorial might be a good idea.
http://lucene.apache.org/solr/tutorial.html
Cheers.
> Hello Team:
>
>
> I am in the process of setting up Solr 1.4 with Magento ENterpris
Hello Team:
I am in the process of setting up Solr 1.4 with Magento ENterprise Edition
1.9.
When I try to index the products I get the following error message.
Jan 24, 2011 3:30:14 PM org.apache.solr.update.processor.LogUpdateProcessor
fini
sh
INFO: {} 0 0
Jan 24, 2011 3:30:14 PM org.apache
: I notice that in the schema, it is only possible to specify a Analyzer class,
: but not a Factory class as for the other elements (Tokenizer, Fitler, etc.).
: This limits the use of this feature, as it is impossible to specify parameters
: for the Analyzer.
: I have looked at the IndexSchema imp
: Having found some code that searches a Lucene index, the only analyzers
: referenced are Lucene.Net.Analysis.Standard.StandardAnalyzer.
:
: How can I map this is Solr? The example schema doesn't seem to mention this,
: and specifying 'text' or 'string' for every field doesn't seem to help.
1)
Hmmm, I don't see any screen shots. Several things:
1> If your stopword file has comments, I'm not sure what the effect would
be.
2> Something's not right here, or I'm being fooled again. Your withresults
xml has this line:
+DisjunctionMaxQuery((meta_text:"ecol d
ingenieur")~0.01) ()
and your nores
: Subject: searching based on grouping result
: In-Reply-To: <913367.31366...@web121705.mail.ne1.yahoo.com>
: References: <913367.31366...@web121705.mail.ne1.yahoo.com>
http://people.apache.org/~hossman/#threadhijack
Thread Hijacking on Mailing Lists
When starting a new discussion on a mailing l
: I'm trying to dynamically add a core to a multi core system using the
: following command:
:
:
http://localhost:8983/solr/admin/cores?action=CREATE&name=items&instanceDir=items&config=data-config.xml&schema=schema.xml&dataDir=data&persist=true
:
: the data-config.xml looks like this:
:
:
I
: this is the error that i'm getting.. no idea of what is it..
Did you follow the instructions in the error message and look at your solr
log file to see what the "severe errors in solr configuration" might be?
: SimplePostTool: FATAL: Solr returned an error:
:
Severe_errors_in_solr_configurat
See below.
On Mon, Jan 24, 2011 at 1:51 PM, johnnyisrael wrote:
>
> Hi,
>
> I am trying out the auto suggest using EdgeNgram.
>
> Using the following tutorial as a reference.
>
>
> http://www.lucidimagination.com/blog/2009/09/08/auto-suggest-from-popular-queries-using-edgengrams/
>
> In the above
Just to add one thing, in case it makes a difference.
Max document size on which highlighting needs to be done is few hundred kb's
(in file system). In index its compressed so should be much smaller. Total
documents are more than 100 million.
On Tue, Jan 25, 2011 at 12:42 AM, Salman Akram <
salma
Hi,
Does anyone have any benchmarks how much highlighting speeds up with Term
Vectors (compared to without it)? e.g. if highlighting on 20 documents take
1 sec with Term Vectors any idea how long it will take without them?
I need to know since the index used for highlighting has a TVF file of
aro
On Mon, Jan 24, 2011 at 10:55:59AM -0800, Em said:
> Could it be possible that your slaves not finished their replicating until
> the new replication-process starts?
> If so, there you got the OOM :).
This was one of my thoughts as well - we're currently running a slave
which has no queries in it
On Mon, Jan 24, 2011 at 08:00:53PM +0100, Markus Jelsma said:
> Are you using 3rd-party plugins?
No third party plugins - this is actually pretty much stock tomcat6 +
solr from Ubuntu. The only difference is that we've adapted the
directory layout to fit in with our house style
On Mon, Jan 24, 2011 at 2:28 PM, Dinesh wrote:
>
> my solrconfig.xml
>
> http://pastebin.com/XDg0L4di
>
> my schema.xml
>
> http://pastebin.com/3Vqvr3C0
>
> my try.xml
>
> http://pastebin.com/YWsB37ZW
[...]
OK, thanks for the above.
You also need to:
* Give us a sample of your log files (for cry
Are you using 3rd-party plugins?
> We have two slaves replicating off one master every 2 minutes.
>
> Both using the CMS + ParNew Garbage collector. Specifically
>
> -server -XX:+UseConcMarkSweepGC -XX:+UseParNewGC
> -XX:+CMSIncrementalMode -XX:+CMSIncrementalPacing
>
> but periodically they bo
UNSUBSCRIBE
On 1/23/11, Papp Richard wrote:
> Hi all,
>
>
>
> I wasted the last few hours trying to serialize some column values (from
> mysql) into a Solr column, but I just can't find such a function. I'll use
> the value in PHP - I don't know if it is possible to serialize in PHP style
> at
Hi Simon,
I got no experiences with a distributed environment.
However, what you are talking about reminds me on another post on the
mailing list.
Could it be possible that your slaves not finished their replicating until
the new replication-process starts?
If so, there you got the OOM :).
Just
Hi Stefan,
yes, this is exactly what I intend - I don't want to search in this field
- just quicly return me the result in a serialized form (the search criteria
is on other fields). Well, if I could serialize the data exactly as like the
PHP serialize() does I would be maximally satisfied, but
Hi Em, Erick
thanks for your feedback.
Em : yes Here is the stopwords.txt I use :
-
http://svn.apache.org/repos/asf/lucene/dev/trunk/modules/analysis/common/src/resources/org/apache/lucene/analysis/snowball/french_stop.txt
On Mon, Jan 24, 2011 at 6:58 PM, Erick Erickson wrote:
> Try submitting
Hi,
I am trying out the auto suggest using EdgeNgram.
Using the following tutorial as a reference.
http://www.lucidimagination.com/blog/2009/09/08/auto-suggest-from-popular-queries-using-edgengrams/
In the above tutorial, The below two lines has been clearly mentioned,
"Note that it’s neces
Hmm, thanks for the response. I'll play around with it and see if that
helps.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Faceting-Question-tp2320542p2321887.html
Sent from the Solr - User mailing list archive at Nabble.com.
We have two slaves replicating off one master every 2 minutes.
Both using the CMS + ParNew Garbage collector. Specifically
-server -XX:+UseConcMarkSweepGC -XX:+UseParNewGC
-XX:+CMSIncrementalMode -XX:+CMSIncrementalPacing
but periodically they both get into a GC storm and just keel over.
Looki
I had difficulties getting this to work, so hopefully this will help others
having the same issue.
My environment:
Solr 3.1
MySQL 5.0.77
Schema:
DIH data-config:
I kept getting build errors similar to this:
org.apache.solr.common.SolrException:
org.apache.lucene.spat
Try submitting your query from the admin page with &debugQuery=on and see
if that helps. The output is pretty dense, so feel free to cut-paste the
results for
help.
Your stemmers have English as the language, which could also be
"interesting".
As Em says, the analysis page may help here, but I'd
There aren't any great general purpose out of the box ways to handle
hieararchical data in Solr. Solr isn't an rdbms.
There may be some particular advice on how to set up a particular Solr
index to answer particular questions with regard to hieararchical data.
I saw a great point made recent
Hi Jerome,
does your fieldtype contains a stopword-filter?
Probably this could be the root of all evil :-).
Could you provide us the fieldtype definition and the explain-content of an
example-query?
Did you check the analysis.jsp to have a look at the produced results?
Regards,
Em
Jerome Rena
Thank you for the advice, Erick!
I will take a look at extending the StandardRequestHandler for such
usecases.
Erick Erickson wrote:
>
> I wasn't thinking about this for adding information to the *request*.
> Rather, in this
> case the autocomplete uses an Ajax call that just uses the TermsCom
Thanks Alexander, what a valuable ressource :).
- Em
--
View this message in context:
http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2321335.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
I have a problem with phrase queries, from times to times I do not get any
result
where as I know I should get returned something.
The search is run against a field of type "text" which definition is
available at the following URL :
- http://pastebin.com/Ncem7M8z
This field is defined with t
Thanks
What I'd really like to do is to exclude an entire group if a certain field is
set to true in any of the documents that make up that group. I can't do it at
index time because some of my users have certain documents hidden from them. So
they shouldn't see the flag as set, while others wo
Hi Otis,
I was implementing Katta because I discovered it before Solr Cloud.
Before replying to your email, I took some time to go through the examples
on the solr cloud wiki.
The examples worked without any issue for me and I now have a better
understanding of what solr cloud is offering.
My e
Steve,
and what exactly do you expect? You can work on the Group itself with
http://wiki.apache.org/solr/FieldCollapsing#Request_Parameters in a limited
way, but of course it's "just" a normal Solr-Result, group by some Values,
nothing really special.
"Can't be referenced in the query" - what do
Hi Rich,
i'm a bit confused after reading your post .. what exactly you wanna try to
achieve? Serializing (like http://php.net/serialize) your complete row into
one field? Don't wanna search in them, just store and deliver them in your
results? Does that make sense? Sounds a bit strange :)
Regard
I wasn't thinking about this for adding information to the *request*.
Rather, in this
case the autocomplete uses an Ajax call that just uses the TermsComponent
to get the autocomplete data and display it. This is just textual, so adding
it to the
request is client-side magic.
If you want your app
> &fq={!tag=tag1}tags:( |1003| |1007|) AND tags:(
>|10015|)&version=2.2&start=0&rows=10&indent=on&facet=on&facet.field={!ex=tag1}category&facet.field=capacity&facet.field=brand
I'm just guessing here, but perhaps {!tag=tag1} is only picking up the 'tags:(
|1003| |1007|) '-part. If so {!ex=tag1} wo
Hi Erick,
in some usecases I really think that your suggestion with some
unique-documents for meta-information is a good approach to solve some
issues.
However there is a hurdle for me and maybe you can help me to clear it:
What is the best way to get such meta-data?
I see three possible approac
Thanks Em and Erick for your answers,
Now, i better understand functioning of Solr.
Damien
Le 24/01/2011 16:23, Erick Erickson a écrit :
First, the redundancy is certainly there, but that's what Solr does, handles
large
amounts of data. 4 million documents is actually a pretty small corpus by
I am attempting to do facets on products similar to how hayneedle does it on
their online stores (they do NOT use Solr). See:
http://www.clockstyle.com/wall-clocks/antiqued/1359+1429+4294885075.cfm
So simple example, my left nav might contain categories and 2 attributes,
brand and capacity:
Ca
Em,
that's correct. You can use 'lsof' to see file handles still in use.
See
http://0xfe.blogspot.com/2006/03/troubleshooting-unix-systems-with-lsof.html,
"Recipe #11".
-Alexander
On Sun, Jan 23, 2011 at 1:52 AM, Em wrote:
>
> Hi Alexander,
>
> thank you for your response.
>
> You said that th
Hi Thanks for the response.
I didn't explain myself well, I am using the field collapsing and things are
working as that page describes.
I think my problem is that as well as field collapsing works, solr is still
just returning a list of documents. There don't seem to be any operations I can
d
First, the redundancy is certainly there, but that's what Solr does, handles
large
amounts of data. 4 million documents is actually a pretty small corpus by
Solr
standards, so you may well be able to do exactly what you propose with
acceptable performance/size. I'd advise just trying it with, say,
Yes, you can =) Prefix & Suffix, both is working fine
On Sun, Jan 23, 2011 at 9:54 PM, Geert-Jan Brits wrote:
> Yep you can. Although I'm not sure you can use a wildcard-prefix. (perhaps
> you can I'm just not sure) . I always use wildcard-suffixes.
>
> Cheers,
> Geert-Jan
>
> 2011/1/23 Dennis G
i got this query from the mailing list.
but i found the problem. wrong query. i dont know why i construct my query
so ... =(
but thanks for your help =)
-
--- System
One Server, 12 GB RAM, 2 Solr Instances, 7 Cores,
You need to get more familiar with debugging, spending the time on it
is well worth the effort.
But assuming the '+' in your pasted query are really URL-encoded spaces
your syntax is really confused.
sender:(name1 name2 name3 OR sender_2:name1 name2 name3)
It *looks* like you intend something li
Just for illustration:
This is your original data:
doc1 : hello world
doc2: hello daniem
doc3: hello pal
Now, Lucene produces something like this from the input:
hello: id_doc1,id_doc2,id_doc3
daniem: id_doc2
pal: id_doc3
Well, it's more complex, but enough for illustration.
As you can see, th
Hi MapButcher,
There are a couple things that are going on here.
1. The spatial functionality is confusing between versions of Solr. I wish
someone would update the solr Spatial Search wiki page.
2. You will want to use the jTDS Driver here instead of the one from
Microsoft. http://jtds.sour
I mean, when you run the DIH, what's the output of the Solr Log ? Probably
there is more info about whats happening...
On Mon, Jan 24, 2011 at 10:28 AM, Dinesh wrote:
>
> its a DHCP log.. i want ti index it
>
> -
> DINESHKUMAR . M
> I am neither especially clever nor especially gifted. I am on
Le 24/01/2011 13:10, Em a écrit :
Hi Daniem,
ahm, the formula I wrote was no definitive guide, just some numbers I
combined to visualize the amount of data - perhaps not even a complete
formula.
Well, when you can use your taxonomy as indexed-only you do not double the
used disk space when yo
its a DHCP log.. i want ti index it
-
DINESHKUMAR . M
I am neither especially clever nor especially gifted. I am only very, very
curious.
--
View this message in context:
http://lucene.472066.n3.nabble.com/please-help-Problem-with-dataImportHandler-tp2318585p2319627.html
Sent from the Solr
And what the logs says about it?
On Mon, Jan 24, 2011 at 7:15 AM, Dinesh wrote:
>
> actually its a log file i seperately created an handler for that... its not
> XML
>
> -
> DINESHKUMAR . M
> I am neither especially clever nor especially gifted. I am only very, very
> curious.
> --
> View thi
Hi Daniem,
ahm, the formula I wrote was no definitive guide, just some numbers I
combined to visualize the amount of data - perhaps not even a complete
formula.
Well, when you can use your taxonomy as indexed-only you do not double the
used disk space when you are indexing two equal documents.
Having found some code that searches a Lucene index, the only analyzers
referenced are Lucene.Net.Analysis.Standard.StandardAnalyzer.
How can I map this is Solr? The example schema doesn't seem to mention
this, and specifying 'text' or 'string' for every field doesn't seem to
help.
Thanks
Le
We can't guess what's wrong with the cores but you need to reindex anyway:
http://svn.apache.org/repos/asf/lucene/solr/branches/branch-1.4/CHANGES.txt
On Monday 24 January 2011 12:06:10 Prasad Joshi wrote:
> Hi,
> I want to migrate from 1.4.0 to 1.4.1 . Tried keeping the same conf for the
> cores
Hi,
I want to migrate from 1.4.0 to 1.4.1 . Tried keeping the same conf for the
cores as in 1.4.0, added the relevant core names in solr.xml and restarted
solr but the old cores dont show up on the browser "localhost:8983". There
were a few cores in examples/multicore/ in the solr1.4.1 source from
that is my query: &q=sender:name1+name2+name3
exaclty the request is:
&q=sender:(name1+name2+name3+OR+sender_2:name1+name2+name3)
so solr is using another field for name2 and name3 ?
debugging cannot help me, or i dont understand the debugging ...
when i search only for "name1 + name2" search
It's all explained on the wiki:
http://wiki.apache.org/solr/SolrReplication#How_does_the_slave_replicate.3F
On Monday 24 January 2011 11:25:45 dhanesh wrote:
> Hi,
> I'm currently facing an issue with SOLR (exactly with the slaves
> replication) and after having spent quite a few time reading onl
This is not the fieldType but your query that is giving you trouble. You only
specify fieldName for value name1, so Solr will use defaultField for values
name2 and name3. You also omitted an operator, so Solr will use
defaultOperator instead. See you schema.xml for the values for the defaults
a
Hello.
my field sender with fieldType=textgen cannot find any documents wich are
more than 2 tokens long.
->q=sender:name1 name2 name3 => 0 Documents found
WHY ???
that is my field (original from default schema.xml)
-
--- System
why should i use Solandra instead of NRT only with Solr and own
configurations !?
-
--- System
One Server, 12 GB RAM, 2 Solr Instances, 7 Cores,
1 Core with 31 Million Documents other Cores < 100.000
- Solr1 for Search-Re
Hi,
I'm currently facing an issue with SOLR (exactly with the slaves
replication) and after having spent quite a few time reading online I
find myself having to ask for some enlightenment.
To be more factual, here is the context that led me to this question.
If the website administrator edited an
Thanks Em,
How i can calculate index time, update time and space disk used by one
taxonomy ?
Le 24/01/2011 10:58, Em a écrit :
100 Entries per taxon?
Well, with Solr you got 100 taxon-entries * 4mio docs * 10 taxons.
If your indexed taxon-versions are looking okay, you could leave out the
DB-
actually its a log file i seperately created an handler for that... its not
XML
-
DINESHKUMAR . M
I am neither especially clever nor especially gifted. I am only very, very
curious.
--
View this message in context:
http://lucene.472066.n3.nabble.com/please-help-Problem-with-dataImportHandle
This may be a dumb question, but Is the xml encoded in UTF-8?
On Mon, Jan 24, 2011 at 7:08 AM, Dinesh wrote:
>
> this is the error that i'm getting.. no idea of what is it..
>
>
> /apache-solr-1.4.1/example/exampledocs# java -jar post.jar sample.txt
> SimplePostTool: version 1.2
> SimplePostTool:
this is the error that i'm getting.. no idea of what is it..
/apache-solr-1.4.1/example/exampledocs# java -jar post.jar sample.txt
SimplePostTool: version 1.2
SimplePostTool: WARNING: Make sure your XML documents are encoded in UTF-8,
other encodings are not currently supported
SimplePostTool:
Thank you for your response.
In what way is 'timestamp' not perfect?
I've looked into the SolrEntityProcessor and added a timestamp field to our
index.
However i'm struggling to work out a query to get the max value od the
timestamp field
and does the SolrEntityProcessor entity appear before the
100 Entries per taxon?
Well, with Solr you got 100 taxon-entries * 4mio docs * 10 taxons.
If your indexed taxon-versions are looking okay, you could leave out the
DB-overhead and could do everything in Solr.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Taxonomy-in-SOLR-t
Yes, i am not obliged to store taxonomies.
My taxonomies are type of
english_taxon_label = Berlin
english_taxon_type = location
english_taxon_hierarchy = 0/world
1/world/europe
2/world/europe/germany
Hi,
I'm a bit of a solr beginner. I have installed Solr 4.0 and I'm trying to
index some spatial data stored in a sql server instance. I'm using the
DataImportHandler here is my data-comfig.xml:
Hi Damien,
why are you storing the taxonomies?
When it comes to faceting, it only depends on indexed values. If there is a
meaningful difference between the indexed and the stored value, I would
prefer to use an RDBMs or something like that to reduce redundancy.
Does this help?
Regards
--
View
My schema :
Le 24/01/2011 09:56, Em a écrit :
Hi Damien,
can you provide a schema sample plus example-data?
Since your information is really general, I think no one can give you a
situation-specific advice.
Regards
Hi Dennis,
thank you for your answer, but didn't understand why you say it doesn't need
serialization. I'm with the option "C".
but the main question is, how to put into one field a result of many fields:
"SELECT * FROM".
thanks,
Rich
-Original Message-
From: Dennis Gearon [mailt
my solrconfig.xml
http://pastebin.com/XDg0L4di
my schema.xml
http://pastebin.com/3Vqvr3C0
my try.xml
http://pastebin.com/YWsB37ZW
-
DINESHKUMAR . M
I am neither especially clever nor especially gifted. I am only very, very
curious.
--
View this message in context:
http://lucene.472066.
Hi Damien,
can you provide a schema sample plus example-data?
Since your information is really general, I think no one can give you a
situation-specific advice.
Regards
--
View this message in context:
http://lucene.472066.n3.nabble.com/Taxonomy-in-SOLR-tp2317955p2318200.html
Sent from the Sol
Hi,
I am trying Solr and i have one question. In the schema that i set up,
there are 10 fields with always same data(hierarchical taxonomies) but
with 4 million
documents, space disk and indexing time must be big. I need this field
for auto complete. Is there another way to do this type of ope
85 matches
Mail list logo