http://10.1.0.10:8983/solr/payment/dataimport?commad=delta-import&debug=on
dont work. no debug is started =(
thanks. i will try mergefactor=2
--
View this message in context:
http://lucene.472066.n3.nabble.com/Best-practice-for-Delta-every-2-Minutes-tp1992714p1997595.html
Sent from the Solr - U
Okay, I'll see what I can do.
Also for what it is worth, if anyone is in London tomorrow, I'm giving a
presentation which covers this topic at the (free) Online Information
2010 exhibition at Kensington Olympia, at 3:20pm. Anyone interested is
welcome to come along. I believe we're hoping to vide
On Tue, 30 Nov 2010 23:11 -0800, "Dennis Gearon"
wrote:
> Wow, would you put a diagram somewhere up on the Solr site?
> Or, here, and I will put it somewhere there.
I'll see what I can do to make a diagram.
> And, what is a VIP?
Virtual IP. It is what a load balancer uses. You assign a 'virtua
Note, all extracted from http://wiki.apache.org/solr/SolrReplication
You'd put:
startup
commit
into every box you want to be able to act as a master, then use:
http://slave_host:port/solr/replication?command=fetchindex&masterUrl=
As the above page says bett
Thanks Upayavira, that sounds very good.
p.s.:
I read that page some weeks ago and didn't get back to check on it.
2010/12/1 Upayavira
> Note, all extracted from http://wiki.apache.org/solr/SolrReplication
>
> You'd put:
>
>
>
>
>startup
>commit
>
>
>
> into e
sorry for lost, following is my schema.xml config and I use IKTokenizer for
Chinese charactor
Hi
It seems work fine again after I change "author" field type from text to
string, could anybody give some info about it? very appriciated.
On Wed, Dec 1, 2010 at 5:20 PM, Jerry Li wrote:
> sorry for lost, following is my schema.xml config and I use IKTokenizer for
> Chinese charactor
>
>
>
Hi ,
I am a newbie of solr. I found it really interesting specially spetial search.
I am very interested to go in its depth but i am facing some problem to use it
as i have 1.4.1 version installed on my machine but the spetial search is a
feature of 4.0 version which is not released yet. I have
Hi,
I have some troubles with forming query for solr.
Here is my task :
I'm indexing objects with 3 fields, for example {field1, field2, filed3}
In solr's response I want to get object in special order :
1. Firstly I want to get objects where all 3 fields are matched
2. Then I want to get object
Hi
I've built a schema for a proof of concept and it is all working fairly
fine, niave maybe but fine.
However I think we might run into trouble in the future if we ever use
facets.
The data models train destination city routes from a origin city:
Doc:City
Name: cityname [uniq key]
CityTy
On Tue, Nov 30, 2010 at 7:51 PM, Martin Grotzke
wrote:
> On Tue, Nov 30, 2010 at 3:09 PM, Yonik Seeley
> wrote:
>> On Tue, Nov 30, 2010 at 8:24 AM, Martin Grotzke
>> wrote:
>>> Still I'm wondering, why this issue does not occur with the plain
>>> example solr setup with 2 indexed docs. Any expla
my thoughts exactly that it may seem fairly straightforward but i fear for
when a client wants a perfectly reasonable new feature to be added to their
report and SOLR simply cannot support this feature.
i am hoping we wont have any real issues with scalability as Loggly because
we dont index and
Hello.
do anyone using Solr-PHP-Client ?
how are you using mutltivalued fields with the method addFields() ?
solr says to me SCHWERWIEGEND: java.lang.NumberFormatException: empty String
when i send a raw xml like this:
24038608
778
reason1
reason1
in schema i defined:
why dont work thi
Thanx for the answer.
Is it possible to remove the QueryNorm??
so all the bf boost became an add of the solr score??
omitNorm is about fieldNorm or queryNorm??
thanx
Gastone
2010/11/30 Jayendra Patil
> fieldNorm is the combination of length of the field with index and query
> time boosts.
>
>
Hi,
also take a look at solandra:
https://github.com/tjake/Lucandra/tree/solandra
I don't have it in prod yet but regarding administration overhead it
looks very promising.
And you'll get some other neat features like (soft) real time, for free.
So its same like A) + C) + X) - Y) ;-)
Rega
check jteam's spatial search plugin.
very easy to install
Aisha Zafar a écrit
> Hi ,
>
> I am a newbie of solr. I found it really interesting specially spetial
> search. I am very interested to go in its depth but i am facing some problem
> to use it as i have 1.4.1 version installed on my
This is good timing I am/was just to embark on a spike if anyone is keen to
help out
On 30 Nov 2010, at 00:37, Mark wrote:
> The DataSource subclass route is what I will probably be interested in. Are
> there are working examples of this already out there?
>
> On 11/29/10 12:32 PM, Aaron Mort
> It seems work fine again after I change "author" field type
> from text to
> string, could anybody give some info about it? very
> appriciated.
http://wiki.apache.org/solr/FAQ#Why_Isn.27t_Sorting_Working_on_my_Text_Fields.3F
And also see Erick's explanation
http://search-lucene.com/m/7fnj1TtNd
Hi again,
actually trying to implement spellcheck on a different way, and had the idea
to access /solr/spellcheck to get all required data, before executing the
final query to /solr/select - but, that seemed to be impossible - since
there is no configuration option to change the /select part of th
oooh, sorry - used the wrong thread for my suggestion ... please, just
ignore this :)
On Wed, Dec 1, 2010 at 2:01 PM, Stefan Matheis <
matheis.ste...@googlemail.com> wrote:
> Hi again,
>
> actually trying to implement spellcheck on a different way, and had the
> idea to access /solr/spellcheck to
I agree with the Master with multiple slaves setup. Very easy using
the built-in java setup in 1.4.1. When we set this up it made our
developers think about how we were writing to Solr. We were using a
Delta Import Handler (DIH?) for most writes but our app was also
writing 'deletes' directly to
I'd think that facet.query would work for you, something like:
&facet=true&facet.query=FareJanStandard:[price1 TO
price2]&facet.query:fareJanStandard[price2 TO price3]
You can string as many facet.query clauses as you want, across as many
fields as you want, they're all
independent and will get the
1.4.1 spatial is pretty much superseded by "geospatial" in the current code,
you can
download a nightly build from here:
https://hudson.apache.org/hudson/
Scroll down to "Solr-trunk" and pick a nightly build that suits you. Follow
the link through
"build artifacts" and checkout/solr/dist and you'l
Hi Erick,
so if i understand you we could do something like:
if Jan is selected in the user interface and we have 10 price ranges
query would be 20 cluases in the query (10 * 2 fare clases)
if first is selected in the user interface and we have 10 price ranges
query would be 120 cluases (12 mont
If your index warmings take longer than two minutes, but you're doing a
commit every two minutes -- you're going to run into trouble with
overlapping index preperations, eventually leading to an OOM. Could
this be it?
On 11/30/2010 11:36 AM, Erick Erickson wrote:
I don't know, you'll have to
On 11/29/2010 5:43 PM, Robert Muir wrote:
On Mon, Nov 29, 2010 at 5:41 PM, Jonathan Rochkind wrote:
* As a tokenizer, I use the WhitespaceTokenizer.
* Then I apply a custom filter that looks for CJK chars, and re-tokenizes
any CJK chars into one-token-per-char. This custom filter was written b
(Jonathan, I apologize for emailing you twice, i meant to hit reply-all)
On Wed, Dec 1, 2010 at 10:49 AM, Jonathan Rochkind wrote:
>
> Wait, standardtokenizer already handles CJK and will put each CJK char into
> it's own token? Really? I had no idea! Is that documented anywhere, or you
> just
"if first is selected in the user interface and we have 10 price ranges
query would be 120 cluases (12 months * 10 price ranges)"
What would you intend to do with the returned facet-results in this
situation? I doubt you want to display 12 categories (1 for each month) ?
When a user hasn't select
Does anyone know how to index a pdf file with very big size (more than 100MB)?
Thanks so much,
Xiaohui
-Original Message-
From: Ma, Xiaohui (NIH/NLM/LHC) [C]
Sent: Tuesday, November 30, 2010 4:22 PM
To: 'solr-user@lucene.apache.org'
Subject: RE: how to set maxFieldLength to unlimitd
I s
Hmmm, that's getting to be a pretty clunky query sure enough. Now you're
going to
have to insure that HTTP request that long get through and stuff like
that
I'm reaching a bit here, but you can facet on a tokenized field. Although
that's not
often done there's no prohibition against it.
So, w
Hello All,
I have a simple problem;
In my "conf/dataimport.properties" i have "last_index_time" with this
format '%Y-%m-%d %H:%M:%S'
for example: last_index_time=2010-12-01 16\:53\:16.
But when i use this propertie in my data-config.conf the value format
began "%Y-%m-%d";
for example:
url="http:
You just can't set it to "unlimited". What you could do, is ignoring the
positions and put a filter in, that sets the token for all but the first token
to 0 (means the field length will be just 1, all tokens "stacked" on the first
position)
You could also break per page, so you put each "page" o
Geert
The UI would be something like:
user selections
for the facet price
max price: £100
fare class: any
city attributes facet
cityattribute1 etc: xxx
results displayed something like
Facet price
Standard fares [10]
First fares [3]
in Jan [9]
in feb [10]
in march [1]
etc
is this compatible wit
If I want to delete an entire index and start over, in previous versions of
Solr, you could stop Solr, delete all files in the index directory and restart
Solr. Solr would then create empty segments files and you could start
indexing. In Solr 3x if I delete all the files in the index directo
Thanks so much for your replay, Jan. I just found I cannot index pdf files with
the file size more than 20MB.
I use curl index them, didn't get any error either. Do you have any suggestions
to index pdf files with more than 20MB?
Thanks,
Xiaohui
-Original Message-
From: jan.kure...@no
Sorry Geert missed of the price value bit from the user interface so we'd
display
Facet price
Standard fares [10]
First fares [3]
When traveling
in Jan [9]
in feb [10]
in march [1]
Fare Price
0 - 25 : [20]
25 - 50: [10]
50 - 100 [2]
cheers lee c
On 1 December 2010 17:00, lee carroll wrote:
It has typically been when query traffic was lowest! We are at 12 GB heap, so
I will try to bump it to 14 GB. We have 64GB main memory installed now. Here
is our settings, do these look OK?
export JAVA_OPTS="-Xmx12228m -Xms12228m -XX:+UseConcMarkSweepGC
-XX:+CMSIncrementalMode"
-Origi
Hi,
thanks all, this has been very instructive. It looks like in the short
term using a combination of replication and sharding, based on Upayavira's
setup, might be the safest thing to do, while in the longer term following the
zookeeper integration and solandra development might provid
On Wed, Dec 1, 2010 at 11:01 AM, Robert Muir wrote:
> (Jonathan, I apologize for emailing you twice, i meant to hit reply-all)
>
> On Wed, Dec 1, 2010 at 10:49 AM, Jonathan Rochkind
> wrote:
> >
> > Wait, standardtokenizer already handles CJK and will put each CJK char
> into
> > it's own token?
On Wed, Dec 1, 2010 at 12:25 PM, Jacob Elder wrote:
>
> What does this mean to those of us on Solr 1.4 and Lucene 2.9.3? Does the
> current stable StandardTokenizer handle CJK?
>
yes
On Nov 30, 2010, at 5:16pm, Robert Petersen wrote:
What would I do with the heap dump though? Run one of those java heap
analyzers looking for memory leaks or something? I have no experience
with thoseI saw there was a bug fix in solr 1.4.1 for a 100 byte
memory
leak occurring on each commi
On 12/1/2010 10:12 AM, Burton-West, Tom wrote:
If I want to delete an entire index and start over, in previous versions of
Solr, you could stop Solr, delete all files in the index directory and restart
Solr. Solr would then create empty segments files and you could start
indexing. In Solr 3
Ok longer answer than anticipated (and good conceptual practice ;-)
Yeah I belief that would work if I understand correctly that:
'in Jan [9]
in feb [10]
in march [1]'
has nothing to do with pricing, but only with availability?
If so you could seperate it out as two seperate issues:
1. ) showi
Also, filtering and sorting on price can be done as well. Just be sure to
use the correct price- field.
Geert-Jan
2010/12/1 Geert-Jan Brits
> Ok longer answer than anticipated (and good conceptual practice ;-)
>
> Yeah I belief that would work if I understand correctly that:
>
> 'in Jan [9]
> in
Hi Geert,
Ok I think I follow. the magic is in the multi-valued field.
The only danger would be complexity if we allow users to multi select
months/prices/fare classes. For example they can search for first prices in
jan, april and november. I think what you describe is possible in this case
just
We are using a recent Solr 3.x (See below for exact version).
We have set the ramBufferSizeMB to 320 in both the indexDefaults and the
mainIndex sections of our solrconfig.xml:
320
20
We expected that this would mean that the index would not write to disk until
it reached somewhere approximate
I don't know about upload limitations, but for sure there are some in
the default settings, this could explain the limit of 20MB. Which
upload mechanism on solr side do you use? I guess this is not a lucene
problem but rather the http-layer of solr.
If you manage to stream your PDF and start
The ram efficiency (= size of segment once flushed divided by size of
RAM buffer) can vary drastically.
Because the in-RAM data structures must be "growable" (to append new
docs to the postings as they are encountered), the efficiency is never
100%. I think 50% is actually a "good" ram efficiency
Not sure how to write that subject line. I'm getting some weird behavior out
of the highlighter in Solr. It seems like an edge case, but I'm curious to
hear if this is known about, or if it's something worth looking into further.
Background:
I'm using Solr's highlighting facility to tag words
On 12/1/2010 12:13 PM, Burton-West, Tom wrote:
We have set the ramBufferSizeMB to 320 in both the indexDefaults and the
mainIndex sections of our solrconfig.xml:
320
20
We expected that this would mean that the index would not write to disk until
it reached somewhere approximately over 300MB
Indeed, selecting the best price for January OR April OR November and
sorting on it isn't possible with this solution (if that's what you mean).
However, any combination of selecting 1 month and/or 1 price-range and/or 1
fare-type IS possible.
2010/12/1 lee carroll
> Hi Geert,
>
> Ok I think I f
Thanks so much, Jan. I use curl to index pdf files. Is there other way to do it?
I changed it the positionIncrement to 0, I didn't get it work either.
Thanks,
Xiaohui
-Original Message-
From: jan.kure...@nokia.com [mailto:jan.kure...@nokia.com]
Sent: Wednesday, December 01, 2010 2:34 P
Thanks Mike,
Yes we have many unique terms due to dirty OCR and 400 languages and probably
lots of low doc freq terms as well (although with the ICUTokenizer and
ICUFoldingFilter we should get fewer terms due to bad tokenization and
normalization.)
Is this additional overhead because each uniq
also try to minimize maxWarming searchers to 1(?) or 2.
And decrease cache usage (especially autowarming) if possible at all.
But again: only if it doesn't affect performance ...
Regards,
Peter.
On Tue, Nov 30, 2010 at 6:04 PM, Robert Petersen wrote:
My question is this. Why in the world
On Wed, Dec 1, 2010 at 3:16 PM, Burton-West, Tom wrote:
> Thanks Mike,
>
> Yes we have many unique terms due to dirty OCR and 400 languages and probably
> lots of low doc freq terms as well (although with the ICUTokenizer and
> ICUFoldingFilter we should get fewer terms due to bad tokenization a
Good idea. Our farm is behind Akamai so that should be ok to do.
-Original Message-
From: Peter Karich [mailto:peat...@yahoo.de]
Sent: Wednesday, December 01, 2010 12:21 PM
To: solr-user@lucene.apache.org
Subject: Re: entire farm fails at the same time with OOM issues
also try to min
On Tue, Nov 30, 2010 at 10:07 AM, Robert Muir wrote:
> On Tue, Nov 30, 2010 at 9:45 AM, Jacob Elder wrote:
> > Right. CJK doesn't tend to have a lot of whitespace to begin with. In the
> > past, we were using a patched version of StandardTokenizer which treated
> > @twitteruser and #hashtag bett
I am trying to get spatial search to work on my Solr installation. I am running
version 1.4.1 with the Jayway Team spatial-solr-plugin. I am performing the
search with the following url:
http://localhost:8080/solr/select?wt=json&indent=true&q=title:Art%20Loft{!spatial%20lat=37.326375%20lng=-121.
Take this with a sizeable grain of salt as I haven't actually tried doing
this. But you might try using an IndexReader which it looks like you can get
from this class:
http://lucene.apache.org/solr/api/org/apache/solr/core/StandardIndexReaderFactory.html
sasank
On Tue, Nov 30, 2010 at 6:45 AM, L
On the face of it, this doesn't make sense, so perhaps you can explain a
bit.The doc IDs
from one Solr instance have no relation to the doc IDs from another Solr
instance. So anything
that uses doc IDs from one Solr instance to create a filter on another
instance doesn't seem
to be something you'd
Got it with thanks.
On Wed, Dec 1, 2010 at 8:02 PM, Ahmet Arslan wrote:
> > It seems work fine again after I change "author" field type
> > from text to
> > string, could anybody give some info about it? very
> > appriciated.
>
>
> http://wiki.apache.org/solr/FAQ#Why_Isn.27t_Sorting_Working_on_m
Try this...
http://localhost:8080/solr/select?wt=json&indent=true&q={!spatial%20lat=37.326375%20lng=-121.892639%20radius=3%20unit=km%20threadCount=3}title:Art%20Loft
- Original Message -
From: "Dennis Gearon"
To:
Sent: Wednesday, December 01, 2010 7:51 PM
Subject: spatial query parin
Is there any way that Windows 7 and disk drivers are not honoring the
fsync() calls? That would cause files and/or blocks to get saved out
of order.
On Tue, Nov 30, 2010 at 3:24 PM, Peter Sturge wrote:
> After a recent Windows 7 crash (:-\), upon restart, Solr starts giving
> LockObtainFailedExce
Thanks Jean-Sebastion. I forwarded it to my partner. His membership is still
being held up.
I'll be the go between until he has access.
Dennis Gearon
Signature Warning
It is always a good idea to learn from your own mistakes. It is usually a
better
idea to learn from others
I just saw the parameter 'lng' in your query... I believe it should be
'long'. Give it a try if the link I sent you is not working
- Original Message -
From: "Dennis Gearon"
To:
Sent: Wednesday, December 01, 2010 11:39 PM
Subject: Re: spatial query parinsg error:
org.apache.lucene.q
hi all,
What's the best way to programmatically-in-java get the 'maxDoc' attribute
(as seen on the stats.jsp page).
I don't see any hooks on the solrj api.
Currently I plan to use an http client to get stats.jsp (which returns xml)
and parse it using xpath.
If anyone can recommend a better appr
Forwarded to my partner, thx, will let you know.
Dennis Gearon
Signature Warning
It is always a good idea to learn from your own mistakes. It is usually a
better
idea to learn from others’ mistakes, so you do not have to make them yourself.
from 'http://blogs.techrepublic.co
(10/12/02 13:51), Will Milspec wrote:
hi all,
What's the best way to programmatically-in-java get the 'maxDoc' attribute
(as seen on the stats.jsp page).
I don't see any hooks on the solrj api.
Currently I plan to use an http client to get stats.jsp (which returns xml)
and parse it using xpath
My project has a couple custom caches that descend from FastLRUCache. These
worked fine in Solr 1.3. Then I started migrating my project to Solr 1.4.1 and
had problems during startup.
I believe the problem is that I attempt to access the core in the init process.
I currently use the deprecated
Hello all,
1)
I want to restrict access to Solr only in localhost. How to acheive that?
2)
If i want to allow the clients to search but not to delete? How to restric the
access?
Any thoughts?
Regards
Ganesh.
Send free SMS to your Friends on Mobile from your Yahoo! Messenger. Download
Now! ht
70 matches
Mail list logo