I have now updated to solr 3.3 but segment.gen is still not replicated.
Any idea why, is it a bug or a feature?
Should I write a jira issue for it?
Regards
Bernd
Am 29.07.2011 14:10, schrieb Bernd Fehling:
Dear list,
is there a deeper logic behind why the segment.gen file is not
replicated w
Hi Erick,
thanks a lot!
This looks like a good idea:
Our queries with the "changeable" fields fits the join-idea from
https://issues.apache.org/jira/browse/SOLR-2272
because
- we do not need relevance ranking
- we can separate in a conjunction of a query with the "changeable" fields and
our oth
This file is actually optional; its there for redundancy in case the
filesystem is not "reliable" when listing a directory. Ie, normally,
we list the directory to find the latest segments_N file; but if this
is wrong (eg the file system might have stale a cache) then we
fallback to reading the seg
The development of the field collapse feature is a long and confusing story.
The main point is that SOLR-236 was never going to scale
and the performance in general was bad. A new approach was needed.
This was implemented in SOLR-1682 and added to the trunk (4.0-dev)
around September last year. Lat
Thank you for the many replies!
Like I said, I couldn't find anything in logs created by solr.
I just had a look at the /var/logs/messages and there wasn't anything
either.
What I mean by crash is that the process is still there and http GET
pings would return 200
but when i try visiting /sol
Am 04.08.2011 12:52, schrieb Michael McCandless:
This file is actually optional; its there for redundancy in case the
filesystem is not "reliable" when listing a directory. Ie, normally,
we list the directory to find the latest segments_N file; but if this
is wrong (eg the file system might ha
Concerning the downtime, we found a solution that works well for us. We
allready implemented an update mechanism so that when authors are changing
some content in the cms, the index regarding this piece of content gets
updated (delete than index again) as well.
All we had to do is:
1. Change the s
Hello folks,
i use solr 1.4.1 and every 2 to 6 hours i have indexing errors in my log
files.
on the client side:
2011-08-04 12:01:18,966 ERROR [Worker-242] IndexServiceImpl - Indexing
failed with SolrServerException.
Details: org.apache.commons.httpclient.ProtocolException: Unbuffered entity
encl
We have 16 shards on 4 physical servers. Shard size was determined by
measuring query response times as a function of doc count. Multiple shards
per server provides parallelism. In a VM environment, I would lean towards 1
shard per VM (with 1/4 the RAM). We implemented our own distributed search
(p
Dumb question time - you are using a 64 bit Java, and not a 32 bit Java?
Bob Sandiford | Lead Software Engineer | SirsiDynix
P: 800.288.8020 X6943 | bob.sandif...@sirsidynix.com
www.sirsidynix.com
> -Original Message-
> From: Bernd Fehling [mailto:bernd.fehl...@uni-bielefeld.de]
> Sent:
java version "1.6.0_21"
Java(TM) SE Runtime Environment (build 1.6.0_21-b06)
Java HotSpot(TM) 64-Bit Server VM (build 17.0-b16, mixed mode)
java: file format elf64-x86-64
Including the -d64 switch.
Am 04.08.2011 14:40, schrieb Bob Sandiford:
Dumb question time - you are using a 64 bit Java,
Ok thank you very much for clearing that up a little. I think another reason
I was confused was that the wiki page for grouping was based around the
original field collapsing plan at the time which led me to the jira and
hence the patch files, rant over!
Perhaps you can help to clarify if the curr
Well, the original page moved to:
http://wiki.apache.org/solr/FieldCollapsingUncommitted
Assuming that you're using Solr 3.3 you can't get the grouped result () with SolrJ.
I added grouping support to SolrJ some time ago and will be in Solr 3.4. You
can use a nightly 3.x build to use the grouping
On Thu, Aug 4, 2011 at 8:09 AM, alexander sulz wrote:
> Thank you for the many replies!
>
> Like I said, I couldn't find anything in logs created by solr.
> I just had a look at the /var/logs/messages and there wasn't anything
> either.
>
> What I mean by crash is that the process is still there a
For anyone who comes across this topic in the future, I "solved" the problem
this way: by agreement with the stakeholders, on the presumption that no one
would look at more than 5000 records, I modified my search code so that, if the
user selected to sort by the name, I set the row count to retu
On 8/4/2011 12:38 AM, Bernd Fehling wrote:
Hi Shawn,
the 0.05 seconds for search time at peek times (3 qps) is my target
for Solr.
The numbers for solr are from Solr's statistic report page. So 39.5
seconds
average per request is definately to long and I have to change to
sharding.
Solr rep
Check out Physcial memory/virtual memory usage.
RAM usage might be less but Physical memory usage goes up as you index more
documents.
It might be because of MMapDirectory which used MappedByteBuffer.
On Thu, Aug 4, 2011 at 7:38 PM, Yonik Seeley wrote:
> On Thu, Aug 4, 2011 at 8:09 AM, alexander
Hi Yonik
So I tested the join using the sample data below and the latest trunk. I still
got the same behaviour.
HOWEVER! In this case it was nothing to do with the patch or solr version. It
was the tokeniser splitting G1 into G and 1.
So thank you for a nice patch and your suggestions.
I do h
I have indexed around 1 million tweets ( using "text" dataType).
when I search the tweet with "#" OR "@" I dont get the exact result.
e.g. when I search for "#ipad" OR "@ipad" I get the result where ipad is
mentioned skipping the "#" and "@".
please suggest me, how to tune or what are filterF
On Thu, Aug 4, 2011 at 11:21 AM, wrote:
> Hi Yonik
>
> So I tested the join using the sample data below and the latest trunk. I
> still got the same behaviour.
>
> HOWEVER! In this case it was nothing to do with the patch or solr version. It
> was the tokeniser splitting G1 into G and 1.
Ah, g
It's the WordDelimiterFactory in your filter chain that's removing the
punctuation entirely from your index, I think.
Read up on what the WordDelimiter filter does, and what it's settings
are; decide how you want things to be tokenized in your index to get the
behavior your want; either get Wo
I'm at the point in my Solr deployment where I want to start using it
for autosuggest, but I've run into a snag. Because the fields that I
want to use for autosuggest are tokenized, I can only get single terms
out of it. I would like to have it find common phrases that are between
two and fiv
We handled similar requirement in our product kitchendaily.com by creating a
list of Search terms which were frequently searched over a period of time
and then building auto-suggestion index from this data. The constant updates
of this will allow you to support a well formed auto-suggest feature. T
On 8/4/2011 10:04 AM, Sethi, Parampreet wrote:
We handled similar requirement in our product kitchendaily.com by creating a
list of Search terms which were frequently searched over a period of time
and then building auto-suggestion index from this data. The constant updates
of this will allow you
Hi,
We are having a requirement where we are having almost 100,000 documents to
be indexed (atleast 20 fields). These fields are not having length greater
than 10 KB.
Also we are running parallel search for the same index.
We found that it is taking almost 3 min to index the entire documents.
S
Sorry for 15k Docs, it is taking 3 mins.
On Thu, Aug 4, 2011 at 10:07 PM, Naveen Gupta wrote:
> Hi,
>
> We are having a requirement where we are having almost 100,000 documents to
> be indexed (atleast 20 fields). These fields are not having length greater
> than 10 KB.
>
> Also we are running p
I think we should fix replication to copy it?
Mike McCandless
http://blog.mikemccandless.com
On Thu, Aug 4, 2011 at 8:16 AM, Bernd Fehling
wrote:
>
>
> Am 04.08.2011 12:52, schrieb Michael McCandless:
>>
>> This file is actually optional; its there for redundancy in case the
>> filesystem is no
Hi Tobias,
sadly, it seems you are right.
After a little bit investigation we also recognized that some names (we use
it for auto-completing author-names), are missing. And since it is a
distributed setup ...
But I am almost sure it worked with Solr 3.2.
Best regards,
Sebastian
--
View
On Thu, Aug 4, 2011 at 10:08 AM, Yonik Seeley
wrote:
>
> "ignores" means what? The request hangs? If so, could you get a thread dump?
>
> Do queries work (like /solr/select?q=*:*) ?
>
>> thous throwing no errors, no 503's.. It's like the server has a blackout and
>> stares blankly into space.
>
Hi Erik,
I have several "types" with different properties, but they are supposed
to be combined to one search.
Imagine a book with property "title" and a journal with property "name".
(the types in my project have of course more complex properties.)
So I created a new core with combined searc
Thanks Eric for your reply. I am aware of facet.sort, but I haven't used it.
I will try that though.
Can it handle the values below in the correct order?
Under 10
10 - 20
20 - 30
Above 30
Or
Small
Medium
Large
XL
...
My second question is that if Solr can't do that for the values above by
using
Hi, guys,
What's the best way (practice) to do index distribution at this moment?
Hadoop? or rsyncd (back to 3 years ago ;-)) ?
Thanks,
Yugang
No, it can not. It just sorts "alphabetically", actually by raw byte-order.
No other facet sorting functionality is available, and it would be
tricky to implement in a performant way because of the way lucene
works. But it would certainly be useful to me too if someone could
figure out a way
I have decided to use solr for indexing as well.
the types of searches im doing are professional/academic.
so for example, i need to match:
all over the following exactly from my source data:
"3.56",
"4 harv. l. rev. 45",
"187-532",
"3 llm 56",
"5 unts 8",
"6 u.n.t.s.
I'm not sure what you mean by "index distribution", that could possibly
mean several things.
But Solr has had a replication feature built into it from 1.4, that can
probably handle the same use cases as rsync, but better. So that may be
what you want.
There are certainly other experiments g
It depends. Okay, the source contains "4 harv. l. rev. 45" .
Do you want a user entered "harv." to ALSO match "harv" (without the
period) in source, and vice versa? Or do you require it NOT match? Or do
you not care?
The default filter analysis chain will index "4 harv. l. rev. 45"
essential
It can be achieved by creating own (app specific) custom comparators for
fields defined in schema.xml and having an extra attribute to specify the
comparator class in the field tag itself. But it will require changes in the
Solr to support this feature. (Not sure if it's feasible though just
throwi
Yes, I am talking about replication feature. I remember I tried rsync 3
years ago with solr 1.2. Just not sure if someone else have done anything
better than that during the last 3 years. ;-) Personally I am thinking about
using Hadoop and ZooKeeper. Has anyone tried those features?
I found a coupl
Hello all
I'm using multiple cores. I there's a directory named by the core and it
contains a subdir named data that contains a subdir named index that
contains a bunch of files that contain the data for my index.
Let's say I want to completely rebuild the index from scratch.
Can I delete the d
I ran into a problem when I deleted just the "index" directory; I deleted the
entire data directory and it was recreated on the next load. BTW, if you're
using the DIH, its default behavior is to remove all records on a full import,
so you can save yourself having to remove any actual files.
--
I am trying to post the json update request using
java.net.HttpURLConnection.
Parameters I am using:
url : http://localhost:8983/solr/update/json?commit=true
String data =
"[{\"id\" : \"TestDoc7\", \"title\" : \"test 7\"}, {\"id\" :
\"TestDoc8\", \"title\" : \"another test 8\"}]";
Hi,
I am using Solr 3.1 with the SolrJ client library. I can see that it is
possible to get the maximum score for your search by using the following:
response.getResults().getMaxScore()
I am wondering is there some simple solution to get the minimum score?
Many thanks.
How does one issue span queries in SOLR (Span, SpanNear, etc)? I've
done a bit of research and it seems that these are not supported. It
would seem that I need to implement a QueryParserPlugin to accomplish
this. Is this the correct path? Surely this has been done before. Does
anybody have links to
I noticed that lucene supports 'Nested Documents'. However - I
couldn't find mention of this feature within SOLR. Does anybody know
how to leverage this lucene feature through SOLR?
Thanks!
Josh Harness
Ok. After analysis, I narrowed the reduced results set to the fact that the
zipcode field is not indexed 'as is'. i.e the zipcode field values are
broken down into tokens and then stored. Hence, if there are 10 documents
with zipcode fields varying from 91000-91009, then the zipcode fields are
not
Off the top of my head you "maybe" you can get the number of results and
then
look at the last document and check its score. I believe the results
will be ordered by score?
On 08/04/2011 05:44 PM, Kissue Kissue wrote:
Hi,
I am using Solr 3.1 with the SolrJ client library. I can see that it is
Never mind, It was some stupid bug.
Figured it out.
Cheers,
Sharath
On Thu, Aug 4, 2011 at 2:35 PM, Sharath Jagannath
wrote:
> I am trying to post the json update request using
> java.net.HttpURLConnection.
>
> Parameters I am using:
>
> url : http://localhost:8983/solr/update/json?commit=true
Hello,
I would like to know the best way to load a huge synonym list into Solr.
I would like to do concept indexing (a.k.a category indexing) with Solr. For
example, I want to be able to index all cities and be able to search for all
of them using a special keyword, say 'CONCEPTcity', where 'CONC
https://issues.apache.org/jira/browse/LUCENE-3233
On Thu, Aug 4, 2011 at 7:24 PM, Arun Atreya wrote:
> Hello,
>
> I would like to know the best way to load a huge synonym list into Solr.
>
> I would like to do concept indexing (a.k.a category indexing) with Solr. For
> example, I want to be able
Hi
I would like to know whether i can add new fields while replicating index on
Slave. E.g. My Master has index with field "F1" which is created with type
"string". Now, i don't want "F1" as a type "string" & also have limitation
that i cannot change the field type at schema level.
Now, if i repl
This is perhaps a 'truly newbie' question. I am processing some files via
DIH handler/XPATH Processor. Some of the date fields in the XML are in
'Java Long format' i.e. just a big long number. I am wondering how to map
them Solr Date field. I used the DIH DateFormatTransformer for some other
'
You might have to do this with an external script. The DIH lets you
process fields with javascript or Groovy.
Also, somewhere in the DIH you can give an XSL stylesheet instead of
just an XPath.
On Thu, Aug 4, 2011 at 10:31 PM, solruser@9913 wrote:
> This is perhaps a 'truly newbie' question. I
you can give it a try with the facet.sort.
We had such a requirement for sorting facets by order determined by
other field and had to resort to a very crude way to get through it.
We pre-pended the facets values with the order in which it had to be
displayed ... and used the facet.sort to sort alp
53 matches
Mail list logo