I've implemented a fairly simple UpdateRequestProcessor much like the example
here: http://wiki.apache.org/solr/UpdateRequestProcessor
I attempted the below configuration in solrconfig.xml (like the above link
shows) but nothing happens, no errorsnothing. Is this configuration
supposed to
> If i analyse this field type in analysis.jsp, the follwoing
> are the results
> if i give "running" its stems word to run which is fine
> If i give "machine" why is that it stems to "machin", now
> from where does
> this word come from
> If i give "revolutionary" it stems to "revolutionari", i
>
Each instance has its own ReplicationHandler instance/MBean. I guess
the problem is with the jmx implementation. both MBeans may be
registered with the same name
On Fri, Aug 28, 2009 at 2:04 AM, J G wrote:
>
> We have multiple solr webapps all running from the same WAR file. Each webapp
> is runn
could you provide more details on what exactly is that you have done?
On Fri, Aug 28, 2009 at 7:08 AM, Erik Earle wrote:
>
> I've read through the wiki for this and it explains most everything except
> where in the solrconfig.xml the goes.
>
> I tried it at the top level but that doesn't seem to
any attribute specified in solrcore.properties can be referenced in
solrconfig.xml/schema.xml. this has nothing specific with replication.
On Fri, Aug 28, 2009 at 8:19 AM, Ilan Rabinovitch wrote:
> Hello,
>
> I noticed the the documentation around Solr Replication in the wiki has
> recently change
On Aug 27, 2009, at 10:35 PM, Paul Tomblin wrote:
Yesterday or the day before, I asked specifically if I would need to
restart the Solr server if somebody else loaded data into the Solr
index using the EmbeddedServer, and I was told confidently that no,
the Solr server would see the new data as
Hello,
I noticed the the documentation around Solr Replication in the wiki has
recently changed to take Paul's patch into account (SOLR-1355).
I now see that with the current trunk of SOLR 1.4 it is possible to use
a single solrconfig.xml to define both master and slave configurations,
with
Yesterday or the day before, I asked specifically if I would need to
restart the Solr server if somebody else loaded data into the Solr
index using the EmbeddedServer, and I was told confidently that no,
the Solr server would see the new data as soon as it was committed.
So today I fired up the Sol
Hi,
In Schema.xml file,I am not able ot find splitOnCaseChange="1".
I am not looking for case sensitive search.
Let me know what file you are refering to?.
I am looking for exact match search only
Moreover for scenario 2 the KeywordTokenizerFactory
and EdgeNGramFilterFactory refers which link in
On Thu, Aug 27, 2009 at 9:24 PM, Paul Tomblin wrote:
>>cwd=/Users/ptomblin/apache-tomcat-6.0.20
>>SolrHome=/Users/ptomblin/src/lucidity/solr/
>
Ok, I've spotted the problem - while SolrHome is in the right place,
it's still looking for the data in
/Users/ptomblin/apache-tomcat-6.0.20/solr/data/
Hi,
We have integrated Solr index with Carrot2 Search Engine and able to get search
results.
In my search results page, by default Total Number of records matched for the
particular query is not getting displayed.
http://localhost:8089/carrot2-webapp-3.0.1/search?source=Solr&view=tree
I've read through the wiki for this and it explains most everything except
where in the solrconfig.xml the goes.
I tried it at the top level but that doesn't seem to do anything.
http://wiki.apache.org/solr/UpdateRequestProcessor
If you have configured multi-core, then all you need to do is use the
following url pattern:
http://:/solr//select?q=...
where is the name of the core you wish to query.
Uri
Khai Doan wrote:
Thanks Uri,
Now my question is: how can I specify which schema to query against?
Thanks!
Khai
On
I've loaded some data into my solr using the embedded server, and I
can see the data using Luke. I start up the web app, and it says
>cwd=/Users/ptomblin/apache-tomcat-6.0.20
>SolrHome=/Users/ptomblin/src/lucidity/solr/
I hit the "schema" button and it shows the correct schema. However,
if I t
Thanks Uri,
Now my question is: how can I specify which schema to query against?
Thanks!
Khai
On Thu, Aug 27, 2009 at 5:43 PM, Uri Boness wrote:
> Not in the same core. You can define multiple cores where each core is a
> separate solr instance except they all run within one container. each c
Hi Ahmet,
Yes, for display purpose. Okay, so I don't have to copy fields then.
Thank you very much.
R
On Fri, Aug 28, 2009 at 4:57 AM, AHMET ARSLAN wrote:
>
>
> --- On Thu, 8/27/09, Rihaed Tan wrote:
>
> > From: Rihaed Tan
> > Subject: Case insensitive search and original string
> > To: sol
Not in the same core. You can define multiple cores where each core is a
separate solr instance except they all run within one container. each
core has its own index, schema and configuration. If you want to compare
it to databases, then I guess a core is to Solr Server what a database
is to it
I guess if you have stored="true" then there is no problem.
2. If you don't use stored="true" you can still get access to term vectors,
which you can probably reuse to create fake field with same term vector in
an updated document... just an idea, may be I am wrong...
Reconstructing a the field
Can I get all the distinct values from the Solr "database", or do I
have to select everything and aggregate it myself?
--
http://www.linkedin.com/in/paultomblin
Shalin, the XML from solr admin for the relevant field is displaying as -
Moncrieff, Joan, Macauley, Peter and Epps, Janine 2006, “My Universe is Here�: Implications
For the Future of Academic Libraries From the Results of a Survey of
Researchers, vol. 38, no. 2, pp. 71-83.
The wei
Funtick wrote:
>
>>then 2) get all P's by ID, including facet counts, etc.
>>The problem I face with this solution is that I can have many matching P's
> (10,000+), so my second query will have many (10,000+) constraints.
>
> SOLR can automatically provide you P's with Counts, and it will be
>
Have you determined if the problem is on the indexing side or the
query side? I don't see any reason you should have to set/change any
encoding in the JVM.
-Yonik
http://www.lucidimagination.com
On Thu, Aug 27, 2009 at 7:03 PM, Bernadette
Houghton wrote:
> Hi Shalin, strangely, things still ar
Hi Shalin, strangely, things still aren't working. I've set the JAVA_OPTS
through either the GUI or to startup.bat, but absolutely no impact. Have tried
reindexing also, but still no impact - results such as -
“My Universe is Here�
bern
-Original Message-
From: Shalin Shekh
i have a field defined in my schema.xml file
If i analyse this field type in analysis.jsp, the follwoing are the results
if i give "running" its stems word to run which is fine
If i give "machine" why is that it stems to "machin", now fro
> Hi all,How would I go about
> implementing a 'tag cloud' with Solr1.3? All I
> want to do is to display a list of most occurring terms in
> the corpus. Is there an easy way to go about that in 1.3?
Yes http://localhost:8983/solr/admin/luke?fl=text&numTerms=100 will give you
top 100 most occur
Check also Liferay trunk and WIKI pages, it had similar problem - and they
have plugin for SOLR now, just a matter of configuration change - and search
implementation is SOLR... They use SolrJ to do this task, and generic
wrappers around search implementation (which could be anything)...
-Fuad
http
--- On Thu, 8/27/09, Rihaed Tan wrote:
> From: Rihaed Tan
> Subject: Case insensitive search and original string
> To: solr-user@lucene.apache.org
> Date: Thursday, August 27, 2009, 10:10 PM
> Hi,
> Totally a Solr newbie here. The docs and list have been
> helpful but I have a
> question on lo
I haven't read all messages in this thread yet, but I probably have an
answer to some questions...
1. You want to change schema.xml and to reindex, but you don't have access
to source documents (stored somewhere on Internet). But you probably use
stored="true" in your schema. Then, use SOLR as you
2009/8/27 Fuad Efendi :
> stored="true" means that this piece of info will be stored in a filesystem.
> So that your index will contain 1Mb of pure log PLUS some info related to
> indexing itself: terms, etc.
>
> Search speed is more important than index size...
Not if you run out of space for the
We have multiple solr webapps all running from the same WAR file. Each webapp
is running under the same Tomcat container and I consider each webapp the same
thing as a "slice" (or "instance"). I've configured the Tomcat container to
enable JMX and when I connect using JConsole I only see the re
I am currently prototyping the use of Alfresco Document Management that has
an internal Lucene to index all the documents managed by Alfresco.
What would I need to understand in order to integrate that Lucene Index into
a separate Solr installation?
I am new to Solr and am trying to use Solr to
Hi Eric,
I think I understand what you are saying but I'm not sure how it would work.
I think you are saying to have two different indexes, each one has the
same documents, but one has the hard-to-get fields and the other has the
easy-to-get fields. Then I would make the same query twice, once
Andrew,
Which version of Solr are you using?
There's an open issue to fix caching filters at the segment
level, which will not clear the caches on each commit, you can
vote to indicate your interest.
http://issues.apache.org/jira/browse/SOLR-1308
-J
On Thu, Aug 27, 2009 at 7:06 AM, Andrew Ingra
Hi,
I have a similar requirement to Matthew (from his post 2 years ago). Is this
still the way to go in storing both the ID and name/value for facet values?
I'm planning to use id#name format if this is still the case and doing a
prefix query. I believe this is a common requirement so I'd apprecia
Hi all,How would I go about implementing a 'tag cloud' with Solr1.3? All I
want to do is to display a list of most occurring terms in the corpus. Is
there an easy way to go about that in 1.3? I saw a couple of postings about
implementing it with TermVectorComponent but thats in 1.4.
I'd really appr
Hi Sébastien,
I've experienced the same issue but when using "range queries". Maybe this
might help you too.
I was trying to filter a query using a range as "[ B TO F ]" being case and
accent insensitive, and still get back the case and accent at results.
The solution have been NOT TOKENIZE the
Hi,
Totally a Solr newbie here. The docs and list have been helpful but I have a
question on lowercase / case insensitive search. Do you really need to have
another field (copied or not) to retain the original casing of a field?
So let's say I have a field with a type that is lowercased during ind
If you have a specific need, ask on this list. That worked for me. I
don't think I would have recognized KeywordAnalyzer as the one I wanted.
wunder
On Aug 27, 2009, at 11:32 AM, Joe Calderon wrote:
is there an online resource or a book that contains a thorough list of
tokenizers and filters
Hmmm.. I see, too bad. So, here's a crazy question: if you had to guess,
how much of these numbers come from Solr nowadays (compared to lucene
java and the other related projects)? (I know.. it is a crazy question,
but I had to ask :-))
Grant Ingersoll wrote:
On Aug 27, 2009, at 4:00 AM, Uri
stored="true" means that this piece of info will be stored in a filesystem.
So that your index will contain 1Mb of pure log PLUS some info related to
indexing itself: terms, etc.
Search speed is more important than index size...
And note this: message field contains actual log, stored="true", so
is there an online resource or a book that contains a thorough list of
tokenizers and filters available and their functionality?
http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters
is very helpful but i would like to go through additional filters to
make sure im not reinventing the wheel
Right... You know, if some of your data needs to updated frequently,
but other is updated once per year, and is really massive dataset,
then maybe splitting it up into separate cores? Since you mentioned
that you can't get the raw data again, you could just duplicate your
existing index by doing
Eric Pugh wrote:
Do you have to "reindex"? Are you meaning an optimize operation? You
can do an "update" by just sending Solr a new record, and letting Solr
deal with the removing and adding of the data.
The problem is that I can't easily create the new record. There is some
data that I no l
On Thu, Aug 27, 2009 at 12:27 PM, Shalin Shekhar
Mangar wrote:
> On Wed, Aug 26, 2009 at 11:53 PM, Ron Ellis wrote:
>
>> Hi Everyone,
>>
>> When trying to utilize the new HTTP based replication built into Solr 1.4 I
>> encounter a problem. When I view the replication admin page on the slave
>> all
Agreed, Solr uses random access bitsets everywhere so I'm thinking
this could be an improvement or at least a great option to enable and
try out. I'll update LUCENE-1536 so we can benchmark.
On Thu, Aug 27, 2009 at 4:06 AM, Michael
McCandless wrote:
> On Thu, Aug 27, 2009 at 6:30 AM, Grant Ingerso
On Thu, Aug 27, 2009 at 1:27 PM, Eric
Pugh wrote:
> You can just query Solr, find the records that you want (including all
> the website data). Update them, and then send the entire record back.
>
Correct me if I'm wrong, but I think you'd end up losing the fields
that are indexed but not stored.
Do you have to "reindex"? Are you meaning an optimize operation? You
can do an "update" by just sending Solr a new record, and letting Solr
deal with the removing and adding of the data.
You can just query Solr, find the records that you want (including all
the website data). Update them, and t
I realize there is no way to update particular fields in a solr record.
I know the recommendation is to delete the record from the index and
re-add it, but in my case, it is difficult to completely reindex, so
that creates problems with my work flow.
That is, the info that I use to create a so
: I have a valid xml document that begins:
how are you inspecting the document?
I suspect that what you actually have is a documenting containing hte
literal bytes "R&D" but some tool you are using to view the document is
displaying the $ to you as &
...OR...
your source document has
:Try running ant example and then run Solr.
right ... on a clean checkout, the solr.war needs to be built and copied
to the example directory, otherwise you are just running an empty jetty
server.
do you see anything in example/webapps?
: > 1 get the latest Solr from svn (R 808058)
: > 2 r
Hi All,
I need to build a Search system using Solr. I need to keep data of 30 days
which will be around 400GB.
I will be using Distributed Search with Master/Slaves (Data will be
published to each shard on round robin basis). My challenge is I need to
delete older than 30 days data (around 12GB)
Great site (fast from Canada), multilingual, hope you will get millions of
ads quickly and share your findings of SOLR faceting performance (don't
forget about SOLR HTTP-caching support!)
I am currently developing similar in Canada, http://www.casaGURU.com (and
hope to improve http://www.zoocasa.co
This looks great! Congratulations!
Feel free to add your site to the "Powered by Solr" page at
http://wiki.apache.org/solr/PublicServers
On Thu, Aug 27, 2009 at 5:34 PM, gwk wrote:
> Hello,
>
> Earlier this your our company decided to (finally :)) upgrade our website
> to something a little fas
Hello,
I started to use solr only recently using the ruby/rails sunspot-solr
client. I use solr on a slovak/czech data set and realized one not wanted
behaviour of the search. When the user searches an expression or word which
contains dicritics, letters like š, č, ť, ä, ô,... usually the spe
Hi,
If I have documents of type a, b and c but when I sort by some criteria,
lets say date,
can I make documents of kind c always appear at the bottom ?
So effectively I want one kind of records always appear at the bottom since
they don't have valid data,
whether sort is ascending or descending;
Yes it will.
Thanks.
On Wed, Aug 26, 2009 at 8:51 PM, Yonik Seeley wrote:
> SortableFloatField works in function queries... it's just that
> everyone goes through SortableFloatField.getValueSource() to create
> them. Will that work for you?
>
> -Yonik
> http://www.lucidimagination.com
>
>
> On W
Hi all,
I'm trying to work out the optimum cache settings for our Solr server, I'll
begin by outlining our usage.
Number of documents: approximately 25,000
Commit frequency: sometimes we do massive amounts of sequential commits,
most of the time its less frequent but still several times an hour
We
: JDPA/JDWP are for remote debugging of SUN JVM...
: It shouldn't be SOLR related... check configs of Resin...
right, it sounds like you probably already have another process that is
listening on that port (and older execution of resin that was never shut
down cleanly?) ...
: then, when we want
I can't reproduce any problem.
Are you using a recent nightly build?
See the example schema of a recent nightly build for the correct way
to define a Trie based field - the article / blog may be out of date.
Here's what I used to test the example data:
http://localhost:8983/solr/select?q=manufact
Dave Searle wrote:
Hi Gwk,
It's a nice clean site, easy to use and seems very fast, well done! How well
does it do in regards to SEO though? I noticed there's a lot of ajax going on
in the background to help speed things up for the user (love the sliders), but
seems to be lacking structure fo
Hi Gwk,
It's a nice clean site, easy to use and seems very fast, well done! How well
does it do in regards to SEO though? I noticed there's a lot of ajax going on
in the background to help speed things up for the user (love the sliders), but
seems to be lacking structure for the search engines.
Hello,
Earlier this your our company decided to (finally :)) upgrade our
website to something a little faster/prettier/maintainable-er. After
some research we decided on using Solr and after indexing our data for
the first time and trying some manual queries we were all amazed at the
speed. T
Hi,
We started a new Lucene user group in The Netherlands. In the last
couple of years we've notice an increasing demand and interest in Lucene
and Solr. We thought it's about time to have a centralize place where
people can have open discussions, trainings, and periodic meet-ups to
share kno
On Thu, Aug 27, 2009 at 6:30 AM, Grant Ingersoll wrote:
>> I am wondering... are new SOLR filtering features faster than standard
>> Lucene queries like
>> {query} AND {filter}???
>
> The new filtering features in Solr are just doing what Lucene started doing
> in 2.4 and that is using skipping wh
On Aug 27, 2009, at 4:00 AM, Uri Boness wrote:
Hi,
Where can I find general statistics about the Solr project. The only
thing I found is statistics about the Lucene project at:
http://people.apache.org/~vgritsenko/stats/projects/lucene.html#Downloads-N1008F
Now the question is whether thes
On Aug 26, 2009, at 10:24 PM, Fuad Efendi wrote:
I am wondering... are new SOLR filtering features faster than standard
Lucene queries like
{query} AND {filter}???
The new filtering features in Solr are just doing what Lucene started
doing in 2.4 and that is using skipping when possible. I
Fuad -
http://www.lucidimagination.com/blog/2009/05/27/filtered-query-performance-increases-for-solr-14/
Use fq=filter instead, generally speaking.
Erik
On Aug 26, 2009, at 10:24 PM, Fuad Efendi wrote:
I am wondering... are new SOLR filtering features faster than standard
Lucene que
>
> In Schema.xml file,I am not able ot find splitOnCaseChange="1".
>
Unless you have modified the stock field type definition of "text" field in
your core's schema.xml you should be able to find this property set for the
WordDelimiterFilterFactory. Read more here -
http://wiki.apache.org/solr/Anal
Hello.
Thanks for the hints. Still some trouble, though.
I added just the HTMLStripCharFilterFactory because, according to
documentation, it should also replace HTML entities. It did, but still
left a space after the entity, so I got two tokens from "Günther".
That seems like a bug?
Adding Mappi
Guys,
Thanks everyone who helped or tried to help me out with this issue.
After talking with a buddy of mine who uses solr, he said that XPath
exception seemed familiar. It turns out that right at the bottom of
the Solr Wiki install page is a troubleshooting section with one
entry... and it was re
Hi,
Where can I find general statistics about the Solr project. The only
thing I found is statistics about the Lucene project at:
http://people.apache.org/~vgritsenko/stats/projects/lucene.html#Downloads-N1008F
Now the question is whether these number include all lucene's
sub-projects (includ
Hello everyone,
after reading Grant's article about TrieRange capabilities on the lucid blog
I did some experimenting, but I have some trouble with the tdate type and I
was hoping that you guys could point me in the right direction.
So, basically I index a regular solr date field and use that for s
when you say a slice you mean one instance of solr? So your JMX
console is connecting to only one solr?
On Thu, Aug 27, 2009 at 3:19 AM, J G wrote:
>
> Thanks for the response.
>
> It's interesting because when I run jconsole all I can see is one
> ReplicationHandler jmx mbean. It looks like it i
On Thu, Aug 27, 2009 at 12:28 PM, Shalin Shekhar Mangar <
shalinman...@gmail.com> wrote:
> On Thu, Aug 27, 2009 at 1:05 AM, Ryan McKinley wrote:
>
>>
>> On Aug 26, 2009, at 3:33 PM, djain101 wrote:
>>
>>
>>> I have one quick question...
>>>
>>> If in solrconfig.xml, if it says ...
>>>
>>>
>>> ${s
74 matches
Mail list logo