Re: NOT operator not working
q=(sharp+AND+-deviceType:Access*) that is not the query type of dismax for q operator (deviceType:Access*). either you use as Filter query or write query like q=(sharp AND -Access*) - Thanx: Grijesh -- View this message in context: http://lucene.472066.n3.nabble.com/NOT-operator-not-working-tp2365831p2365952.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: A Maven archetype that helps packaging Solr as a standalone application embedded in Apache Tomcat
Hi Stefan, thanks a lot, I'm joining now! :) http://people.apache.org/~simonetripodi/ http://www.99soft.org/ On Thu, Jan 27, 2011 at 4:27 PM, Stefan Matheis wrote: > Simo, it's freenode.net > > On Thu, Jan 27, 2011 at 4:16 PM, Simone Tripodi > wrote: > >> Hi Paul, >> sorry I'm late but I've been in the middle of a conf call :( On which >> IRC server the #solr channel is? I'll reach you ASAP. >> Thanks a lot! >> Simo >> >> http://people.apache.org/~simonetripodi/ >> http://www.99soft.org/ >> >> >> >> On Thu, Jan 27, 2011 at 4:00 PM, Paul Libbrecht wrote: >> > >> > Le 27 janv. 2011 à 12:42, Simone Tripodi a écrit : >> >> thanks a lot for your feedbacks, much more than appreciated! :) >> > >> > One more anomaly I find: the license is in the output of the pom.xml. >> > I think this should not be the case. >> > *my* license should be there, not the license of the archetype. Or? >> > >> > paul >> >
Re: Does solr supports indexing of files other than UTF-8
Thanks paul. However I want to support local encoding files to be indexed. How would I achieve it? On Thu, Jan 27, 2011 at 2:46 PM, Paul Libbrecht wrote: > At least in java utf-8 transcoding is done on a stream basis. No issue > there. > > paul > > > Le 27 janv. 2011 à 09:51, prasad deshpande a écrit : > > > The size of docs can be huge, like suppose there are 800MB pdf file to > index > > it I need to translate it in UTF-8 and then send this file to index. Now > > suppose there can be any number of clients who can upload file. at that > time > > it will affect performance. and already our product support localization > > with local encoding. > > > > Thanks, > > Prasad > > > > On Thu, Jan 27, 2011 at 2:04 PM, Paul Libbrecht > wrote: > > > >> Why is converting documents to utf-8 not feasible? > >> Nowadays any platform offers such services. > >> > >> Can you give a detailed failure description (maybe with the URL to a > sample > >> document you post)? > >> > >> paul > >> > >> > >> Le 27 janv. 2011 à 07:31, prasad deshpande a écrit : > >>> I am able to successfully index/search non-Engilsh data(like Hebrew, > >>> Japnese) which was encoded in UTF-8. > >>> However, When I tried to index data which was encoded in local encoding > >> like > >>> Big5 for Japanese I could not see the desired results. > >>> The contents after indexing looked garbled for Big5 encoded document > when > >> I > >>> searched for all indexed documents. > >>> > >>> Converting a complete document in UTF-8 is not feasible. > >>> I am not very clear about how Solr support these localizations with > other > >>> than UTF-8 encoding. > >>> > >>> > >>> I verified below links > >>> 1. http://lucene.apache.org/java/3_0_3/api/all/index.html > >>> 2. http://wiki.apache.org/solr/LanguageAnalysis > >>> > >>> Thanks and Regards, > >>> Prasad > >> > >> > >
Execute a second request in a component
Hi, I would execute a second solr request in a component. Someone has already done that ? Regards, Damien
Re: configure httpclient to access solr with user credential on third party host
On Thu, 27 Jan 2011 13:47 -0800, "Darniz" wrote: > > thanks exaclty i asked my domain hosting provider and he provided me with > some other port > > i am wondering can i specify credentials without the port > > i mean when i open the browser and i type > www.mydomainmame/solr i get the tomcat auth login screen. > > in the same way can i configure the http client so that i dont have to > specify the port Sure. This likely means your traffic is going via Apache (on the default port 80) but there's no real problem with that. Upayavira --- Enterprise Search Consultant at Sourcesense UK, Making Sense of Open Source
Re: Solr for noSQL
On Thu, 27 Jan 2011 21:38 -0800, "Dennis Gearon" wrote: > Why not make one's own DIH handler, Lance? Personally, I don't like that approach. Solr is best related to as something of a black box that you configure, then push content to. Having Solr know about your data sources, and pull content in seems to me to be mixing concerns. I relate to the DIH as a useful tool for smaller sites or for prototyping, but would expect anything more substantial to require an indexing application that gives you full control over the indexing process. It could be a lightweight app that uses a MongoDB java client and SolrJ, and simply pulls from one and pushes to the other. If you don't want to run another JVM, it could run as a separate webapp within your Solr JVM. >From an architectural point of view, do you configure Mysql, or MongoDB for that matter, to pull content into itself? Likewise, Solr should be a service that listens, waiting to be given data. Upayavira --- Enterprise Search Consultant at Sourcesense UK, Making Sense of Open Source
Re: DismaxParser Query
yes. I know it's been done by other users, but haven't done it myself... Best Erick On Fri, Jan 28, 2011 at 12:37 AM, Isan Fulia wrote: > Hi all, > I am currently using solr1.4.1 .Do I need to apply patch for extended > dismax parser. > > On 28 January 2011 03:42, Erick Erickson wrote: > > > In general, patches are applied to the source tree and it's re-compiled. > > See: http://wiki.apache.org/solr/HowToContribute#Working_With_Patches > > > > This is pretty easy, and I do know that "some people" have applied the > > eDismax > > patch to the 1.4 code line, but I haven't done it myself. > > > > Best > > Erick > > > > On Thu, Jan 27, 2011 at 10:27 AM, Jonathan Rochkind > >wrote: > > > > > Yes, I think nested queries are the only way to do that, and yes, > nested > > > queries like Daniel's example work (I've done it myself). I haven't > > really > > > tried to get into understanding/demonstrating _exactly_ how the > relevance > > > ends up working on the overall master query in such a situation, but it > > sort > > > of works. > > > > > > (Just note that Daniel's example isn't quite right, I think you need > > double > > > quotes for the nested _query_, just check the wiki page/blog post on > > nested > > > queries). > > > > > > Does eDismax handle parens for order of operation too? If so, eDismax > is > > > probably the best/easiest solution, especially if you're trying to > parse > > an > > > incoming query from some OTHER format and translate it to something > that > > can > > > be sent to Solr, which is what I often do. > > > > > > I haven't messed with eDismax myself yet. Does anyone know if there's > > any > > > easy (easy!) way to get eDismax in a Solr 1.4? Any easy way to compile > > an > > > eDismax query parser on it's own that works with Solr 1.4, and then > just > > > drop it into your local lib/ for use with an existing Solr 1.4? > > > > > > Jonathan > > > > > > > > > From: Daniel Pötzinger [daniel.poetzin...@aoemedia.de] > > > Sent: Thursday, January 27, 2011 9:26 AM > > > To: solr-user@lucene.apache.org > > > Subject: AW: DismaxParser Query > > > > > > It may also be an option to mix the query parsers? > > > Something like this (not tested): > > > > > > q={!lucene}field1:test OR field2:test2 _query_:{!dismax qf=fields}+my > > > dismax -bad > > > > > > So you have the benefits of lucene and dismax parser > > > > > > -Ursprüngliche Nachricht- > > > Von: Erick Erickson [mailto:erickerick...@gmail.com] > > > Gesendet: Donnerstag, 27. Januar 2011 15:15 > > > An: solr-user@lucene.apache.org > > > Betreff: Re: DismaxParser Query > > > > > > What version of Solr are you using, and could you consider either 3x or > > > applying a patch to 1.4.1? Because eDismax (extended dismax) handles > the > > > full Lucene query language and probably works here. See the Solr > > > JIRA 1553 at https://issues.apache.org/jira/browse/SOLR-1553 > > > > > > Best > > > Erick > > > > > > On Thu, Jan 27, 2011 at 8:32 AM, Isan Fulia > > >wrote: > > > > > > > It worked by making mm=0 (it acted as OR operator) > > > > but how to handle this > > > > > > > > field1:((keyword1 AND keyword2) OR (keyword3 AND keyword4)) OR > > > > field2:((keyword1 AND keyword2) OR (keyword3 AND keyword4)) OR > > > > field3:((keyword1 AND keyword2) OR (keyword3 AND keyword4)) > > > > > > > > > > > > > > > > > > > > On 27 January 2011 17:06, lee carroll > > > > wrote: > > > > > > > > > sorry ignore that - we are on dismax here - look at mm param in the > > > docs > > > > > you can set this to achieve what you need > > > > > > > > > > On 27 January 2011 11:34, lee carroll < > lee.a.carr...@googlemail.com> > > > > > wrote: > > > > > > > > > > > the default operation can be set in your config to be "or" or on > > the > > > > > query > > > > > > something like q.op=OR > > > > > > > > > > > > > > > > > > > > > > > > On 27 January 2011 11:26, Isan Fulia > > > wrote: > > > > > > > > > > > >> but q="keyword1 keyword2" does AND operation not OR > > > > > >> > > > > > >> On 27 January 2011 16:22, lee carroll < > > lee.a.carr...@googlemail.com > > > > > > > > > >> wrote: > > > > > >> > > > > > >> > use dismax q for first three fields and a filter query for the > > 4th > > > > and > > > > > >> 5th > > > > > >> > fields > > > > > >> > so > > > > > >> > q="keyword1 keyword 2" > > > > > >> > qf = field1,feild2,field3 > > > > > >> > pf = field1,feild2,field3 > > > > > >> > mm=something sensible for you > > > > > >> > defType=dismax > > > > > >> > fq=" field4:(keyword3 OR keyword4) AND field5:(keyword5)" > > > > > >> > > > > > > >> > take a look at the dismax docs for extra params > > > > > >> > > > > > > >> > > > > > > >> > > > > > > >> > On 27 January 2011 08:52, Isan Fulia < > isan.fu...@germinait.com> > > > > > wrote: > > > > > >> > > > > > > >> > > Hi all, > > > > > >> > > The query for standard request handler is as follows > > > > > >> > > field1:(keyword1 OR keyword2) OR field2:(keyword1 OR > k
Re: Execute a second request in a component
What are you trying to do? That is, what's the use-case you're trying to realize? Because there's not enough information here to give a helpful reply... Best Erick On Fri, Jan 28, 2011 at 4:25 AM, Damien Fontaine wrote: > Hi, > > I would execute a second solr request in a component. Someone has already > done that ? > > Regards, > Damien >
Re: Solr for noSQL
I'll reply for Lance because I'm awake earlier ... To make your own DIH, you have to solve all the problems you'd have to solve to use a Java program connect to your datasource via JDBC, PLUS fit it into the DIH framework. Why do the extra work? The other thing is that writing your own code gives you much greater control over, say, error handling, exception handling, continue-or-abort decisions, etc. DIH is a good tool, don't get me wrong, but I prefer more control in production situations. Plus, connecting to Solr via SolrJ AND connecting to your database takes about 20 lines of code, it's not very complex. You can have that done pretty quickly... But if you'd rather make your own DIH, it's up to you. Best Erick On Fri, Jan 28, 2011 at 12:38 AM, Dennis Gearon wrote: > Why not make one's own DIH handler, Lance? > > Dennis Gearon > > > Signature Warning > > It is always a good idea to learn from your own mistakes. It is usually a > better > idea to learn from others’ mistakes, so you do not have to make them > yourself. > from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036' > > > EARTH has a Right To Life, > otherwise we all die. > > > > - Original Message > From: Lance Norskog > To: solr-user@lucene.apache.org > Sent: Thu, January 27, 2011 9:33:25 PM > Subject: Re: Solr for noSQL > > There no special connectors available to read from the key-value > stores like memcache/cassandra/mongodb. You would have to get a Java > client library for the DB and code your own dataimporthandler > datasource. I cannot recommend this; you should make your own program > to read data and upload to Solr with one of the Solr client libraries. > > Lance > > On 1/27/11, Jianbin Dai wrote: > > Hi, > > > > > > > > Do we have data import handler to fast read in data from noSQL database, > > specifically, MongoDB I am thinking to use? > > > > Or a more general question, how does Solr work with noSQL database? > > > > Thanks. > > > > > > > > Jianbin > > > > > > > > > > > -- > Lance Norskog > goks...@gmail.com > >
Re: Solr for noSQL
Have you tried indexing using HTTP POST, you just call your information or documents from your DB and store it in a variable, next you just loop the POST as many register you have, and problem solve. With this method it doesn't matter what kind of DB you are using... On 1/28/11 7:43 AM, "Erick Erickson" wrote: > I'll reply for Lance because I'm awake earlier ... > > To make your own DIH, you have to solve all the > problems you'd have to solve to use a Java program > connect to your datasource via JDBC, PLUS > fit it into the DIH framework. Why do the extra work? > > The other thing is that writing your own code gives > you much greater control over, say, error handling, > exception handling, continue-or-abort decisions, etc. > DIH is a good tool, don't get me wrong, but I prefer > more control in production situations. > > Plus, connecting to Solr via SolrJ AND > connecting to your database takes about 20 lines > of code, it's not very complex. You can have that > done pretty quickly... > > But if you'd rather make your own DIH, it's up to you. > > Best > Erick > > On Fri, Jan 28, 2011 at 12:38 AM, Dennis Gearon wrote: > >> Why not make one's own DIH handler, Lance? >> >> Dennis Gearon >> >> >> Signature Warning >> >> It is always a good idea to learn from your own mistakes. It is usually a >> better >> idea to learn from others¹ mistakes, so you do not have to make them >> yourself. >> from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036' >> >> >> EARTH has a Right To Life, >> otherwise we all die. >> >> >> >> - Original Message >> From: Lance Norskog >> To: solr-user@lucene.apache.org >> Sent: Thu, January 27, 2011 9:33:25 PM >> Subject: Re: Solr for noSQL >> >> There no special connectors available to read from the key-value >> stores like memcache/cassandra/mongodb. You would have to get a Java >> client library for the DB and code your own dataimporthandler >> datasource. I cannot recommend this; you should make your own program >> to read data and upload to Solr with one of the Solr client libraries. >> >> Lance >> >> On 1/27/11, Jianbin Dai wrote: >>> Hi, >>> >>> >>> >>> Do we have data import handler to fast read in data from noSQL database, >>> specifically, MongoDB I am thinking to use? >>> >>> Or a more general question, how does Solr work with noSQL database? >>> >>> Thanks. >>> >>> >>> >>> Jianbin >>> >>> >>> >>> >> >> >> -- >> Lance Norskog >> goks...@gmail.com >> >>
field=string with value: 0, 1 and 2
Hello- i am using shard-requests over several cores. each core has his own index and own schema. but every core have the field "status" ! regularly the status is 0 or 1. but one core can have the status: 0, 1 OR 2 --> the status field type i used is "string" but string make the cores (with only 0 and 1 values) to boolean: true and false. the core wich can have status 2 only indexing the value "0" "1" "2" ... so i cannot filter with shard request over these cores, because solr cannor find if you using status:true, when the field is "1" ... how can i say to solr that he index the 0,1 values as integer and not boolean ??? (fieldtype int shows the same behaviour) ?? thx - --- System One Server, 12 GB RAM, 2 Solr Instances, 7 Cores, 1 Core with 31 Million Documents other Cores < 100.000 - Solr1 for Search-Requests - commit every Minute - 4GB Xmx - Solr2 for Update-Request - delta every 2 Minutes - 4GB Xmx -- View this message in context: http://lucene.472066.n3.nabble.com/field-string-with-value-0-1-and-2-tp2367038p2367038.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: field=string with value: 0, 1 and 2
i think DIH is the problem !? when i am using "int" as type for the fields, solr NumberFormatException occurs - For input string: "false" - --- System One Server, 12 GB RAM, 2 Solr Instances, 7 Cores, 1 Core with 31 Million Documents other Cores < 100.000 - Solr1 for Search-Requests - commit every Minute - 4GB Xmx - Solr2 for Update-Request - delta every 2 Minutes - 4GB Xmx -- View this message in context: http://lucene.472066.n3.nabble.com/field-string-with-value-0-1-and-2-tp2367038p2367053.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Does solr supports indexing of files other than UTF-8
Use ICONV library in your server side language. Convert it to utf-8, store it with a filed describing what incoding it was in, and re encode it if you wish. Dennis Gearon Signature Warning It is always a good idea to learn from your own mistakes. It is usually a better idea to learn from others’ mistakes, so you do not have to make them yourself. from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036' EARTH has a Right To Life, otherwise we all die. - Original Message From: prasad deshpande To: solr-user@lucene.apache.org Sent: Fri, January 28, 2011 12:41:29 AM Subject: Re: Does solr supports indexing of files other than UTF-8 Thanks paul. However I want to support local encoding files to be indexed. How would I achieve it? On Thu, Jan 27, 2011 at 2:46 PM, Paul Libbrecht wrote: > At least in java utf-8 transcoding is done on a stream basis. No issue > there. > > paul > > > Le 27 janv. 2011 à 09:51, prasad deshpande a écrit : > > > The size of docs can be huge, like suppose there are 800MB pdf file to > index > > it I need to translate it in UTF-8 and then send this file to index. Now > > suppose there can be any number of clients who can upload file. at that > time > > it will affect performance. and already our product support localization > > with local encoding. > > > > Thanks, > > Prasad > > > > On Thu, Jan 27, 2011 at 2:04 PM, Paul Libbrecht > wrote: > > > >> Why is converting documents to utf-8 not feasible? > >> Nowadays any platform offers such services. > >> > >> Can you give a detailed failure description (maybe with the URL to a > sample > >> document you post)? > >> > >> paul > >> > >> > >> Le 27 janv. 2011 à 07:31, prasad deshpande a écrit : > >>> I am able to successfully index/search non-Engilsh data(like Hebrew, > >>> Japnese) which was encoded in UTF-8. > >>> However, When I tried to index data which was encoded in local encoding > >> like > >>> Big5 for Japanese I could not see the desired results. > >>> The contents after indexing looked garbled for Big5 encoded document > when > >> I > >>> searched for all indexed documents. > >>> > >>> Converting a complete document in UTF-8 is not feasible. > >>> I am not very clear about how Solr support these localizations with > other > >>> than UTF-8 encoding. > >>> > >>> > >>> I verified below links > >>> 1. http://lucene.apache.org/java/3_0_3/api/all/index.html > >>> 2. http://wiki.apache.org/solr/LanguageAnalysis > >>> > >>> Thanks and Regards, > >>> Prasad > >> > >> > >
Re: Solr for noSQL
Personally, I just create a view that flattens out the database and renames the fields as I desire. Then I call the view with the DIH to import it. Solr doesn't knwo anything about the databsae, except how to get a connection and fetch rows. And that's pretty darn useful, just that much less code to write. Dennis Gearon Signature Warning It is always a good idea to learn from your own mistakes. It is usually a better idea to learn from others’ mistakes, so you do not have to make them yourself. from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036' EARTH has a Right To Life, otherwise we all die. - Original Message From: Upayavira To: solr-user@lucene.apache.org Sent: Fri, January 28, 2011 1:41:42 AM Subject: Re: Solr for noSQL On Thu, 27 Jan 2011 21:38 -0800, "Dennis Gearon" wrote: > Why not make one's own DIH handler, Lance? Personally, I don't like that approach. Solr is best related to as something of a black box that you configure, then push content to. Having Solr know about your data sources, and pull content in seems to me to be mixing concerns. I relate to the DIH as a useful tool for smaller sites or for prototyping, but would expect anything more substantial to require an indexing application that gives you full control over the indexing process. It could be a lightweight app that uses a MongoDB java client and SolrJ, and simply pulls from one and pushes to the other. If you don't want to run another JVM, it could run as a separate webapp within your Solr JVM. From an architectural point of view, do you configure Mysql, or MongoDB for that matter, to pull content into itself? Likewise, Solr should be a service that listens, waiting to be given data. Upayavira --- Enterprise Search Consultant at Sourcesense UK, Making Sense of Open Source
Re: Does solr supports indexing of files other than UTF-8
On Thu, Jan 27, 2011 at 3:51 AM, prasad deshpande wrote: > The size of docs can be huge, like suppose there are 800MB pdf file to index > it I need to translate it in UTF-8 and then send this file to index. PDF is binary AFAIK... you shouldn't need to do any charset translation before sending it to solr, or any other extraction library. If you're using solr-cell then it's the Tika component that is responsible for pulling out the text in the right format. -Yonik http://lucidimagination.com
Re: Searching for negative numbers very slow
On Thu, Jan 27, 2011 at 6:32 PM, Simon Wistow wrote: > If I do > > qt=dismax > fq=uid:1 > > (or any other positive number) then queries are as quick as normal - in > the 20ms range. > > However, any of > > fq=uid:\-1 > > or > > fq=uid:[* TO -1] > > or > > fq=uid:[-1 to -1] > > or > > fq=-uid:[0 TO *] > > then queries are incredibly slow - in the 9 *second* range. That's odd - there should be nothing special about negative numbers. Here are a couple of ideas: - if you have a really big index and querying by a negative number is much more rare, it could just be that part of the index wasn't cached by the OS and so the query needs to hit the disk. This can happen with any term and a really big index - nothing special for negatives here. - if -1 is a really common value, it can be slower. is fq=uid:\-2 or other negative numbers really slow also? -Yonik http://lucidimagination.com
DIH populate multiple fields with one column
Hi All, I attempting to parse xml stored in a database as blobs and then store pieces of this xml such as titles and authors in separate fields. The above configuration works, but the issue is that I have to parse the XML again and again for each new field. The indexing time increases linearly. What I would like to do is get the Blob once, parse the xml into a DOM document and populate all the fields at once, in a single transformer. Is this possible? Could anyone point me in the right direction? Regards, Alexei -- View this message in context: http://lucene.472066.n3.nabble.com/DIH-populate-multiple-fields-with-one-column-tp2367379p2367379.html Sent from the Solr - User mailing list archive at Nabble.com.
Http Connection is hanging while deleteByQuery
Hi, I am trying to delete Solr documents using deleteByQuery(). Unfortunately it simply hangs, for a long time. I have tried using both SimpleHttpConnectionManager and MultiThreaded...(default) 1. SolrServer solr; SimpleHttpConnectionManager cm = new SimpleHttpConnectionManager(true); the below options were tried as nothing worked //cm.closeIdleConnections(0L); //cm.getParams().setConnectionTimeout(1); //cm.getParams().setMaxTotalConnections(10); HttpClient httpClient = new HttpClient(cm); solr = new CommonsHttpSolrServer(solrUrl, httpClient); 2. SolrServer solr; solr = new CommonsHttpSolrServer(solrUrl); Do anyone have any idea? Thank You. Regards, Shan -- View this message in context: http://lucene.472066.n3.nabble.com/Http-Connection-is-hanging-while-deleteByQuery-tp2367405p2367405.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Http Connection is hanging while deleteByQuery
Though it may not be needed, just to ad.. this is how I delete by query UpdateResponse updateResponse = solr.deleteByQuery(new SolrQuery("my_query")); -- View this message in context: http://lucene.472066.n3.nabble.com/Http-Connection-is-hanging-while-deleteByQuery-tp2367405p2367410.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: filter update by IP
On 01/24/2011 02:02 AM, Jonathan Rochkind wrote: Kind of personally curious_why_ it keeps coming up on the list so much. Is everyone trying to go into business vending Solr in the cloud to customers who will write their own apps, or are there some other less obvious (to me) use cases? In my case, it's to index stuff (content, csv ...) and to use it in javascript apps. I wrote a thin wrapper around solr to embed it in Lutece portal : http://goo.gl/f3pGA Lutece portal administrator can write and read solr and the rest can read it.
Re: configure httpclient to access solr with user credential on third party host
hi I just wrote a plugin for Lutece portal to restrict access to read/write of solr to Lutece admin and read access for the rest: http://goo.gl/f3pGA t.
Sending binary data as part of a query
Hi all, Here is what I am interested in doing: I would like to send a compressed integer bitset as a query to solr. The bitset integers represent my document ids and the results I want to get back is the facet data for those documents. I have successfully created a QueryComponent class that, assuming it has the integer bitset, can turn that into the necessary DocSetFilter to pass to the searcher, get back the facets, etc. That part all works right now because I'm using either canned or randomly generated bitsets on the server side. What I'm unsure how to do is actually send this compressed bitset from a client to solr as part of the query. From what I can tell, the Solr API classes that are involved in handling binary data as part of a request assume that the data is a document to be added. For instance, extending ContentStreamHandlerBase requires implementing some kind of document loader and an UpdateRequestProcessorChain and a bunch of other stuff that I don't really think I should need. Is there a simpler way? Anyone tried or succeeded in doing anything similar to this? Thanks, --jay
Filter Query, Filter Cache and Hit Ratio
Hi, I am looking for some more information on how the filter cache is working, and how the hit are incremented. We are using filter queries for certain predefined value, such as the timestamp:[2011-01-21T00:00:00Z+TO+NOW] (which is the current day). From what I understand from the documentation: "the filter cache stores the results of any filter queries ("fq" parameters) that Solr is explicitly asked to execute. (Each filter is executed and cached separately. When it's time to use them to limit the number of results returned by a query, this is done using set intersections.)" So, we were imagining that is two consecutive queries (as the one above) was using the same timestamp filter query, the second query will take advantage of the filter cache, and we would see the number of hits increasing (hit on the cached timestamp filter query) . However, this is not the case, the number of hits on the filter cache does not increase and stays very low. Is it normal ? INFO: [] webapp=/siren path=/select params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=true&fq=timestamp:[2011-01-21T00:00:00Z+TO+NOW]&fq=domain:my.wordpress.com&fsv=true} hits=0 status=0 QTime=139 INFO: [] webapp=/siren path=/select params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=true&fq=timestamp:[2011-01-21T00:00:00Z+TO+NOW]&fq=domain:syours.wordpress.com&fsv=true} hits=0 status=0 QTime=138 -- Renaud Delbru
RE: Filter Query, Filter Cache and Hit Ratio
Hi, You've used NOW in the range query which will give a date/time accurate to the millisecond, try using NOW\DAY Colin. > -Original Message- > From: Renaud Delbru [mailto:renaud.del...@deri.org] > Sent: Friday, January 28, 2011 2:22 PM > To: solr-user@lucene.apache.org > Subject: Filter Query, Filter Cache and Hit Ratio > > Hi, > > I am looking for some more information on how the filter cache is > working, and how the hit are incremented. > > We are using filter queries for certain predefined value, such as the > timestamp:[2011-01-21T00:00:00Z+TO+NOW] (which is the current day). > From > what I understand from the documentation: > "the filter cache stores the results of any filter queries ("fq" > parameters) that Solr is explicitly asked to execute. (Each filter is > executed and cached separately. When it's time to use them to limit the > number of results returned by a query, this is done using set > intersections.)" > So, we were imagining that is two consecutive queries (as the one > above) > was using the same timestamp filter query, the second query will take > advantage of the filter cache, and we would see the number of hits > increasing (hit on the cached timestamp filter query) . However, this > is > not the case, the number of hits on the filter cache does not increase > and stays very low. Is it normal ? > > INFO: [] webapp=/siren path=/select > params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=t > rue&fq=timestamp:[2011-01- > 21T00:00:00Z+TO+NOW]&fq=domain:my.wordpress.com&fsv=true} > hits=0 status=0 QTime=139 > INFO: [] webapp=/siren path=/select > params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=t > rue&fq=timestamp:[2011-01- > 21T00:00:00Z+TO+NOW]&fq=domain:syours.wordpress.com&fsv=true} > hits=0 status=0 QTime=138 > > -- > Renaud Delbru
RE: Filter Query, Filter Cache and Hit Ratio
Ooops, I meant NOW/DAY > -Original Message- > From: cbenn...@job.com [mailto:cbenn...@job.com] > Sent: Friday, January 28, 2011 3:37 PM > To: solr-user@lucene.apache.org > Subject: RE: Filter Query, Filter Cache and Hit Ratio > > Hi, > > You've used NOW in the range query which will give a date/time accurate > to > the millisecond, try using NOW\DAY > > Colin. > > > -Original Message- > > From: Renaud Delbru [mailto:renaud.del...@deri.org] > > Sent: Friday, January 28, 2011 2:22 PM > > To: solr-user@lucene.apache.org > > Subject: Filter Query, Filter Cache and Hit Ratio > > > > Hi, > > > > I am looking for some more information on how the filter cache is > > working, and how the hit are incremented. > > > > We are using filter queries for certain predefined value, such as the > > timestamp:[2011-01-21T00:00:00Z+TO+NOW] (which is the current day). > > From > > what I understand from the documentation: > > "the filter cache stores the results of any filter queries ("fq" > > parameters) that Solr is explicitly asked to execute. (Each filter is > > executed and cached separately. When it's time to use them to limit > the > > number of results returned by a query, this is done using set > > intersections.)" > > So, we were imagining that is two consecutive queries (as the one > > above) > > was using the same timestamp filter query, the second query will take > > advantage of the filter cache, and we would see the number of hits > > increasing (hit on the cached timestamp filter query) . However, this > > is > > not the case, the number of hits on the filter cache does not > increase > > and stays very low. Is it normal ? > > > > INFO: [] webapp=/siren path=/select > > > params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=t > > rue&fq=timestamp:[2011-01- > > 21T00:00:00Z+TO+NOW]&fq=domain:my.wordpress.com&fsv=true} > > hits=0 status=0 QTime=139 > > INFO: [] webapp=/siren path=/select > > > params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=t > > rue&fq=timestamp:[2011-01- > > 21T00:00:00Z+TO+NOW]&fq=domain:syours.wordpress.com&fsv=true} > > hits=0 status=0 QTime=138 > > > > -- > > Renaud Delbru > > >
Thoughts on USING dynamic fields for extending objects
Well, mid, next month we're going to start using dynamic fields as the relate to our business rules. Basically, in involves have a basic set of objects in code/database, and flattened for search in Solr. The MAIN business object is to be extendable by the custoemer, will still having to supply the required fields in the base object. We will use defined type, dynamic fields I had a question for those more experienced than I. We are thinking about two possible usage patterns: A/ User can add any field they want, as long as they use the right suffix for the field. Changing the schema can be done at will, and updating past objects is totally on the user. They get: 1/ Find within the field. 2/ Range queries 3/ other future, single field functionality later 4 B/ User can NOT add any field they want, but they must submit a schema, hopefully automated. The data still goes into the Solr index as dynamically accepted fields as long as they use the right suffix for the field. Changing the schema can be done at by submitting the new schema. Updating past objects is STILL totally on the user. They get: 1/ Find within the field. 2/ Range queries 3/ Various filter functions like: mandatory fields, acceptable ranges, minimum lengths on strings, and other processing. 4/ Other future, single field functionality later 5/ The ability to make their own copyfields for 'grouping' of their own fields. 'A' I see as most simple to administer, but possible has security holes? THAT's my main question, all thoughts welcom. 'B' is better as a value added service, but has a LOT more work on our site's end, I believe. We could also possibly do non acceptance of sensitive field names for security? Any thoughts much appreciateed. Dennis Gearon Signature Warning It is always a good idea to learn from your own mistakes. It is usually a better idea to learn from others’ mistakes, so you do not have to make them yourself. from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036' EARTH has a Right To Life, otherwise we all die.
Re: DIH populate multiple fields with one column
Sorry about this post. I'll RTFM more carefully next time. Resolved: Regards, Alexei -- View this message in context: http://lucene.472066.n3.nabble.com/DIH-populate-multiple-fields-with-one-column-tp2367379p2368071.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: edismax vs dismax
On Fri, Jan 28, 2011 at 3:00 PM, Thumuluri, Sai wrote: > I recently upgraded to Solr 1.4.1 from Solr 1.3 and with the upgrade > used edismax query parser. Here is my solrconfig.xml . When I search for > "mw verification and payment information" - I get no results with > defType set to edismax, It's probably a bit of natural language query parsing in edismax... - and is treated as AND (the lucene operator) in the appropriate context (i.e. we won't if it's at the start or end of the query, etc) - or is treated as OR in the appropriate context The lowercaseOperators parameter can control this, so try setting lowercaseOperators=false -Yonik http://lucidimagination.com > if I switch the deftype to dismax - I get the results I am looking for > > Can anyone explain, why this would be the case? I thought edismax is > dismax and more. > > Thank you, > > For 1.4.1 > default="true"> > > dismax > explicit > 0.01 > > body^1.0 title^10.0 name^3.0 taxonomy_names^2.0 tags_h1^5.0 > tags_h2_h3^3.0 tags_h4_h5_h6^2.0 tags_inline^1.0 > > > body^10.0 > > 4 > > 2<-25% > > *:* > > true > body > 3 > true > > body > 256 > > Sai Thumuluri
deleteById throwing SocketTimeoutException
Hello, We have a core with about 900K docs. Recently I have noticed that the deleteById query seems to always give me a SocketTimeoutException(stack trace is shown below). I cannot figure out why only deletion fails but not add/update. The SOLR client instance is created via spring wiring (configuration given below). Did anybody face the same issue ? How can I solve this issue ? Increasing the timeout did not help. Configuration -- http://:8080/solr-admin/core Code - String filename = delfile.getName(); String id = filename.replace("_search.xml", ""); log.debug("Deleting id " + id); UpdateResponse response = solrServer.deleteById(id); log.info("Deleting response for " + id + " is " + response); boolean success = Util.moveFile(delfile, delprocessedpath); /** * Now delete old successfully processed files so that full reindex * from processed and transformed folders will not process unwanted/deleted documents */ File transformedFile = new File(transformedpath + (filename.replace("_search.xml", "_fast.xml"))); if(transformedFile.exists()) { log.info("Deleting archived Transformed file: " + transformedFile.getAbsolutePath()); transformedFile.delete(); } File processedFile = new File(processedpath+filename); if(processedFile.exists()) { log.info("Deleting archived Processed file: " + processedFile.getAbsolutePath()); processedFile.delete(); } Stack Trace -- 2011-01-28 15:51:18,842-0500 ERROR [com.search.service.topics.feedprocessor.DeleteFeedProcessor] - Error deleting from Solr server for -AR2011011403385_search.xml org.apache.solr.client.solrj.SolrServerException: java.net.SocketTimeoutException: Read timed out at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:472) at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:243) at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105) at org.apache.solr.client.solrj.SolrServer.deleteById(SolrServer.java:102) at com.search.service.topics.feedprocessor.DeleteFeedProcessor.processDelete(DeleteFeedProcessor.java:76) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.springframework.util.MethodInvoker.invoke(MethodInvoker.java:276) at org.springframework.scheduling.quartz.MethodInvokingJobDetailFactoryBean$MethodInvokingJob.executeInternal(MethodInvokingJobDetailFactoryBean.java:260) at org.springframework.scheduling.quartz.QuartzJobBean.execute(QuartzJobBean.java:86) at org.quartz.core.JobRunShell.run(JobRunShell.java:202) at org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:525) Caused by: java.net.SocketTimeoutException: Read timed out at java.net.SocketInputStream.socketRead0(Native Method) at java.net.SocketInputStream.read(SocketInputStream.java:129) at java.io.BufferedInputStream.fill(BufferedInputStream.java:218) at java.io.BufferedInputStream.read(BufferedInputStream.java:237) at org.apache.commons.httpclient.HttpParser.readRawLine(HttpParser.java:78) at org.apache.commons.httpclient.HttpParser.readLine(HttpParser.java:106) at org.apache.commons.httpclient.HttpConnection.readLine(HttpConnection.java:1116) at org.apache.commons.httpclient.MultiThreadedHttpConnectionManager$HttpConnectionAdapter.readLine(MultiThreadedHttpConnectionManager.java:1413) at org.apache.commons.httpclient.HttpMethodBase.readStatusLine(HttpMethodBase.java:1973) at org.apache.commons.httpclient.HttpMethodBase.readResponse(HttpMethodBase.java:1735) at org.apache.commons.httpclient.HttpMethodBase.execute(HttpMethodBase.java:1098) at org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:398) at org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:171) at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397) at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:323) at org.apache.solr.cli
[Failure] to start Solr 4.0
All, I've checked out the latest code and built the root directory with ant compile and then I build the solr directory again using the ant dist command which gives me the lucene-libs directory and a couple others. Now Solr won't start. What am I missing??? This is as far as it gets. mini:example Adam$ java -jar start.jar 2011-01-28 17:14:23.402:INFO::Logging to STDERR via org.mortbay.log.StdErrLog 2011-01-28 17:14:23.605:INFO::jetty-6.1.26 2011-01-28 17:14:23.638:INFO::Started SocketConnector@0.0.0.0:8983 What couple possibly be the problem? Adam
Re: [Failure] to start Solr 4.0
I found the problem...You HAVE to build the Solr directory using "ant example" in order for the web application to start properly. Sorry to post so many times. Adam On Jan 28, 2011, at 5:20 PM, Adam Estrada wrote: > All, > > I've checked out the latest code and built the root directory with ant > compile and then I build the solr directory again using the ant dist command > which gives me the lucene-libs directory and a couple others. Now Solr won't > start. What am I missing??? This is as far as it gets. > > mini:example Adam$ java -jar start.jar > 2011-01-28 17:14:23.402:INFO::Logging to STDERR via org.mortbay.log.StdErrLog > 2011-01-28 17:14:23.605:INFO::jetty-6.1.26 > 2011-01-28 17:14:23.638:INFO::Started SocketConnector@0.0.0.0:8983 > > What couple possibly be the problem? > > Adam
Controlling Tika's metadata
Just getting my feet wet with the text extraction using both schema and solrconfig settings from the example directory in the 1.4 distribution, so I might miss something obvious. Trying to provide my own title (and discarding the one received through Tika's metadata) wasn't straightforward. I had to use the following: fmap.title=tika_title (to discard the Tika title) literal.attr_title=New Title (to provide the correct one) fmap.attr_title=title (to map it back to the field as I would like to use title in searches) Is there anything easier than the above? How can this best be generalized to other metadata provided by Tika (which in our use case will be mostly ignored, as it is provided separately)? Thanks in advance for your responses.
DeletepkQuery is not actually deleting the records, please help
Hello, I am trying to delete some records from my index with delta-import using deletePkQuery with the below config, the log prints the deleted documents and says delta import successfully, but when I search, my search results still have those deleted documents, i have already spent so much time researching it but couldn't find any solution. All my database updates are getting updated with deltaQuery and deltaImportQuery but not the deletes. Could any one suggest me any solution? URL: http://localhost:8983/dataimport?command=delta-import dataConfig.xml: Log file: [2011/01/28 16:56:26.498] Completed ModifiedRowKey for Entity: item rows obtained : 0 [2011/01/28 16:56:26.499] Completed DeletedRowKey for Entity: item rows obtained : 6 [2011/01/28 16:56:26.499] Completed parentDeltaQuery for Entity: item [2011/01/28 16:56:32.563] Deleting stale documents . . [2011/01/28 16:58:00.319] Deleting document: BAAH [2011/01/28 17:06:50.537] Deleting document: BAAI [2011/01/28 17:07:28.470] Deleting document: BAAL [2011/01/28 17:08:13.187] Deleting document: BAAM [2011/01/28 17:08:27.011] Deleting document: BAAJ [2011/01/28 17:08:44.218] Deleting document: BAAK [2011/01/28 17:09:13.487] Delta Import completed successfully [2011/01/28 17:09:32.174] Import completed successfully [2011/01/28 17:09:32.175] start commit(optimize=true,waitFlush=false,waitSearcher=true,expungeDeletes=false) .. [2011/01/28 17:09:32.212] autowarming Searcher@a44b35 main from Searcher@f41f34 main [2011/01/28 17:09:32.215] Read dataimport.properties [2011/01/28 17:09:32.217] Wrote last indexed time to dataimport.properties [2011/01/28 17:09:33.791] Time taken = 0:13:45.366 Any suggestions would be highly appreciated. Thanks, Prad. -- View this message in context: http://lucene.472066.n3.nabble.com/DeletepkQuery-is-not-actually-deleting-the-records-please-help-tp2368463p2368463.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: How to group result when search on multiple fields
I don't think this field collapsing can simply solve my problem after second thought. As I mentioned, user only type in a search phrase, and click on search. Underlying that, the application logic is going to compose search query against multiple fields (title_name, actors, directors, ...) by the search phrase/term. Therefore, search result can match search term with any of the fields above. However, for all results due to actor name match, i want to make a group which only list the first two actors. for that group, i want to put it to the right spot of the result based on the relevancy score of the best actor match. For example, if i search a keyword "Tom" (term match as well fuzzy match). there are matching result based on video title name as well as actor name result 1: <-- title match score: 1.0 title_name: tom's story actor: jamie lee result 2: <-- title match score: 0.95 title_name: tom green's store actor: joanne anderson result 3: <-- actor match actor 2: tommy jackson -- score: 0.5 actor 1: tim robin -- score: 0.4 See all actors match "Tom" result 4: <-- title match score: 0.333 title_name: atom theory actor: kevin sheen in this case, field collapsing can only achieve this: out of search result, It will list out all actors as long as there is title or actor match. For example: Assume only show top 2 result in each group group value: Anglelina Jolie numFound:13 id:1, title_name:tom's story id:2, title_name:tom green's store group value: tommy jackson numFound:1 id: 201, title_name: ... group value: kevin sheen numFound:1 id: 30, title_name: atom theory <-- even if Angelina Jolie not in the result because of actor name match. but because her movie title matching "tom" with highest relevance. she will still be number 1 group. This is different than what i expected. -- View this message in context: http://lucene.472066.n3.nabble.com/How-to-group-result-when-search-on-multiple-fields-tp2358441p2368496.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: How to group result when search on multiple fields
Some typo out there in my example: I mean the first 2 movies by angelina jolie. This is the correct example: result 1: <-- title match score: 1.0 title_name: tom's story actor: Angelina Jolie result 2: <-- title match score: 0.95 title_name: tom green's store actor: Angelina Jolie result 3: <-- actor match actor 2: tommy jackson -- score: 0.5 actor 1: tim robin -- score: 0.4 See all actors match "Tom" result 4: <-- title match score: 0.333 title_name: atom theory actor: kevin sheen Here is the corresponding result if field collasping (result grouping) is used: group value: Anglelina Jolie numFound:13 id:1, title_name:tom's story id:2, title_name:tom green's store group value: tommy jackson numFound:1 id: 201, title_name: ... group value: tim robin numFound:1 id: 202, title_name: ... group value: kevin sheen numFound:1 id: 30, title_name: atom theory -- View this message in context: http://lucene.472066.n3.nabble.com/How-to-group-result-when-search-on-multiple-fields-tp2358441p2368512.html Sent from the Solr - User mailing list archive at Nabble.com.
velocity tools in solr-contrib-velocity?
Hello list, can anyone tell me how I can plug the velocity tools into my solr? Do I understand correctly the following comment in the source: // Velocity context tools - TODO: make these pluggable that it's only hard-coded thus far? thanks in advance paul
solr - uima error
I tried to do the uima integration with solr I followed the steps in the readme file using the solr example folder as my solr instance but when I started solr, I got this error message HTTP ERROR 500 Problem accessing /solr/admin/. Reason: Severe errors in solr configuration. Check your log files for more detailed information on what may be wrong. - org.apache.solr.common.SolrException: Error loading class 'org.apache.solr.uima.processor.UIMAProcessorFactory' at org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:408) at org.apache.solr.core.SolrCore.createInstance(SolrCore.java:399) at org.apache.solr.core.SolrCore.createInitInstance(SolrCore.java:422) at org.apache.solr.core.SolrCore.initPlugins(SolrCore.java:1486) at org.apache.solr.update.processor.UpdateRequestProcessorChain.init(UpdateRequestProcessorChain.java:57) at org.apache.solr.core.SolrCore.createInitInstance(SolrCore.java:424) at org.apache.solr.core.SolrCore.initPlugins(SolrCore.java:1470) at org.apache.solr.core.SolrCore.initPlugins(SolrCore.java:1464) at org.apache.solr.core.SolrCore.loadUpdateProcessorChains(SolrCore.java:572) at org.apache.solr.core.SolrCore.(SolrCore.java:514) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:647) at org.apache.solr.core.CoreContainer.load(CoreContainer.java:399) at org.apache.solr.core.CoreContainer.load(CoreContainer.java:283) at org.apache.solr.core.CoreContainer$Initializer.initialize(CoreContainer.java:232) at org.apache.solr.servlet.SolrDispatchFilter.init(SolrDispatchFilter.java:85) at org.mortbay.jetty.servlet.FilterHolder.doStart(FilterHolder.java:97) at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50) at org.mortbay.jetty.servlet.ServletHandler.initialize(ServletHandler.java:713) at org.mortbay.jetty.servlet.Context.startContext(Context.java:140) at org.mortbay.jetty.webapp.WebAppContext.startContext(WebAppContext.java:1282) at org.mortbay.jetty.handler.ContextHandler.doStart(ContextHandler.java:518) at org.mortbay.jetty.webapp.WebAppContext.doStart(WebAppContext.java:499) at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50) at org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerCollection.java:152) at org.mortbay.jetty.handler.ContextHandlerCollection.doStart(ContextHandlerCollection.java:156) at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50) at org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerCollection.java:152) at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50) at org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrapper.java:130) at org.mortbay.jetty.Server.doStart(Server.java:224) at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50) at org.mortbay.xml.XmlConfiguration.main(XmlConfiguration.java:985) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.mortbay.start.Main.invokeMain(Main.java:194) at org.mortbay.start.Main.start(Main.java:534) at org.mortbay.start.Main.start(Main.java:441) at org.mortbay.start.Main.main(Main.java:119) Caused by: java.lang.ClassNotFoundException: org.apache.solr.uima.processor.UIMAProcessorFactory at java.net.URLClassLoader$1.run(Unknown Source) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(Unknown Source) at java.lang.ClassLoader.loadClass(Unknown Source) at java.net.FactoryURLClassLoader.loadClass(Unknown Source) at java.lang.ClassLoader.loadClass(Unknown Source) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Unknown Source) at org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:392) ... 39 more -- *Powered by Jetty://* I Searched for the class "IMAProcessorFactory" but I didn't find it . any body can help ? By the way I'm using solr trunk checked out today Jan 29th