Re: NOT operator not working

2011-01-28 Thread Grijesh

q=(sharp+AND+-deviceType:Access*)

that is not the query type of dismax for q operator (deviceType:Access*).
either you use as Filter query or write query like

q=(sharp AND -Access*)

-
Thanx:
Grijesh
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/NOT-operator-not-working-tp2365831p2365952.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: A Maven archetype that helps packaging Solr as a standalone application embedded in Apache Tomcat

2011-01-28 Thread Simone Tripodi
Hi Stefan,
thanks a lot, I'm joining now! :)

http://people.apache.org/~simonetripodi/
http://www.99soft.org/



On Thu, Jan 27, 2011 at 4:27 PM, Stefan Matheis
 wrote:
> Simo, it's freenode.net
>
> On Thu, Jan 27, 2011 at 4:16 PM, Simone Tripodi 
> wrote:
>
>> Hi Paul,
>> sorry I'm late but I've been in the middle of a conf call :( On which
>> IRC server the #solr channel is? I'll reach you ASAP.
>> Thanks a lot!
>> Simo
>>
>> http://people.apache.org/~simonetripodi/
>> http://www.99soft.org/
>>
>>
>>
>> On Thu, Jan 27, 2011 at 4:00 PM, Paul Libbrecht  wrote:
>> >
>> > Le 27 janv. 2011 à 12:42, Simone Tripodi a écrit :
>> >> thanks a lot for your feedbacks, much more than appreciated! :)
>> >
>> > One more anomaly I find: the license is in the output of the pom.xml.
>> > I think this should not be the case.
>> > *my* license should be there, not the license of the archetype. Or?
>> >
>> > paul
>>
>


Re: Does solr supports indexing of files other than UTF-8

2011-01-28 Thread prasad deshpande
Thanks paul.

However I want to support local encoding files to be indexed. How would I
achieve it?

On Thu, Jan 27, 2011 at 2:46 PM, Paul Libbrecht  wrote:

> At least in java utf-8 transcoding is done on a stream basis. No issue
> there.
>
> paul
>
>
> Le 27 janv. 2011 à 09:51, prasad deshpande a écrit :
>
> > The size of docs can be huge, like suppose there are 800MB pdf file to
> index
> > it I need to translate it in UTF-8 and then send this file to index. Now
> > suppose there can be any number of clients who can upload file. at that
> time
> > it will affect performance. and already our product support localization
> > with local encoding.
> >
> > Thanks,
> > Prasad
> >
> > On Thu, Jan 27, 2011 at 2:04 PM, Paul Libbrecht 
> wrote:
> >
> >> Why is converting documents to utf-8 not feasible?
> >> Nowadays any platform offers such services.
> >>
> >> Can you give a detailed failure description (maybe with the URL to a
> sample
> >> document you post)?
> >>
> >> paul
> >>
> >>
> >> Le 27 janv. 2011 à 07:31, prasad deshpande a écrit :
> >>> I am able to successfully index/search non-Engilsh data(like Hebrew,
> >>> Japnese) which was encoded in UTF-8.
> >>> However, When I tried to index data which was encoded in local encoding
> >> like
> >>> Big5 for Japanese I could not see the desired results.
> >>> The contents after indexing looked garbled for Big5 encoded document
> when
> >> I
> >>> searched for all indexed documents.
> >>>
> >>> Converting a complete document in UTF-8 is not feasible.
> >>> I am not very clear about how Solr support these localizations with
> other
> >>> than UTF-8 encoding.
> >>>
> >>>
> >>> I verified below links
> >>> 1. http://lucene.apache.org/java/3_0_3/api/all/index.html
> >>> 2.  http://wiki.apache.org/solr/LanguageAnalysis
> >>>
> >>> Thanks and Regards,
> >>> Prasad
> >>
> >>
>
>


Execute a second request in a component

2011-01-28 Thread Damien Fontaine

Hi,

I would execute a second solr request in a component. Someone has 
already done that ?


Regards,
Damien


Re: configure httpclient to access solr with user credential on third party host

2011-01-28 Thread Upayavira


On Thu, 27 Jan 2011 13:47 -0800, "Darniz" 
wrote:
> 
> thanks exaclty i asked my domain hosting provider and he provided me with
> some other port
> 
> i am wondering can i specify credentials without the port
> 
> i mean when i open the browser and i type 
> www.mydomainmame/solr i get the tomcat auth login screen.
> 
> in the same way can i configure the http client so that i dont have to
> specify the port 

Sure. This likely means your traffic is going via Apache (on the default
port 80) but there's no real problem with that.

Upayavira
--- 
Enterprise Search Consultant at Sourcesense UK, 
Making Sense of Open Source



Re: Solr for noSQL

2011-01-28 Thread Upayavira


On Thu, 27 Jan 2011 21:38 -0800, "Dennis Gearon" 
wrote:
> Why not make one's own DIH handler, Lance?

Personally, I don't like that approach. Solr is best related to as
something of a black box that you configure, then push content to.
Having Solr know about your data sources, and pull content in seems to
me to be mixing concerns.

I relate to the DIH as a useful tool for smaller sites or for
prototyping, but would expect anything more substantial to require an
indexing application that gives you full control over the indexing
process. It could be a lightweight app that uses a MongoDB java client
and SolrJ, and simply pulls from one and pushes to the other. If you
don't want to run another JVM, it could run as a separate webapp within
your Solr JVM.

>From an architectural point of view, do you configure Mysql, or MongoDB
for that matter, to pull content into itself? Likewise, Solr should be a
service that listens, waiting to be given data.

Upayavira
--- 
Enterprise Search Consultant at Sourcesense UK, 
Making Sense of Open Source



Re: DismaxParser Query

2011-01-28 Thread Erick Erickson
yes. I know it's been done by other users, but haven't done it myself...

Best
Erick

On Fri, Jan 28, 2011 at 12:37 AM, Isan Fulia wrote:

> Hi all,
> I am currently using solr1.4.1 .Do  I need to apply patch for extended
> dismax parser.
>
> On 28 January 2011 03:42, Erick Erickson  wrote:
>
> > In general, patches are applied to the source tree and it's re-compiled.
> > See: http://wiki.apache.org/solr/HowToContribute#Working_With_Patches
> >
> > This is pretty easy, and I do know that "some people" have applied the
> > eDismax
> > patch to the 1.4 code line, but I haven't done it myself.
> >
> > Best
> > Erick
> >
> > On Thu, Jan 27, 2011 at 10:27 AM, Jonathan Rochkind  > >wrote:
> >
> > > Yes, I think nested queries are the only way to do that, and yes,
> nested
> > > queries like Daniel's example work (I've done it myself).  I haven't
> > really
> > > tried to get into understanding/demonstrating _exactly_ how the
> relevance
> > > ends up working on the overall master query in such a situation, but it
> > sort
> > > of works.
> > >
> > > (Just note that Daniel's example isn't quite right, I think you need
> > double
> > > quotes for the nested _query_, just check the wiki page/blog post on
> > nested
> > > queries).
> > >
> > > Does eDismax handle parens for order of operation too?  If so, eDismax
> is
> > > probably the best/easiest solution, especially if you're trying to
> parse
> > an
> > > incoming query from some OTHER format and translate it to something
> that
> > can
> > > be sent to Solr, which is what I often do.
> > >
> > > I haven't messed with eDismax myself yet.  Does anyone know if there's
> > any
> > > easy (easy!) way to get eDismax in a Solr 1.4?  Any easy way to compile
> > an
> > > eDismax query parser on it's own that works with Solr 1.4, and then
> just
> > > drop it into your local lib/ for use with an existing Solr 1.4?
> > >
> > > Jonathan
> > >
> > > 
> > > From: Daniel Pötzinger [daniel.poetzin...@aoemedia.de]
> > > Sent: Thursday, January 27, 2011 9:26 AM
> > > To: solr-user@lucene.apache.org
> > > Subject: AW: DismaxParser Query
> > >
> > > It may also be an option to mix the query parsers?
> > > Something like this (not tested):
> > >
> > > q={!lucene}field1:test OR field2:test2 _query_:{!dismax qf=fields}+my
> > > dismax -bad
> > >
> > > So you have the benefits of lucene and dismax parser
> > >
> > > -Ursprüngliche Nachricht-
> > > Von: Erick Erickson [mailto:erickerick...@gmail.com]
> > > Gesendet: Donnerstag, 27. Januar 2011 15:15
> > > An: solr-user@lucene.apache.org
> > > Betreff: Re: DismaxParser Query
> > >
> > > What version of Solr are you using, and could you consider either 3x or
> > > applying a patch to 1.4.1? Because eDismax (extended dismax) handles
> the
> > > full Lucene query language and probably works here. See the Solr
> > > JIRA 1553 at https://issues.apache.org/jira/browse/SOLR-1553
> > >
> > > Best
> > > Erick
> > >
> > > On Thu, Jan 27, 2011 at 8:32 AM, Isan Fulia  > > >wrote:
> > >
> > > > It worked by making mm=0 (it acted as OR operator)
> > > > but how to handle this
> > > >
> > > > field1:((keyword1 AND keyword2) OR (keyword3 AND keyword4)) OR
> > > > field2:((keyword1 AND keyword2) OR (keyword3 AND keyword4)) OR
> > > > field3:((keyword1 AND keyword2) OR (keyword3 AND keyword4))
> > > >
> > > >
> > > >
> > > >
> > > > On 27 January 2011 17:06, lee carroll 
> > > > wrote:
> > > >
> > > > > sorry ignore that - we are on dismax here - look at mm param in the
> > > docs
> > > > > you can set this to achieve what you need
> > > > >
> > > > > On 27 January 2011 11:34, lee carroll <
> lee.a.carr...@googlemail.com>
> > > > > wrote:
> > > > >
> > > > > > the default operation can be set in your config to be "or" or on
> > the
> > > > > query
> > > > > > something like q.op=OR
> > > > > >
> > > > > >
> > > > > >
> > > > > > On 27 January 2011 11:26, Isan Fulia 
> > > wrote:
> > > > > >
> > > > > >> but q="keyword1 keyword2"  does AND operation  not OR
> > > > > >>
> > > > > >> On 27 January 2011 16:22, lee carroll <
> > lee.a.carr...@googlemail.com
> > > >
> > > > > >> wrote:
> > > > > >>
> > > > > >> > use dismax q for first three fields and a filter query for the
> > 4th
> > > > and
> > > > > >> 5th
> > > > > >> > fields
> > > > > >> > so
> > > > > >> > q="keyword1 keyword 2"
> > > > > >> > qf = field1,feild2,field3
> > > > > >> > pf = field1,feild2,field3
> > > > > >> > mm=something sensible for you
> > > > > >> > defType=dismax
> > > > > >> > fq=" field4:(keyword3 OR keyword4) AND field5:(keyword5)"
> > > > > >> >
> > > > > >> > take a look at the dismax docs for extra params
> > > > > >> >
> > > > > >> >
> > > > > >> >
> > > > > >> > On 27 January 2011 08:52, Isan Fulia <
> isan.fu...@germinait.com>
> > > > > wrote:
> > > > > >> >
> > > > > >> > > Hi all,
> > > > > >> > > The query for standard request handler is as follows
> > > > > >> > > field1:(keyword1 OR keyword2) OR field2:(keyword1 OR
> k

Re: Execute a second request in a component

2011-01-28 Thread Erick Erickson
What are you trying to do? That is, what's the use-case you're
trying to realize?

Because there's not enough information here to give a
helpful reply...

Best
Erick

On Fri, Jan 28, 2011 at 4:25 AM, Damien Fontaine wrote:

> Hi,
>
> I would execute a second solr request in a component. Someone has already
> done that ?
>
> Regards,
> Damien
>


Re: Solr for noSQL

2011-01-28 Thread Erick Erickson
I'll reply for Lance because I'm awake earlier ...

To make your own DIH, you have to solve all the
problems you'd have to solve to use a Java program
connect to your datasource via JDBC, PLUS
fit it into the DIH framework. Why do the extra work?

The other thing is that writing your own code gives
you much greater control over, say, error handling,
exception handling, continue-or-abort decisions, etc.
DIH is a good tool, don't get me wrong, but I prefer
more control in production situations.

Plus, connecting to Solr via SolrJ AND
connecting to your database takes about 20 lines
of code, it's not very complex. You can have that
done pretty quickly...

But if you'd rather make your own DIH, it's up to you.

Best
Erick

On Fri, Jan 28, 2011 at 12:38 AM, Dennis Gearon wrote:

> Why not make one's own DIH handler, Lance?
>
>  Dennis Gearon
>
>
> Signature Warning
> 
> It is always a good idea to learn from your own mistakes. It is usually a
> better
> idea to learn from others’ mistakes, so you do not have to make them
> yourself.
> from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036'
>
>
> EARTH has a Right To Life,
> otherwise we all die.
>
>
>
> - Original Message 
> From: Lance Norskog 
> To: solr-user@lucene.apache.org
> Sent: Thu, January 27, 2011 9:33:25 PM
> Subject: Re: Solr for noSQL
>
> There no special connectors available to read from the key-value
> stores like memcache/cassandra/mongodb. You would have to get a Java
> client library for the DB and code your own dataimporthandler
> datasource.  I cannot recommend this; you should make your own program
> to read data and upload to Solr with one of the Solr client libraries.
>
> Lance
>
> On 1/27/11, Jianbin Dai  wrote:
> > Hi,
> >
> >
> >
> > Do we have data import handler to fast read in data from noSQL database,
> > specifically, MongoDB I am thinking to use?
> >
> > Or a more general question, how does Solr work with noSQL database?
> >
> > Thanks.
> >
> >
> >
> > Jianbin
> >
> >
> >
> >
>
>
> --
> Lance Norskog
> goks...@gmail.com
>
>


Re: Solr for noSQL

2011-01-28 Thread Alejandro Delgadillo
Have you tried indexing using HTTP POST, you just call your information or
documents from your DB and store it in a variable, next you just loop the
POST as many register you have, and problem solve.

With this method it doesn't matter what kind of DB you are using...


On 1/28/11 7:43 AM, "Erick Erickson"  wrote:

> I'll reply for Lance because I'm awake earlier ...
> 
> To make your own DIH, you have to solve all the
> problems you'd have to solve to use a Java program
> connect to your datasource via JDBC, PLUS
> fit it into the DIH framework. Why do the extra work?
> 
> The other thing is that writing your own code gives
> you much greater control over, say, error handling,
> exception handling, continue-or-abort decisions, etc.
> DIH is a good tool, don't get me wrong, but I prefer
> more control in production situations.
> 
> Plus, connecting to Solr via SolrJ AND
> connecting to your database takes about 20 lines
> of code, it's not very complex. You can have that
> done pretty quickly...
> 
> But if you'd rather make your own DIH, it's up to you.
> 
> Best
> Erick
> 
> On Fri, Jan 28, 2011 at 12:38 AM, Dennis Gearon wrote:
> 
>> Why not make one's own DIH handler, Lance?
>> 
>>  Dennis Gearon
>> 
>> 
>> Signature Warning
>> 
>> It is always a good idea to learn from your own mistakes. It is usually a
>> better
>> idea to learn from others¹ mistakes, so you do not have to make them
>> yourself.
>> from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036'
>> 
>> 
>> EARTH has a Right To Life,
>> otherwise we all die.
>> 
>> 
>> 
>> - Original Message 
>> From: Lance Norskog 
>> To: solr-user@lucene.apache.org
>> Sent: Thu, January 27, 2011 9:33:25 PM
>> Subject: Re: Solr for noSQL
>> 
>> There no special connectors available to read from the key-value
>> stores like memcache/cassandra/mongodb. You would have to get a Java
>> client library for the DB and code your own dataimporthandler
>> datasource.  I cannot recommend this; you should make your own program
>> to read data and upload to Solr with one of the Solr client libraries.
>> 
>> Lance
>> 
>> On 1/27/11, Jianbin Dai  wrote:
>>> Hi,
>>> 
>>> 
>>> 
>>> Do we have data import handler to fast read in data from noSQL database,
>>> specifically, MongoDB I am thinking to use?
>>> 
>>> Or a more general question, how does Solr work with noSQL database?
>>> 
>>> Thanks.
>>> 
>>> 
>>> 
>>> Jianbin
>>> 
>>> 
>>> 
>>> 
>> 
>> 
>> --
>> Lance Norskog
>> goks...@gmail.com
>> 
>> 




field=string with value: 0, 1 and 2

2011-01-28 Thread stockii

Hello-


i am using shard-requests over several cores.
each core has his own index and own schema. but every core have the field
"status" !

regularly the status is 0 or 1. but one core can have the status: 0, 1 OR 2

--> the status field type i used is "string" but string make the cores (with
only 0 and 1 values) to boolean: true and false. the core wich can have
status 2 only indexing the value "0" "1" "2" ...

so i cannot filter with shard request over these cores, because solr cannor
find if you using status:true, when the field is "1" ...

how can i say to solr that he index the 0,1 values as integer and not
boolean ??? (fieldtype int shows the same behaviour)

?? thx


-
--- System


One Server, 12 GB RAM, 2 Solr Instances, 7 Cores, 
1 Core with 31 Million Documents other Cores < 100.000

- Solr1 for Search-Requests - commit every Minute  - 4GB Xmx
- Solr2 for Update-Request  - delta every 2 Minutes - 4GB Xmx
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/field-string-with-value-0-1-and-2-tp2367038p2367038.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: field=string with value: 0, 1 and 2

2011-01-28 Thread stockii

i think DIH is the problem !? 

when i am using "int" as type for the fields, solr NumberFormatException
occurs - For input string: "false"


-
--- System


One Server, 12 GB RAM, 2 Solr Instances, 7 Cores, 
1 Core with 31 Million Documents other Cores < 100.000

- Solr1 for Search-Requests - commit every Minute  - 4GB Xmx
- Solr2 for Update-Request  - delta every 2 Minutes - 4GB Xmx
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/field-string-with-value-0-1-and-2-tp2367038p2367053.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Does solr supports indexing of files other than UTF-8

2011-01-28 Thread Dennis Gearon
Use ICONV library in your server side language.

Convert it to utf-8, store it with a filed describing what incoding it was in, 
and re encode it if you wish.

 Dennis Gearon


Signature Warning

It is always a good idea to learn from your own mistakes. It is usually a 
better 
idea to learn from others’ mistakes, so you do not have to make them yourself. 
from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036'


EARTH has a Right To Life,
otherwise we all die.



- Original Message 
From: prasad deshpande 
To: solr-user@lucene.apache.org
Sent: Fri, January 28, 2011 12:41:29 AM
Subject: Re: Does solr supports indexing of files other than UTF-8

Thanks paul.

However I want to support local encoding files to be indexed. How would I
achieve it?

On Thu, Jan 27, 2011 at 2:46 PM, Paul Libbrecht  wrote:

> At least in java utf-8 transcoding is done on a stream basis. No issue
> there.
>
> paul
>
>
> Le 27 janv. 2011 à 09:51, prasad deshpande a écrit :
>
> > The size of docs can be huge, like suppose there are 800MB pdf file to
> index
> > it I need to translate it in UTF-8 and then send this file to index. Now
> > suppose there can be any number of clients who can upload file. at that
> time
> > it will affect performance. and already our product support localization
> > with local encoding.
> >
> > Thanks,
> > Prasad
> >
> > On Thu, Jan 27, 2011 at 2:04 PM, Paul Libbrecht 
> wrote:
> >
> >> Why is converting documents to utf-8 not feasible?
> >> Nowadays any platform offers such services.
> >>
> >> Can you give a detailed failure description (maybe with the URL to a
> sample
> >> document you post)?
> >>
> >> paul
> >>
> >>
> >> Le 27 janv. 2011 à 07:31, prasad deshpande a écrit :
> >>> I am able to successfully index/search non-Engilsh data(like Hebrew,
> >>> Japnese) which was encoded in UTF-8.
> >>> However, When I tried to index data which was encoded in local encoding
> >> like
> >>> Big5 for Japanese I could not see the desired results.
> >>> The contents after indexing looked garbled for Big5 encoded document
> when
> >> I
> >>> searched for all indexed documents.
> >>>
> >>> Converting a complete document in UTF-8 is not feasible.
> >>> I am not very clear about how Solr support these localizations with
> other
> >>> than UTF-8 encoding.
> >>>
> >>>
> >>> I verified below links
> >>> 1. http://lucene.apache.org/java/3_0_3/api/all/index.html
> >>> 2.  http://wiki.apache.org/solr/LanguageAnalysis
> >>>
> >>> Thanks and Regards,
> >>> Prasad
> >>
> >>
>
>



Re: Solr for noSQL

2011-01-28 Thread Dennis Gearon
Personally, I just create a view that flattens out the database and renames the 
fields as I desire. Then I call the view with the DIH to import it.

Solr doesn't knwo anything about the databsae, except how to get a connection 
and fetch rows. And that's pretty darn useful, just that much less code to 
write.

 Dennis Gearon


Signature Warning

It is always a good idea to learn from your own mistakes. It is usually a 
better 
idea to learn from others’ mistakes, so you do not have to make them yourself. 
from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036'


EARTH has a Right To Life,
otherwise we all die.



- Original Message 
From: Upayavira 
To: solr-user@lucene.apache.org
Sent: Fri, January 28, 2011 1:41:42 AM
Subject: Re: Solr for noSQL



On Thu, 27 Jan 2011 21:38 -0800, "Dennis Gearon" 
wrote:
> Why not make one's own DIH handler, Lance?

Personally, I don't like that approach. Solr is best related to as
something of a black box that you configure, then push content to.
Having Solr know about your data sources, and pull content in seems to
me to be mixing concerns.

I relate to the DIH as a useful tool for smaller sites or for
prototyping, but would expect anything more substantial to require an
indexing application that gives you full control over the indexing
process. It could be a lightweight app that uses a MongoDB java client
and SolrJ, and simply pulls from one and pushes to the other. If you
don't want to run another JVM, it could run as a separate webapp within
your Solr JVM.

From an architectural point of view, do you configure Mysql, or MongoDB
for that matter, to pull content into itself? Likewise, Solr should be a
service that listens, waiting to be given data.

Upayavira
--- 
Enterprise Search Consultant at Sourcesense UK, 
Making Sense of Open Source


Re: Does solr supports indexing of files other than UTF-8

2011-01-28 Thread Yonik Seeley
On Thu, Jan 27, 2011 at 3:51 AM, prasad deshpande
 wrote:
> The size of docs can be huge, like suppose there are 800MB pdf file to index
> it I need to translate it in UTF-8 and then send this file to index.

PDF is binary AFAIK... you shouldn't need to do any charset
translation before sending it to solr, or any other extraction
library.  If you're using solr-cell then it's the Tika component that
is responsible for pulling out the text in the right format.

-Yonik
http://lucidimagination.com


Re: Searching for negative numbers very slow

2011-01-28 Thread Yonik Seeley
On Thu, Jan 27, 2011 at 6:32 PM, Simon Wistow  wrote:
> If I do
>
>        qt=dismax
>    fq=uid:1
>
> (or any other positive number) then queries are as quick as normal - in
> the 20ms range.
>
> However, any of
>
>        fq=uid:\-1
>
> or
>
>    fq=uid:[* TO -1]
>
> or
>
>    fq=uid:[-1 to -1]
>
> or
>
>    fq=-uid:[0 TO *]
>
> then queries are incredibly slow - in the 9 *second* range.

That's odd - there should be nothing special about negative numbers.
Here are a couple of ideas:
  - if you have a really big index and querying by a negative number
is much more rare, it could just be that part of the index wasn't
cached by the OS and so the query needs to hit the disk.  This can
happen with any term and a really big index - nothing special for
negatives here.
 - if -1 is a really common value, it can be slower.  is fq=uid:\-2 or
other negative numbers really slow also?

-Yonik
http://lucidimagination.com


DIH populate multiple fields with one column

2011-01-28 Thread alexei

Hi All,

I attempting to parse xml stored in a database as blobs 
and then store pieces of this xml such as titles and authors
in separate fields.





   

The above configuration works, but the issue is that I have to parse 
the XML again and again for each new field. The indexing time increases
linearly.

What I would like to do is get the Blob once, parse the xml into a DOM
document
and populate all the fields at once, in a single transformer.

Is this possible? Could anyone point me in the right direction?

Regards,
Alexei
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/DIH-populate-multiple-fields-with-one-column-tp2367379p2367379.html
Sent from the Solr - User mailing list archive at Nabble.com.


Http Connection is hanging while deleteByQuery

2011-01-28 Thread shan2812

Hi,

I am trying to delete Solr documents using deleteByQuery(). Unfortunately it
simply hangs, for a long time.

I have tried using both SimpleHttpConnectionManager and
MultiThreaded...(default)
1. 
SolrServer solr;
SimpleHttpConnectionManager cm = new SimpleHttpConnectionManager(true);

the below options were tried as nothing worked
//cm.closeIdleConnections(0L);
//cm.getParams().setConnectionTimeout(1);
//cm.getParams().setMaxTotalConnections(10);
HttpClient httpClient = new HttpClient(cm);

solr = new CommonsHttpSolrServer(solrUrl, httpClient);

2.
SolrServer solr;
solr = new CommonsHttpSolrServer(solrUrl);

Do anyone have any idea?

Thank You.

Regards,
Shan
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Http-Connection-is-hanging-while-deleteByQuery-tp2367405p2367405.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Http Connection is hanging while deleteByQuery

2011-01-28 Thread shan2812

Though it may not be needed, just to ad..

this is how I delete by query

UpdateResponse updateResponse = solr.deleteByQuery(new
SolrQuery("my_query"));
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Http-Connection-is-hanging-while-deleteByQuery-tp2367405p2367410.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: filter update by IP

2011-01-28 Thread Thibaut

On 01/24/2011 02:02 AM, Jonathan Rochkind wrote:

Kind of personally curious_why_  it keeps coming up on the list so much. Is 
everyone trying to go into business vending Solr in the cloud to customers who 
will write their own apps, or are there some other less obvious (to me) use 
cases?


In my case, it's to index stuff (content, csv ...) and to use it in 
javascript apps.


I wrote a thin wrapper around solr to embed it in Lutece portal : 
http://goo.gl/f3pGA


Lutece portal administrator can write and read solr and the rest can 
read it.




Re: configure httpclient to access solr with user credential on third party host

2011-01-28 Thread Thibaut

hi

I just wrote a plugin for Lutece portal to restrict access to read/write 
of solr to Lutece admin and read access for the rest: http://goo.gl/f3pGA


t.



Sending binary data as part of a query

2011-01-28 Thread Jay Luker
Hi all,

Here is what I am interested in doing: I would like to send a
compressed integer bitset as a query to solr. The bitset integers
represent my document ids and the results I want to get back is the
facet data for those documents.

I have successfully created a QueryComponent class that, assuming it
has the integer bitset, can turn that into the necessary DocSetFilter
to pass to the searcher, get back the facets, etc. That part all works
right now because I'm using either canned or randomly generated
bitsets on the server side.

What I'm unsure how to do is actually send this compressed bitset from
a client to solr as part of the query. From what I can tell, the Solr
API classes that are involved in handling binary data as part of a
request assume that the data is a document to be added. For instance,
extending ContentStreamHandlerBase requires implementing some kind of
document loader and an UpdateRequestProcessorChain and a bunch of
other stuff that I don't really think I should need. Is there a
simpler way? Anyone tried or succeeded in doing anything similar to
this?

Thanks,
--jay


Filter Query, Filter Cache and Hit Ratio

2011-01-28 Thread Renaud Delbru

Hi,

I am looking for some more information on how the filter cache is 
working, and how the hit are incremented.


We are using filter queries for certain predefined value, such as the 
timestamp:[2011-01-21T00:00:00Z+TO+NOW] (which is the current day). From 
what I understand from the documentation:
"the filter cache stores the results of any filter queries ("fq" 
parameters) that Solr is explicitly asked to execute. (Each filter is 
executed and cached separately. When it's time to use them to limit the 
number of results returned by a query, this is done using set 
intersections.)"
So, we were imagining that is two consecutive queries (as the one above) 
was using the same timestamp filter query, the second query will take 
advantage of the filter cache, and we would see the number of hits 
increasing (hit on the cached timestamp filter query) . However, this is 
not the case, the number of hits on the filter cache does not increase 
and stays very low. Is it normal ?


INFO: [] webapp=/siren path=/select 
params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=true&fq=timestamp:[2011-01-21T00:00:00Z+TO+NOW]&fq=domain:my.wordpress.com&fsv=true} 
hits=0 status=0 QTime=139
INFO: [] webapp=/siren path=/select 
params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=true&fq=timestamp:[2011-01-21T00:00:00Z+TO+NOW]&fq=domain:syours.wordpress.com&fsv=true} 
hits=0 status=0 QTime=138


--
Renaud Delbru



RE: Filter Query, Filter Cache and Hit Ratio

2011-01-28 Thread cbennett
Hi,

You've used NOW in the range query which will give a date/time accurate to
the millisecond, try using NOW\DAY

Colin.

> -Original Message-
> From: Renaud Delbru [mailto:renaud.del...@deri.org]
> Sent: Friday, January 28, 2011 2:22 PM
> To: solr-user@lucene.apache.org
> Subject: Filter Query, Filter Cache and Hit Ratio
> 
> Hi,
> 
> I am looking for some more information on how the filter cache is
> working, and how the hit are incremented.
> 
> We are using filter queries for certain predefined value, such as the
> timestamp:[2011-01-21T00:00:00Z+TO+NOW] (which is the current day).
> From
> what I understand from the documentation:
> "the filter cache stores the results of any filter queries ("fq"
> parameters) that Solr is explicitly asked to execute. (Each filter is
> executed and cached separately. When it's time to use them to limit the
> number of results returned by a query, this is done using set
> intersections.)"
> So, we were imagining that is two consecutive queries (as the one
> above)
> was using the same timestamp filter query, the second query will take
> advantage of the filter cache, and we would see the number of hits
> increasing (hit on the cached timestamp filter query) . However, this
> is
> not the case, the number of hits on the filter cache does not increase
> and stays very low. Is it normal ?
> 
> INFO: [] webapp=/siren path=/select
> params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=t
> rue&fq=timestamp:[2011-01-
> 21T00:00:00Z+TO+NOW]&fq=domain:my.wordpress.com&fsv=true}
> hits=0 status=0 QTime=139
> INFO: [] webapp=/siren path=/select
> params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=t
> rue&fq=timestamp:[2011-01-
> 21T00:00:00Z+TO+NOW]&fq=domain:syours.wordpress.com&fsv=true}
> hits=0 status=0 QTime=138
> 
> --
> Renaud Delbru






RE: Filter Query, Filter Cache and Hit Ratio

2011-01-28 Thread cbennett
Ooops,

I meant NOW/DAY 

> -Original Message-
> From: cbenn...@job.com [mailto:cbenn...@job.com]
> Sent: Friday, January 28, 2011 3:37 PM
> To: solr-user@lucene.apache.org
> Subject: RE: Filter Query, Filter Cache and Hit Ratio
> 
> Hi,
> 
> You've used NOW in the range query which will give a date/time accurate
> to
> the millisecond, try using NOW\DAY
> 
> Colin.
> 
> > -Original Message-
> > From: Renaud Delbru [mailto:renaud.del...@deri.org]
> > Sent: Friday, January 28, 2011 2:22 PM
> > To: solr-user@lucene.apache.org
> > Subject: Filter Query, Filter Cache and Hit Ratio
> >
> > Hi,
> >
> > I am looking for some more information on how the filter cache is
> > working, and how the hit are incremented.
> >
> > We are using filter queries for certain predefined value, such as the
> > timestamp:[2011-01-21T00:00:00Z+TO+NOW] (which is the current day).
> > From
> > what I understand from the documentation:
> > "the filter cache stores the results of any filter queries ("fq"
> > parameters) that Solr is explicitly asked to execute. (Each filter is
> > executed and cached separately. When it's time to use them to limit
> the
> > number of results returned by a query, this is done using set
> > intersections.)"
> > So, we were imagining that is two consecutive queries (as the one
> > above)
> > was using the same timestamp filter query, the second query will take
> > advantage of the filter cache, and we would see the number of hits
> > increasing (hit on the cached timestamp filter query) . However, this
> > is
> > not the case, the number of hits on the filter cache does not
> increase
> > and stays very low. Is it normal ?
> >
> > INFO: [] webapp=/siren path=/select
> >
> params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=t
> > rue&fq=timestamp:[2011-01-
> > 21T00:00:00Z+TO+NOW]&fq=domain:my.wordpress.com&fsv=true}
> > hits=0 status=0 QTime=139
> > INFO: [] webapp=/siren path=/select
> >
> params={wt=javabin&rows=0&version=2&fl=id,score&start=0&q=*:*&isShard=t
> > rue&fq=timestamp:[2011-01-
> > 21T00:00:00Z+TO+NOW]&fq=domain:syours.wordpress.com&fsv=true}
> > hits=0 status=0 QTime=138
> >
> > --
> > Renaud Delbru
> 
> 
> 






Thoughts on USING dynamic fields for extending objects

2011-01-28 Thread Dennis Gearon
Well, mid, next month we're going to start using dynamic fields as the relate 
to 
our business rules. Basically, in involves have a basic set of objects in 
code/database, and flattened for search in Solr. The MAIN business object is to 
be extendable by the custoemer, will still having to supply the required fields 
in the base object. We will use defined type, dynamic fields

I had a question for those more experienced than I. We are thinking about  two 
possible usage patterns:

A/ User can add any field they want, as long as they use the right suffix for 
the field.
Changing the schema can be done at will, and updating past objects is 
totally on the user.
They get:
  1/ Find within the field.
  2/ Range queries
  3/ other future, single field functionality later
  4
B/ User can NOT add any field they want, but they must submit a schema, 
hopefully automated. The data still
 goes into the Solr index as dynamically accepted fields as long as they 
use 
the right suffix for the field.
 Changing the schema can be done at by submitting the new schema. Updating 
past objects is STILL
 totally on the user.
They get:
  1/ Find within the field.
  2/ Range queries
  3/ Various filter functions like: mandatory fields, acceptable ranges, 
minimum lengths on strings, and other processing.
  4/ Other future, single field functionality later
  5/ The ability to make their own copyfields for 'grouping' of their own 
fields.


'A' I see as most simple to administer, but possible has security holes? THAT's 
my main question, all thoughts welcom.
'B' is better as a value added service, but has a LOT more work on our site's 
end, I believe. We could also possibly do non acceptance of sensitive field 
names for security?

Any thoughts much appreciateed.
 



 Dennis Gearon


Signature Warning

It is always a good idea to learn from your own mistakes. It is usually a 
better 
idea to learn from others’ mistakes, so you do not have to make them yourself. 
from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036'


EARTH has a Right To Life,
otherwise we all die.



Re: DIH populate multiple fields with one column

2011-01-28 Thread alexei

Sorry about this post. I'll RTFM more carefully next time. 
Resolved:
 
 

Regards,
Alexei


-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/DIH-populate-multiple-fields-with-one-column-tp2367379p2368071.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: edismax vs dismax

2011-01-28 Thread Yonik Seeley
On Fri, Jan 28, 2011 at 3:00 PM, Thumuluri, Sai
 wrote:
> I recently upgraded to Solr 1.4.1 from Solr 1.3 and with the upgrade
> used edismax query parser. Here is my solrconfig.xml . When I search for
> "mw verification and payment information" - I get no results with
> defType set to edismax,

It's probably a bit of natural language query parsing in edismax...
- and is treated as AND (the lucene operator) in the appropriate
context (i.e. we won't if it's at the start or end of the query, etc)
- or is treated as OR in the appropriate context

The lowercaseOperators parameter can control this, so try setting
lowercaseOperators=false

-Yonik
http://lucidimagination.com



> if I switch the deftype to dismax - I get the results I am looking for
>
> Can anyone explain, why this would be the case? I thought edismax is
> dismax and more.
>
> Thank you,
>
> For 1.4.1
>  default="true">
>    
>     dismax
>     explicit
>     0.01
>     
>        body^1.0 title^10.0 name^3.0 taxonomy_names^2.0 tags_h1^5.0
> tags_h2_h3^3.0 tags_h4_h5_h6^2.0 tags_inline^1.0
>     
>     
>        body^10.0
>     
>     4
>     
>        2<-25%
>     
>     *:*
>
>     true
>     body
>     3
>     true
>   
>     body
>     256
>      
> Sai Thumuluri


deleteById throwing SocketTimeoutException

2011-01-28 Thread Ravi Kiran
Hello,
We have a core with about 900K docs. Recently I have noticed that
the deleteById query seems to always give me a SocketTimeoutException(stack
trace is shown below). I cannot figure out why only deletion fails but not
add/update. The SOLR client instance is created via spring wiring
(configuration given below). Did anybody face the same issue ? How can I
solve this issue ? Increasing the timeout did not help.

Configuration
--


http://:8080/solr-admin/core











Code
-
String filename = delfile.getName();
String id = filename.replace("_search.xml", "");

log.debug("Deleting id " + id);
UpdateResponse response = solrServer.deleteById(id);
log.info("Deleting response for " + id + " is " +
response);

boolean success = Util.moveFile(delfile,
delprocessedpath);

/**
 * Now delete old successfully processed files so that
full reindex
 * from processed and transformed folders will not
process unwanted/deleted documents
 */
File transformedFile = new File(transformedpath +
(filename.replace("_search.xml", "_fast.xml")));
if(transformedFile.exists()) {
log.info("Deleting archived Transformed file: " +
transformedFile.getAbsolutePath());
transformedFile.delete();
}

File processedFile = new File(processedpath+filename);
if(processedFile.exists()) {
log.info("Deleting archived Processed file: " +
processedFile.getAbsolutePath());
processedFile.delete();
}

Stack Trace
--
2011-01-28 15:51:18,842-0500 ERROR
[com.search.service.topics.feedprocessor.DeleteFeedProcessor] - Error
deleting from Solr server for -AR2011011403385_search.xml
org.apache.solr.client.solrj.SolrServerException:
java.net.SocketTimeoutException: Read timed out
at
org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:472)
at
org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:243)
at
org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105)
at
org.apache.solr.client.solrj.SolrServer.deleteById(SolrServer.java:102)
at
com.search.service.topics.feedprocessor.DeleteFeedProcessor.processDelete(DeleteFeedProcessor.java:76)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at
org.springframework.util.MethodInvoker.invoke(MethodInvoker.java:276)
at
org.springframework.scheduling.quartz.MethodInvokingJobDetailFactoryBean$MethodInvokingJob.executeInternal(MethodInvokingJobDetailFactoryBean.java:260)
at
org.springframework.scheduling.quartz.QuartzJobBean.execute(QuartzJobBean.java:86)
at org.quartz.core.JobRunShell.run(JobRunShell.java:202)
at
org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:525)
Caused by: java.net.SocketTimeoutException: Read timed out
at java.net.SocketInputStream.socketRead0(Native Method)
at java.net.SocketInputStream.read(SocketInputStream.java:129)
at java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
at java.io.BufferedInputStream.read(BufferedInputStream.java:237)
at
org.apache.commons.httpclient.HttpParser.readRawLine(HttpParser.java:78)
at
org.apache.commons.httpclient.HttpParser.readLine(HttpParser.java:106)
at
org.apache.commons.httpclient.HttpConnection.readLine(HttpConnection.java:1116)
at
org.apache.commons.httpclient.MultiThreadedHttpConnectionManager$HttpConnectionAdapter.readLine(MultiThreadedHttpConnectionManager.java:1413)
at
org.apache.commons.httpclient.HttpMethodBase.readStatusLine(HttpMethodBase.java:1973)
at
org.apache.commons.httpclient.HttpMethodBase.readResponse(HttpMethodBase.java:1735)
at
org.apache.commons.httpclient.HttpMethodBase.execute(HttpMethodBase.java:1098)
at
org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:398)
at
org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:171)
at
org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397)
at
org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:323)
at
org.apache.solr.cli

[Failure] to start Solr 4.0

2011-01-28 Thread Adam Estrada
All,

I've checked out the latest code and built the root directory with ant compile 
and then I build the solr directory again using the ant dist command which 
gives me the lucene-libs directory and a couple others. Now Solr won't start.  
What am I missing???  This is as far as it gets.

mini:example Adam$ java -jar start.jar 
2011-01-28 17:14:23.402:INFO::Logging to STDERR via org.mortbay.log.StdErrLog
2011-01-28 17:14:23.605:INFO::jetty-6.1.26
2011-01-28 17:14:23.638:INFO::Started SocketConnector@0.0.0.0:8983

What couple possibly be the problem?

Adam

Re: [Failure] to start Solr 4.0

2011-01-28 Thread Adam Estrada
I found the problem...You HAVE to build the Solr directory using "ant example" 
in order for the web application to start properly. Sorry to post so many times.

Adam

On Jan 28, 2011, at 5:20 PM, Adam Estrada wrote:

> All,
> 
> I've checked out the latest code and built the root directory with ant 
> compile and then I build the solr directory again using the ant dist command 
> which gives me the lucene-libs directory and a couple others. Now Solr won't 
> start.  What am I missing???  This is as far as it gets.
> 
> mini:example Adam$ java -jar start.jar 
> 2011-01-28 17:14:23.402:INFO::Logging to STDERR via org.mortbay.log.StdErrLog
> 2011-01-28 17:14:23.605:INFO::jetty-6.1.26
> 2011-01-28 17:14:23.638:INFO::Started SocketConnector@0.0.0.0:8983
> 
> What couple possibly be the problem?
> 
> Adam



Controlling Tika's metadata

2011-01-28 Thread Andreas Kemkes
Just getting my feet wet with the text extraction using both schema and 
solrconfig settings from the example directory in the 1.4 distribution, so I 
might miss something obvious.

Trying to provide my own title (and discarding the one received through Tika's 
metadata) wasn't straightforward. I had to use the following:

fmap.title=tika_title (to discard the Tika title)
literal.attr_title=New Title (to provide the correct one)
fmap.attr_title=title (to map it back to the field as I would like to use title 
in searches)

Is there anything easier than the above?

How can this best be generalized to other metadata provided by Tika (which in 
our use case will be mostly ignored, as it is provided separately)?

Thanks in advance for your responses.


  

DeletepkQuery is not actually deleting the records, please help

2011-01-28 Thread makeyourrules

Hello, 
I am trying to delete some records from my index with delta-import using
deletePkQuery with the below config, the log prints the deleted documents
and says delta import successfully, but when I search, my search results
still have those deleted documents, i have already spent so much time
researching it but couldn't find any solution. All my database updates are
getting updated with deltaQuery and deltaImportQuery but not the deletes.
Could any one suggest me any solution? 

URL: 
http://localhost:8983/dataimport?command=delta-import

dataConfig.xml:










Log file:



[2011/01/28 16:56:26.498] Completed ModifiedRowKey for Entity: item rows
obtained : 0
[2011/01/28 16:56:26.499] Completed DeletedRowKey for Entity: item rows
obtained : 6
[2011/01/28 16:56:26.499] Completed parentDeltaQuery for Entity: item
[2011/01/28 16:56:32.563] Deleting stale documents 
.
.
[2011/01/28 16:58:00.319] Deleting document: BAAH
[2011/01/28 17:06:50.537] Deleting document: BAAI
[2011/01/28 17:07:28.470] Deleting document: BAAL
[2011/01/28 17:08:13.187] Deleting document: BAAM
[2011/01/28 17:08:27.011] Deleting document: BAAJ
[2011/01/28 17:08:44.218] Deleting document: BAAK
[2011/01/28 17:09:13.487] Delta Import completed successfully
[2011/01/28 17:09:32.174] Import completed successfully
[2011/01/28 17:09:32.175] start
commit(optimize=true,waitFlush=false,waitSearcher=true,expungeDeletes=false)
..
[2011/01/28 17:09:32.212] autowarming Searcher@a44b35 main from
Searcher@f41f34 main
[2011/01/28 17:09:32.215] Read dataimport.properties
[2011/01/28 17:09:32.217] Wrote last indexed time to dataimport.properties
[2011/01/28 17:09:33.791] Time taken = 0:13:45.366


Any suggestions would be highly appreciated.

Thanks,
Prad.

-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/DeletepkQuery-is-not-actually-deleting-the-records-please-help-tp2368463p2368463.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: How to group result when search on multiple fields

2011-01-28 Thread cyang2010

I don't think this field collapsing can simply solve my problem after second
thought.

As I mentioned, user only type in a search phrase, and click on search. 
Underlying that, the application logic is going to compose search query
against multiple fields (title_name, actors, directors, ...) by the search
phrase/term.

Therefore, search result can match search term with any of the fields above. 
However, for all results due to actor name match, i want to make a group
which only list the first two actors.  for that group, i want to put it to
the right spot of the result based on the relevancy score of the best actor
match.  

For example, if i search  a keyword "Tom" (term match as well fuzzy match). 
there are matching result  based on video title name as well as actor name 

result 1:  <--  title match
score: 1.0
title_name: tom's story
actor: jamie lee

result 2:  <--  title match
score: 0.95
title_name: tom green's store
actor: joanne anderson

result 3:  <-- actor match
actor 2: tommy jackson  -- score: 0.5
actor 1:  tim robin   -- score: 0.4
See all actors match "Tom"

result 4:   <-- title match
score: 0.333
title_name: atom theory
actor: kevin sheen
  


in this case, field collapsing can only achieve this:  out of search result,
It will list out all actors as long as there is title or actor match.  For 
example:

Assume only show top 2 result in each group
group value: Anglelina Jolie  
  numFound:13
  
id:1,
title_name:tom's story
  
id:2,
title_name:tom green's store

group value: tommy jackson
  numFound:1  
id: 201,
title_name: ...  



group value: kevin sheen
  numFound:1  
id: 30,
title_name:  atom theory



<-- even if Angelina Jolie not in the result because of actor name match. 
but because her movie title  matching "tom" with highest relevance.  she
will still be number 1 group.   This is different than what i expected.  
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/How-to-group-result-when-search-on-multiple-fields-tp2358441p2368496.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: How to group result when search on multiple fields

2011-01-28 Thread cyang2010

Some typo out there in my example:  I mean the first 2 movies by angelina
jolie.  This is the correct example:

result 1:  <--  title match
score: 1.0
title_name: tom's story
actor: Angelina Jolie

result 2:  <--  title match
score: 0.95
title_name: tom green's store
actor: Angelina Jolie

result 3:  <-- actor match
actor 2: tommy jackson  -- score: 0.5
actor 1:  tim robin   -- score: 0.4
See all actors match "Tom"

result 4:   <-- title match
score: 0.333
title_name: atom theory
actor: kevin sheen



Here is the corresponding result if field collasping (result grouping) is
used:

group value: Anglelina Jolie  
  numFound:13
  
id:1,
title_name:tom's story
  
id:2,
title_name:tom green's store

group value: tommy jackson
  numFound:1  
id: 201,
title_name: ...  

group value: tim robin
  numFound:1  
id: 202,
title_name: ...

group value: kevin sheen
  numFound:1  
id: 30,
title_name:  atom theory

  
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/How-to-group-result-when-search-on-multiple-fields-tp2358441p2368512.html
Sent from the Solr - User mailing list archive at Nabble.com.


velocity tools in solr-contrib-velocity?

2011-01-28 Thread Paul Libbrecht

Hello list,

can anyone tell me how I can plug the velocity tools into my solr?
Do I understand correctly the following comment in the source:

// Velocity context tools - TODO: make these pluggable

that it's only hard-coded thus far?

thanks in advance

paul


solr - uima error

2011-01-28 Thread Darx Oman
I tried to do the uima integration with solr
I followed the steps in the readme file
using the solr example folder as my solr instance but when I started solr, I
got this error message

HTTP ERROR 500

Problem accessing /solr/admin/. Reason:

Severe errors in solr configuration.

Check your log files for more detailed information on what may be wrong.

-
org.apache.solr.common.SolrException: Error loading class
'org.apache.solr.uima.processor.UIMAProcessorFactory'
at 
org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:408)
at org.apache.solr.core.SolrCore.createInstance(SolrCore.java:399)
at org.apache.solr.core.SolrCore.createInitInstance(SolrCore.java:422)
at org.apache.solr.core.SolrCore.initPlugins(SolrCore.java:1486)
at 
org.apache.solr.update.processor.UpdateRequestProcessorChain.init(UpdateRequestProcessorChain.java:57)
at org.apache.solr.core.SolrCore.createInitInstance(SolrCore.java:424)
at org.apache.solr.core.SolrCore.initPlugins(SolrCore.java:1470)
at org.apache.solr.core.SolrCore.initPlugins(SolrCore.java:1464)
at 
org.apache.solr.core.SolrCore.loadUpdateProcessorChains(SolrCore.java:572)
at org.apache.solr.core.SolrCore.(SolrCore.java:514)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:647)
at org.apache.solr.core.CoreContainer.load(CoreContainer.java:399)
at org.apache.solr.core.CoreContainer.load(CoreContainer.java:283)
at 
org.apache.solr.core.CoreContainer$Initializer.initialize(CoreContainer.java:232)
at 
org.apache.solr.servlet.SolrDispatchFilter.init(SolrDispatchFilter.java:85)
at org.mortbay.jetty.servlet.FilterHolder.doStart(FilterHolder.java:97)
at 
org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at 
org.mortbay.jetty.servlet.ServletHandler.initialize(ServletHandler.java:713)
at org.mortbay.jetty.servlet.Context.startContext(Context.java:140)
at 
org.mortbay.jetty.webapp.WebAppContext.startContext(WebAppContext.java:1282)
at 
org.mortbay.jetty.handler.ContextHandler.doStart(ContextHandler.java:518)
at 
org.mortbay.jetty.webapp.WebAppContext.doStart(WebAppContext.java:499)
at 
org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at 
org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerCollection.java:152)
at 
org.mortbay.jetty.handler.ContextHandlerCollection.doStart(ContextHandlerCollection.java:156)
at 
org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at 
org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerCollection.java:152)
at 
org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at 
org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrapper.java:130)
at org.mortbay.jetty.Server.doStart(Server.java:224)
at 
org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at org.mortbay.xml.XmlConfiguration.main(XmlConfiguration.java:985)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.mortbay.start.Main.invokeMain(Main.java:194)
at org.mortbay.start.Main.start(Main.java:534)
at org.mortbay.start.Main.start(Main.java:441)
at org.mortbay.start.Main.main(Main.java:119)
Caused by: java.lang.ClassNotFoundException:
org.apache.solr.uima.processor.UIMAProcessorFactory
at java.net.URLClassLoader$1.run(Unknown Source)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(Unknown Source)
at java.lang.ClassLoader.loadClass(Unknown Source)
at java.net.FactoryURLClassLoader.loadClass(Unknown Source)
at java.lang.ClassLoader.loadClass(Unknown Source)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Unknown Source)
at 
org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:392)
... 39 more

 --
 *Powered by Jetty://*

I Searched for the class  "IMAProcessorFactory" but I didn't find it  .
any body can help ?
By the way I'm using solr trunk checked out today  Jan 29th