Re: Is it possible to writing solr result on disk from the server side?

2009-05-08 Thread arno13

Tanks Paul!
Yes, I saw it. In fact, if I well understand, it could be a solution, but a
little beat too complicate for what I want to do. Currently my client part
is not in Java and I still need a kind of client and server model because
it's a Web application and solr have to be running and waiting query
continuously.
So even if it seems possible with EmbeddedSolrServer, sharing the xml
results on a file will be a faster solution for me regarding also the
development time.



Noble Paul നോബിള്‍  नोब्ळ्-2 wrote:
> 
> did you consider using an EmbeddedSolrServer?
> 
> On Thu, May 7, 2009 at 8:25 PM, arno13 
> wrote:
>>
>> Do you know if it's possible to writing solr results directly on a hard
>> disk
>> from server side and not to use an HTTP connection to transfer the
>> results?
>>
>> While the query time is very fast for solr, I want to do that, cause of
>> the
>> time taken during the transfer of the results between the client and the
>> solr server when you have lot of 'rows'.
>> For instance for 10'000 rows, the query time could be 50 ms and 19s to
>> get
>> the results from the server. As my client and server are on the same
>> system,
>> I could get the results faster directly on the hard disk (or better in a
>> ram
>> disk), is it possible configuring solr for that?
>>
>> Regards,
>>
>>
>>
>>
>> --
>> View this message in context:
>> http://www.nabble.com/Is-it-possible-to-writing-solr-result-on-disk-from-the-server-side--tp23428509p23428509.html
>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
>>
> 
> 
> 
> -- 
> -
> Noble Paul | Principal Engineer| AOL | http://aol.com
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Is-it-possible-to-writing-solr-result-on-disk-from-the-server-side--tp23428509p23441651.html
Sent from the Solr - User mailing list archive at Nabble.com.



Organizing multiple searchers around overlapping subsets of data

2009-05-08 Thread Michael Ludwig

I have one type of document, but different searchers, each of
which is interested in a different subset of the documents,
which are different configurations of TV channels {A,B,C,D}.

* Application S1 is interested in all channels, i.e. {A,B,C,D}.
* Application S2 is interested in {A,B,C}.
* Application S3 is interested in {A,C,D}.
* Application S4 is interested in {B,D}.

As can be seen from this simplified example, the subsets are
not disjoint, but do have considerable overlaps.

The total data volume is only about 200 MB. There are four
searchers, and they may become ten or a dozen.

The set elements an application may or may not be interested
in, however, i.e. the channels, which are {A,B,C,D} in this
example, are not just four, but about 150, each of which has
about 1000 documents.

What is the best way to organize this?

(a) Set up different cores for each application, i.e. going
multi-core, thereby incurring a good deal of redundancy, but
simplifying searches?

(b) Apply filter queries to select documents from only, say
60, 80 or 110 out of 150 channels.

(c) Something else I'm not aware of.

Am I right in suspecting that multi-core makes less sense with
increasing overlaps and hence redundancy?

Michael Ludwig


Re: StatsComponent and 1.3

2009-05-08 Thread Eric Pugh
I'm guessing that manipulating the client end, acts_as_solr, is an  
easier approach then backporting server side functionality.   
Especially as you will have to forward migrate at some point.


Out of curiosity, which version of acts_as_solr are you using?  The  
plugin has moved homes a couple of times, and I have heard and found  
that the version by Mathias Meyer at http://github.com/mattmatt/acts_as_solr/tree/master 
 is the best.  I've used it with 1.4 trunk with no issues, and  
Mathias has been very responsive.


Eric


On May 7, 2009, at 10:25 PM, David Shettler wrote:


Foreword:  I'm not a java developer :)

OSVDB.org and datalossdb.org make use of solr pretty extensively via
acts_as_solr.

I found myself with a real need for some of the StatsComponent stuff
(mainly the sum feature), so I pulled down a nightly build and played
with it.  StatsComponent proved perfect, but... the nightly build
output seems to be different, and thus incompatible with acts_as_solr.

Now, I realize this is more or less an acts_as_solr issue, but...

Is it possible, with some degree of effort (obviously) for me to
essentially port some of the functionality of StatsComponent to 1.3
myself?  It's that, or waiting for 1.4 to come out and someone
developing support for it into acts_as_solr, or myself fixing what I
have for acts_as_solr to work with the output.  I'm just trying to
gauge the easiest solution :)

Any feedback or suggestions would be grand.

Thanks,

Dave
Open Security Foundation


-
Eric Pugh | Principal | OpenSource Connections, LLC | 434.466.1467 | 
http://www.opensourceconnections.com
Free/Busy: http://tinyurl.com/eric-cal






Re: Core Reload issue

2009-05-08 Thread Noble Paul നോബിള്‍ नोब्ळ्
As long as your indexed documents have the stop words you will
continue to see the stop words in the results.

On Fri, May 8, 2009 at 11:24 AM, Sagar Khetkade
 wrote:
>
>
>
> From my understanding re-indexing the documents is a different thing. If you 
> have the stop word filter for field type say "text" then after reloading the 
> core if i type in a query which is stop word only it would get parsed from 
> stop word filter which would eventually will not serach against the index.
>
> But in my case i am getting the results having the stop word; so the issue.
>
>
>
> ~Sagar
>
>
>
>
>> From: noble.p...@gmail.com
>> Date: Tue, 5 May 2009 10:09:29 +0530
>> Subject: Re: Core Reload issue
>> To: solr-user@lucene.apache.org
>>
>> If you change the the conf files and if you reindex the documents it
>> must be reflected are you sure you re-indexed?
>>
>> On Tue, May 5, 2009 at 10:00 AM, Sagar Khetkade
>>  wrote:
>> >
>> > Hi,
>> >
>> > I came across a strange problem while reloading the core in multicore 
>> > scenario. In the config of one of the core I am making changes in the 
>> > synonym and stopword files and then reloading the core. The core gets 
>> > reloaded but the changes in stopword and synonym fiels does not get 
>> > reflected when I query in. The filters for index and query are the same. I 
>> > face this problem even if I reindex the documents. But when I restart the 
>> > servlet container in which the solr is embedded I problem does not 
>> > resurfaces.
>> > My ultimate goal is/was to reflect the changes made in the text files 
>> > inside the config folder.
>> > Is this the expected behaviour or some problem at my side. Could anyone 
>> > suggest me the possible work around?
>> >
>> > Thanks in advance!
>> >
>> > Regards,
>> > Sagar Khetkade
>> > _
>> > More than messages–check out the rest of the Windows Live™.
>> > http://www.microsoft.com/india/windows/windowslive/
>>
>>
>>
>> --
>> -
>> Noble Paul | Principal Engineer| AOL | http://aol.com
>
> _
> Planning the weekend ? Here’s what is happening in your town.
> http://msn.asklaila.com/events/



-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com


Re: Is it possible to writing solr result on disk from the server side?

2009-05-08 Thread Noble Paul നോബിള്‍ नोब्ळ्
Solr does not need any special configuration to do this. Just fire
your query once and save the results xml/json into a file or in
memory. When you need them again , just read it from disk.memory

On Fri, May 8, 2009 at 1:21 PM, arno13  wrote:
>
> Tanks Paul!
> Yes, I saw it. In fact, if I well understand, it could be a solution, but a
> little beat too complicate for what I want to do. Currently my client part
> is not in Java and I still need a kind of client and server model because
> it's a Web application and solr have to be running and waiting query
> continuously.
> So even if it seems possible with EmbeddedSolrServer, sharing the xml
> results on a file will be a faster solution for me regarding also the
> development time.
>
>
>
> Noble Paul നോബിള്‍  नोब्ळ्-2 wrote:
>>
>> did you consider using an EmbeddedSolrServer?
>>
>> On Thu, May 7, 2009 at 8:25 PM, arno13 
>> wrote:
>>>
>>> Do you know if it's possible to writing solr results directly on a hard
>>> disk
>>> from server side and not to use an HTTP connection to transfer the
>>> results?
>>>
>>> While the query time is very fast for solr, I want to do that, cause of
>>> the
>>> time taken during the transfer of the results between the client and the
>>> solr server when you have lot of 'rows'.
>>> For instance for 10'000 rows, the query time could be 50 ms and 19s to
>>> get
>>> the results from the server. As my client and server are on the same
>>> system,
>>> I could get the results faster directly on the hard disk (or better in a
>>> ram
>>> disk), is it possible configuring solr for that?
>>>
>>> Regards,
>>>
>>>
>>>
>>>
>>> --
>>> View this message in context:
>>> http://www.nabble.com/Is-it-possible-to-writing-solr-result-on-disk-from-the-server-side--tp23428509p23428509.html
>>> Sent from the Solr - User mailing list archive at Nabble.com.
>>>
>>>
>>
>>
>>
>> --
>> -
>> Noble Paul | Principal Engineer| AOL | http://aol.com
>>
>>
>
> --
> View this message in context: 
> http://www.nabble.com/Is-it-possible-to-writing-solr-result-on-disk-from-the-server-side--tp23428509p23441651.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>



-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com


Re: What are the Unicode encodings supported by Solr?

2009-05-08 Thread Michael Ludwig

KK schrieb:


I'd like to know about the different Unicode[/any other?] encodings
supported by Solr for posting docs [thru Solrj in my case]. Is it that
just UTF-8, UCN  supported or other character encodings like
NCR(decimal), NCR(hex) etc are supported as well?


Any numerical character reference (NCR), decimal or hexadecimal, is
valid UTF-8 as long as it maps to a valid Unicode character.


I found that for most of the pages the encoding is UTF-8[in this case
searching works fine] but for others the encoding is some other
character encoding[like NCR(dec), NCR(hex) or might be something else,
don't have much idea on this].


Whatever the encoding is, your application needs to know what it is when
dealing with bytes read from the network.


So when I fetch the page content thru java methods using
InputSteamReaders and after stripping various tags what I obtained
is raw text with some encoding not getting supported by Solr.


Did you make sure to not rely on your platform default encoding
(Charset) when constructing the InputStreamReader? If in doubt, take
a look at the InputStreamReader constructors.

Michael Ludwig


Re: Solr spring application context error

2009-05-08 Thread Raju444us

Paul/Erik,
  
   Thanks for your reply.I have the jar file containing the plugin code and
the applicationContext.xml in solr/home/lib directory.It is instantiating
the plugin code.But it is not loading the application context.xml file from
solr/home/lib dir.But when i copied the jar file containing the
applicationContext.xml file into the solr.war file's WEB-Inf/lib dir and
placed the solr.war file in tomcat's web-apps dir ,it worked.

As Erik said solr may only load the xml from solr.war file?

Please let me know if there is any way to do this by placing the
applicationContext.xml file in solr/home/lib.

Thanks,
Raju

   

Noble Paul നോബിള്‍  नोब्ळ्-2 wrote:
> 
> a point to keep in mind is that all the plugin code and everything
> else must be put into the solrhome/lib directory.
> 
> where have you placed the file com/mypackage/applicationContext.xml ?
> 
> On Fri, May 8, 2009 at 12:19 AM, Raju444us  wrote:
>>
>> I have configured solr using tomcat.Everything works fine.I overrode
>> QParserPlugin and configured it.The overriden QParserPlugin has a
>> dependency
>> on another project say project1.So I made a jar of the project and copied
>> the jar to the solr/home lib dir.
>>
>> the project1 project is using spring.It has a factory class which loads
>> the
>> beans.Iam using this factory calss in QParserPlugin to get a bean.When I
>> start my tomcat the factory class is loading fine.But the problem is its
>> not
>> loading the beans.And Iam getting exception
>>
>> org.springframework.beans.factory.BeanDefinitionStoreException:
>> IOException
>> parsing XML document from class path resource
>> [com/mypackage/applicationContext.xml]; nested exception is
>> java.io.FileNotFoundException: class path resource
>> [com/mypackage/applicationContext.xml] cannot be opened because it does
>> not
>> exist
>>
>> Do I need to do something else?. Can anybody please help me.
>>
>> Thanks,
>> Raju
>>
>>
>> --
>> View this message in context:
>> http://www.nabble.com/Solr-spring-application-context-error-tp23432901p23432901.html
>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
>>
> 
> 
> 
> -- 
> -
> Noble Paul | Principal Engineer| AOL | http://aol.com
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Solr-spring-application-context-error-tp23432901p23444847.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Solr spring application context error

2009-05-08 Thread Mark Miller
I've run into this in the past as well. Its fairly annoying. Anyone know 
why the limitation? Why aren't we passing the ClassLoader thats loading 
Solr classes as the parent to the lib dir plugin classloader?


- Mark

Erik Hatcher wrote:
This is probably because Solr loads its extensions from a custom class 
loader, but if that class then needs to access things from the 
classpath, it is only going to see the built-in WEB-INF/lib classes, 
not solr/home lib JAR files.  Maybe there is a Spring way to point it 
at that lib directory also?  This is the kinda pain we get, it seems, 
when reinventing a container, unfortunately.


Erik

On May 7, 2009, at 2:49 PM, Raju444us wrote:



I have configured solr using tomcat.Everything works fine.I overrode
QParserPlugin and configured it.The overriden QParserPlugin has a 
dependency
on another project say project1.So I made a jar of the project and 
copied

the jar to the solr/home lib dir.

the project1 project is using spring.It has a factory class which 
loads the

beans.Iam using this factory calss in QParserPlugin to get a bean.When I
start my tomcat the factory class is loading fine.But the problem is 
its not

loading the beans.And Iam getting exception

org.springframework.beans.factory.BeanDefinitionStoreException: 
IOException

parsing XML document from class path resource
[com/mypackage/applicationContext.xml]; nested exception is
java.io.FileNotFoundException: class path resource
[com/mypackage/applicationContext.xml] cannot be opened because it 
does not

exist

Do I need to do something else?. Can anybody please help me.

Thanks,
Raju


--
View this message in context: 
http://www.nabble.com/Solr-spring-application-context-error-tp23432901p23432901.html 


Sent from the Solr - User mailing list archive at Nabble.com.





--
- Mark

http://www.lucidimagination.com





Re: Is it possible to writing solr result on disk from the server side?

2009-05-08 Thread arno13

It's what I do from the client side however I don't know how to do this from
the server side (solr).
Sorry if I wasn't clear enough.



Noble Paul നോബിള്‍  नोब्ळ्-2 wrote:
> 
> Solr does not need any special configuration to do this. Just fire
> your query once and save the results xml/json into a file or in
> memory. When you need them again , just read it from disk.memory
> 
> On Fri, May 8, 2009 at 1:21 PM, arno13 
> wrote:
>>
>> Thanks Paul!
>> Yes, I saw it. In fact, if I well understand, it could be a solution, but
>> a
>> little beat too complicate for what I want to do. Currently my client
>> part
>> is not in Java and I still need a kind of client and server model because
>> it's a Web application and solr have to be running and waiting query
>> continuously.
>> So even if it seems possible with EmbeddedSolrServer, sharing the xml
>> results on a file will be a faster solution for me regarding also the
>> development time.
>>
>>
>>
>> Noble Paul നോബിള്‍  नोब्ळ्-2 wrote:
>>>
>>> did you consider using an EmbeddedSolrServer?
>>>
>>> On Thu, May 7, 2009 at 8:25 PM, arno13 
>>> wrote:

 Do you know if it's possible to writing solr results directly on a hard
 disk
 from server side and not to use an HTTP connection to transfer the
 results?

 While the query time is very fast for solr, I want to do that, cause of
 the
 time taken during the transfer of the results between the client and
 the
 solr server when you have lot of 'rows'.
 For instance for 10'000 rows, the query time could be 50 ms and 19s to
 get
 the results from the server. As my client and server are on the same
 system,
 I could get the results faster directly on the hard disk (or better in
 a
 ram
 disk), is it possible configuring solr for that?

 Regards,




 --
 View this message in context:
 http://www.nabble.com/Is-it-possible-to-writing-solr-result-on-disk-from-the-server-side--tp23428509p23428509.html
 Sent from the Solr - User mailing list archive at Nabble.com.


>>>
>>>
>>>
>>> --
>>> -
>>> Noble Paul | Principal Engineer| AOL | http://aol.com
>>>
>>>
>>
>> --
>> View this message in context:
>> http://www.nabble.com/Is-it-possible-to-writing-solr-result-on-disk-from-the-server-side--tp23428509p23441651.html
>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
>>
> 
> 
> 
> -- 
> -
> Noble Paul | Principal Engineer| AOL | http://aol.com
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Is-it-possible-to-writing-solr-result-on-disk-from-the-server-side--tp23428509p23445157.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Wildcard Search

2009-05-08 Thread Erick Erickson
Are you by any chance stemming the field when you index?

Erick

On Fri, May 8, 2009 at 2:29 AM, dabboo  wrote:

>
> Hi,
>
> I am facing a n wierd issue while searching.
>
> I am searching for word *sytem*, it displays all the records which contains
> system, systems etc. But when I tried to search *systems*, it only returns
> me those records, which have systems-, systems/ etc etc. It is considering
> wildcard as 1 or more character and not zero character.
>
> So, it is not returning records which has systems has one word. Is there
> any
> way to resolve this.
>
> Please suggest.
>
> Thanks,
> Amit Garg
> --
> View this message in context:
> http://www.nabble.com/Wildcard-Search-tp23440795p23440795.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>


Re: Wildcard Search

2009-05-08 Thread dabboo

Yes, thats correct. I have applied EnglishPorterFilterFactory at index time
as well. Do you think, I should remove it and do the indexing again.


Erick Erickson wrote:
> 
> Are you by any chance stemming the field when you index?
> 
> Erick
> 
> On Fri, May 8, 2009 at 2:29 AM, dabboo  wrote:
> 
>>
>> Hi,
>>
>> I am facing a n wierd issue while searching.
>>
>> I am searching for word *sytem*, it displays all the records which
>> contains
>> system, systems etc. But when I tried to search *systems*, it only
>> returns
>> me those records, which have systems-, systems/ etc etc. It is
>> considering
>> wildcard as 1 or more character and not zero character.
>>
>> So, it is not returning records which has systems has one word. Is there
>> any
>> way to resolve this.
>>
>> Please suggest.
>>
>> Thanks,
>> Amit Garg
>> --
>> View this message in context:
>> http://www.nabble.com/Wildcard-Search-tp23440795p23440795.html
>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
>>
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Wildcard-Search-tp23440795p23445966.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Wildcard Search

2009-05-08 Thread Erick Erickson
My *guess* is that what you're seeing is that wildcard searches are
not analyzed, in this case not run through the stemmer. So your
index only contains "system" and the funky variants (e.g. "systems/").
I don't really understand why you'd get "systems/" in your index, but
I'm assuming that your filter chain doesn't remove things like slashes.

So, you have "system" and "systems/" in your index, but not
"systems" due to stemming, so searching for "systems*"
translates into "systems" OR "systems/" OR and since
no documents have "systems", you don't get them as hits.

All that said, you need to revisit your indexing parameters to make
what happens fit your expectations. I'd advise getting a copy of Luke
and pointing it at your index in order to see what *really* gets put in
it.

Best
Erick



You need to either introduce filters that remove odd stuff like slashes

On Fri, May 8, 2009 at 9:25 AM, dabboo  wrote:

>
> Yes, thats correct. I have applied EnglishPorterFilterFactory at index time
> as well. Do you think, I should remove it and do the indexing again.
>
>
> Erick Erickson wrote:
> >
> > Are you by any chance stemming the field when you index?
> >
> > Erick
> >
> > On Fri, May 8, 2009 at 2:29 AM, dabboo  wrote:
> >
> >>
> >> Hi,
> >>
> >> I am facing a n wierd issue while searching.
> >>
> >> I am searching for word *sytem*, it displays all the records which
> >> contains
> >> system, systems etc. But when I tried to search *systems*, it only
> >> returns
> >> me those records, which have systems-, systems/ etc etc. It is
> >> considering
> >> wildcard as 1 or more character and not zero character.
> >>
> >> So, it is not returning records which has systems has one word. Is there
> >> any
> >> way to resolve this.
> >>
> >> Please suggest.
> >>
> >> Thanks,
> >> Amit Garg
> >> --
> >> View this message in context:
> >> http://www.nabble.com/Wildcard-Search-tp23440795p23440795.html
> >> Sent from the Solr - User mailing list archive at Nabble.com.
> >>
> >>
> >
> >
>
> --
> View this message in context:
> http://www.nabble.com/Wildcard-Search-tp23440795p23445966.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>


Re: Solr spring application context error

2009-05-08 Thread sachin78

Iam having the same problem.Please let me know if anyone finds answer to
this.

Thank you,
Sachin


markrmiller wrote:
> 
> I've run into this in the past as well. Its fairly annoying. Anyone know 
> why the limitation? Why aren't we passing the ClassLoader thats loading 
> Solr classes as the parent to the lib dir plugin classloader?
> 
> - Mark
> 
> Erik Hatcher wrote:
>> This is probably because Solr loads its extensions from a custom class 
>> loader, but if that class then needs to access things from the 
>> classpath, it is only going to see the built-in WEB-INF/lib classes, 
>> not solr/home lib JAR files.  Maybe there is a Spring way to point it 
>> at that lib directory also?  This is the kinda pain we get, it seems, 
>> when reinventing a container, unfortunately.
>>
>> Erik
>>
>> On May 7, 2009, at 2:49 PM, Raju444us wrote:
>>
>>>
>>> I have configured solr using tomcat.Everything works fine.I overrode
>>> QParserPlugin and configured it.The overriden QParserPlugin has a 
>>> dependency
>>> on another project say project1.So I made a jar of the project and 
>>> copied
>>> the jar to the solr/home lib dir.
>>>
>>> the project1 project is using spring.It has a factory class which 
>>> loads the
>>> beans.Iam using this factory calss in QParserPlugin to get a bean.When I
>>> start my tomcat the factory class is loading fine.But the problem is 
>>> its not
>>> loading the beans.And Iam getting exception
>>>
>>> org.springframework.beans.factory.BeanDefinitionStoreException: 
>>> IOException
>>> parsing XML document from class path resource
>>> [com/mypackage/applicationContext.xml]; nested exception is
>>> java.io.FileNotFoundException: class path resource
>>> [com/mypackage/applicationContext.xml] cannot be opened because it 
>>> does not
>>> exist
>>>
>>> Do I need to do something else?. Can anybody please help me.
>>>
>>> Thanks,
>>> Raju
>>>
>>>
>>> -- 
>>> View this message in context: 
>>> http://www.nabble.com/Solr-spring-application-context-error-tp23432901p23432901.html
>>>  
>>>
>>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
> 
> 
> -- 
> - Mark
> 
> http://www.lucidimagination.com
> 
> 
> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Solr-spring-application-context-error-tp23432901p23446518.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: bug? No highlighting results with dismax and q.alt=*:*

2009-05-08 Thread Peter Wolanin
Possibly this issue is related:  https://issues.apache.org/jira/browse/SOLR-825

Though it seems that might affect the standard handler, while what I'm
seeing is more sepcific to the dismax handler.

-Peter

On Thu, May 7, 2009 at 8:27 PM, Peter Wolanin  wrote:
> For the Drupal Apache Solr Integration module, we are exploring the
> possibility of doing facet browsing  - since we are using dismax as
> the default handler, this would mean issuing a query with an empty q
> and falling back to to q.alt='*:*' or some other q.alt that matches
> all docs.
>
> However, I notice when I do this that we do not get any highlights
> back in the results despite defining a highlight alternate field.
>
> In contrast, if I force the standard request handler then I do get
> text back from the highlight alternate field:
>
> select/?q=*:*&qt=standard&hl=true&hl.fl=body&hl.alternateField=body&hl.maxAlternateFieldLength=256
>
> However, I then loose the nice dismax features of weighting the
> results using bq and bf parameters.  So, is this a bug or the intended
> behavior?
>
> The relevant fragment of the solrconfig.xml is this:
>
>  
>    
>     dismax
>
>     *:*
>
>   
>     true
>     body
>     3
>     true
>   
>     body
>     256
>
>
> Full solrconfig.xml and other files:
> http://cvs.drupal.org/viewvc.py/drupal/contributions/modules/apachesolr/?pathrev=DRUPAL-6--1
>
> --
> Peter M. Wolanin, Ph.D.
> Momentum Specialist,  Acquia. Inc.
> peter.wola...@acquia.com
>



-- 
Peter M. Wolanin, Ph.D.
Momentum Specialist,  Acquia. Inc.
peter.wola...@acquia.com


Re: Backups using Java-based Replication (forced snapshot)

2009-05-08 Thread pjaol

Thinking the same last week, as I was tailoring the snapshooter.sh script.
The data directory should be kept for the temp snapshot, as a way to ensure
linking is occurring on the same device.
snapshooter.sh
 87 name=${data_dir}/${snap_name}
I think only this needs to be configurable for the final move.


Grant Ingersoll-6 wrote:
> 
> On the page http://wiki.apache.org/solr/SolrReplication, it says the  
> following:
> "Force a snapshot on master.This is useful to take periodic  
> backups .command :  http://master_host:port/solr/replication? 
> command=snapshoot"
> 
> This then puts the snapshot under the data directory.  Perfectly  
> reasonable thing to do.  However, is it possible to have it take in a  
> directory location and store the snapshot there?  For instance, I may  
> want to have it write to a specific directory that is being watched  
> for backup data.
> 
> Thanks,
> Grant
> 
> 
> --
> Grant Ingersoll
> http://www.lucidimagination.com/
> 
> Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids)  
> using Solr/Lucene:
> http://www.lucidimagination.com/search
> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Backups-using-Java-based-Replication-%28forced-snapshot%29-tp23434396p23447792.html
Sent from the Solr - User mailing list archive at Nabble.com.



solr + wordpress

2009-05-08 Thread Noble Paul നോബിള്‍ नोब्ळ्
Somebody has writte an articles on integrating Solr with wordpress

http://www.ipros.nl/2008/12/15/using-solr-with-wordpress/

-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com


Re: bug? No highlighting results with dismax and q.alt=*:*

2009-05-08 Thread Marc Sturlese

I have experienced it before... maybe you can manage something similar to
your q.alt using the params q and qf. Highlight will work in that case (I
sorted it out doing that)

Peter Wolanin-2 wrote:
> 
> Possibly this issue is related: 
> https://issues.apache.org/jira/browse/SOLR-825
> 
> Though it seems that might affect the standard handler, while what I'm
> seeing is more sepcific to the dismax handler.
> 
> -Peter
> 
> On Thu, May 7, 2009 at 8:27 PM, Peter Wolanin 
> wrote:
>> For the Drupal Apache Solr Integration module, we are exploring the
>> possibility of doing facet browsing  - since we are using dismax as
>> the default handler, this would mean issuing a query with an empty q
>> and falling back to to q.alt='*:*' or some other q.alt that matches
>> all docs.
>>
>> However, I notice when I do this that we do not get any highlights
>> back in the results despite defining a highlight alternate field.
>>
>> In contrast, if I force the standard request handler then I do get
>> text back from the highlight alternate field:
>>
>> select/?q=*:*&qt=standard&hl=true&hl.fl=body&hl.alternateField=body&hl.maxAlternateFieldLength=256
>>
>> However, I then loose the nice dismax features of weighting the
>> results using bq and bf parameters.  So, is this a bug or the intended
>> behavior?
>>
>> The relevant fragment of the solrconfig.xml is this:
>>
>>  > default="true">
>>    
>>     dismax
>>
>>     *:*
>>
>>   
>>     true
>>     body
>>     3
>>     true
>>   
>>     body
>>     256
>>
>>
>> Full solrconfig.xml and other files:
>> http://cvs.drupal.org/viewvc.py/drupal/contributions/modules/apachesolr/?pathrev=DRUPAL-6--1
>>
>> --
>> Peter M. Wolanin, Ph.D.
>> Momentum Specialist,  Acquia. Inc.
>> peter.wola...@acquia.com
>>
> 
> 
> 
> -- 
> Peter M. Wolanin, Ph.D.
> Momentum Specialist,  Acquia. Inc.
> peter.wola...@acquia.com
> 
> 

-- 
View this message in context: 
http://www.nabble.com/bug--No-highlighting-results-with-dismax-and-q.alt%3D*%3A*-tp23438048p23450189.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: preImportDeleteQuery

2009-05-08 Thread wojtekpia

I'm using full-import, not delta-import. I tried it with delta-import, and it
would work, except that I'm querying for a large number of documents so I
can't afford the cost of deltaImportQuery for each document.

It sounds like $deleteDocId will work. I just need to update from 1.3 to
trunk. Thanks!


Noble Paul നോബിള്‍  नोब्ळ्-2 wrote:
> 
> are you doing a full-import or a delta-import?
> 
> for delta-import there is an option of deletedPkQuery which should
> meet your needs
> 
> 

-- 
View this message in context: 
http://www.nabble.com/preImportDeleteQuery-tp23437674p23450308.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: solr + wordpress

2009-05-08 Thread Matt Weber

I actually wrote a plugin that integrates Solr with WordPress.

http://www.mattweber.org/2009/04/21/solr-for-wordpress/
http://wordpress.org/extend/plugins/solr-for-wordpress/
https://launchpad.net/solr4wordpress

Thanks,

Matt Weber
eSr Technologies
http://www.esr-technologies.com




On May 8, 2009, at 10:10 AM, Noble Paul നോബിള്‍  
नोब्ळ् wrote:



Somebody has writte an articles on integrating Solr with wordpress

http://www.ipros.nl/2008/12/15/using-solr-with-wordpress/

--
-
Noble Paul | Principal Engineer| AOL | http://aol.com




how to pronounce solr

2009-05-08 Thread Charles Federspiel
Hi,
My company is evaluating different open-source indexing and search software
and we are seriously considering Solr.
One of my collegues pronounces it differently than I do and I have no basis
of correcting him.
Is Solr pronounced "SOLerrr"(emphasis on first syllable), or pirate-like,
"SolAhhRrr" (emphasis on the R).

This coworker has just come from a big meeting with various managers where
the technology came up and I'm afraid my battle over this very important
matter may already have been lost.
thank you,
Charles


RE: Initialising of CommonsHttpSolrServer in Spring framwork

2009-05-08 Thread sachin78

Ranjeeth,

Di you figured aout how to do this.If yes, can you share with me how you
did it? Example bean definition in xml will be helpful.

--Sachin


Funtick wrote:
> 
> Use constructor and pass URL parameter. Nothing SPRING related... 
> 
> Create a Spring bean with attributes 'MySolr', 'MySolrUrl', and 'init'
> method... 'init' will create instance of CommonsHttpSolrServer. Configure
> Spring...
> 
> 
> 
>> I am using Solr 1.3 and Solrj as a Java Client. I am 
>> Integarating Solrj in Spring framwork, I am facing a problem, 
>> Spring framework is not inializing CommonsHttpSolrServer 
>> class, how can  I define this class to get the instance of 
>> SolrServer to invoke furthur method on this. 
>> 
> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Initialising-of-CommonsHttpSolrServer-in-Spring-framwork-tp18808743p23451795.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: how to pronounce solr

2009-05-08 Thread Sean Timm
This is the funniest e-mail I've had all day.  SOLer is the typical 
pronunciation, but I've heard solAR as well.  It's the description of 
"pirate-like" that made me chuckle.


-Sean

Charles Federspiel wrote:

Hi,
My company is evaluating different open-source indexing and search software
and we are seriously considering Solr.
One of my collegues pronounces it differently than I do and I have no basis
of correcting him.
Is Solr pronounced "SOLerrr"(emphasis on first syllable), or pirate-like,
"SolAhhRrr" (emphasis on the R).

This coworker has just come from a big meeting with various managers where
the technology came up and I'm afraid my battle over this very important
matter may already have been lost.
thank you,
Charles

  


Re: StatsComponent and 1.3

2009-05-08 Thread Erik Hatcher


On May 7, 2009, at 10:25 PM, David Shettler wrote:

I found myself with a real need for some of the StatsComponent stuff
(mainly the sum feature), so I pulled down a nightly build and played
with it.  StatsComponent proved perfect, but... the nightly build
output seems to be different, and thus incompatible with acts_as_solr.


Could you give some more details on what seems different and  
incompatible with acts_as_solr?


You can query the StatsComponent from Ruby using the solr-ruby  
library.  Using the example from the wiki at http://wiki.apache.org/solr/StatsComponent 
, it points to http://localhost:8983/solr/select?q=*:*&stats=true&stats.field=price&stats.field=popularity&rows=0&indent=true


require 'solr'
solr = Solr::Connection.new
solr.send(Solr::Request::Select.new(:standard, :q => '*:*', :stats =>  
true, 'stats.field' => ['price','popularity'], :rows=>0))


Which output this (in irb):

=> #1,  
"params"=>{"stats"=>"true", "qt"=>"standard", "stats.field"=>["price",  
"popularity"], "q"=>"*:*", "rows"=>"0", "wt"=>"ruby"}, "status"=>0},  
@raw_response 
= 
"{'responseHeader 
'= 
> 
{'status 
'= 
> 
0 
,'QTime 
'= 
> 
1 
,'params 
'= 
> 
{'stats 
'= 
> 
'true 
','q 
'= 
> 
'*:*','stats 
.field 
'= 
> 
['price 
','popularity 
'],'qt 
'= 
> 
'standard 
','wt 
'= 
> 
'ruby 
','rows 
'= 
> 
'0 
'}},'response 
'= 
> 
{'numFound 
'= 
> 
26 
,'start 
'= 
> 
0 
,'docs 
'= 
> 
[]},'stats 
'= 
> 
{'stats_fields 
'= 
> 
{'price 
'= 
> 
{'min 
'= 
> 
0.0 
,'max 
'= 
> 
2199.0 
,'sum 
'= 
> 
5251.26995 
,'count 
'= 
> 
15 
,'missing 
'= 
> 
11 
,'sumOfSquares 
'= 
> 
6038619.160315 
,'mean 
'= 
> 
350.084664 
,'stddev 
'= 
> 
547.737557906113 
},'popularity 
'= 
> 
{'min 
'= 
> 
0.0 
,'max 
'= 
> 
10.0 
,'sum 
'= 
> 
90.0 
,'count 
'= 
> 
26 
,'missing 
'= 
> 
0 
,'sumOfSquares 
'=>628.0,'mean'=>3.4615384615384617,'stddev'=>3.5578731762756157",  
@data={"response"=>{"start"=>0, "docs"=>[], "numFound"=>26},  
"stats"=>{"stats_fields"=>{"price"=>{"sumOfSquares"=>6038619.1603,  
"sum"=>5251.27, "max"=>2199.0, "mean"=>350.0847, "count"=>15,  
"stddev"=>547.737557906113, "min"=>0.0, "missing"=>11},  
"popularity"=>{"sumOfSquares"=>628.0, "sum"=>90.0, "max"=>10.0,  
"mean"=>3.46153846153846, "count"=>26, "stddev"=>3.55787317627562,  
"min"=>0.0, "missing"=>0}}}, "responseHeader"=>{"QTime"=>1,  
"params"=>{"stats"=>"true", "qt"=>"standard", "stats.field"=>["price",  
"popularity"], "q"=>"*:*", "rows"=>"0", "wt"=>"ruby"}, "status"=>0}}>



Is it possible, with some degree of effort (obviously) for me to
essentially port some of the functionality of StatsComponent to 1.3
myself?  It's that, or waiting for 1.4 to come out and someone
developing support for it into acts_as_solr, or myself fixing what I
have for acts_as_solr to work with the output.  I'm just trying to
gauge the easiest solution :)


I'm unclear on what what the discrepancies are, so not quite sure how  
to help just yet.


As Eric asked, what version/branch of acts_as_solr are you using?

Erik



Re: how to pronounce solr

2009-05-08 Thread Jon Gorman
On Fri, May 8, 2009 at 2:07 PM, Charles Federspiel
 wrote:
> Hi,
> My company is evaluating different open-source indexing and search software
> and we are seriously considering Solr.
> One of my collegues pronounces it differently than I do and I have no basis
> of correcting him.
> Is Solr pronounced "SOLerrr"(emphasis on first syllable), or pirate-like,
> "SolAhhRrr" (emphasis on the R).
>
> This coworker has just come from a big meeting with various managers where
> the technology came up and I'm afraid my battle over this very important
> matter may already have been lost.
> thank you,


It's pronounced Solar.  However you choose to pronounce solar, of
course, is up to you or your regionalism.  But that's what explains
the sun logo and people who make puns about solr energy. ;).

It's also the third question in the FAQ
http://wiki.apache.org/solr/FAQ#head-0076d43a3911cf40a231e9ecf7df5303ccee0dc7.
 Just in case you need documented proof to argue your point.  Unless
of course you wanted it to be a pirate word.

Perhaps you should send him or her an eye-patch in case a correction
in this matter would hurt the collegues' feelings.  On second thought,
maybe not.

Jon Gorman


Re: JVM exception_access_violation

2009-05-08 Thread wojtekpia

I updated to Java 6 update 13 and have been running problem free for just
over a month. I'll continue this thread if I run into any problems that seem
to be related.


Yonik Seeley-2 wrote:
> 
> I assume that you're not using any Tomcat native libs?  If you are,
> try removing them... if not (and the crash happened more than once in
> the same place) then it looks like a JVM bug rather than flakey
> hardware and the easiest path forward would be to try the latest Java6
> (update 12).
> 

-- 
View this message in context: 
http://www.nabble.com/JVM-exception_access_violation-tp22623667p23451994.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Is it possible to writing solr result on disk from the server side?

2009-05-08 Thread Yonik Seeley
On Thu, May 7, 2009 at 10:55 AM, arno13  wrote:
> Do you know if it's possible to writing solr results directly on a hard disk
> from server side and not to use an HTTP connection to transfer the results?

If you have something like a CSV file (or any other file type that
Solr accepts over HTTP), you can instruct that the body be read
directly from disk instead.

http://wiki.apache.org/solr/UpdateCSV

-Yonik
http://www.lucidimagination.com


Re: Autocommit blocking adds? AutoCommit Speedup?

2009-05-08 Thread Jim Murphy

Any pointers to this newer more concurrent behavior in lucene?  I can try an
experiment where I downgrade the iwCommit lock to the iwAccess lock to allow
updates to happen during commit.  

Would you expect that to work?

Thanks for bootstrapping me on this. 

Jim



Yonik Seeley-2 wrote:
> 
> On Thu, May 7, 2009 at 8:37 PM, Jim Murphy  wrote:
>> Interesting.  So is there a JIRA ticket open for this already? Any chance
>> of
>> getting it into 1.4?
> 
> No ticket currently open, but IMO it could make it for 1.4.
> 
>> Its seriously kicking out butts right now.  We write
>> into our masters with ~50ms response times till we hit the autocommit
>> then
>> add/update response time is 10-30 seconds.  Ouch.
> 
> It's probably been made a little worse lately since Lucene now does
> fsync on index files before writing the segments file that points to
> those files.  A necessary evil to prevent index corruption.
> 
>> I'd be willing to work on submitting a patch given a better understanding
>> of
>> the issue.
> 
> Great, go for it!
> 
> -Yonik
> http://www.lucidimagination.com
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Autocommit-blocking-adds---AutoCommit-Speedup--tp23435224p23452011.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Autocommit blocking adds? AutoCommit Speedup?

2009-05-08 Thread Yonik Seeley
On Fri, May 8, 2009 at 4:27 PM, Jim Murphy  wrote:
>
> Any pointers to this newer more concurrent behavior in lucene?

At the API level we care about, IndexWriter.commit() instead of close()

Also, we shouldn't have to worry about other parts of the code closing
the writer on us since things like deleteByQuery no longer need to
close the writer to work.

core.getSearcher()... if we don't lock until it's finished, then what
could currently happen is that you could wind up with a newer version
of the index than you thought you might.  I think this should be fine
though.

We'd need to think about what type of synchronization may be needed
for postCommit and postOptimize hooks too.

Here's the relevant code:

iwCommit.lock();
try {
  log.info("start "+cmd);

  if (cmd.optimize) {
openWriter();
writer.optimize(cmd.maxOptimizeSegments);
  }

  closeWriter();

  callPostCommitCallbacks();
  if (cmd.optimize) {
callPostOptimizeCallbacks();
  }
  // open a new searcher in the sync block to avoid opening it
  // after a deleteByQuery changed the index, or in between deletes
  // and adds of another commit being done.
  core.getSearcher(true,false,waitSearcher);



-Yonik
http://www.lucidimagination.com


Re: aka Replication Stall

2009-05-08 Thread Yonik Seeley
2009/5/7 Noble Paul നോബിള്‍  नोब्ळ् :
> BTW, if the timeout occurs it resumes from the point where the failure
> happened. It retries 5 times before giving up.

Sweet!  I was just going to ask what happens on a timeout.
Have you tested this out (say kill the master in the middle of
replicating a big file)?

-Yonik
http://www.lucidimagination.com


Re: Autocommit blocking adds? AutoCommit Speedup?

2009-05-08 Thread jayson.minard

Created issue:

https://issues.apache.org/jira/browse/SOLR-1155



Jim Murphy wrote:
> 
> Any pointers to this newer more concurrent behavior in lucene?  I can try
> an experiment where I downgrade the iwCommit lock to the iwAccess lock to
> allow updates to happen during commit.  
> 
> Would you expect that to work?
> 
> Thanks for bootstrapping me on this. 
> 
> Jim
> 
> 
> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Autocommit-blocking-adds---AutoCommit-Speedup--tp23435224p23453693.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Autocommit blocking adds? AutoCommit Speedup?

2009-05-08 Thread Jim Murphy



Yonik Seeley-2 wrote:
> 
> ...your code snippit elided and edited below ...
> 



Don't take this code as correct (or even compiling) but is this the essence? 
I moved shared access to the writer inside the read lock and kept the other
non-commit bits to the write lock.  I'd need to rethink the locking in a
more fundamental way but is this close to idea? 



 public void commit(CommitUpdateCommand cmd) throws IOException {

if (cmd.optimize) {
  optimizeCommands.incrementAndGet();
} else {
  commitCommands.incrementAndGet();
}

Future[] waitSearcher = null;
if (cmd.waitSearcher) {
  waitSearcher = new Future[1];
}

boolean error=true;
iwCommit.lock();
try {
  log.info("start "+cmd);

  if (cmd.optimize) {
closeSearcher();
openWriter();
writer.optimize(cmd.maxOptimizeSegments);
  }
finally {
  iwCommit.unlock();
 }


  iwAccess.lock(); 
  try
 {
  writer.commit();
 }
 finally
 {
  iwAccess.unlock(); 
 }

  iwCommit.lock(); 
  try
 {
  callPostCommitCallbacks();
  if (cmd.optimize) {
callPostOptimizeCallbacks();
  }
  // open a new searcher in the sync block to avoid opening it
  // after a deleteByQuery changed the index, or in between deletes
  // and adds of another commit being done.
  core.getSearcher(true,false,waitSearcher);

  // reset commit tracking
  tracker.didCommit();

  log.info("end_commit_flush");

  error=false;
}
finally {
  iwCommit.unlock();
  addCommands.set(0);
  deleteByIdCommands.set(0);
  deleteByQueryCommands.set(0);
  numErrors.set(error ? 1 : 0);
}

// if we are supposed to wait for the searcher to be registered, then we
should do it
// outside of the synchronized block so that other update operations can
proceed.
if (waitSearcher!=null && waitSearcher[0] != null) {
   try {
waitSearcher[0].get();
  } catch (InterruptedException e) {
SolrException.log(log,e);
  } catch (ExecutionException e) {
SolrException.log(log,e);
  }
}
  }



-- 
View this message in context: 
http://www.nabble.com/Autocommit-blocking-adds---AutoCommit-Speedup--tp23435224p23454419.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Autocommit blocking adds? AutoCommit Speedup?

2009-05-08 Thread jayson.minard

Can we move this to patch files within the JIRA issue please.  Will make it
easier to review and help out a as a patch to current trunk.

--j


Jim Murphy wrote:
> 
> 
> 
> Yonik Seeley-2 wrote:
>> 
>> ...your code snippit elided and edited below ...
>> 
> 
> 
> 
> Don't take this code as correct (or even compiling) but is this the
> essence?  I moved shared access to the writer inside the read lock and
> kept the other non-commit bits to the write lock.  I'd need to rethink the
> locking in a more fundamental way but is this close to idea? 
> 
> 
> 
>  public void commit(CommitUpdateCommand cmd) throws IOException {
> 
> if (cmd.optimize) {
>   optimizeCommands.incrementAndGet();
> } else {
>   commitCommands.incrementAndGet();
> }
> 
> Future[] waitSearcher = null;
> if (cmd.waitSearcher) {
>   waitSearcher = new Future[1];
> }
> 
> boolean error=true;
> iwCommit.lock();
> try {
>   log.info("start "+cmd);
> 
>   if (cmd.optimize) {
> closeSearcher();
> openWriter();
> writer.optimize(cmd.maxOptimizeSegments);
>   }
> finally {
>   iwCommit.unlock();
>  }
> 
> 
>   iwAccess.lock(); 
>   try
>  {
>   writer.commit();
>  }
>  finally
>  {
>   iwAccess.unlock(); 
>  }
> 
>   iwCommit.lock(); 
>   try
>  {
>   callPostCommitCallbacks();
>   if (cmd.optimize) {
> callPostOptimizeCallbacks();
>   }
>   // open a new searcher in the sync block to avoid opening it
>   // after a deleteByQuery changed the index, or in between deletes
>   // and adds of another commit being done.
>   core.getSearcher(true,false,waitSearcher);
> 
>   // reset commit tracking
>   tracker.didCommit();
> 
>   log.info("end_commit_flush");
> 
>   error=false;
> }
> finally {
>   iwCommit.unlock();
>   addCommands.set(0);
>   deleteByIdCommands.set(0);
>   deleteByQueryCommands.set(0);
>   numErrors.set(error ? 1 : 0);
> }
> 
> // if we are supposed to wait for the searcher to be registered, then
> we should do it
> // outside of the synchronized block so that other update operations
> can proceed.
> if (waitSearcher!=null && waitSearcher[0] != null) {
>try {
> waitSearcher[0].get();
>   } catch (InterruptedException e) {
> SolrException.log(log,e);
>   } catch (ExecutionException e) {
> SolrException.log(log,e);
>   }
> }
>   }
> 
> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Autocommit-blocking-adds---AutoCommit-Speedup--tp23435224p23455432.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Initialising of CommonsHttpSolrServer in Spring framwork

2009-05-08 Thread Avlesh Singh
I am giving you a detailed sample of my spring usage.













http://localhost/solr/core1"/>





http://localhost/solr/core2"/>





Hope this helps.

Cheers
Avlesh

On Sat, May 9, 2009 at 12:39 AM, sachin78 wrote:

>
> Ranjeeth,
>
>Did you figured aout how to do this? If yes, can you share with me how
> you did it? Example bean definition in xml will be helpful.
>
> --Sachin
>
>
> Funtick wrote:
> >
> > Use constructor and pass URL parameter. Nothing SPRING related...
> >
> > Create a Spring bean with attributes 'MySolr', 'MySolrUrl', and 'init'
> > method... 'init' will create instance of CommonsHttpSolrServer. Configure
> > Spring...
> >
> >
> >
> >> I am using Solr 1.3 and Solrj as a Java Client. I am
> >> Integarating Solrj in Spring framwork, I am facing a problem,
> >> Spring framework is not inializing CommonsHttpSolrServer
> >> class, how can  I define this class to get the instance of
> >> SolrServer to invoke furthur method on this.
> >>
> >
> >
> >
>
> --
> View this message in context:
> http://www.nabble.com/Initialising-of-CommonsHttpSolrServer-in-Spring-framwork-tp18808743p23451795.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>


Re: Solrconfig.xml

2009-05-08 Thread Shalin Shekhar Mangar
On Thu, May 7, 2009 at 10:06 AM, Francis Yakin  wrote:

>
> No error, attached is solrconfig.xml files( one is from 1.2.0 that works
> and the other is 1.3.0 that doesn't work)
>
>
Francis, it seems the attached files were eaten up by the mailing list. Can
you re-send or put them up online somewhere (e.g. on pastebin.com)?
-- 
Regards,
Shalin Shekhar Mangar.


Re: Control segment size

2009-05-08 Thread Shalin Shekhar Mangar
On Fri, May 8, 2009 at 1:30 AM, vivek sar  wrote:

>
> I did set the maxMergeDocs to 10M, but I still see couple of index
> files over 30G which do not match with max number of documents. Here
> are some numbers,
>
> 1) My total index size = 66GB
> 2) Number of total documents = 200M
> 3) 1M doc = 300MB
> 4) 10M doc should be roughly around 3-4GB.
>
> As you can see couple of files are huge. Are those documents or index
> files? How can I control the file size so no single file grows more
> than 10GB.
>

No, there is no way to limit an individual file to a specific size.

-- 
Regards,
Shalin Shekhar Mangar.


Re: aka Replication Stall

2009-05-08 Thread Noble Paul നോബിള്‍ नोब्ळ्
On Sat, May 9, 2009 at 2:23 AM, Yonik Seeley  wrote:
> 2009/5/7 Noble Paul നോബിള്‍  नोब्ळ् :
>> BTW, if the timeout occurs it resumes from the point where the failure
>> happened. It retries 5 times before giving up.
>
> Sweet!  I was just going to ask what happens on a timeout.
> Have you tested this out (say kill the master in the middle of
> replicating a big file)?

Actually yes,

not by killing the master (then the replication will abort). I
modified the master code to close the connection after transferring x
MB of data. The slave retried an completed the replication .
>
> -Yonik
> http://www.lucidimagination.com
>



-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com


Re: Autocommit blocking adds? AutoCommit Speedup?

2009-05-08 Thread jayson.minard

First cut of updated handler now in:
https://issues.apache.org/jira/browse/SOLR-1155

Needs review from those that know Lucene better, and double check for errors
in locking or other areas of the code.  Thanks.

--j


jayson.minard wrote:
> 
> Can we move this to patch files within the JIRA issue please.  Will make
> it easier to review and help out a as a patch to current trunk.
> 
> --j
> 
> 
> Jim Murphy wrote:
>> 
>> 
>> 
>> Yonik Seeley-2 wrote:
>>> 
>>> ...your code snippit elided and edited below ...
>>> 
>> 
>> 
>> 
>> Don't take this code as correct (or even compiling) but is this the
>> essence?  I moved shared access to the writer inside the read lock and
>> kept the other non-commit bits to the write lock.  I'd need to rethink
>> the locking in a more fundamental way but is this close to idea? 
>> 
>> 
>> 
>>  public void commit(CommitUpdateCommand cmd) throws IOException {
>> 
>> if (cmd.optimize) {
>>   optimizeCommands.incrementAndGet();
>> } else {
>>   commitCommands.incrementAndGet();
>> }
>> 
>> Future[] waitSearcher = null;
>> if (cmd.waitSearcher) {
>>   waitSearcher = new Future[1];
>> }
>> 
>> boolean error=true;
>> iwCommit.lock();
>> try {
>>   log.info("start "+cmd);
>> 
>>   if (cmd.optimize) {
>> closeSearcher();
>> openWriter();
>> writer.optimize(cmd.maxOptimizeSegments);
>>   }
>> finally {
>>   iwCommit.unlock();
>>  }
>> 
>> 
>>   iwAccess.lock(); 
>>   try
>>  {
>>   writer.commit();
>>  }
>>  finally
>>  {
>>   iwAccess.unlock(); 
>>  }
>> 
>>   iwCommit.lock(); 
>>   try
>>  {
>>   callPostCommitCallbacks();
>>   if (cmd.optimize) {
>> callPostOptimizeCallbacks();
>>   }
>>   // open a new searcher in the sync block to avoid opening it
>>   // after a deleteByQuery changed the index, or in between deletes
>>   // and adds of another commit being done.
>>   core.getSearcher(true,false,waitSearcher);
>> 
>>   // reset commit tracking
>>   tracker.didCommit();
>> 
>>   log.info("end_commit_flush");
>> 
>>   error=false;
>> }
>> finally {
>>   iwCommit.unlock();
>>   addCommands.set(0);
>>   deleteByIdCommands.set(0);
>>   deleteByQueryCommands.set(0);
>>   numErrors.set(error ? 1 : 0);
>> }
>> 
>> // if we are supposed to wait for the searcher to be registered, then
>> we should do it
>> // outside of the synchronized block so that other update operations
>> can proceed.
>> if (waitSearcher!=null && waitSearcher[0] != null) {
>>try {
>> waitSearcher[0].get();
>>   } catch (InterruptedException e) {
>> SolrException.log(log,e);
>>   } catch (ExecutionException e) {
>> SolrException.log(log,e);
>>   }
>> }
>>   }
>> 
>> 
>> 
>> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Autocommit-blocking-adds---AutoCommit-Speedup--tp23435224p23457422.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: French and SpellingQueryConverter

2009-05-08 Thread Shalin Shekhar Mangar
On Fri, May 8, 2009 at 2:14 AM, Jonathan Mamou  wrote:

> Hi
> It does not seem to be related to FrenchStemmer, the stemmer does not split
> a word into 2 words. I have checked with other words and
> SpellingQueryConverter always splits words with special character.
> I think that the issue is in SpellingQueryConverter class
> Pattern.compile.("(?:(?!(\\w+:|\\d+)))\\w+");?:
> According to
> http://java.sun.com/javase/6/docs/api/java/util/regex/Pattern.html,
> \w A word character: [a-zA-Z_0-9]
> I think that special character should also be added to the regex.
>

If you use spellcheck.q parameter for specifying the spelling query, then
the field's analyzer will be used (in this case, FrenchAnalyzer). If you use
the q parameter, then the SpellingQueryConverter is used.

-- 
Regards,
Shalin Shekhar Mangar.