Fergus McMenemie wrote:
Hi all,
I am attempting to test some changes I made to my DIH based
indexing process. The changes only affect the way I
describe my fields in data-config.xml, there should be no
changes to the way the data is indexed or stored.
As a QA check I was wanting to compare th
Hi all,
I am attempting to test some changes I made to my DIH based
indexing process. The changes only affect the way I
describe my fields in data-config.xml, there should be no
changes to the way the data is indexed or stored.
As a QA check I was wanting to compare the results from
indexing the
Thanks Hoss!
The approach that I explained in my subsequent email works like a charm.
Cheers
Avlesh
On Wed, Sep 30, 2009 at 3:45 AM, Chris Hostetter
wrote:
>
> : The question was either non-trivial or heavily uninteresting! No replies
> yet
>
> it's pretty non-trivial, and pretty interesting, bu
: Now I've got a completely different error:
: "org.apache.lucene.index.CorruptIndexException: Unknown format version: -9".
: I think it might be time for a fresh install...
I've added a FAQ for this...
http://wiki.apache.org/solr/FAQ#What_does_.22CorruptIndexException:_Unknown_format_version.22
: Hi all, I'm having problems getting Solr to start on Tomcat 6.
which version of Solr?
: Tomcat is installed in /opt/apache-tomcat , solr is in
: /opt/apache-tomcat/webapps/solr , and my Solr home directory is /opt/solr .
if "solr is in /opt/apache-tomcat/webapps/solr" means that you put the
Yep, super straight-forward, thanks a bunch!
Guess I missed this piece of the wiki, looks like its going through alot of
updates towards
solr 1.4 release..
thanks,
-Chak
ryguasu wrote:
>
> The documentation could maybe be improved, but the basics of backup
> snapshots with the in-process (Java
: Is it possible to tell Solr or Lucene, when optimizing, to write the files
: that constitute the optimized index to somewhere other than
: SOLR_HOME/data/index or is there something about the optimize that requires
: the final segment to be created in SOLR_HOME/data/index?
For what purp
: I've been using q.alt=-*:* because *:* is said to be the most efficient way
of
: querying for every document. is -*:* the most efficient way of querying for
: no document?
I don't think so ... solr internally reverse pure negative queries so that
they are combined with a matchalldocsquery t
I'll just allow the user to pass in the port via a param for now.
Thx!
On Tue, Sep 29, 2009 at 4:13 PM, Mark Miller wrote:
> Unfortunately, because they don't want you counting on access to the
> servlet request due to embedded Solr and what not, to get that type of
> info you have to override a
Unfortunately, because they don't want you counting on access to the
servlet request due to embedded Solr and what not, to get that type of
info you have to override and use your own SolrDispatchFilter:
protected void execute( HttpServletRequest req, SolrRequestHandler
handler, SolrQueryRequest
: 2009-09-23T19:25:03.400Z
:
: 2009-09-23T19:25:19.951
:
: 2009-09-23T20:10:07.919Z
is that a cut/paste error, or did you really get a date back from Solr
w/o the trailing "Z" ?!?!?!
...
: So, not only is the date sorting wrong, but the exact same document
: shows up on the next page
Yah, I just found it, and was going to reply to my own message with
that exactly!
My next question is how to get the port the request was on?
On Tue, Sep 29, 2009 at 4:01 PM, Mark Miller wrote:
> Jason Rutherglen wrote:
>> Howdy,
>>
>> I was wondering what the best way is to access the current
>
Jason Rutherglen wrote:
> Howdy,
>
> I was wondering what the best way is to access the current
> instance of CoreContainer? It seems like the only way to do this
> is to extend CoreAdminHandler. I'd prefer a way via a way to
> access CoreContainer from SolrCore or RequestHandlerBase.
>
> The use c
Howdy,
I was wondering what the best way is to access the current
instance of CoreContainer? It seems like the only way to do this
is to extend CoreAdminHandler. I'd prefer a way via a way to
access CoreContainer from SolrCore or RequestHandlerBase.
The use case is, I want to implement a SearchHa
: For a particular requirement we have - we need to do a query that is a
: combination of multiple dismax queries behind the scenes. (Using solr 1.4
: nightly ).
...
: Creating a custom QParser works right away as below.
...
: Curious to see if we have an alternate method to im
: I have been trying to deploy Solr on websphere but no luck yet.
: I was trying to deploy the war file under "dist" folder, but I kept getting
: errors. (recent one is that it couldn't find the configuration file). When I
Did you start by going through the tutorial using the instance of jetty
in
: Does the following query has any performance impact over
: the second query?
: +title:lucene +(title:lucene -name:sid)
: +(title:lucene -name:sid)
the second should in theory be faster then the first just because of
reduced number of comparisons needed -- but wether or not yo
Hi Shalin,
Can you please elaborate, why we need to do unload after create? So, if we
do a create, will it modify the solr.xml everytime? Can it be avoided in
subsequent requests for create?
Also, if we want to implement Load, can you please give some directions to
implement load action?
Than
: I am working on an XSD document for all the types in the response xml
: version 2.2
:
: Do you think there is a need for this?
we haven't had one yet, and it doesn't seem like it's really caused any
problems for people (plus the lack of response to this question suggests
no one is super exci
: The question was either non-trivial or heavily uninteresting! No replies yet
it's pretty non-trivial, and pretty interesting, but i'm also pretty
behind on my solr-user email.
I don't think there's anyway to do what you wanted without a custom
plugin, so your efforts weren't in vain ... if w
The documentation could maybe be improved, but the basics of backup
snapshots with the in-process (Java-based) replication handler
actually seem pretty straightforward to me, now that I understand it:
1. You can make a snapshot whenever you want by hitting
http://master_host:port/solr/replication?
So, you want to display 5 results from each category and still know
how many results are in each category. This is a perfect situation
for the field collapsing patch:
https://issues.apache.org/jira/browse/SOLR-236
http://wiki.apache.org/solr/FieldCollapsing
Here is how I would do it.
Add a
On Wed, Sep 30, 2009 at 12:42 AM, djain101 wrote:
>
> Thanks Shalin for quick response. On the wiki link you mentioned, it is
> saying "not implemented yet!". Can you please confirm again? If yes, then
> in
> which release it is available?
>
Ah, I'm sorry. You are right. Load is not implemented
A description and examples of both parameters can be found here:
http://wiki.apache.org/solr/CommonQueryParameters
Thanks,
Matt Weber
On Sep 29, 2009, at 4:10 AM, Avlesh Singh wrote:
/?q=*:*&fq:category:animal&sort=child_count%20asc
Search for all documents (of animals), and filter the ones
Thanks Shalin for quick response. On the wiki link you mentioned, it is
saying "not implemented yet!". Can you please confirm again? If yes, then in
which release it is available?
Appreciate your quick response.
Regards,
Dharmveer
Shalin Shekhar Mangar wrote:
>
> On Tue, Sep 29, 2009 at 10:0
On Tue, Sep 29, 2009 at 10:01 PM, djain101 wrote:
>
> Is the LOAD action implemented yet?
>
Yes, see http://wiki.apache.org/solr/CoreAdmin
--
Regards,
Shalin Shekhar Mangar.
On Tue, Sep 29, 2009 at 8:14 PM, Bill Dueber wrote:
> Is this possible? I can't figure out a syntax that works, and all the
> examples show using last_index_time as an argument to an SQL query.
>
>
It is possible but it doesn't work right now. I've created an issue and I
will give a patch shortly
Hey,
I noticed with new in-process replication, it is not as straightforward to
have
(production serving) solr index snapshots for backup (it used to be a
natural byproduct
of the snapshot taking process.)
I understand there are some command-line utilities for this (abc..)
Can someone please expla
Dear All,
[Apologies for cross-posting.]
This is just to let you know that we've released version 3.1.0 of Carrot2
Search Results Clustering Engine.
The 3.1.0 release comes with:
* Experimental support for clustering Chinese Simplified content (based on
Lucene's Smart Chinese Analyzer)
* Docume
Hi,
We are also facing the same issue. Is the LOAD action implemented yet? If
not then what should we do to achieve the same functionality?
Thanks,
djain
ryantxu wrote:
>
> The LOAD method will load a core from a schema/config file -- it will
> not need to be in multicore.xml (the persist=
Grant,
Thanks for the link. Based on the example, I think this is what I need.
If effeciency is a problem, I will consider it. I see the note that
tv.df can be expensive.
I guess it all depends on how big the collection is.
I'm a proponent of not reinvientin the wheel if it has already been
inven
Is this possible? I can't figure out a syntax that works, and all the
examples show using last_index_time as an argument to an SQL query.
--
Bill Dueber
Library Systems Programmer
University of Michigan Library
It's harder debugging stuff like this with custom code (you say that
you're using EmbeddedSolrServer) and different servlet containers.
Perahps try putting your config files and index into the example jetty
server, and then do a single request from curl or your web browser to
see if the times are
Constantijn Visinescu wrote:
>
> This might be a bit of a hack but i got this in the web.xml of my
> applicatin
> and it works great.
>
>
>
>solr/home
>/Solr/WebRoot/WEB-INF/solr
>java.lang.String
>
>
>
That worked, thanks. You're right though, it is a
I just want to clarify here that I understand my memory allocation might be
less given the load on the system. The response times were only slightly
better when we ran the test on a Solaris box with 12CPU, 24G RAM and with
3.2 GB allocated for the JVM. I know that I have a performance
problem.
On Tue, Sep 29, 2009 at 3:19 AM, Mark Miller wrote:
> Looks like a bug to me. I don't see the commit point being reserved in
> the backup code - which means its likely be removed before its done
> being copied. Gotto reserve it using the delete policy to keep around
> for the full backup duration
Sorry for the delayed response
**
*How big are your documents?*
I have totally 1 million documents. I have totally 1950 fields in the index.
Every document would probably have values for around 20 - 50 fields.
*What is the total size of the index?*
1 GB
*What's the amout of RAM on your box? How bi
Hi Users...
i have a Problem
I have a lot of fields, (type=text) for search in all fields i copy all
fields in the default text field and use this for default search.
Now i will search...
This is into a Field
"RI-MC500034-1"
when i search "RI-MC500034-1" i found it...
if i seacht "RI-M
This might be a bit of a hack but i got this in the web.xml of my applicatin
and it works great.
solr/home
/Solr/WebRoot/WEB-INF/solr
java.lang.String
On Tue, Sep 29, 2009 at 2:32 PM, Andrew Clegg wrote:
>
> Hi all, I'm having problems getting Solr to start on
Hi all, I'm having problems getting Solr to start on Tomcat 6.
Tomcat is installed in /opt/apache-tomcat , solr is in
/opt/apache-tomcat/webapps/solr , and my Solr home directory is /opt/solr .
My config file is in /opt/solr/conf/solrconfig.xml .
I have a Solr-specific context file in
/opt/apach
On Tue, Sep 29, 2009 at 11:36 AM, Varun Gupta wrote:
> ...
>
> One way that I can think of doing this is by making as many queries as there
> are categories and show these results under each category. But this will be
> very inefficient. Is there any way I can do this ?
Hi Varun!
I think that d
/?q=*:*&fq:category:animal&sort=child_count%20asc
Search for all documents (of animals), and filter the ones that belong to
the category "animal" and sort ascending by a field called child_count that
contains number of children for each animal.
You can pass multiple fq's with more "&fq=..." param
Hi,
Can some one let me know how to use sort and fq parameters in Solr.
Any examples woould be appreciated.
Regards
Bhaskar
Hi,
I am looking for a way to do the following in solr:
When somebody does a search, I want to show results by category (facet) such
that I display 5 results from each category (along with showing the total
number of results in each category which I can always do using the facet
search). This is k
Seems like the post in the SolrNet group: http://groups.google.com/group/solrnet/browse_thread/thread/7e3034b626d3e82d?pli=1
helped me get trough.
Thanks you solr-user's for helping out too!
Steinar
Videresendt melding:
Fra: Steinar Asbjørnsen
Dato: 28. september 2009 17.07.15 GMT+02.00
> Actually the CPU usage of the solr servers is almost insignificant (it was
> like that before).
>>The time spent on collecting memory dropped from 11% to 3.81%
I even think that 3.81% from 5% is nothing (suspecting that SOLR uses 5%
CPU, mostly loading large field values in memory) :)))
(woul
I am using Solr terms for auto suggest and I have 4 millions document in
index and Its working fine. I want to know which will be more faster and
efficient from 'MySql FULLTEXT index and AGAINST' and Solr terms search.
Or Is there any other way in solr for auto suggest. I have separate
applicati
47 matches
Mail list logo