Hi Chris
I did a proper checkout of TIKA 0.9 and built the jars as specified in the
"http://tika.apache.org/0.9/gettingstarted.html"; and replaced the existing
tika0.4 jars with 0.9 jars. I don't see any difference. The documents are
getting indexed but the fmap.content(attr_content) is still not
Thanks it works!!
I want to change the format of the NOW in SOLR. Is it possible? Now date
format looks like this:
-MM-dd T HH:mm:sss Z
In my db the format is dd-MM. How can i fix the NOW so i can do something
like * TO NOW(dd-mm)??
--
View this message in context:
http://lucene.472066.n3.
Hi,
I have an idea how to switching master and slave in case of one server
is crashing:
Setting up two server as repeater but disabling master and slave
config on both with false.
Now you can dynamically disable and enable master or slave option by url:
enable / disable replication on master:
h
Hi all.
I follow the steps of creating a LogTransformer in DataImportHandler wiki:
The java statement that start Solr:
java "-Dremarks=solr:8983"
"-Djava.util.logging.config.file=logging.properties" -jar start.jar
logging.properties file content
# Default global logging level:
.level = DEBU
it will be in the solr logs
On Tue, Jun 21, 2011 at 2:18 PM, Alucard wrote:
> Hi all.
>
> I follow the steps of creating a LogTransformer in DataImportHandler wiki:
>
> pk="office_add_Key" transformer="LogTransformer" logLevel="debug"
> logTemplate="office_add_Key:
> ${office_
Thank you, but what do you mean by "solr logs"?
Actually I cannot find my message in "Solr logs", which is resided in:
/logs/solr_log-%g.log
2011/6/21 Noble Paul നോബിള് नोब्ळ्
> it will be in the solr logs
>
> On Tue, Jun 21, 2011 at 2:18 PM, Alucard wrote:
> > Hi all.
> >
> > I follow the s
OK that sounds like a good solution!
You can also have CMS limit how many merges are allowed to run at
once, if your IO system has trouble w/ that much concurrency.
Mike McCandless
http://blog.mikemccandless.com
On Mon, Jun 20, 2011 at 6:29 PM, Shawn Heisey wrote:
> On 6/20/2011 3:18 PM, Micha
Hi,
I have the following situation:
1. I am using Solr 3.1
2. I am using the edismax query handler for my queries
3. I am using the SolrJ client library
4. Currently i have configured the fields i want to search on and the bosst
factors in solr config.
But i have just been told that we would nee
--- On Tue, 6/21/11, Kissue Kissue wrote:
> From: Kissue Kissue
> Subject: Applying boost factors at run time
> To: solr-user@lucene.apache.org
> Date: Tuesday, June 21, 2011, 1:31 PM
> Hi,
>
> I have the following situation:
>
> 1. I am using Solr 3.1
> 2. I am using the edismax query handl
What is it you want help with? You haven't told us what the
problem you're trying to solve is. Are you asking how to
speed up indexing? What have you tried? Have you
looked at: http://wiki.apache.org/solr/FAQ#Performance?
Best
Erick
On Tue, Jun 21, 2011 at 2:16 AM, Jonty Rhods wrote:
> I am usin
No, you can't as far as I know. The time format in Solr is fixed. Besides,
I don't know what NOW(dd-mm) would mean The day represented
by dd-mm in the current year?
You can probably make your db select emit the dates in the Solr format
Best
Erick
On Tue, Jun 21, 2011 at 3:37 AM, roySolr wro
It should, but there are a couple of issues.
1> you have to make your remaining slaves poll the new master for index updates.
2> your indexing process has to be pointed the new master (if it's external)
3> you have to make sure anything you've indexed to the master that has NOT
been replicated
Yes, current year. I understand that something like dd-mm-yy isn't possible.
I will fix this in my db,
Thanks for your help!
--
View this message in context:
http://lucene.472066.n3.nabble.com/Complex-situation-tp3071936p3090247.html
Sent from the Solr - User mailing list archive at Nabble.com.
we are testing the upgrade to solr 3.2. a quick test look good. solr 3.2. comes
up and we can do searches with our configs (using the "old" dismax handler,
which i have inserted in solrconfig.xml). only problem is, that i am not able
to set up user auth in jetty.
i took the same config files tha
Hi Andreas
I tried solr 3.1 as well as 3.2... i was not able to overcome these issues with
the newer versions too. For me, I need the attr_content:* should return me
results (with 1.4.1 this is successful) which is not happening . It indexes well
in 3.1 but in 3.2 i have the following issue.
Invali
Many thanks for the tip. I will give it a go.
On Tue, Jun 21, 2011 at 11:48 AM, Ahmet Arslan wrote:
>
>
> --- On Tue, 6/21/11, Kissue Kissue wrote:
>
> > From: Kissue Kissue
> > Subject: Applying boost factors at run time
> > To: solr-user@lucene.apache.org
> > Date: Tuesday, June 21, 2011, 1
I am enabling spell checking using solr in search application. i also want
to run wild card queries.
the problem i am facing is when i search for for example diam* then it gives
me a suggestion for diamond and search results for diamond. while i have
some other words in my document say for exampl
On 6/20/2011 12:31 PM, Michael McCandless wrote:
For back-compat, mergeFactor maps to both of these, but it's better to
set them directly eg:
10
20
(and then remove your mergeFactor setting under indexDefaults)
When I did this and ran a reindex, it merged once it rea
I'm attempting to work through the configuration for the home folder for solr
running on a standalone jetty 7.4.2 setup, this is being used on a Mac OS X
10.6.7
I have this working currently under one condition and that is that I specify
the system property when I start Jetty either in the term
Hi Surendra,
Thanks. Besides replacing the tika-*-0.9.jar files, you also need to replace
the dependency jar files for the other libs as well since they have been
upgraded. It's also possible that b/c of API changes, Solr 1.4.1 won't work
with Tika 0.9 without modifying the ExtractingRequestHan
On Tue, Jun 21, 2011 at 2:15 AM, Rafał Kuć wrote:
> Hello!
>
> Once again thanks for the response ;) So the solution is to generate
> the data files once again and either adding the space after doubled
> encapsulator
Maybe...
I can't tell if the file is encoded correctly or not since I don't
know
Are you working on some changes to support earlier versions of PHP?
--
View this message in context:
http://lucene.472066.n3.nabble.com/ANNOUNCEMENT-PHP-Solr-Extension-1-0-1-Stable-Has-Been-Released-tp3024040p3090702.html
Sent from the Solr - User mailing list archive at Nabble.com.
I accidentally rename a core to the same name of existing core, e.g. using
example-DIH:
http://localhost:8983/solr/admin/cores?action=RENAME&core=db&other=tika
I expected solr throws an exception, but it worked, and the existing core
(tika) is gone.
Does it a known bug (but I couldn't find open
On Tue, Jun 21, 2011 at 9:42 AM, Shawn Heisey wrote:
> On 6/20/2011 12:31 PM, Michael McCandless wrote:
>>
>> For back-compat, mergeFactor maps to both of these, but it's better to
>> set them directly eg:
>>
>>
>> 10
>> 20
>>
>>
>> (and then remove your mergeFactor setting u
the problem is that before
https://issues.apache.org/jira/browse/SOLR-2567, Solr invoked the
TieredMergePolicy "setters" *before* it tried to apply these 'global'
mergeFactor etc params.
So, even if you set them explicitly inside the , they
would then get clobbered by these 'global' params / defau
Hello Everyone,
While trying to index a set of documents on remote Solr instance, the
connection broke and it left the index in an inconsistent state. Now, when I
start the instance, it fails while getting the searcher with the following
exception-
Caused by: java.io.IOException: read past EOF
at
On 6/21/2011 9:09 AM, Robert Muir wrote:
the problem is that before
https://issues.apache.org/jira/browse/SOLR-2567, Solr invoked the
TieredMergePolicy "setters" *before* it tried to apply these 'global'
mergeFactor etc params.
So, even if you set them explicitly inside the, they
would then get
We are successfully extracting PDF content with Solr 3.1 and Tika 0.9.
Replace
fontbox-1.3.1.jar jempbox-1.3.1.jar pdfbox-1.3.1.jar tika-core-0.8.jar
tika-parsers-0.8.jar
with
fontbox-1.4.0.jar jempbox-1.4.0.jar pdfbox-1.4.0.jar tika-core-0.9.jar
tika-parsers-0.9.jar
I'm not entirely certa
Sorry, here are some details:
requestHandler: XmlUpdateRequesetHandler
protocol: http (10 concurrend threads)
document: 1kb size, 15 fields
cpu load: 20%
memory usage: 50%
But generally speaking, is that normal or must be something wrong with my
configuration, ...
2011/6/17 Erick Erickson
>
We (Etsy.com) are currently using a version of trunk from mid-October 2010
(SVN tag 1021515, to be exact). We'd like to upgrade to the current trunk
and are wondering if this is a good time. Is the new stuff (esp. DocValues)
stable? Are any other major features or performance improvements about to
Hi !
I want to integrate Solr (Solr 1.4) in a Jonas server. However, I get this
error on jonas :
java.lang.NoSuchMethodError: org.apache.xpath.XPathContext.(Z)V
at org.apache.xpath.jaxp.XPathImpl.eval(XPathImpl.java:207)
at org.apache.xpath.jaxp.XPathImpl.evaluate(XPathImpl.java:281)
at o
Hi everyone,
I'm trying to get Velocity running in Solr 1.4.0, and I'm having a weird
problem. When I navigate to http://localhost:8983/solr/itas , I get an error
message which I'll paste to the end of this email. It says it can't find
velocity.properties, despite the fact that I have this f
You can use prefix with TermsComponent, which may do what you
need.
Best
Erick
On Tue, Jun 21, 2011 at 9:40 AM, Romi wrote:
> I am enabling spell checking using solr in search application. i also want
> to run wild card queries.
> the problem i am facing is when i search for for example diam*
Solr3.1 using SolrJ
So I have a gui that allows folks to search my solr repository and I want to
show appropriate errors when something bad happens, but my problem is that
the Solr exception are not very pretty and sometimes are not very
descriptive.
For instance if I enter a bad query the messag
hello,
i'm not sure of the correct velocity syntax to link, let's say a title
field, to the actual document itself. i have a hostname, a category (which
is also the directory where the file sits) and filename fields in my schema.
can i potentially use these fields to get at the document itself?
Is it possible to do case insensitive searches but return the original
case? So for instance the original field is:
John Smith
I need to be able to do case insensitive searches tokenized searches, but
when the value is returned for faceting I'd like the value to be just "John
Smith", not "john"
There is information
http://wiki.apache.org/solr/DataImportHandler#Scheduling here about
Scheduling but I don't understand how to use them. I am not a Java developer
so maybe I am missing something obvious.
Based on instructions
http://stackoverflow.com/questions/3206171/how-can-i-schedule-dat
Not really. The problem here is that facets are done on terms. To
search effectively, Solr needs tokenized, lower-cased etc. terms.
But since faceting is really just faceting on terms, this is incompatible
with returning multi-term facets like "John Smith" so about all you can
do is to copyfield t
Hi,
I have a "X" field in my index, which is a feature hash I would like to use
to remove the duplicates in my result.
I cant keep this as the unique id field. Is there any method or any
parameter at the search time to remove the duplicates on a particular
field(hash in this case)?
Thanks in adva
I think this is what you're looking for:
http://wiki.apache.org/solr/Deduplication
Best
Erick
On Tue, Jun 21, 2011 at 1:40 PM, Pooja Verlani wrote:
> Hi,
>
> I have a "X" field in my index, which is a feature hash I would like to use
> to remove the duplicates in my result.
> I cant keep this as
Hi Eric,
Thanks for the quick reply.
I had looked at the deduplication but I found it to deduplication at the
index time, right? I would prefer to do deduplication at the search time!
Regards,
Pooja
On Tue, Jun 21, 2011 at 11:15 PM, Erick Erickson wrote:
> I think this is what you're looking fo
Thanks for the reply, I did see that but I am displaying the information in
that field as well so I'll need to store them for this case. For fields I
don't need to display I know that I can just tell it not to store it.
On Tue, Jun 21, 2011 at 1:34 PM, Erick Erickson wrote:
> Not really. The pro
Well, in trunk and the soon-to-be-released Solr 3.3, you could use grouping,
what is the use-case here? Are you going to show all the docs (even duplicates)
some of the time?
Best
Erick
On Tue, Jun 21, 2011 at 1:53 PM, Pooja Verlani wrote:
> Hi Eric,
>
> Thanks for the quick reply.
> I had looke
Right. I'm saying that you can store one or the other, but there
is no good reason to store both. The facet values are the
values retrieved from the index, not the stored values. So you
can pull the stored values from either the searchable author field
just fine
Best
Erick
On Tue, Jun 21, 201
I am fine to remove the duplicates and not show them up for this use case.
But grouping can also help me show one representative from the group.
At present I am using solr 1.4. Any idea how to achieve it otherwise if not
by using solr 3.3.
Regards,
Pooja
On Tue, Jun 21, 2011 at 11:55 PM, Erick Er
: I'm working with Solrj, and I like to use the SolrResponseBase.toString()
: method, as it seems to return JSON. However, the JSON returned is not
many of the toString methods on internal solr objects use {} to show
encapsulation when recursively calling toString() on sub objects, but they
ar
Hi,
Currently, only the clustering of search results is implemented in Solr,
clustering of the whole index is not possible out of the box. In other
words, clustering applies only to the records you fetch during searching.
For example, if you set rows=10, only the 10 returned documents will be
clus
I have a field with a RecordSeparator in it, how can i go about searching on
this field using solrj and solr?
> I have a field with a RecordSeparator
> in it, how can i go about searching on
> this field using solrj and solr?
What do you mean by RecordSeparator?
ASCII RecordSeparator http://www.bbdsoft.com/ascii.html
(char)30 will create it in Java
On Tue, Jun 21, 2011 at 4:41 PM, Ahmet Arslan wrote:
> > I have a field with a RecordSeparator
> > in it, how can i go about searching on
> > this field using solrj and solr?
>
> What do you mean by RecordS
I guess you mean from the /browse view?
You can override/replace hit.vm (in conf/velocity/hit.vm) with whatever you
like. Here's an example from a demo I recently did using the open Best Buy
data where I mapped their url value for a product into a url_s field in Solr
and rendered a link to it:
On Tue, Jun 21, 2011 at 10:41 PM, sabman wrote:
> There is information
> http://wiki.apache.org/solr/DataImportHandler#Scheduling here about
> Scheduling but I don't understand how to use them. I am not a Java developer
> so maybe I am missing something obvious.
[...]
Depending on what operating
Thanks. Using curl would be an option but ideally I want to implement it
using this scheduler. I want to add Solr as part of another application
package and send it to clients. So rather than asking them run a cron job it
would be easier to have Solr configured to run the scheduler.
--
View this m
: After switching to solr 3.2 and building a new index from scratch I ran
: check_index which reports:
: Segments file=segments_or numSegments=1 version=FORMAT_3_1 [Lucene 3.1]
:
: Why do I get FORMAT_3_1 and Lucene 3.1, anything wrong with my index?
That's just because the index format didn't c
: This is for debugging purposes, so I am sending the exact same data that are
: already stored in Solr's index.
...
: ERROR: [288400] multiple values encountered for non multiValued field
: "field2" [fieldvalue, fieldvalue]
:
: The scenario:
: - "field1" is implicitly single value, type
: It seems like the problem is when different fields in the 'qf' produce a
: different number of tokens for a given query. dismax needs to know the number
: of tokens in the input in order to calculate 'mm', when 'mm' is expressed as a
: percentage, or when different mm's are given for different
:
http://localhost:8983/solr/select?q=life&qf=description_text&defType=dismax&sort=scores:rails_f+desc
...
: If I insert the same document into solr 3.1 and run the same query I get the
: error:
:
: Problem accessing /solr/select. Reason:
:
: undefined field scores
:
: For some rea
: I am not sure if I can use function queries this way. I have a query
: like this"attributeX:[* TO ?]" in my DB. I replace the ? with input from
: the front end. Obviously, this works fine. However, what I really want
: to do is "attributeX:[* TO (3 * ?)]" Is there anyway to embed the
: resul
Thanks, that's helpful.
It still seems like current behavior does the "wrong" thing in _many_ cases (I
know a lot of people get tripped up by it, sometimes on this list) -- but I
understand your cases where it does the right thing, and where what I'm
suggesting would be the wrong thing.
> Ul
I have a field: specialties that is multiValued.
It indicates the doctor's specialties: cardiologist, internist, etc.
When someone does a search: "Cardiologist", I use
q=cardiologist&defType=dismax&qf=specialties&facet=true&facet.field=specialt
ies
What I want to come out in the facet is the Car
: not other) setups/intentions. It's counter-intuitive to me that adding
: a field to the 'qf' set results in _fewer_ hits than the same 'qf' set
agreed .. but that's where looking the debug info comes in to understand
the reason for that behavior is that your old qf treated part of your
inp
So are you saying that for all results for "cardiologist",
you don't want facets not matching "Cardiologist" to be
returned as facets?
what happens when you make q=specialities:Cardiologist?
instead of just q=Cardiologist?
Seems that if you make the query on the field, then all
your results will
Doing it with q=specialities:Cardiologist or
q=Cardiologist&defType=dismax&qf=specialties
does not matter, the issue is how I see facets. I want the facets to only
show the one match,
and not all the multiValued fields in specialties that match...
Example,
Name|specialties
Bell|Cardiologist
Smith
Hi
thanks Alot,
can you please help me how can i implement the filter of topic cluster
like Model(10) when i will click on model then i need to get 10 docs .
Regards
Nilay Tiwari
On Wed, Jun 22, 2011 at 1:14 AM, Stanislaw Osinski-4 [via Lucene] <
ml-node+3092594-1426669115-405...@n3.nabble.c
First commit and then try again to search.
You can also use lucene's CheckIndex tool to check & fix your index (it may
remove some corrupt segments in your index)
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Read-past-EOF-error-due-to-broken-connection-tp309
*$("#submit").click(function(){
var query=getquerystring() ; //get the query string entered by user
// get the JSON response from solr server
var newquery=query;
$.getJSON("http://192.168.1.9:8983/solr/db/select/?wt=json&&start=0&rows=100&q="+query+"&json.wrf=?";,
func
Thanks Alot . I was thinking i am not doing in correct way .
-
Regards
Nilay Tiwari
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-Clustering-For-Multiple-Pages-tp3085507p3094379.html
Sent from the Solr - User mailing list archive at Nabble.com.
Can you please tell me how can i apply filter in cluster data in Solr ?
Currently i storing docid and topic name in Map and get the ids by topic
from Map and then pass into solr separating by OR condition
Is there any other way to do this
-
Regards
Nilay Tiwari
--
View this message
68 matches
Mail list logo