Hi,
Many time while starting solr I see the below message and then the solr
is not reachable.
|debraj@boutique3:~/solr5$ sudo bin/solr start -p 8789 Waiting to see
Solr listening on port 8789 [-] Still not seeing Solr listening on 8789
after 30 seconds!|
However when I try to start solr ag
You can add the debug flag to the end of the request and see exactly what
the scoring is and why things are happening.
&debug=ALL will show you everything including the scoring.
Showing the result of the debug query should help you, or adding that into
your question here, decipher what is going o
member?"],
> "strtitle":"healthcare description",
> "id":"
> http://localhost:4503/content/uhcdotcom/en/home/waysin/poc/upendra-custon/healthcare-description.html
> ",
> "tstamp":"2016-05-09T17:15:5
product of:
> 0.074107975 = idf(docFreq=6, docCount=6)
> 1.853831 = tfNorm, computed from:
> 4.0 = termFreq=4.0
> 1.2 = parameter k1
> 0.75 = parameter b
> 168.3 = avgFieldLength
> 83.591835 = fieldLength
> ',
>
You can use a combination of ngram or edgengram fields and possibly the
shingle factory if you want to combine words. Also might want to have it as
exact text with no query sloop if the two words, even the partial text,
need to be right next to each other. Edge is great for left to right ngram
is g
cept is the same with regular old ngram:
2001 N Drive Derek Fullerton
2
[32]
0
1
1
word
1
20
[32 30]
0
2
1
word
1
200
[32 30 30]
0
3
1
word
1
2001
[32 30 30 31]
0
4
1
word
1
n
[6e]
5
6
1
word
2
d
[64]
7
8
1
word
3
dr
[64 72]
7
9
1
word
3
dri
[64 72 69]
7
10
1
word
3
driv
[64 72 69 76]
7
11
1
word
3
d
There are only two ways I can think of to accomplish this and neither of
them are dynamically setting the suggester field as is looks according to
the Doc (which does sometimes have lacking info so I might be wrong) you
cannot set something like *suggest.fl=combo_box_field* at query time. But
maybe
https://wiki.apache.org/solr/MoreLikeThisHandler
Bottom of the page, using context streams. I believe this still works in
newer versions of Solr. Although I have not tested it on a new version of
Solr.
But if you plan on indexing the document anyways then just indexing and
then passing the ID to
Although you did mention that you wont need to sort and you are using
mutlivalued=true. On the off chance you do change something like
multivalued=false docValues=false then this will come in to play:
https://issues.apache.org/jira/browse/SOLR-7495
This has been a rather large pain to deal with i
t; >
> > Best,
> > Erick
> >
> > On Fri, May 27, 2016 at 6:25 AM, Steven White
> > wrote:
> > > Thank you Erick for pointing out about DocValues. I re-indexed my data
> > > with it set to true and my index size grew by 20%. Is this expecte
Hi Anshum,
I'm not having any issue connecting.
-Aravind
--
View this message in context:
http://lucene.472066.n3.nabble.com/Trouble-connecting-to-IRC-tp4343512p4343515.html
Sent from the Solr - User mailing list archive at Nabble.com.
I want to weighted average the Lucene score with an additional score i have,
i.e. (W1 * Lucene score + W2 * Other score) / (W1 + W2) .
What is the easiest way to do this?
Also, is the Lucene score normalized.
Thanks,
t; Otis
> --
> Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
>
>
>
> - Original Message
> > From: s d <[EMAIL PROTECTED]>
> > To: solr-user@lucene.apache.org
> > Sent: Tuesday, August 26, 2008 1:55:38 PM
> > Subject: Weighting the Lice
Hi,Is there a recommended index size (on disk, number of documents) for when
to start partitioning it to ensure good response time?
Thanks,
S
Hello, I have solr 1.4 deployed in websphere 6.1. Im trying to add a url
based security constraint to my project but if I specify the core name in
the constraint the path to the admin of each core give a 404 error. Does
anyone have any experience of this or suggestions of how I can work around
it?
We have a large index, separated
into multiple shards, that consists of records exported from a database. One
requirement is to support near real-time
synchronization with the database. To accomplish this we are considering
creating
a "daily" shard where create and update documents
(records ne
There seems to be some code out for Tika now (not packaged/announced yet,
but...). Could someone please take a look at it and see if that could fit
in? I am eagerly waiting for a reply back from tika-dev, but no luck yet.
http://svn.apache.org/repos/asf/incubator/tika/trunk/src/main/java/org/apach
va programmer so if you have questions about the internals
> of the code, definitely direct those to Eric as I cannot help. I have
> only implemented it in web applications. If you have any other
> questions about the use of the patch I can answer those questions.
>
> Enjoy!
>
>
(create extra
elements, create '' for pdf files and '' for html files, etc..),
etc... This strips out having to deal with if/else scripting outside of
Solr.
Rao
On 8/21/07, Vish D. <[EMAIL PROTECTED]> wrote:
> > Pete,
> >
> > Thanks for the gr
On 8/21/07, Vish D. <[EMAIL PROTECTED]> wrote:
>
> On 8/21/07, Peter Manis <[EMAIL PROTECTED]> wrote:
> >
> > I am a little confused how you have things setup, so these meta data
> > files contain certain information and there may or may not be a pdf,
>
How can I retrieve the "analyzed tokens" (e.g. the stemmed values) of a
specific field?
Is there a way to retrieve the "analyzed tokens" (e.g. the stemmed
values) of a field from the SOLR index ?
Almost like using SOLR as a utility for generating the tokens.
Thanks !
Is there a way to import a Lucene index (as is) into SOLR? Basically, I'm
looking to enjoy the "web context" and caching provided by SOLR but keep the
index under my control in Lucene.
Is there a way to use RAMDirectory with SOLR?If you can point me to
documentation that would be great.
Thanks,
S
Is there a simpler way to write this query (I'm using the standard handler)
?
field1:t1 field1:t2 field1:"t1 t2" field2:t1 field2:t2 field2:"t1 t2"
Thanks,
but i want to sum the scores and not use max, can i still do it with the
DisMax? am i missing anything ?
On Jan 4, 2008 2:32 AM, Erik Hatcher <[EMAIL PROTECTED]> wrote:
>
> On Jan 4, 2008, at 4:40 AM, s d wrote:
> > Is there a simpler way to write this query (I'm using t
What is the best approach to tune queryResultCache ?For example the default
size is: size="512" but since a document id is just an int (it is an int,
right?) ,i.e 4 bytes why not set size to 10,000,000 for example (it's only
~38Mb).
I sense there is something that I'm missing here :). any help wou
How do i boost a field (not a term) using the standard handler syntax? I
know i can do that with the DisMax but I'm trying to keep myself in the
standard one.Can this be done ?
Thanks,
n 6, 2008 12:59 AM, s d <[EMAIL PROTECTED]> wrote:
> > What is the best approach to tune queryResultCache ?For example the
> default
> > size is: size="512" but since a document id is just an int (it is an
> int,
> > right?) ,i.e 4 bytes why not set size
Got it. Smart.
Thx
On 1/6/08, Chris Hostetter <[EMAIL PROTECTED]> wrote:
>
> : number than the default one and i was wondering is there any disadvantage
> in
> : having a big number/ cache?BTW, where is the TTL controlled ?
>
> no disadvantage as long as you've got the RAM ... NOTE: the magic "512
e.g. if the index is field1 and field2 and documents of type (A) always have
information for field1 AND information for field2 while document of type (B)
always have information for field1 but NEVER information for field2.
The problem is that the formula will sum field1 and field2 hence skewing in
ote:
>
> On 7-Jan-08, at 9:02 PM, s d wrote:
>
> > e.g. if the index is field1 and field2 and documents of type (A)
> > always have
> > information for field1 AND information for field2 while document of
> > type (B)
> > always have information for field1 b
Adding a FunctionQuery made the query response time slower by ~300ms, adding
a 2ndFunctionQuery added another ~300ms so overall i got over 0.5sec for a
response time (slow).Is this expected or am i doing something wrong ?
Thx
Is there a way or a point in filtering all results bellow a certain score?
e.g. exclude all results bellow score Y.Thanks
Got it (
http://wiki.apache.org/solr/DisMaxRequestHandler#head-cfa8058622bce1baaf98607b197dc906a7f09590)
.
thx !
On Jan 8, 2008 12:11 AM, Chris Hostetter < [EMAIL PROTECTED]> wrote:
>
> : Isn't there a better way to take the information into account but still
> : normalize? taking the score of on
User Query: x1 x2
Desired query (Lucene): field:x1 x2 field:"x1 x2"~a^b
In the standard handler the only way i saw how to make this work was:
field:x1 field:x2 field:"x1 x2"!a^b
Now that i want to try the DisMax is there a way to implement this without
having duplicate fields? i.e. since the fiel
I may be mistaken, but this is not equivalent to my query.In my query i have
matches for x1, matches for x2 without slope and/or boosting and then match
to "x1 x2" (exact match) with slope (~) a and boost (b) in order to have
results with exact match score better.
The total score is the sum of all
Hi,I use SOLR with standard handler and when i send the same exact query to
solr i get different results every time (i.e. refresh the page with the
query and get different results).
Any ideas?
Thx,
We have an index of documents from different sources and we want to make
sure the results we display are interleaved from the different sources and
not only ranked based on relevancy.Is there a way to do this ?
Thanks,
S.
Hi,I have a query similar to: x OR y OR z and i want to know if there is a
way to make sure i get 1 result with x, 1 result with y and one with z ?
Alternatively, is it possible to achieve through facets?
Thanks,
S.
Can i use RAMDirectory in SOLR?Thanks,
S
Mike,
I am currently evaluating different search engine technologies (esp., open
source ones), and this is very interesting to me, for the following reasons:
Our data is much like yours in that we have different types of data
(abstracts, fulltext, music, etc...), which eventually fall under diffe
Yonik/Chris,
Do we have a eta on " Allow multiple independent Solr *webapps* in the same
app server"?
After reading up, silently, on the many emails on this topic, I agree with
you that it would be worthwhile to test out the current implementation and
see how it performs. But, it makes sense to r
Hi all,
I am trying to figure out how I can have some type of faceted browsing
working. I am also in need of a way to get a list of unique field values
within a query's results set (for filtering, etc...). When I say trying, I
mean having it up and running without much coding, b/c of time reasons
Thank you Chris and Erik. That makes it a bit clearer, but I might need to
sit down and look at the code (nines + DisMax...) a bit closer to see how it
all works in Solr.
Erik, when do you plan on having your implementation refactored with "good"
use of code? Or, in general, when is Solr planning
Thanks!
On 6/24/06, Erik Hatcher <[EMAIL PROTECTED]> wrote:
On Jun 24, 2006, at 12:38 PM, Vish D. wrote:
> Erik, when do you plan on having your implementation refactored
> with "good"
> use of code?
This weekend :) I have imported more data than my hacked
implementat
Erik,
Any update on your progress? Eager to get my hands on on your latest code...
:=)
Thanks!
On 6/28/06, Chris Hostetter <[EMAIL PROTECTED]> wrote:
: > well, the most obvious solution i can think of would be a patch adding
an
: > invert() method to DocSet, HashDocSet and BitDocSet. :)
:
Are there any plans on implementing a MultiSearcher into solr?
I have been following the list for a while, and read quite a few topic on
multiple instances of solr, in order to accomodate multiple schemas as well
as break down index sizes for performance reasons. I have a use case that
sits right
I just noticed that link on the first reply from Yonik about
FederatedSearch. I see that a lot of thought went in to it. I guess the
question to ask would be, any progress on it, Yonik? :)
On 9/27/06, Vish D. <[EMAIL PROTECTED]> wrote:
Are there any plans on implementing a MultiSearche
Anyone have experience working with LIUS (
http://sourceforge.net/projects/lius/)? I can't seem to find any real
documentation on it, even though it seems 'active' @ sourceforge. I need a
way to index various types of fulltext, and LIUS seems very promising at
first glance. What do you guys think?
Sounds interesting. I can't seem to find any clear dates on the project
website. Do you know? ...V1 shipping date?
Thanks!
On 6/12/07, Bertrand Delacretaz <[EMAIL PROTECTED]> wrote:
On 6/12/07, Yonik Seeley <[EMAIL PROTECTED]> wrote:
>... I think Tika will be the way forward (some of the code
Wonder if TOM could be useful to integrate?
http://tom.library.upenn.edu/convert/sofar.html
On 6/12/07, Bertrand Delacretaz <[EMAIL PROTECTED]> wrote:
On 6/12/07, Vish D. <[EMAIL PROTECTED]> wrote:
> ...Sounds interesting. I can't seem to find any clear dates on the
proje
I have run into an issue with multi-word synonyms and a min-should-match
(MM) of anything other than `0`, *Solr version 6.6.0*.
Here is my example query, first with mm set to zero and the second with a
non-zero value:
With MM set to 0
select?fl=*&indent=on&wt=json&debug=ALL&q=EIB&qf=ngs_title%20n
Thanks Steve,
I'll test out that version.
Nick
On Feb 6, 2018 6:23 AM, "Steve Rowe" wrote:
> Hi Nick,
>
> I think this was fixed by https://issues.apache.org/
> jira/browse/LUCENE-7878 in Solr 6.6.1.
>
> --
> Steve
> www.lucidworks.com
>
>
To whom it may concern:
I have a Windows based system running Java 8. I have installed SOLR 7.7.2 ( I
also tried this with version 8.1.1 as well with the same results ). I have SQL
Server 2018 with 1 table that contains 22+ columns and a few thousand rows. I
am attempting to index the SQL Se
Hello All,
We recently upgraded from Solr 6.6 to Solr 7.7.2 and recently had spikes in
memory that eventually caused either an OOM or almost 100% utilization of
the available memory. After trying a few things, increasing the JVM heap,
making sure docValues were set for all Sort, facet fields (thou
to be effectively ignored (see SOLR-12243)
> 3. upgrade to 8.0, which will restore the failsafe maxBooleanClauses,
> avoiding OOM but returning an error code for affected queries (which
> in your case sounds like most queries?) (see SOLR-13336)
>
> Michael
>
> On Tue, Dec 17, 20
Hi all,
I have a question regarding field collapsing.
Has anyone successfully used SOLR to roll up more information than just
"numFound" in the response to a group query?
For some background, I'm trying to see if any artifact in the group has been
marked with a certain priority. Ideally I can
I've put the question nicely formatted on StackOverflow here
http://stackoverflow.com/questions/8360257/solr-lucene-why-is-this-or-query-failing-when-the-two-individual-queries-suc
Here is that question Verbatim:
I have a Solr document schema with with a solr.TrieDateField and noticed
this boolea
Hi,
Lakh or Lac - 100,000
Crore - 100,00,000 (ten million)
Commonly used in India
Sincerely,
Sithu D Sudarsan
-Original Message-
From: Walter Underwood [mailto:wun...@wunderwood.org]
Sent: Monday, November 16, 2009 5:22 PM
To: solr-user@lucene.apache.org
Subject: Re: Solr
I'll recommend setting -Xms and -Xmx to the same value.
Sincerely,
Sithu D Sudarsan
-Original Message-
From: Jason Rutherglen [mailto:jason.rutherg...@gmail.com]
Sent: Tuesday, December 22, 2009 2:31 PM
To: solr-user@lucene.apache.org
Subject: Re: java.lang.OutOfMemoryError
available which shows using tomcat as the web
engine rather than jetty?
Your help is appreciated,
Sincerely,
Sithu D Sudarsan
ORISE Fellow, DESE/OSEL/CDRH
WO62 - 3209
&
GRA, UALR
[EMAIL PROTECTED]
[EMAIL PROTECTED]
Hi All:
Is there any study / research done on using scanned paper documents as
images (may be PDF), and then use some OCR or other technique for
extracting text, and the resultant index quality?
Thanks in advance,
Sithu D Sudarsan
sithu.sudar...@fda.hhs.gov
sdsudar...@ualr.edu
Thanks Hannes,
The tool looks good.
Sincerely,
Sithu D Sudarsan
sithu.sudar...@fda.hhs.gov
sdsudar...@ualr.edu
-Original Message-
From: hannesc...@googlemail.com [mailto:hannesc...@googlemail.com] On
Behalf Of Hannes Carl Meyer
Sent: Thursday, February 26, 2009 11:35 AM
To: solr-user
ject called ocropus done by the DFKI, check the online
demo
> > here: http://demo.iupr.org/cgi-bin/main.cgi
> >
> > And also http://sites.google.com/site/ocropus/
> >
> > Regards
> >
> > Hannes
> >
> > m...@hcmeyer.com
> > http://mimblog.de
>
Hi All,
We have solr 6.4.2 currently deployed as war in tomcat. It doesn't have
authentication now. I want to setup the authentication for solr. When it
deployed as war in tomcat, The process specified in below page is not working,
Even after adding security.json in solr home directory, curl re
calculations already done and just retrieving
them from the Database?
Regards,
Joe
Joseph Costello
Chief Information Officer
F&D Reports | Creditntell | ARMS
===
Information Clearinghouse Inc. & Market Service Inc.
310 East Shore Road, Great Neck, NY 11023
ema
calculations already done and just retrieving
them from the Database? If it could I would not need to run a nightly process
that pre calculates the distances. Thougths?
Regards,
Joe
Joseph Costello
Chief Information Officer
F&D Reports | Creditntell |
that pulling
pre-calculate data values directly would be faster. Just looking for
confirmation before I go down a rabbit hole seeing if Solr can be faster.
Thoughts?
#SaveMeFromTheRabbitHole. Thx in advance
Regards,
Joe
Joseph Costello
Chief Information Officer
F&D Reports | Creditn
If the process pre-calculates without issue the speed at which the SQL query
returns data is super-fast and not the issue at all. Thx again and I look
forward to any addition replies and/or feedback.
Regards,
Joe
Joseph Costello
Chief Information Officer
F&D Reports | Creditnte
produce the best
performance results. Correct?
Regards,
Joe
Joseph Costello
Chief Information Officer
F&D Reports | Creditntell | ARMS
===
Information Clearinghouse Inc. & Market Service Inc.
310 East Shore Road, Great Neck, NY 11023
email: jose...@fdreports.c
71 matches
Mail list logo