Its 7,200,000 hits == number of documents found by all 10K queries.
We have RHEL tikanga version.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-Index-linear-growth-Performance-degradation-tp4000934p4001069.html
Sent from the Solr - User mailing list archive at Nabble.
Its 7.2Gig Hits. (GB was typo)
This is the total number of index hits - calculated by summing each
"numFound" attribute from solr query response.
We have RHEL Tikanga version.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-Index-linear-growth-Performance-degradation-tp
index hits == total number of documents found by search query.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-Index-linear-growth-Performance-degradation-tp4000934p4001063.html
Sent from the Solr - User mailing list archive at Nabble.com.
It looks like reducing the jvm heap allocation did help in lowering the
response time to some extent.
Thanks for the pointer.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-Index-linear-growth-Performance-degradation-tp4000934p4001056.html
Sent from the Solr - User mai
How many documents does each search find? What does this mean: "number
of index hits: 7.2GB."
Above a threshold, the more memory you give Java, the more time it
spends collecting. You want to start with very little memory and
gradually increase memory size until the program stops using it all,
and
Hi,
I have a beefy box with 24Gb RAM (12GB for Tomcat7 which houses SOLR3.6) 2
Processors Intel Xeon 64 bit Server, 30TB HDD. JDK 1.7.0_03 x64 bit
Data Index Dir Size: 400GB
Metadata of files is stored in it. I have around 15 schema fields.
Total number of items:150million approx.
I have a sce
HI All,
I have thousands file on some folder which i want to index using solr.
Firstly my file only 9 until 20 file, so i upload them manually into solr
using curl.
But Now, my file is thousands file, how i can index it using solr ? should i
upload them one by one ?
i've tried using curl command
There is a per segment faceting option - but I think just for single value
fields right now?
On Mon, Aug 13, 2012 at 2:38 PM, Fuad Efendi wrote:
> SOLR-4.0
>
> I am trying to implement this; funny idea to share:
>
> 1. http://wiki.apache.org/solr/HierarchicalFaceting
> unfortunately it does not
: We're passing multiple Arbitrary Faceting Query (facet.query) to get the
: number of matching documents (the facet count) evaluated over the search
: results in a *single* Solr query. My use case demands the actual matching
: facet results/documents/fields also along with facet count.
:
: My q
1. So we have 24.5GB assigned to jvm which is half of the total memory, which
is 48GB RAM.(If that's what you meant, and if i am getting that right ?)
2. Size of *.fdt and *fdx is around 300m and 50m respectively.So that's
definitely less that 5%.
Do you see a problem there ?
Is there a way that w
Instant reactions:
1> that's probably too much memory. Try, as Lance said, 1/2 of your
memory. Uwe Schindler wrote an excellent blog about this issue as it
relates to MMapDirectory
http://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html
2> You've doubled the number of docs on the
Rahul Warawdekar wrote
>
> Hi Divakar,
>
> Try making your updateRequestProcessorChain as default. Simply add
> default="true" as follows and check if that works.
>
>
>
>
Rahul,
This fixed my problem, you saved my week!
I was following the README.txt instructions and they didn't work, aft
Here's few list of queries
---
parallel zur xml beschreibungsdatei gibt es eine
die verbindung zwischen beiden sei ten geschieht
die owner klasse muss sich aus der
benutzer ein oder mehrere lieblingsfarben ausw hlen kann
found sample questions at http bjs ojp
Hi,
I saw some old posts on the Solr vs. ES topic, but they were about
performance/benchmarks only, and even those were not done correctly.
We have a couple of posts on that topic pending over on Sematext Blog. We can
publish them next week, so keep an eye on http://blog.sematext.com/ and/or
> Now I want to split things up. Is there a way that I can
> only get
> suggestions of a query without getting the normal search
> results? I may need
> to create a new handler for this. Can anyone please give me
> some ideas on
> that?
Appending &query=false disables QueryComponent. I am not sure
I haven't tried it but i'd try to use spellcheck.q as input and specifiy the
spellcheck component in the components section, not the last-components section
because components because it has (iirc) the five default components, query,
debug, mlt, highlighter and facet.
-Original message--
Does querying with rows=0 work?
Michael Della Bitta
Appinions | 18 East 41st St., Suite 1806 | New York, NY 10017
www.appinions.com
Where Influence Isn’t a Game
On Mon, Aug 13, 2012 at 4:21 PM, Bing Hua wrote:
>
>
> Now I want to split things up
Hi,
I'm having a spell check component that does auto-complete suggestions. It
is part of "last-components" of my /select search handler. So apart from
normal search results I also get a list of suggestions.
Now I want to split things up. Is there a way that I can only get
suggestions of a query
How much ram do you assign to the JVM? The JVM should be allocated
maybe 1/2 gb more than it needs to run "comfortably". Also, how large
are your caches?
How large are the documents? How many search terms are there? If you
add more documents are there new search terms?
On Mon, Aug 13, 2012 at 11:
We have 4 shards with 14GB index on each of them
Each shard has a master and 3 slaves(each of them with 32GB RAM)
We're expecting that the index size will grow to double or triple in near
future.
So we thought of merging our indexes to 28GB index so that each shard has
28GB index and also increase
SOLR-4.0
I am trying to implement this; funny idea to share:
1. http://wiki.apache.org/solr/HierarchicalFaceting
unfortunately it does not support date ranges. However, workaround: use
"String" type instead of "*_tdt" and define fields such as
published_hour
published_day
published_week
Of cou
On Sun, Aug 12, 2012 at 12:31 PM, Alexey Serba wrote:
> > It would be vastly preferable if Solr could just exit when it gets a
> memory
> > error, because we have it running under daemontools, and that would cause
> > an automatic restart.
> -XX:OnOutOfMemoryError="; "
> Run user-defined commands
Sujatha,
As the API of the classes you're compiling against may have changed
with a different Solr version, it's always a good idea to build
against the new version of Solr, otherwise you might see weird issues
at runtime.
You wouldn't have to do anything special other than to drop your src
file
What I would be doing is this ..
Create a custom class that refer to all org,apache.* classes (import stt)
,the custom file's location is independent of the solr core class files.
compile this separately
package this as a jar
move this to lib dir of each solr core
refer to this in lib directory
Hi,
I am trying to figure out if this is the expected behaviour of spell check
component. (when using collate=true)
I am searching for keyword 'high tet', the suggester returns expected result
'test' but I expected the collated results to return 'high test' (corrected
word returned by suggester)
No, the jar would be exactly the same, with the caveat that you'd have
to build against the newer Solr version of course.
Michael Della Bitta
Appinions | 18 East 41st St., Suite 1806 | New York, NY 10017
www.appinions.com
Where Influence Isn’t a Ga
Thanks ,I am going to try this on solr 1.3 version .Would the approach be
any different for the recent sorl versions?
Regards
Sujatha
On Mon, Aug 13, 2012 at 8:53 PM, Michael Della Bitta <
michael.della.bi...@appinions.com> wrote:
> Then you're on the right track.
>
> 1. You'd either have to res
Hi,
I am trying to configure spell check component in SOLR. I just want to
confirm if I am on right path
I have a text field - field name: name_spell (with no analyzers, uses
solr.TextField field type). This field will be used for building terms for
spell check. I have copied necessary data (
Hello,
Has anybody compared feature set of SOLR 4 with Elastic Search? I saw
some earlier comparisons and they talked about sharding and
distributed service, etc. Seems to me, most of those are addressed in
version 4.
The only big issue I see is a better support from ES for nested items
and/or pa
: We would like to use multiple jvm's to host solr cores but can not
: because the queries ignore the jetty.port settings. The following is
: they query generated using the admin interface, solr is running in jetty
: under port 8080.
:
: http://solr-cluster-1.issuu.com:8983/solr/core0/select?q
Did you try this:
http://lucene.472066.n3.nabble.com/How-to-change-a-port-td490375.html
-- Jack Krupansky
-Original Message-
From: Raghav Karol
Sent: Monday, August 13, 2012 11:49 AM
To: solr-user@lucene.apache.org
Subject: Solr 4.0.0, query, default port not changeable
Hello *,
Runn
Hello *,
Running Solr 4.0.0-ALHPA we have an issue with queries.
We would like to use multiple jvm's to host solr cores but can not because the
queries ignore the jetty.port settings. The following is they query generated
using the admin interface, solr is running in jetty under port 8080.
ht
Got it at
https://svn.apache.org/repos/asf/lucene/dev/trunk/solr/solrj/src/test/org/apache/solr/client/solrj/SolrExampleTests.java
Problem solved.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr4-0-Partially-update-document-tp4000875p4000878.html
Sent from the Solr -
--- On Mon, 8/13/12, Videnova, Svetlana wrote:
> From: Videnova, Svetlana
> Subject: RE: multi-searching problem
> Thank you for your answer, finally it was only my bad
> between copyfield and copyField. Now all good.
> I don't know how copyField and edismax working exactly, but
> can I do bo
Hi,
Several days ago I came across some solrj test code on partially updating
document field values. Sadly I forgot where that was. In Solr 4.0, "/update"
is able to take in document id and fields as hashmaps like
"id": "doc1"
"field1": {"set":"new_value"}
Just trying to figure out what's the so
Then you're on the right track.
1. You'd either have to restart Tomcat or in the case of Multicore
setups, reload the core.
2. If the jar has dependencies outside of the Solr provided classes,
you'll have to include those as well. If it only depends on Solr stuff
or things that are in the servlet
The wiki page show how to use the -H option of curl to set the Content-Type.
See:
http://wiki.apache.org/solr/ExtractingRequestHandler
SolrJ requires some extra coding.
-- Jack Krupansky
-Original Message-
From: 122jxgcn
Sent: Monday, August 13, 2012 5:49 AM
To: solr-user@lucene.apac
Hi,
I'm using Solr 4.0.0-ALPHA and the EmbeddedSolrServer.
Within my SolrJ application, the documents are added to the server
using the commitWithin parameter (in my case 60s). After 1 day my 125
millions document are all added to the server and I can see 89G of
index data files. I stop my SolrJ
Hi,
We are using Apache Solr 1.4.1 since last year and we are very happy about
it.
We are now looking into the upgrade to Solr 3.6.1 but we have stumbled
against a critical (for us) issue for which a workaround seems to be the use
of
LUCENE_33
in the Solr configuration.
The issu
Finally i found it:
In the dataconfig my sql request wasn't good.
-Message d'origine-
De : Videnova, Svetlana [mailto:svetlana.viden...@logica.com]
Envoyé : lundi 13 août 2012 10:41
À : solr-user@lucene.apache.org
Objet : RE: solr indexing problem
Some ideas?
-Message d'origine
Adding a new class
Regards
Sujatha
On Mon, Aug 13, 2012 at 5:54 PM, Michael Della Bitta <
michael.della.bi...@appinions.com> wrote:
> Michael Della Bitta
> Hi Sujatha,
>
> Are you adding a new class, or modifying one of the provided Solr classes?
>
> Michael
>
>
> ---
Hello,
I'd like to set Content-Type of the file while I'm using
ExtractRequestHandler to pass file to Tika.
As I'm indexing custom file type, it seems that Tika is not matching my file
to the right custom parser.
So I really need to explicitly declare Content-Type of my custom file so
that it cann
Hi Arslan,
Thank you for your answer, finally it was only my bad between copyfield and
copyField. Now all good.
I don't know how copyField and edismax working exactly, but can I do both?
Currently I copyed all fields in "all"
all.
So can I use edismax as well in the solrconfig.xml side?
Thank
Ah, okay, I see the usage now. In SQL the right operand of LIKE can be
either a literal wildcard pattern or an expression which is evaluated
per-row during the query. Solr/Lucene has the former, but not the latter.
The wildcard pattern will be fixed at the start of the search.
-- Jack Krupansk
Michael Della Bitta
Hi Sujatha,
Are you adding a new class, or modifying one of the provided Solr classes?
Michael
Appinions | 18 East 41st St., Suite 1806 | New York, NY 10017
www.appinions.com
Where Influence Isn’t a Game
On Mon, Aug 13, 2012
> stored="true" multiValued="true"/>
> dest="all"/>
> dest="all"/>
>
>
> article_id
> all
It is always good idea to edit example schema.xml according to your needs.
See copyField declarations.
http://svn.apache.org/repos/asf/lucene/dev/trunk/solr/example/solr/collection1/conf/sch
>
>
> ...
> type="text" indexed="true" stored="true"
> multiValued="true"/>
> &defType = edismax &qf =
> article_id article_nom
>
>
> article_id
>
>
>
>
>
>
> But i have this error:
>
> ###
> org.xml.sax.SAXParseException: The reference to entity
> "defType" must end with the
Hi ,
I would like to write a custom component for solr to address a particular
issue.
This is what I have been doing ,write the custom code directly in the
downloaded code base and rebuild the war file and deploy the same. We
currently have multiple cores ,hence I want to approach this in a cor
Hi,
I want to create a very simple geocoder for returning co-ordinates of a
place if a user enters in a town or city. There seems to be very little
information about doing it the way I suggest, so I hope I am on a good path.
My first decision was to divide SOLR into two cores, since I am already
Some ideas?
-Message d'origine-
De : Videnova, Svetlana [mailto:svetlana.viden...@logica.com]
Envoyé : vendredi 10 août 2012 11:05
À : solr-user@lucene.apache.org
Objet : RE: solr indexing problem
This is schema.xml
###
I follow this exemple
https://github.com/boonious/misc/blob/master/fedora-solr-integration-conf/schema.xml
but still no results
-Message d'origine-
De : Videnova, Svetlana [mailto:svetlana.viden...@logica.com]
Envoyé : lundi 13 août 2012 08:59
À : solr-user@lucene.apache.org
Objet : RE
Here is what we do in SQL:
mysql> select * from _tbl;
+++
| id | field |
+++
| 1 | plain text |
| 2 | wil_c% |
+++
2 rows in set (0.14 sec)
mysql> SELECT * FROM _TBL WHERE 'wildcard' LIKE FIELD;
+++
| id | field |
++-
On Fri, Aug 10, 2012 at 12:38:46PM -0400, Jack Krupansky wrote:
> "Doc1 has the pattern "AB%CD%" associated with it (somehow?!)."
>
> You need to clarify what you mean by that.
I'm not the OP, but I think (s)he means the patterns are in the
database and the string to match is given in the query.
I always tried this in my schema.xml:
###
article_id
all
###
I have no errors with that code but when I search for a term who is present in
article_nom this give me 0 results ... I don't know why and where im doing
wrong :s
Thank you for your help
-Message d'or
54 matches
Mail list logo