Hi All,
I'm having a problem while searching for some string with a word defined in
stopwords.txt.
eg: I have 'of' defined in stopwords.txt
My schema analyzer's defined as follows:
I have defined the filed
Thanks Shawn and Erick.
I followed SOLR configuration document and modified index strategy.
Looks good now. I haven't seen any problems in last 1 week.
Thanks for your suggestions.
--
View this message in context:
http://lucene.472066.n3.nabble.com/SOLR-4-4-Slave-always-replicates-full-index
: Yes, the memory and cpu spiked for that machine. Another issue I found in
: the log was "SolrException: Too many values for UnInvertedField faceting on
: field".
: I was using the fc method. Will changing the method/params help?
the fc/fcs faceting methods really aren't going to work well with
(cross posted, please keep any replies to general@lucene)
ApacheCon Denver is coming up and registration is currently open.
In addition to a solid 3 day track of Lucene & Solr related talks, there
are also some post confrence events that are open to anyone even if you
don't attend the confren
It depends on hardware, your latency requirements and such.
We've helped customers with several billion documents, so big numbers alone
are not a problem.
Otis
Solr & ElasticSearch Support
http://sematext.com/
On Feb 27, 2014 6:47 AM, "Vineet Mishra" wrote:
> Hi All
>
> What is the Scalability
: As far as I am aware, you can't specify the default core with the new solr.xml
: format. The default is hard-coded to what's in the example - collection1. If
: you happen to have a core named collection1, it will work without a core name
: in the URL.
Correct, this is for backcompat with sing
Bumping this.
I'm seeing the error mentioned earlier in the thread - "Unable to download
completely. Downloaded 0!=" often in my logs. I'm
dealing with a situation where maxDoc count is growing at a faster rate
than numDocs and is now almost twice as large. I'm not optimizing but
rather relying o
Hi,
We are using 'solr.TrieDateField' type for indexing a date column in Solr.
By default triedate will index date columns as UTC format. But we need the
date as it is in the source(DB table) with time associated with that date.
Do we need to use DateFormatTransformer to get the right date format.
Done, thanks for contributing!
Erick
On Thu, Feb 27, 2014 at 1:59 PM, Eva Andreasson wrote:
> I was thinking I could help with some wiki changes/updates. Would be great
> if I could be added to the ContributorsGroup for the Solr wiki.
>
> My wiki user name is EvaAndreasson
>
> Cheers,
> Eva
>
Hi,
http://localhost:8983/solr/collection1/select?facet=true&fq=timestamp:[2013-01-01T00:00:00Z%20TO%202013-12-02T00:00:00Z]&facet.prefix=google
You url misses these parameters : q=*:* facet.field=content_autosuggest
And if you are getting numFound=0 because of your filter query, we cannot
ge
Hi,
Your example response () looks like customized. Your
SolrJ program also has Element class that I have never seen.
By the way, fq=-author:Dickens, Janet is parsed as -author:Dickens
defaultField:Janet
Assuming it is String type, either use quotes fq=-author:"Dickens, Janet" or
term query
I was thinking I could help with some wiki changes/updates. Would be great
if I could be added to the ContributorsGroup for the Solr wiki.
My wiki user name is EvaAndreasson
Cheers,
Eva
Thanks for the quick response,
I have configured suggest and its working great but its searching all the
documents.
From my field definition:
I need to search between specific timestamp and by username fields for
autocomplete by field content/content_autosuggest.
I need to autocomplete word
Hi epnRui,
I don't full follow your e-mail (I think you need to describe your use case)
but here are some answers,
- Is it possible to have facets of two or more words?
Yes. For example if you use ShingleFilterFactory at index time you will see two
or more words in facets.
- Can I tokenize
Hi,
What do you mean by 'date range' ? Faceting component respects q or fq
parameters, so if you set fq=date:[somedate TO otherdate] yes faceting values
will change. facet.prefix can be used for auto complete.
On Thursday, February 27, 2014 5:03 PM, Bala Iyer wrote:
Hi,
I am new to solr, Is
Hi,
I would index sentences as solr documents. If you don't have existing sentence
tokenizer, you can Apache OpenNLP for this.
https://opennlp.apache.org/documentation/1.5.3/manual/opennlp.html#tools.sentdetect
With this strategy you can retrieve highlighted full sentences.
On Thursday, Feb
You would get more room for disk cache by reducing your large heap.
Otherwise, you'd have to add more RAM to your systems or shard your index
to more nodes to gain more RAM that way.
The Linux VM subsystem actually has a number of tuning parameters (like
vm.bdflush, vm.swappiness and vm.pagecache)
On 2/27/2014 1:09 PM, Joshi, Shital wrote:
If page cache is the issue, what is the solution?
What operating system are you using, and what tool are you looking at to
see your memory usage? Can you share a screenshot with us? Use a file
sharing website for that - the list generally doesn't l
On 2/27/2014 11:32 AM, solr2020 wrote:
We are using DefaultHttpClient 4.2 .3 on solrj side to send parallel
queries to Solr. But the connection manager associated with this
(PoolingHttpClientConnectionManager) by default allows 2 concurrent
connections only.How to send more than 2 parallel queri
Hi Michael,
If page cache is the issue, what is the solution?
Thanks!
-Original Message-
From: Michael Della Bitta [mailto:michael.della.bi...@appinions.com]
Sent: Monday, February 24, 2014 9:54 PM
To: solr-user@lucene.apache.org
Subject: Re: Solr4 performance
I'm not sure how you're
On 2/27/2014 4:46 AM, Vineet Mishra wrote:
What is the Scalability Limit of CloudSolr, can it reach to index Billions
of Documents and each document containing 400-500 Number Field(probably
Float or Double).
Is it possible and feasible to go with current CloudSolr Architecture or
are there some o
On 2/27/2014 7:45 AM, Cynthia Park wrote:
I may have missed this but, how do you specify a default core when using
the new-style
for the solr.xml? When I view the status of my Solr core setup (
http://localhost:8983/solr/admin/cores?action=STATUS) I see a
isDefaultCore speficiation
but, i'm not s
Hi,
I have a use case where I need a document to be send to SOLR in multiple
post requests (by curl/libcurl). Basically, trying to stream documents into
SOLR (via curl). Now what seems to be happening is as soon as one request
comes in SOLR tries to parse that chunk and extract field values, hence
Backslashes are used to escape special characters in queries, but the
backslash must in turn be encoded in the URL as %5C.
-- Jack Krupansky
-Original Message-
From: deniz
Sent: Thursday, February 27, 2014 1:36 AM
To: solr-user@lucene.apache.org
Subject: Searching with special chars
Thanks Jack.
We are using DefaultHttpClient 4.2 .3 on solrj side to send parallel
queries to Solr. But the connection manager associated with this
(PoolingHttpClientConnectionManager) by default allows 2 concurrent
connections only.How to send more than 2 parallel queries/how to establish
more th
: Problem here is, I have indexed a huge amount of data into solr in
: production. It would be a pain to edit the schema file and re-index all the
: values.
There's not much you can do about it in an already built index -- if the
field doesn't exist, it doesn't exist. Nothing magically/secretl
I agree with Erick, but if you want the special characters to count in
searches, you might consider not just stripping them out but replacing them
with textual placeholders (which would also have to be done at indexing time).
For instance, I replace C# with csharp and C++ with cplusplus during
Hi Joe,
With 10,000 documents the CollapsingQParserPlugin will likely not have any
performance advantages. The CollapsingQParserPlugin will be faster then
standard grouping when you have a higher number of distinct groups and
large result sets. For the scale you are working at you will be just fin
Good luck! You'll need it.
Problem is this is such a sticky wicket. You can
move the cleaning up to the PHP layer, that is
strip out the parens.
You could write a Solr component that got the
query _very_ early and transformed it. You'd
have to get here before parsing.
Either way, though, you'll
I think it is just a side-effect of the current implementation that
the ranges are assigned linearly. You can also verify this by choosing
a document from each shard and running it's uniqueKey against the
CompositeIdRouter's sliceHash method and verifying that it is included
in the range.
I couldn
Hi Markus;
I am already using existing functionality at Nutch. I have calculated the
batch size effect and I think that map task should be tune up.
Thanks;
Furkan KAMACI
2014-02-27 17:21 GMT+02:00 Markus Jelsma :
> Something must be eating your memory in your solrcloud indexer in Nutch.
> We h
I noticed group queries are not getting cached in SOLR, is that normal.
I would like to enable caching if possible, any quick pointers would be
helpful.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Group-query-not-cached-in-SOLR-tp4120159.html
Sent from the Solr - User m
Something must be eating your memory in your solrcloud indexer in Nutch. We
have our own SolrCloud indexer in Nutch and it uses extremely little memory.
You either have a leak or your batch size is too large.
-Original message-
> From:Furkan KAMACI
> Sent: Thursday 27th February 2014
Hi,
I am new to solr, Is it possible to accomplish autocomplete through ( Terms or
Suggest or Facet ) over a date range.
I need to have a autocomplete over a field between a date range from another
field.
Thanks
newbie
Hi,
The target here is to use the facets to generate tag clouds, whose set can
have go upto trigrams. This works fine in dev, but our production is having
issues with faceting..
Regards,
On Wed, Feb 26, 2014 at 10:00 PM, David Miller wrote:
> Hi Jack,
>
> Ya, the requirement is like that. I
Hi;
I'm trying to index 2 million documents into SolrCloud via Map Reduce Jobs
(really small number of documents for my system). However I get that error
at tasks when I increase the added document size:
java.lang.ClassCastException: java.lang.OutOfMemoryError cannot be
cast to java.lang.Exceptio
Please tell me how to get the full sentence as output of search result in
SOLR. For example if i am searching for "Sensor" and there is a sentence in
the document which has sentence as "This Sensor is working fine." It should
provide me the output as This Sensor is working fine. I am trying to use
Hello,
I may have missed this but, how do you specify a default core when using
the new-style
for the solr.xml? When I view the status of my Solr core setup (
http://localhost:8983/solr/admin/cores?action=STATUS) I see a
isDefaultCore speficiation
but, i'm not sure where it can from and and where
I can see what you mean, what you are expecting is a single host:port
combination for "The Cloud" that always works, and you can call from your
UI. That is perfectly possible, but its really not within the scope of
Solr itself.
What you should understand is that Solr provides is a cloud that has
I recommend you to attach your log to the issue and commend the process that
you did to run in this error. Maybe your logs can have some valuable
information.
--
Yago Riveiro
Sent with Sparrow (http://www.sparrowmailapp.com/?sig)
On Thursday, February 27, 2014 at 12:03 PM, Yago Riveiro wrot
I remember tat I need to empty my queue before restart the cluster too.
This bug is a little scary because if you have a schedule system to deletes
collections on the fly your cluster can blow and you don't know why …
--
Yago Riveiro
Sent with Sparrow (http://www.sparrowmailapp.com/?sig)
On
Interesting. I deleted all old collections, configs, and clusterstate.json from
Zookeeper and I still had that problem.
I’m quite new to Zookeeper, so some of what I say might be wrong. It seems
there were some outstanding changes in Zookeeper. Or at least I found some
queue items in a queue no
I had some problems with DELETE action too.
I reported this some time ago https://issues.apache.org/jira/browse/SOLR-5559
The overseer fail to delete a collection and the solr cluster becomes unstable,
I reloaded my boxes and my cluster never went online.
After some debug, I found some shard f
Hi All
What is the Scalability Limit of CloudSolr, can it reach to index Billions
of Documents and each document containing 400-500 Number Field(probably
Float or Double).
Is it possible and feasible to go with current CloudSolr Architecture or
are there some other alternative or replacement.
Reg
Hi Per
Thanks for your response, got it working.
But moreover I was more interested in querying the same Cloud from UI in a
case of one of the server down and querying the same server to get
collection result. But I guess thats not possible.
Thanks!
On Mon, Feb 24, 2014 at 7:36 PM, Per Steffen
I didn't know of the default parameter. Thanks for sharing.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Know-indexing-time-of-a-document-tp4120051p4120108.html
Sent from the Solr - User mailing list archive at Nabble.com.
Problem here is, I have indexed a huge amount of data into solr in
production. It would be a pain to edit the schema file and re-index all the
values.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Know-indexing-time-of-a-document-tp4120051p4120103.html
Sent from the Solr
Hi everyone!
I'm having a problem and I have searched and Haven't found a solution yet
and am rather confused at the moment.
I have an application that stores human readable texts in my Solr index.
It finds the most relevant terms in that human readable text, I think using
termvectors and facets,
Hi,
Upgraded a small cluster from 4.3.1 to 4.7 in Solr Cloud mode.
I deleted the old data, replaced the solr.xml with the example solr.xml with
auto-discovery, but it seems there is still some old data somewhere, probably
in Zookeeper that keep my machine from starting.
It loops over the same
Thanks Shawn,
thanks for the pointer.
I will try to checkout latest trunk from solrmeter and will see if i can
patch it myself.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solrmeter-with-solr-4-6-1-tp4120082p4120092.html
Sent from the Solr - User mailing list archive
Ok so I cannot move forward with this,
If I use format like q=a&fq=category:(value1 value2 value3)
this gives me results with first category.
What i want is top n results per filter category and I dont want to use
group as performance seems to be very bad for groups my observation is group
query
On 2/27/2014 3:02 AM, Shawn Heisey wrote:
> Looks like we sometimes have a different data type on some of the
> statistics that Solr exposes via admin handlers. There is an issue
> against solrmeter for this:
>
> http://code.google.com/p/solrmeter/issues/detail?id=113
I left out the trail of how
Hi Jan,
I think that in your overall function you messed up a little with your time
units. In your product function you try to get years from miliseconds:
product(ms(NOW,date_field),3.16e-11)
... to have that value comparable with the value of -100 in your map function:
map(product(ms(NOW,date_f
On 2/27/2014 2:43 AM, search engn dev wrote:
> Has anybody used latest solrmeter with solr 4.6.1? because I am getting
> following error whenever i try to connect to solr.
>
>
> [Thread-384]|ERROR|[com.plugtree.solrmeter.model.statistic.CacheHistoryStatistic]|15:02:52,509
> | Could not update sta
Has anybody used latest solrmeter with solr 4.6.1? because I am getting
following error whenever i try to connect to solr.
[Thread-384]|ERROR|[com.plugtree.solrmeter.model.statistic.CacheHistoryStatistic]|15:02:52,509
| Could not update statistic
com.plugtree.solrmeter.model.exception.StatisticCo
On 2/27/2014 12:49 AM, Thomas Scheffler wrote:
>> What problems have you seen with mixing 4.6.0 and 4.6.1? It's possible
>> that I'm completely ignorant here, but I have not heard of any.
>
> Actually bug reports arrive me that sound like
>
> "Unknown type 19"
Aha! I found it! It was caused b
56 matches
Mail list logo