On 9 July 2013 12:08, Mysurf Mail wrote:
> Hi,
> A general question:
>
>
> Let's say I have Car And CarParts 1:n relation.
>
> And I have discovered that the user had entered in the search field instead
> of car name - a part serial number (SKU).
> (I discovered it useing regex)
>
> Is there a way
Hi,
A general question:
Let's say I have Car And CarParts 1:n relation.
And I have discovered that the user had entered in the search field instead
of car name - a part serial number (SKU).
(I discovered it useing regex)
Is there a way to fetch different types of answers in Solr?
Is there a way
Hi,
Maybe you can describe how you are using Solr? Which version exactly?
Can you share the errors you are seeing? etc.
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring -- http://sematext.com/spm
On Tue, Jul 9, 2013 at 2:07 AM, Cool Techi wrote:
> Hi,
>
>
Hi,
We are frequently getting issues of index corruption on the cloud, this used to
not happen in our master slave setup with solr 3.6. I have tried to check the
logs, but don't see an exact reason.
I have run the index checker and it recovers, but I am not able to understand
as to why this is
I wrote a custom data import handler to import data from files. I am trying
to figure out a way to make asynchronous call instead of waiting for the
data import response. Is there an easy way to invoke asynchronously (other
than using futures and callables) ?
public class CustomFileImportHandler
OK, thank you Otis, I *think* this should be easy to add - I can try. We
were calling them 'private library' searches
roman
On Mon, Jul 8, 2013 at 11:58 PM, Otis Gospodnetic <
otis.gospodne...@gmail.com> wrote:
> Hi Roman,
>
> I referred to something I called ""server-side named filters". It
>
Hi Roman,
I referred to something I called ""server-side named filters". It
matches the feature described at
http://www.elasticsearch.org/blog/terms-filter-lookup/
Would be a cool addition, IMHO.
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring -- http://semat
Hi,
I don't recall any such changes around replication...
Why not run the backup on the master?
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring -- http://sematext.com/spm
On Sat, Jul 6, 2013 at 3:05 AM, Cool Techi wrote:
> Hi,
>
> We migrated from solr 3.6
Not sure... but if you need Collocations/SIPs, you can try the
non-free-but-cheaper-than-DIY
http://sematext.com/products/key-phrase-extractor/index.html
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring -- http://sematext.com/spm
On Sun, Jul 7, 2013 at 12:42 A
I think nobody's itching enough, but I think it would be great to have
facet.method=lucene :)
Paste from http://search-lucene.com/m/NzVKPC5C8x1 from Yonik:
"
Would it make sense to add Lucene's faceting as an *additional* Solr
faceting method? Maybe?
I don't really know though - I haven't do
Woho, I love inconsistency in code! Not. :)
Any idea why this is, Jack?
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring -- http://sematext.com/spm
On Mon, Jul 8, 2013 at 12:49 PM, Jack Krupansky wrote:
> Nope.
>
>
> -- Jack Krupansky
>
> -Original M
Hi,
70 GB heap and still OOMing? H sure, 14 fields for faceting,
but still - 70 GB heap!
Don't have source handy, but I quickly looked at FC src here -
http://search-lucene.com/c/Lucene:core/src/java/org/apache/lucene/search/FieldCache.java
- I see mentions of "delete" there, so I would
It's simply a sugar method that no one has gotten to yet. I almost have once or
twice, but I always have moved onto other things before even starting.
It's fairly simple to just start another replica on the TO node and then delete
the replica on the FROM node, so not a lot of urgency.
- Mark
O
Hi,
Solr(Cloud) currently doesn't have any facility to move a specific
replica from one node to the other.
How come? Is there a technical or philosophical reason, or "just" the
"24 hours/day reason"?
Thanks,
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring --
I'm 99% sure that the deleted docs will indeed use up space in the field cache,
at least until the segments that those documents are in are merged - that is
what an optimize will do. Of course, these segments will automatically be
merged eventually, but it might take days for this to happen, dep
Per Jack's suggestion, I changed the heading in the .xml file to and it worked. Thanks so much guys!
From: Shawn Heisey
To: solr-user@lucene.apache.org
Sent: Monday, July 8, 2013 7:43 PM
Subject: Re: Indexing fails for docs with high Latin1 chars
On 7/8/2013
I tried that. It didn't work. I forgot to mention in my first email that I'm
using Solr 3.6. Would that make a difference?
From: Jack Krupansky
To: solr-user@lucene.apache.org; John Randall
Sent: Monday, July 8, 2013 7:22 PM
Subject: Re: Indexing fails for do
Right, the charset must agree with the charset of the program that wrote the
file.
-- Jack Krupansky
-Original Message-
From: Shawn Heisey
Sent: Monday, July 08, 2013 7:43 PM
To: solr-user@lucene.apache.org
Subject: Re: Indexing fails for docs with high Latin1 chars
On 7/8/2013 4:43
(NOTE: cross-posted to variuous lists, please reply only to general@lucene
w/ any questions or follow ups)
The Call for Papers for Lucene/Solr Revolution EU 2013 is currently open.
http://www.lucenerevolution.org/2013/call-for-papers
Lucene/Solr Revolution is the biggest open source conferen
On 7/8/2013 4:43 PM, John Randall wrote:
I'm new to Solr, so I'm probably missing something. So far I've successfully
indexed .xml docs with low Ascii chars. However when I try to add a doc that
has Latin1 chars with diacritics, it fails. I've tried using the Jetty
exampledocs post.jar, as wel
Maybe you need to add "; charset=UTF-8" to your Content-type:
curl
"http://localhost:8080/solr/update/?commit=true&stream.file=c:/solr/tml/exampledocs/57917486.xml&stream.contentType=application/xml;
charset=UTF-8”
-- Jack Krupansky
-Original Message-
From: John Randall
Sent: Monda
I'm new to Solr, so I'm probably missing something. So far I've successfully
indexed .xml docs with low Ascii chars. However when I try to add a doc that
has Latin1 chars with diacritics, it fails. I've tried using the Jetty
exampledocs post.jar, as well as using curl and directly from a browser
What is the actual use case? In other words, why is the list so long?
Maybe exclusion by keyword is not the proper solution... but we need what
the underlying problem is.
Is this for document access control?
-- Jack Krupansky
-Original Message-
From: Angela Zhu
Sent: Monday, July 0
Hi,
We have Solr 3.6 set up with master and two slaves, each one with 70GB JVM. We
run into java.lang.OutOfMemoryError when we cross 250 million documents. Every
time this happens we purge documents, bring it below 200 million and bounce
both slaves. We have facets on 14 fields. We usually don
Roman,
The video was very clarifying and I realized block joins would be a
great fit for my problem. However, I got worried about the size of the
block... I could have 10 million childs for 1 parent, for instance. Althout
this could stay in the same shard, do you guys think it would be a huge
Hello,
Does anyone have step-by-step instructions for running SolrCloud on Jboss?
Thanks
One of the approaches is to index create a new field based on the stopwords
(ie accept only stopwords :)) - ie. if the documents contains them, you
index 1 - and use a q=apple&fq=bad_apple:0
This has many limitations (in terms of flexibility), but it will be
superfast
roman
On Mon, Jul 8, 2013 a
Is there a solr way to remove any result from the list search results that
contain a term in a excluding list?
For example, suppose I search for apple and get 5 documents contains it,
and my excluding list is something like ['bad', 'wrong', 'staled'].
Out of the 5 documents, 3 has a word in this l
Hello,
The joins are not the only idea, you may want to write your own function
(ValueSource) that can implement your logic. However, I think you should
not throw away the regex idea (as being slow), before trying it out -
because it can be faster than the joins. Your problem is that the number of
Hi,
I am curious about the functional query, did you try it and it didn't work?
or was it too slow?
idf(other_field,field(term))
Thanks!
roman
On Mon, Jul 8, 2013 at 4:34 PM, Kathryn Mazaitis wrote:
> Hi All,
>
> Resolution: I ended up cheating. :P Though now that I look at it, I think
>
Hi All,
Resolution: I ended up cheating. :P Though now that I look at it, I think
this was Roman's second suggestion. Thanks!
Since the application that will be processing the IDF figures is located on
the same machine as SOLR, I opened a second IndexReader on the lucene index
and used
reader.nu
That 1024 limit of the DataStax Enterprise packaging of Solr is going to be
relaxed in a coming release - you will be able to have more dynamic fields,
but... "going wild" has memory and performance implications anyway. That
limit is the number of populated fields in a single document - differen
Hello all,
I am using Solr Cloud today and I have the following need:
- My queries focus on counting how many users attend to some criteria.
So my main document is "user" (parent table)
- Each user can access several web pages (a child table) and each web
page might have several a
Jack,
Thanks a lot for your answers. I guess I just had heard on Cassandra
Summit that Solr can't support more than 1024 dynamic fields and it might
be possible in my case, that's why I asked this question. However, your
answer was very complete and made me think in a lot of things.
The
For reference
https://issues.apache.org/jira/browse/SOLR-5019
--
View this message in context:
http://lucene.472066.n3.nabble.com/Concurrent-Modification-Exception-tp4074371p4076330.html
Sent from the Solr - User mailing list archive at Nabble.com.
Thanks Mark!
On Mon, Jul 8, 2013 at 10:46 AM, Mark Miller wrote:
>
> On Jul 8, 2013, at 1:40 PM, "Ali, Saqib" wrote:
>
> > Hello all,
> >
> > We have an app that uses the SolrJ and instantiates using HttpSolrServer.
> >
> > Now that we would like to move to SolrCloud, can we still use the same
On Jul 8, 2013, at 1:40 PM, "Ali, Saqib" wrote:
> Hello all,
>
> We have an app that uses the SolrJ and instantiates using HttpSolrServer.
>
> Now that we would like to move to SolrCloud, can we still use the same app,
> or do we HAVE to switch to
>
> CloudSolrServer server = new CloudSolrSer
Hello all,
We have an app that uses the SolrJ and instantiates using HttpSolrServer.
Now that we would like to move to SolrCloud, can we still use the same app,
or do we HAVE to switch to
CloudSolrServer server = new CloudSolrServer("?");
right away?
Or will point to one instance using Htt
Nope.
-- Jack Krupansky
-Original Message-
From: Alexandre Rafalovitch
Sent: Monday, July 08, 2013 7:20 AM
To: solr-user@lucene.apache.org
Subject: Re: Are the XML element names in schema.xml case sensitive?
But not dynamicField or any others?
Regards,
Alex
On 7 Jul 2013 23:39, "J
On 7/6/2013 4:27 PM, Steven Glass wrote:
Thanks for your response.
But it seems like there should be a way to issue the equivalent of
http://localhost:8983/solr/schema/version
which returns
{
"responseHeader":{
"status":0,
"QTime":4},
"version":1.5}
from the s
You can mix and match the data sources in nested entities, yes. Just make
sure that you declare your data sources at the top and refer to them
properly. As per documentation:
"Ensure that the dataSource is of type DataSource (FileDataSource,
URLDataSource)". So you need to declare one at the top of
The strategy doesn't require putting all the recent data on a single node.
What has been suggested is collection based - the most recent data will simply
be in it's own collection, that may or may not be on a single node.
This is pretty much always going to be advantageous for time series data.
It's a known bug, fix coming in 4.4, 4.4 likely coming within a couple weeks.
https://issues.apache.org/jira/browse/SOLR-4805
- Mark
On Jul 8, 2013, at 10:30 AM, adfel70 wrote:
> Hi
>
> each time I reload a collection via collections API, zookeeper thinks that
> all the shards in the collecti
ok great.
can I use this EntityProcessor within JdbcDataSource?
Like this:
http://localhost/project_1/files/a.txt";
dataSource="data-source-name">
By the way, I currently load the field into "text_en_splitting" as defined
in schema.xml...
On Mon, Jul 8, 2013 at 7:
http://wiki.apache.org/solr/DataImportHandler#PlainTextEntityProcessor or
http://wiki.apache.org/solr/DataImportHandler#LineEntityProcessor ?
The file name gets exposed as a ${entityname.fieldname} variable. You can
probably copy/manipulate it with a transformer on the external entity
before it hi
On this page (http://wiki.apache.org/solr/DataImportHandler), I cant see
how its possible. Perhaps there is another guide..
Basically, this is what I am doing:
Index data from multiple tables into Solr (see here
http://wiki.apache.org/solr/DIHQuickStart). I need to skip 1 very big heavy
table as i
Yes, you should be able to used nested query parsers to mix the queries.
Solr 4.1(?) made it easier.
-- Jack Krupansky
-Original Message-
From: Abeygunawardena, Niran
Sent: Monday, July 08, 2013 7:00 AM
To: solr-user@lucene.apache.org
Subject: Re: Surround query parser not working?
H
Consider keeping your stored/updatable fields in a separate, parallel
collection. It makes queries a multi-step operation, but gives you a lot
more flexibility.
In some cases (but not all), "external file fields" can eliminate the need
to directly update indexed documents.
Or, consider a hyb
Hi
each time I reload a collection via collections API, zookeeper thinks that
all the shards in the collection are down.
It marks them as down and I can't send requests.
Why "thinks"? because if I manually edit clusterstate.json file and set
'state' value to 'active', they come back up and reque
Did you have a chance to look at DIH with nested entities yet? That's
probably the way to go to start out.
Or a custom client, of course. Or, ETL solutions that support Solr (e.g.
Apache Flume - not personally tested yet).
Regards,
Alex.
Personal website: http://www.outerthoughts.com/
LinkedI
Other that the per-node/per-collection limit of 2 billion documents per
Lucene index, most of the limits of Solr are performance-based limits - Solr
can handle it, but the performance may not be acceptable. Dynamic fields are
a great example. Nothing prevents you from creating a document with, s
see: https://issues.apache.org/jira/browse/LUCENE-4258
I'm sure the people working on this would gladly get all
the help they can. WARNING: I suspect (although I haven't
looked myself) that this is very hairy code .
Ah excellent! Thanks! Exactly what I was looking for. Looks like this
has been in
Hi everyone,
I am looking for a way to import/index data such that i load data from
table_1 and instead of joining from table_2, i import the rest of the
"joined" data from a file instead. The name of the file comes from a field
from table_1
Is it possible? and is it easily possible?
--
Reg
Hello everyone,
I am trying to search information about possible solr limitations I
should consider in my architecture. Things like max number of dynamic
fields, max number o documents in SolrCloud, etc.
Does anyone know where I can find this info?
Best regards,
--
Marcelo Elias Del Vall
It seems that I lost the formatting of the data.
parent
idKEY_SCOMMENT_TTYPE_S
11ventilation testParent
22comment2Parent
33comment3Parent
child1
idKEY_SCOMMENT_TTYPE_S
41comment4Child1
52ventilation testChild1
63comment5Child1
child2
idKEY_SCOMMENT_TTYPE_S
71comment6Child2
82comment7Child2
93ven
Thanks for the detailed response Erik, you helped me a lot in clarifying
many Solr concepts!
Best,
Flavio
On Mon, Jul 8, 2013 at 1:59 PM, Erick Erickson wrote:
> Flavio:
>
> I think you're missing a critical bit about SolrCloud,
> namely Zookeeper (ZK), see here on the SolrCloud page
> for a st
On Mon, Jul 8, 2013 at 8:18 AM, adfel70 wrote:
> I have a field that's only indexed in some of the documents.
> Can I create a boolean facet on this field by its existence?
> for instance:
> yes(124)
> no(479)
>
> Note that the fields' value is not facetable because all its values are
> unique mos
Here is exactly the data that I'm working with and the results for some tests
that I performed:
parent child1 child2
id KEY_S COMMENT_T TYPE_S id KEY_S COMMENT_T TYPE_S id KEY_S COMMENT_T TYPE_S
1 1 ventilation test Parent 4 1 comment4 Child1 7 1 comment6 Child2
2 2 comment2
Hi All.
I have some solr 3.5.0 instances on CentOS 6 x86_64. Currently logging is
by going to syslog (logger), for all my cores to one shared logfile. I
would like to have a separate logfile for each core.
My solr process is started from a simple shell script:
daemon $JAVA -Xms25g -Xmx25g -DSTOP
I have a field that's only indexed in some of the documents.
Can I create a boolean facet on this field by its existence?
for instance:
yes(124)
no(479)
Note that the fields' value is not facetable because all its values are
unique most of the time.
I just want to facet on the question whether thi
Just specify successively larger "&start=..." parameters,
When you want page 1 you specify
&start=0&rows=10
page 2:
&start=10&rows=10
and so on.
see: http://wiki.apache.org/solr/CommonQueryParameters#start
Best
Erick
On Mon, Jul 8, 2013 at 5:36 AM, aniljayanti wrote:
> Hi Ralf,
>
> thanks for
Bram:
see: https://issues.apache.org/jira/browse/LUCENE-4258
I'm sure the people working on this would gladly get all
the help they can. WARNING: I suspect (although I haven't
looked myself) that this is very hairy code .
bq: Making all fields stored is simply not possible from
a performance poin
Flavio:
I think you're missing a critical bit about SolrCloud,
namely Zookeeper (ZK), see here on the SolrCloud page
for a start:
http://wiki.apache.org/solr/SolrCloud#ZooKeeper
You'll notice that each Solr node, when it is started,
requires the address of your ZK ensemble, NOT a
solr node. That
You have to tell us _how_ it's not working to get a
meaningful answer. Perhaps you could review:
http://wiki.apache.org/solr/UsingMailingLists
Best
Erick
On Sun, Jul 7, 2013 at 3:05 PM, lampa24 wrote:
> Hello ,
> I use Solr Cloud 4.3.1.
>
> We use query join , so we need index some documents to
Hi Ralf,
thanks for reply,
I have 540 records as a solr result. In that i want only 120 records
(), based on the "numFound"
node value i am calculating paging concept. i want to show 12 records per
page then (120/12 = 10 pages) i would have 10 pages. Right. So i want that
scenario. . i do not wa
Hi Erick,
thanks for reply, after querying using solr getting 540 records, but i had a
requirement to get only 120 records even though it may have greater than
120. After getting results i am calculating the paging on the basis of
*"numFound=120"*.That is why i am checking this. if i take 12 recor
But not dynamicField or any others?
Regards,
Alex
On 7 Jul 2013 23:39, "Jack Krupansky" wrote:
> Yes, the XML element names (tags) and attribute names are all case
> sensitive, but... Solr has a special hack for "fieldtype" as well as
> "fieldType".
>
> -- Jack Krupansky
>
> -Original Me
Hi,
Thanks. I found out that my issue was the default field (df) was being ignored
and I had to specify the parameter by adding &df=text in the URL.
Thank you for updating the wiki page on the surround parser:
http://wiki.apache.org/solr/SurroundQueryParser
Hopefully, ordered proximity searches
Howdy,
Atomic updates only work on stored fields. When submitting an update,
any non-stored fields are apparently emptied. Making all fields stored
is simply not possible from a performance point of view in my case.
Neither is resubmitting all fields.
Are there any plans to change this behav
Hi to all,
I started following this mailing list about 1 month ago and I read many
threads about SolrCloud and distributed Solr. I just want to check if I
understood correctly and, if so, ask for some architectural decision I have
to take:
1) At the moment, in order to design a scalable Solr dep
William:
Geosearch: coming soon!
F
On Sun, Jul 7, 2013 at 6:29 AM, William Bell wrote:
> Can it do Geo Spatial searching? (i.e. Find documents within 10 miles of a
> lat,long?)
>
>
>
> On Fri, Jul 5, 2013 at 12:53 PM, Fergus McDowall
> wrote:
>
> > Here is some news that might be of interest
71 matches
Mail list logo