We have a search system based on Solr using the Solrnet library in C# which
supports some advanced search features like Fuzzy, Synonym and Stemming.
While all of these work, *the expectation from the Stemming Search seems to
be a combination of Stemming by reduction as well as stemming by expansion
I have checked the FSDirectory, it will create "MMapDirectory" or
"NIOFSDirectory" for Directory.
This two directory only supply IndexInput extend for read file (MMapIndexInput
extends ByteBufferIndexInput),
why not there is not MMap/NIO IndexOutput extend for file write. It only use
FSIndexOut
On 27 June 2013 21:13, Michael Della Bitta
wrote:
> Hi,
>
> I haven't used it yet, but I believe you can do this using the
> FileDataSource feature of DataImportHandler:
>
> http://wiki.apache.org/solr/DataImportHandler#FileDataSource
[...]
Please see other recent threads on similar topics
in thi
Hi,
I currently have a query as below. I am using the fq only if latlong value
(using switch plugin) is not empty else I am not using fq at all.
Whenever latlong value is empty, I just use value of $where (in "q")
parameter to return the results based on location.
Now whenever latlong value is a
Thanks Mark, might give it a go, or probably just wait for 4.4 :)
On Thu, Jun 27, 2013 at 4:06 PM, Mark Miller wrote:
> There were a variety of little bugs - it will just be a bit of a land mine
> situation if you try and do it with 4.3.1.
>
> If it ends up working for you, that's that.
>
> - M
Thanks a lot for your response. I created a multithreaded program to create
/submit the documents in batch of 100 to Embedded SOLR server but for some
reason it takes more time to index the data when compared with
ConcurrentUpdateeSOLR server. I was under assumption that embedded server
would take
Might not be useful but a work around would be to divide all scores by max
score to get scores between 0 and 1.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Normalizing-Returning-solr-scores-between-0-to-1-tp4073797p4073829.html
Sent from the Solr - User mailing list arch
There is no way that I am of aware of to have Solr return between 0 and a
1. Perhaps there is some way tom implement a custom Scorer, but that is
overkill and would probably have adverse affects. Instead, just normalize
it in your results. Of course, since you read the link you included, you
realiz
Hello,
I am using Solr 4.3.2 and Oracle DB. The sub entity is using
CachedSqlEntityProcessor. The dataSource is having batchSize="500". The
full-import is failed with 'ORA-01555: snapshot too old: rollback segment
number with name "" too small ORA-22924: snapshot too old' Exception after
5 hours.
I've created a JIRA and applied a patch for it:
https://issues.apache.org/jira/browse/SOLR-4973
2013/6/12 Furkan KAMACI
> Ok, I will create a JIRA for it.
>
>
> 2013/6/11 Mark Miller
>
>>
>> On Jun 11, 2013, at 4:51 AM, Furkan KAMACI
>> wrote:
>>
>> > Why there is no getter method for defaultC
You might be seeing https://issues.apache.org/jira/browse/SOLR-4923 ?
The commit true part of the request that add documents? If so, it might be
SOLR-4923 and you should try the commit in a request after adding the docs.
- Mark
On Jun 27, 2013, at 4:42 PM, "Joshi, Shital" wrote:
> Hi,
>
> We
Hi,
We have a need where we would want normalized scores from score ranging
between 0 to 1 rather than a free range.
I read about it @ http://wiki.apache.org/lucene-java/ScoresAsPercentages and
seems like thats not something that is recommended.
However, is there still a way to set some config i
Shawn,
On Thu, Jun 27, 2013 at 5:03 PM, Shawn Heisey wrote:
> On 6/27/2013 5:59 AM, Robert Krüger wrote:
>> sometime forcing oneself to describe a problem is the first step to a
>> solution. I just realized that I also had an autocommit statement in
>> my config with the exact same amount of time
Hi,
We finally decided on using custom sharding (implicit document routing) for our
project. We will have ~3 mil documents per shardkey. We're maintaining
shardkey -> shardid mapping in a database table. While adding documents we
always specify _shard_ parameter in update URL but while queryin
There were a variety of little bugs - it will just be a bit of a land mine
situation if you try and do it with 4.3.1.
If it ends up working for you, that's that.
- Mark
On Jun 27, 2013, at 3:22 PM, shikhar wrote:
> Can anyone (Eric?) outline what's changing between 4.3.1 and 4.4 wrt
> http://
HTTP?
Otis
--
Solr & ElasticSearch Support -- http://sematext.com/
Performance Monitoring -- http://sematext.com/spm
On Thu, Jun 27, 2013 at 7:40 AM, Furkan KAMACI wrote:
> There is a low level protocol that defines client–server protocol for
> searching and retrieving information from remote
http://wiki.apache.org/solr/SolrCloud#Distributed_Requests
- Mark
On Jun 27, 2013, at 2:34 PM, Otis Gospodnetic
wrote:
> Hi,
>
> This "search across multiple collections" question has come up a few
> times recently:
>
> http://search-lucene.com/m/2Q1BE0IT4Y/&subj=Search+across+multiple+colle
For the record, in case anybody else hits this, I think the ClassCastException
problem had to do with which class loader first loads the class, which is a
side affect of which directory(ies!) you put the jar file in.
I can't reproduce the problem any more, but I believe it went away when I
remo
Can anyone (Eric?) outline what's changing between 4.3.1 and 4.4 wrt
http://wiki.apache.org/solr/Solr.xml%204.4%20and%20beyond, and what makes
the new solr.xml format usable in 4.4 but not 4.3.1?
If one didn't care about sharedLib or solr.xml persistence (the only
solr.xml changes we care about ar
Wow! That's been a while back, and it appears that my journal didn't
carry a good trace of what I did. Here's a reconstruction:
>From my earlier attempt, which is reflected in this solrconfig.xml entry
notice that I am calling solrDirectUpdateHandler2 directly in defining
a requestHandler
I do
Hi,
This "search across multiple collections" question has come up a few
times recently:
http://search-lucene.com/m/2Q1BE0IT4Y/&subj=Search+across+multiple+collections
http://search-lucene.com/m/5JQrXIyhQQ1/&subj=Querying+multiple+collections+in+SolrCloud
One important variation of this Q is - c
On Thu, Jun 27, 2013, at 06:48 PM, abillav...@innoventsolutions.com
wrote:
> Dear List Managers
> I've changed my email that I'd like to use for the solr-user list, as
> it's filling up my work email to the point of insanity.
>
> Regardless of the change in the solr-user community, it still kee
Shawn,
Thanks a lot for your reply.
I have pasted my entire code below, it would be great if you can let me know
if I am doing anything wrong in terms of running the code in multithreaded
environment.
http://pastebin.com/WRLn3yWn
--
View this message in context:
http://lucene.472066.n3.nabbl
Haven't tried this, but I _think_ you can use the
"confFiles" trick with relative paths, see:
http://wiki.apache.org/solr/SolrReplication
Or just put your EFF files in the data dir?
Best
Erick
On Wed, Jun 26, 2013 at 9:01 PM, Arun Rangarajan
wrote:
> From https://wiki.apache.org/solr/SolrRepli
Dear List Managers
I've changed my email that I'd like to use for the solr-user list, as
it's filling up my work email to the point of insanity.
Regardless of the change in the solr-user community, it still keeps
sending the emails of all threads and replies to my work email. Would
you pleas
I'd _guess_ that this is unsupported across collections if
for no other reason than scores really aren't comparable
across collections and the default ordering within groups
is score. This is really a "federated search" type problem.
But if it makes sense to use N collections for other reasons,
it
I stand corrected, you're absolutely right about string types. But I still
don't think text types are supported, at least in my quick test of the
stock Solr distro, trying to gather stats on the "subject" field produced
the error below. Note that string is a completely unanalyzed type, no
tokenizat
Hello,
I've struggling to find a way to query after collapse.field is performed and
I'm hoping someone can help.
I'm doing a multiple core(index) search which generates results that can
have varying fields.
ex.
entry_id, entry_starred
entry_id, entry_read
I perform a collapse.field on entry_id w
Right, string fields are a little tricky, they're easy to confuse with
fields that actually _do_ something.
By default, norms and term frequencies are turned off for types based on '
class="solr.StrField" '. So any field length normalization (i.e. terms that
appear in shorter fields count more) an
On 6/27/2013 9:32 AM, Michael Della Bitta wrote:
Are you using blockUntilFinished() and/or shutdown()?
One of the things to note is that a commit is just another "document," so
writing a commit into the queue of the ConcurrentUpdateSolrServer isn't
enough to get it flushed out.
ConcurrentUpdat
On 6/27/2013 9:19 AM, Learner wrote:
I was using ConcurrentUpdateSOLR for indexing documents to Solr. Later I had
a need to do portable indexing hence started using Embedded solr server.
I created a multithreaded program to create /submit the documents in batch
of 100 to Embedded SOLR server (ru
Jack,
Did you ever find a fix for this?
I'm having similar issues (different parts of solrconfig) and my guess is it's
a config issue somewhere, vs. a proper casting problem, some nested init issue.
Was curious what you found?
On Mar 13, 2013, at 11:52 AM, Jack Park wrote:
> I can safely sa
Hi Michael,
I realized that I might have to use blockUntilFinished before commit, but do
I have to use shutdown as well??
Thanks,
Qun
--
View this message in context:
http://lucene.472066.n3.nabble.com/ConcurrentUpdateSolrServer-hanging-tp4073620p4073651.html
Sent from the Solr - User mailing
Hi,
I haven't used it yet, but I believe you can do this using the
FileDataSource feature of DataImportHandler:
http://wiki.apache.org/solr/DataImportHandler#FileDataSource
HTH,
Michael Della Bitta
Applications Developer
o: +1 646 532 3062 | c: +1 917 477 7906
appinions inc.
“The Science
Qun,
Are you using blockUntilFinished() and/or shutdown()?
One of the things to note is that a commit is just another "document," so
writing a commit into the queue of the ConcurrentUpdateSolrServer isn't
enough to get it flushed out.
Michael Della Bitta
Applications Developer
o: +1 646 532 3
It is all new, and as yet unreleased. It still has more work needed on
formatting, etc, so I guess you could say, make of it what you will, and
don't yet assume it will always be up and available.
Upayavira
On Thu, Jun 27, 2013, at 04:25 PM, Luis Lebolo wrote:
> This page never came up on any of
This page never came up on any of my Google searches, so thanks for the
heads up! Looks good.
-Luis
On Tue, Jun 25, 2013 at 12:32 PM, Learner wrote:
> I just came across a wonderful online reference wiki for SOLR and thought
> of
> sharing it with the community..
>
>
> https://cwiki.apache.org
I was using ConcurrentUpdateSOLR for indexing documents to Solr. Later I had
a need to do portable indexing hence started using Embedded solr server.
I created a multithreaded program to create /submit the documents in batch
of 100 to Embedded SOLR server (running inside Solrj indexing process) bu
On 6/27/2013 5:59 AM, Robert Krüger wrote:
> sometime forcing oneself to describe a problem is the first step to a
> solution. I just realized that I also had an autocommit statement in
> my config with the exact same amount of time the seemed to be between
> the warnings.
>
> I removed that, beca
Hi,
I'm using concurrentUpdateSolrServer to do my incremental indexing nightly.
I have 50 shards to index into, about 10,000 documents each night. I start
one concurrentUpdateSolrServer on each shards and start to send documents.
The queue size for concurrentUpdateSolrServer is 100, and 4 threads.
Odd - looks like it's stuck waiting to be notified that a new searcher is ready.
- Mark
On Jun 27, 2013, at 8:58 AM, Neal Ensor wrote:
> Okay, I have done this (updated to 4.3.1 across master and four slaves; one
> of these is my own PC for experiments, it is not being accessed by clients).
>
On Thu, Jun 27, 2013 at 12:14 PM, Upayavira wrote:
> can you give an example?
>
Thank you. This is an example query:
select
?q=search_field:iraq
&fq={!cache=false}search_field:love%20obama
&defType=edismax
--
Dotan Cohen
http://gibberish.co.il
http://what-is-what.com
On 27 June 2013 13:42, Venter, Scott wrote:
> Hi all,
>
> I am new to SOLR. I have been working through the SOLR 4 Cookbook and my
> experiences so far have been great.
>
> I have worked through the extraction of PDF data recipe, and the Data import
> recipe. I would now like to join these two t
Forgive my ignorance but I want to be sure, do I add to solrindex-mapping.xml?
so that my solrindex-mapping.xml looks like this:
Hello everyone,
I saw that the ticket regarding this issue is still open (
https://issues.apache.org/jira/browse/SOLR-4479). There is last comment
that suggests to reindex documents with solr 4.2. I did reindex with 4.3
version but term vector still doesn't work producing null pointer
exception.
Hi Sri,
This depends on how the fields (that hold the value) are defined and how
the query is generated.
Try running the query in solr console and use &debug=true to see how the
query string is getting parsed.
If that doesn't help then could you answer following 3 questions relating
to your quest
Hi team,
When the user enter search term as "h.e.r.b.a.l" in the search textbox and
click on search button then SOLR search engine is not returning any results
found. As I can see SOLR is accepting the request parameter as h.e.r.b.a.l.
However we have many records with the string h.e.r.b.a.l
Okay, I have done this (updated to 4.3.1 across master and four slaves; one
of these is my own PC for experiments, it is not being accessed by clients).
Just had a minor replication this morning, and all three slaves are "stuck"
again. Replication supposedly started at 8:40, ended 30 seconds late
The URL Classify Update Processor can take a URL and split it into pieces,
including the host name.
http://lucene.apache.org/solr/4_3_0/solr-core/org/apache/solr/update/processor/URLClassifyProcessorFactory.html
Unfortunately, the Javadoc is sparse, not even one example.
I have some examples i
Atomic update. For example:
curl http://localhost:8983/solr/update?commit=true \
-H "Content-Type: application/json" -d '
[{"id": "text-1", "text_ss": {"set": null}}]'
(From the book!)
That's for one document. If you want to do that for all documents, you will
have to iterate yourself.
Just from the string field to a "text" field and use standard
tokenization, then you can search the text field for "youtube" or even
"something" that is a component of the URL path. No wildcard required.
-- Jack Krupansky
-Original Message-
From: Amit Sela
Sent: Thursday, June 27, 2
The stored and indexed string is actually a url like "
http://www.youtube.com/somethingsomething";.
It looks like removing the quotes does the job: iframe:*youtube* or am I
wrong ? For now, performance is not an issue, but accuracy is and I would
like to know for example how many URLS have iframe s
Extract the domain (the "main URL" you mention) into its own indexed field and
use field collapsing/grouping: http://wiki.apache.org/solr/FieldCollapsing
Erik
On Jun 27, 2013, at 08:18 , Wojciech Kapelinski wrote:
> I'm looking for a neat solution to replace default multiple results fro
In my solr schema there is one dynamic field.
So I have one doc value,
docs: [
{
last_name: "Jain",
state_name: "rajasthan",
mobile_no: "234534564621",
id: "4",
jobs_6554: "6554",
},...]
Now I just want to delete one column, means jobs_6554 not the complete doc.
How it can possible in solr.
No, you cannot use wildcards within a quoted term.
Tell us a little more about what your strings look like. You might want to
consider tokenizing or using ngrams to avoid the need for wildcards.
-- Jack Krupansky
-Original Message-
From: Amit Sela
Sent: Thursday, June 27, 2013 3:33
I'm looking for a neat solution to replace default multiple results from
single domain in SERP
somepage.com/contact.html
somepage.com/aboutus.html
otherpage.net/info.html
somepage.com/directions.html etc
with only one result per each domain [main URL by default]
somepage.com
otherpage.net
compl
Hi,
On Thu, Jun 27, 2013 at 12:23 PM, Robert Krüger wrote:
> Hi,
>
> I have a desktop application where I am "abusing" solr as an embedded
> database accessing it and I am quite happy with everything.
> Performance is more than goog enough for my use case and Solr's query
> capabilities match the
On 27 June 2013 14:12, Mysurf Mail wrote:
> I just configured with the caching and it works mighty fast now.
> Instead of unbelievable amount queries it queris only 4 times.
> CPU usage has moved from the db to the solr computer but only for a very
> short time.
>
> Problem :
> I dont see the mult
There is a low level protocol that defines client–server protocol for
searching and retrieving information from remote computer databases called
as Z39.50. Due to Solr is a commonly used search engine (beside being a
NoSQL database) is there any protocol for (I don't mean a low level
protocol, z39.
As much as possible, use new configs. Take fieldType definitions from
your 4.x example dir, don't use the old ones. e.g. if you use the old
date field type, it won't be usable in various ways (e.g. in the MS()
function).
Upayavira
On Thu, Jun 27, 2013, at 11:00 AM, Sandeep Gupta wrote:
> Thanks a
Hi,
I have a desktop application where I am "abusing" solr as an embedded
database accessing it and I am quite happy with everything.
Performance is more than goog enough for my use case and Solr's query
capabilities match the requirements of my app quite well. However, I
have the well-known perfo
Thanks again Shawn for your comments.
I am little worried about the multi threading of web application which uses
servlets.
I also found one of your explanation (please confirm the same whether its
your comment only) in
http://lucene.472066.n3.nabble.com/Memory-problems-with-HttpSolrServer-td4060
* I have created a new RequestHandler and added the list of the shards :
...
localhost:8780/apache-solr/leg0,localhost:8780/apache-solr/leg1,localhost:8780/apache-solr/leg2,localhost:8780/apache-solr/leg3,localhost:8780/apache-solr/leg4,localhost:8780/apache-solr/leg5
...
* In the url, I replac
can you give an example?
On Thu, Jun 27, 2013, at 09:08 AM, Dotan Cohen wrote:
> On a Solr 4.1 install I see that queries with use the fq parameter
> take a long time (upwards of 120 seconds), both on the standard Lucene
> query parser and also with edismax. I have added the {!cache=false}
> local
On thinking about this, isn't it a potentially more serious problem,
especially in view of the NRT support which Solr now offers?
If a server crashes (hard), ZK detects this using the heartbeat, and would
remove the "/live_node", which would trigger a leader election for this
shard.
But if we soft
If what you want to do is create an index that can later be used by
Solr, then create the index with Solr. Solr has constraints about how a
Lucene index is created that you would replicate and would create a huge
amount of work.
SolrJ does have the 'embedded mode' in which the Solr itself runs in
I just configured with the caching and it works mighty fast now.
Instead of unbelievable amount queries it queris only 4 times.
CPU usage has moved from the db to the solr computer but only for a very
short time.
Problem :
I dont see the multi value fields (Inner Entities) anymore
This is my conf
I'm not a hibernate fan either to be honest, but in the Java world if
you have a good model oriented design I'm sure you prefer to map it to a
DB using JPA2 for example, in our case, we use EclipseLink which for
JPA2 I find it simpler and faster than Hibernate, now, I'm not sure of
how many JPA
Hi all,
I am new to SOLR. I have been working through the SOLR 4 Cookbook and my
experiences so far have been great.
I have worked through the extraction of PDF data recipe, and the Data import
recipe. I would now like to join these two things, i.e. I would like to do a
data import from a Data
On a Solr 4.1 install I see that queries with use the fq parameter
take a long time (upwards of 120 seconds), both on the standard Lucene
query parser and also with edismax. I have added the {!cache=false}
localparam to the filter query, but this does not speed up the query.
Putting all the search
I have done this - upgraded a 1.4 index to 3.x then on to 4.x. It
worked, but...
New field types have been introduced over time that facilitate new
functionality. To continue to use an upgraded index, you need to
continue using the old field types, and thus loose some of the coolness
of newer vers
I'm looking to search (in the solr admin search screen) a certain field
for:
*youtube*
I know that leading wildcards takes a lot of resources but I'm not worried
with that
My only question is about the syntax, would this work:
field:"*youtube*" ?
Thanks,
I'm using Solr 3.6.2
On 27 June 2013 12:32, Mysurf Mail wrote:
>
> I have a relational database model
> This is the basics of my data-config.xml
>
>
>
>
> query="select SKU
> FROM [TableB]
> INNER JOIN ...
> ON ...
> INNER JOIN ...
> ON ...
> WHERE ... AND ...'">
>
>
>
> Now, this takes a lot
I have a relational database model
This is the basics of my data-config.xml
Now, this takes a lot of time.
1 rows in the first query and then each other inner entities are
fetched later (around 10 rows each).
If I use a db profiler I see a the three inner entities que
74 matches
Mail list logo