Thank you very much, I found all the information I needed!
Kind regards,
Federico
On 07/15/2013 04:53 PM, Shawn Heisey wrote:
On 7/15/2013 3:08 AM, Federico Ragona wrote:
> Hi,
>
> I'm trying to write a validation test that reads some statistics by
> querying
> Solr 4.3 via HTTP, namely the numb
You've specified bootstrap_confdir and the same collection.configName on
all your cores, so as each of them start, each will be uploading its own
configuration to the collection1_conf area of ZK, so they will all be
overwriting each other.
Are your 4 cores replicas of the same collection or are th
Hey Guys,
I've finally finished my Spring Java application that uses SOLR for
searches and just had performance related question about SOLR. I'm indexing
exactly 1000 *OR* 2000 records every second. Every record having 13 fields
including 'id'. Majority of the fields are solr.StrField (no filters)
Indeed we are using UNLOAD of cores before shutting down extra replica
nodes, works well but already said, it needs such nodes to be up.
Once UNLOADed it is possible to stop them, works well for our use case.
But if nodes are already down, maybe it is possible to manually create
and upload a clea
Originally i was running a single solr 4.3 instance with 4 cores ... and now
starting to learn about solrCloud and thought I could setup number of
shards=1 (since its a single instance) and same 4 cores can be converted to
4 collections on the same single shard same single instance.
How do I defi
If you are not defining global Java startup parameters do not include them
at setenv.sh. Pass that arguments as parameters when you startup your jar.
2013/7/17 smanad
> Originally i was running a single solr 4.3 instance with 4 cores ... and
> now
> starting to learn about solrCloud and thought
Hi
It will not affect the performance. We are doing this regularly. If you do
optimize and search then there may be some impact.
Regards
Aditya
www.findbestopensource.com
On Wed, Jul 17, 2013 at 12:52 PM, Ayman Plaha wrote:
> Hey Guys,
>
> I've finally finished my Spring Java application th
I run a query at my Solr 4.2.1 SolrCloud:
/solr/select?q=*:*&rows=300&wt=csv&fl=url&sort=url asc
result is as follows:
http://goethetc.blogspot.com/
http://about.deviantart.com/contact/
http://browse.deviantart.com/designbattle/
http://browse.deviantart.com/digitalart/
http://hayathepbahar.blogs
Remove the WDF from the analysis chain, it's not going to work with multiple
tokens.
-Original message-
> From:Furkan KAMACI
> Sent: Wednesday 17th July 2013 11:55
> To: solr-user@lucene.apache.org
> Subject: Why "Sort" Doesn't Work?
>
> I run a query at my Solr 4.2.1 SolrCloud:
>
> /
Hi Markus;
This is default schema at Nutch. Do you mean there is a bug with schema?
2013/7/17 Markus Jelsma
> Remove the WDF from the analysis chain, it's not going to work with
> multiple tokens.
>
> -Original message-
> > From:Furkan KAMACI
> > Sent: Wednesday 17th July 2013 11:55
I can now approach http://localhost:8080/solr-4.3.1/#/, thanks!!
I also noticed you mentioning something about a data import handler.
Now, what I will be requiring after I've completed the basic setup of
Tomcat6 and Solr431 I want to migrate my Solr350 (now running on Cygwin)
cores to that envir
Yes, thanks Shawn. I know I can use collections HTTP API to set number of
shards, but the problem with that is it is not easily scriptable so that the
entire cluster can be setup in automated fashion - the script(s) will need to
wait until the SOLR nodes are up and running before using the coll
No, there is no bug in the schema, it is just an example and provides the most
common usage only; sort by score.
-Original message-
> From:Furkan KAMACI
> Sent: Wednesday 17th July 2013 12:10
> To: solr-user@lucene.apache.org
> Subject: Re: Why "Sort" Doesn't Work?
>
> Hi Markus;
>
>
I am out of the office until 18/07/2013.
I will respond to your query on my return,
Thanks
Siobhan
Note: This is an automated response to your message "Re: Solr is not
responding on deployment in tomcat" sent on 17/07/2013 1:33:56.
This is the only notification you will receive while this per
Hi,
I think I will also wait for other people reply as I do not have much idea
now.
I suggested the things because I did it recently but I have only one
collection (default one) .
As you said and I can guess...
you have multiple collections like tt, shop and home in one solr instance..
By default
Thanks Aditya, can I also please get some advice on hosting.
- What *hosting specs* should I get ? How much RAM ? Considering my
- client application is very simple that just register users to database
and queries SOLR and displays SOLR results.
- simple batch program adds the 1000 OR
Hi Markus;
What is that score? It is not listed at schema. Is it document boost?
2013/7/17 Markus Jelsma
> No, there is no bug in the schema, it is just an example and provides the
> most common usage only; sort by score.
>
> -Original message-
> > From:Furkan KAMACI
> > Sent: Wednesda
No, just the usual score calculated by Lucene's Similarity impl.
-Original message-
> From:Furkan KAMACI
> Sent: Wednesday 17th July 2013 13:39
> To: solr-user@lucene.apache.org
> Subject: Re: Why "Sort" Doesn't Work?
>
> Hi Markus;
>
> What is that score? It is not listed at schema.
Dear Usergroup
I am getting an out of memory exception in the following scenario.
I have 4 sql tables: patient, visit, study and image that will be denormalized
for the solr index
The solr index looks like the following
|p_id |p_lastname|v_id |v_na
Hi Erick,
I modified the SOLR schema file for the field as follows and re-indexed the
schema,
My previous scenario seems to be working fine i.e., when I
I don't know the code well, but anything that
mentions "index based spellcheck" would
presumably require re-indexing.
But I'd also guess it depends on the changes.
Any changes to _how_ the index is _used_
shouldn't require re-indexing. But changing
how the tokens are put _into_ the index should.
Where are you getting the syntax
freq:termfreq(product,'spider')
? Try just
termfreq(product,'spider')
you'll get an element in the doc labeled 'termfreq', at least
I do.
Best
Erick
On Tue, Jul 16, 2013 at 1:03 PM, Tony Mullins wrote:
> OK, So thats why I cannot see the FunctionQuery fields in
Roman:
I think that SOLR-1913 is completely different. It's
about having a field in a document and being able
to do bitwise operations on it. So say I have a
field in a Solr doc with the value 6 in it. I can then
form a query like
{!bitwise field=myfield op=AND source=2}
and it would match.
You'r
bq: "Rocket Banana (Single)" should be first because its the closest to "Rocket
Banana".
OK, you've given us nothing to go on here. "it's closest" doesn't mean
anything, it's just someone waving their hands and saying "because I
like it better".
I'm being deliberately obtuse here and trying to th
Yes, the Word Delimiter filter does in fact break up a token into discrete
words. In fact it seems antithetical that you are combining the keyword
tokenizer that doesn't break up a string into words with the WDF that does.
Maybe you should drop back to standard tokenization coupled with the Edg
Hi list,
I have set up two cores (=collections):
http://localhost:8983/solr/core0
http://localhost:8983/solr/core1
In addition the following has been set up:
http://localhost:8984/solr/core0
http://localhost:8984/solr/core1
I'm trying to query the first via the second like this:
http://localho
In general, sorting doesn't work well for multivalued and tokenized fields.
You need to copy your tokenized url to a "utl_str" string field and then
sort that field.
-- Jack Krupansky
-Original Message-
From: Furkan KAMACI
Sent: Wednesday, July 17, 2013 5:54 AM
To: solr-user@lucene.a
iorixxx,
Thanks for pointing me in the direction of the QueryElevation component. If
it did not require that the target documents be keyed by the unique key
field it would be ideal, but since our Sku field is not the Unique field (we
have an internal id which serves as the key while this is the cl
Hi all,
What I find very 'sad' is that Lucene/SOLR contain all the necessary
components for handling multi-token synonyms; the Finite State Automaton
works perfectly for matching these items; the biggest problem is IMO the
old query parser which split things on spaces and doesn't know to be
smarte
On 7/17/2013 1:22 AM, Ayman Plaha wrote:
>*will this effect the query performance of the client website if the
>index grew to 10 million records ? I mean while the commit is happening
>does that *effect the performance of queries* and how will this effect
>the queries if the index g
At my indexing process to my SolrCloud(Solr 4.2.1) from Hadoop I got an
error. What is the reason, does it results with document loss for indexing?
ERROR - 2013-07-17 16:30:01.453; org.apache.solr.common.SolrException;
java.lang.RuntimeException: [was class org.eclipse.jetty.io.EofException]
early
I have crawled some web pages and indexed them at my SolrCloud(Solr 4.2.1).
However before I index them there was already some indexes. I can calculate
the difference between current and previous document count. However it
doesn't mean that I have indexed that count of documents. Because urls of
we
Another problem in addition to dynamicField being declared in the wrong
place, is that you've declared that your geoFindspot field is
multi-valued. LatLonType can't handle that. Use location_rpt in the
example schema to get a multi-value capable geo field.
~ David
On 7/15/13 5:10 PM, "Scott Vand
On 7/17/2013 8:02 AM, Furkan KAMACI wrote:
> At my indexing process to my SolrCloud(Solr 4.2.1) from Hadoop I got an
> error. What is the reason, does it results with document loss for indexing?
>
> ERROR - 2013-07-17 16:30:01.453; org.apache.solr.common.SolrException;
> java.lang.RuntimeException
To the best of my knowledge, there is no patch or collection of patches
which constitutes a "working solution" - just partial solutions.
Yes, it is true, there is some FST work underway (active??) that shows
promise depending on query parser implementation, but again, this is all a
longer-term
I don't think that breakdown is readily available from Solr.
Sounds like a good Jira request for improvement in the response.
-- Jack Krupansky
-Original Message-
From: Furkan KAMACI
Sent: Wednesday, July 17, 2013 10:06 AM
To: solr-user@lucene.apache.org
Subject: How can I learn the t
I would also note that the LucidWorks Search query parser implements
query-time synonym phrases. I don't know if anybody has anything better than
that. Unfortunately, that is proprietary and is kind of a workaround for
current Lucene/Solr limitations than a long-term solution.
-- Jack Krupansk
DirectSorlSpellChecker does not create a dictionary. It uses the field you
specify and uses the Lucene term dictionary. It uses the some of the same code
Fuzzy Search uses to calculate distance between user input and indexed terms.
If you're wondering about the affect of configuration changes
I will open a Jira for it and apply a patch, thanks.
2013/7/17 Jack Krupansky
> I don't think that breakdown is readily available from Solr.
>
> Sounds like a good Jira request for improvement in the response.
>
> -- Jack Krupansky
>
> -Original Message- From: Furkan KAMACI
> Sent: Wedne
Wow! Thanks Shawn. That's great info and helped and thanks for the
SolrPerformance article link, great article, helped a lot :)
I can't use Cloud hosting now since they charge on basis of the memory used
and it will be too expensive and like you said RAM and SSD is what I need
for SOLR performance
All clear. There seems to be a mis-config on my side as the vanilla solr
4.3.1 package works just fine with the described setup.
On Wed, Jul 17, 2013 at 4:21 PM, Dmitry Kan wrote:
> Hi list,
>
> I have set up two cores (=collections):
>
> http://localhost:8983/solr/core0
> http://localhost:8983
Maybe as a first step, it would be nice to have logging that summarized the
count of actual inserts, replacements, actual deletions, and even
atomic/partial updates.
The LogUpdateProcessor outputs some information, like a subset of the
document IDs, but not the insert vs. replace/update counts
OK, let's do a simple test instead of making claims - take your solr
instance, anything bigger or equal to version 4.0
In your schema.xml, pick a field and add the synonym filter
in your synonyms.txt, add these entries:
hubble\0space\0telescope, HST
ATTENTION: the \0 is a null byte, you must
Remember, this is the "users" list, not the "dev" list. Users want to know
what they can do and use off the shelf today, not what "could" be developed.
Hopefully, the situation will be brighter in six months or a year, but
today... is today, not tomorrow.
(And, in fact, users can use LucidWork
As I don't see in the heads of the users, I can make different assumptions
- but OK, seems reasonable that only minority of users here are actually
willing to do more (btw, I've received coding advice in the past here in
this list). I am working under the assumption that Lucene/SOLR devs are
swampe
On 7/17/2013 9:35 AM, Ayman Plaha wrote:
> In my solrconfig.xml I've got these caching config by default which I
don't
> think I will need. Since my index is updated with new documents every 3
> minutes caching anything would be pointless. Am I on the right ?
>
>
That depends on how many queri
Hi guys.
I need a lot of pdf, doc, txt files.
Now I index manually below command.
# PDF INDEX
curl
"http://localhost:8983/solr/update/extract?stream.file=/opt/solr/documents/test.pdf&literal.doc_id=pdf_1&commit=true";
# TXT INDEX
curl
"http://localhost:8983/solr/update/extract
You don't seem to be too creative with your doc_id values, so perhaps you
can use Solr 4's post.jar recursive option:
http://wiki.apache.org/solr/ExtractingRequestHandler#SimplePostTool_.28post.jar.29
Otherwise, you need to correlate the ID and the source file somehow, so you
probably need a file
This is not a problem per se, just want to verify that we're not able
to specify which server shard splits are created as of 4.3.1? From
what I've seen, the new cores for the sub-shards are created on the
leader of the shard being split.
Of course it's easy enough to migrate the new sub-shards to
By all means, feel free to write about how users can in fact do custom code
for Solr, but just keep a clear distinction between what could be developed
and what is actually available off the shelf.
Yes, this list does have a mix of pure users and those who are willing to
customize code as well
On Wed, Jul 17, 2013 at 12:26 PM, Timothy Potter wrote:
> This is not a problem per se, just want to verify that we're not able
> to specify which server shard splits are created as of 4.3.1? From
> what I've seen, the new cores for the sub-shards are created on the
> leader of the shard being spl
Ok, thanks for the answer Yonik. After looking closer at the index
splitting code, definitely seems like you wouldn't want to pay the
network I/O cost when creating the sub-shard indexes. Might be cool to
be able to specify a different local disk path for the new cores so
that we can get some extra
Solr User Group,
I would like to return a hierarchical data relationship when somebody
queries for a parent doc in solr. This sort of relationship doesn't currently
exist in our core as the use-case has been to search for a specific document
only. However, here's kind of an example
You've asked this question several times w/o really providing a clear
concrete set of examples of what you are trying to do ... in several of
your duplicated threads, people have suggested using hte switch qparser,
and you've dismissed that suggestion for various reasons that you also
haven't
Roman,
As a developer, I understand where you are coming from. My issue is that I
specialize in .NET, haven't done java dev in over 10 years. As an
organization we're new to solr (coming from endeca) and we're looking to use
it more across the organization, so for us, we are looking to do the clas
LucidWorks Search:
http://docs.lucidworks.com/display/lweug21/Synonyms%2C+Stop+Words%2C+and+Stemming
"There can be an unlimited number of terms and phrases which are defined as
synonyms. If the Lucid query parser encounters any of those terms or phrases
in a query term list, additional (option
Hello,
I was wondering if people had experience using solr with jetty and a java
service wrapper for automatic deployment? I thought a service wrapper might
be included in the solr download, but I didn't see one.
How does one search the mailing list archive? Are there any previous topics
about th
Which Operating System? I have a write up for Windows:
http://blog.outerthoughts.com/2013/07/setting-up-apache-solr-on-windows-as-a-service/
To search the mailing list, there is several options, try
http://search-lucene.com/ and narrow down by mailing lists/keywords.
Regards,
Alex.
Personal w
Hello Roman and all,
> sorry, haven't the previous thread in its entirety, but few weeks back
that
> Yonik's proposal got implemented, it seems ;)
http://search-lucene.com/m/Fa3Dg14mqoj/bitset&subj=Re+Solr+large+boolean+filter
In that post I see a reference to your plugin BitSetQParserPlugin, rig
How i use fuzzy? Could you give an example, please?
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-optimize-a-search-tp4077531p4078708.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
There was a thread about viewing Solr Wiki offline, About 6 months ago. I'm
intersted, too.
It seems that a manual (cron?) dump will do the work...
Would it be too much to ask that one of the admins will manually create
such a dump? (http://moinmo.in/HelpOnMoinCommand/ExportDump)
Otis, is t
Erick,
Awesome answer, buddy. I totally agree with you.
Right now, i'm facing this problem...just someone waving their hands and
saying "because I
like it better"..
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-optimize-a-search-tp4077531p4078711.html
Sent from th
Not fully following the problem, but is it similar to:
http://robotlibrarian.billdueber.com/boosting-on-exactish-anchored-phrase-matching-in-solr-sst-4/
?
Regards,
Alex.
Personal website: http://www.outerthoughts.com/
LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch
- Time is the qua
I'm using Solr 3!
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-optimize-a-search-tp4077531p4078715.html
Sent from the Solr - User mailing list archive at Nabble.com.
So does the example! Anyway, this is just an attempt to give additional
options.
Regards,
Alex.
Personal website: http://www.outerthoughts.com/
LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch
- Time is the quality of nature that keeps events from happening all at
once. Lately, it doe
Thanks Erick!
I have added the instructions for running SolrCloud on Jboss:
http://wiki.apache.org/solr/SolrCloud%20using%20Jboss
I will refine the instructions further, and also post some screenshots.
Thanks.
On Sun, Jul 14, 2013 at 5:05 AM, Erick Erickson wrote:
> Done, sorry it took so lon
Hi Dave,
On Wed, Jul 17, 2013 at 2:03 PM, dmarini wrote:
> Roman,
>
> As a developer, I understand where you are coming from. My issue is that I
> specialize in .NET, haven't done java dev in over 10 years. As an
> organization we're new to solr (coming from endeca) and we're looking to
> use
Hello Oleg,
On Wed, Jul 17, 2013 at 3:49 PM, Oleg Burlaca wrote:
> Hello Roman and all,
>
> > sorry, haven't the previous thread in its entirety, but few weeks back
> that
> > Yonik's proposal got implemented, it seems ;)
>
> http://search-lucene.com/m/Fa3Dg14mqoj/bitset&subj=Re+Solr+large+bool
: There was a thread about viewing Solr Wiki offline, About 6 months ago. I'm
: intersted, too.
:
: It seems that a manual (cron?) dump will do the work...
:
: Would it be too much to ask that one of the admins will manually create
: such a dump? (http://moinmo.in/HelpOnMoinCommand/ExportDump)
: Yes that worked, thanks Alan. The consistency of this api is "challenging".
It's important to understand what's happening here.
fl, by default, only returns "stored" fields -- but you can also request
"psuedo-fields" such as the results of functions, or the result of a "Doc
Transformer" ...
On 7/17/2013 8:06 AM, Furkan KAMACI wrote:
> I have crawled some web pages and indexed them at my SolrCloud(Solr 4.2.1).
> However before I index them there was already some indexes. I can calculate
> the difference between current and previous document count. However it
> doesn't mean that I have
Chris, the confusion from my perspective is the general inconsistency and
natural growth of the API which is somewhat expected based upon its history.
Obviously this isnt sql, there is no ansi body defining the query language. I
understand well the difference between stored, indexed etc.
Goin
You've found a general bug in the grouping code, and i've opened SOLR-5046
to track it (no idea how hard it is to fix) but in general keep in mind
the major caveat assocaited with grouping and distributed search ...
https://wiki.apache.org/solr/SolrCloud#Known_Limitations
"The Grouping feature
: reference to a few "special" values, such as "id" and "score". Neither
: of them are technically "stored" fields either, but afaik you dont need
: to use "field(id), field(score)" for those. Can you honestly say that
: is consistent?
Nope.
I wasn't defending the quirks of the API, or tryi
I
have a newbie question on integrating Solr with Hadoop.
There are some vendors like Cloudera/MapR who have announced Solr Search
for Hadoop.
If I use the Apache distro, how can I use Solr Search on docs in HDFS/Hadoop
Is there a tutorial on how to use it or getting started.
I am using Flume
I am trying to use Filebased and index based spell checker and getting this
exception "All checkers need to use the same StringDistance."
They work fine as expected individually but not together.
Any pointers?
-Manasi
--
View this message in context:
http://lucene.472066.n3.nabble.com/Except
Eric ,
In freq:termfreq(product,'spider') , freq is alias for 'termfreq' function
query so I could have that field with name 'freq' in document response.
this is my code which I am using to get document object and there is no
termfreq field in its fields collection.
DocList docs = rb.getResults().
77 matches
Mail list logo