Hi Christopher,
when all needed jars are included, you can only have wrong paths in
your solrconfig.xml
Regards
Vadim
2012/1/26 Stanislaw Osinski :
> Hi,
>
> Can you paste the logs from the second run?
>
> Thanks,
>
> Staszek
>
> On Wed, Jan 25, 2012 at 00:12, Christopher J. Bottaro > wrote:
>
>
I'm talking about 2 GB files. It means that I'll have to allocate something
bigger than that for the JVM? Something like 2,5 GB?
Thanks,
Augusto Camarotti
>>> Erick Erickson 1/25/2012 1:48 pm >>>
Mostly it depends on your container settings, quite often that's
where the limits are. I don't t
Hi
If I have added some document to solr, but not done explicit commit yet,
and I get a power outage, will I then loose data? Or asked in another
way, does data go into persistent store before commit? How to avoid
possibility of loosing data?
Does solr have some kind of session concept, so t
Hi,
Yep, anything added between two commits must be regarded as lost in case of
crash.
You can of course minimize this interval by using a low "commitWithin". But
after a crash you should always investigate whether the last minutes of adds
made it.
A transaction log feature is being developed,
On Fri, Jan 27, 2012 at 3:25 PM, Jan Høydahl wrote:
> Hi,
>
> Yep, anything added between two commits must be regarded as lost in case of
> crash.
> You can of course minimize this interval by using a low "commitWithin". But
> after a crash you should always investigate whether the last minutes
Hi, all,
I could use a little education here, if you'd be so kind. My queries
without a field-name qualifier (such as "schedule a pickup", no quotes)
don't return any values UNLESS I've defined copyFields as illustrated
below. The queries work sufficiently well when those fields are defined,
so I
Hello!
If you don't specify the field, the query will be made against the
default search field defined in the schema.xml file. So, when the
default search field is empty (no copy fields) then there are no
search results.
--
Regards,
Rafał Kuć
> Hi, all,
> I could use a little education here,
Hi List,
We use Solr 4.0.2011.12.01.09.59.41 and have a dataset of roughly 40 GB.
Every day we produce a new dataset of 40 GB and have to switch one for
the other.
Once the index switch over has taken place, it takes roughly 30 min for Solr
to reach maximum performance. Are there any hardware or
Rafał,
Thanks for your response.
I defined what I think you're referring to as "the default search field" as
text
I'm confused about how this works. I defined that field "text" to be of
fieldType "text_en_splitting". I don't understand how associating "text"
with anything can work unless I ALSO
Hello!
When you don't specify a field or fields you want to search against,
Solr will use the one set as the default in schema.xml file (the one
defined with ).
So, you have the following field:
When you don't specify copyField's this field won't have any values.
So when searching without speci
Rafal,
Thanks for your response.
I defined what I think you're referring to as "the default search field" as
text
I'm confused about how this works. I defined that field "text" to be of
fieldType "text_en_splitting". I don't understand how associating "text"
with anything can work unless I ALSO
Would sorting by distance work or are you just looking to say something
"only give me all the places in New York"? Might frange work as
a filter query in that case where the distance you provide is XXX
kilometers so you're effectively excluding everything over, say, 160
kilometers from your city-of
Why not just sort on date and take the first doc returned in the list?
Best
Erick
On Thu, Jan 26, 2012 at 10:33 AM, Peter Markey wrote:
> Hello,
>
> I am building a custom component in Solr and I am trying to construct a
> query to get the latest (based on a date field) DocID using SolrIndexSear
You need to provide the relevant bits of your configuration
file for anyone to help I think In particular the
sharding-relevant configurations.
Best
Erick
On Thu, Jan 26, 2012 at 11:29 AM, ramin wrote:
> Hello,
>
> I've gone through the list and have not found the answer but if it is a
> rep
Hmmm, I'd go considerably higher than 2.5G. Problem is you the Tika
processing will need memory, I have no idea how much. Then you'll
have a bunch of stuff for Solr to index it etc.
But I also suspect that this will be about useless to index (assuming
you're talking lots of data, not say just the
Dan,
I can suggest a solution that should help. VeloBit enables you to add SSDs
to your servers as a cache (SSD will cost you $200, per server should be
enough). Then, assuming a 100MB/s read speed from your SAS disks, you can
read 50GB data into the VeloBit HyperCache cache in about 9 mins (this
You say warming queries didn't help? How do those look like? Make sure you
facet and sort in all of the fields that your application allow
faceting/sorting. The same with the filters. Uninversion of fields is done
only when you commit, but warming queries should help you here.
Tomás
On Fri, Jan 27
Hi Mark,
Did you get a chance to look into the issues with running the embedded
Zookeeper ensemble, as per Example C, from the
http://wiki.apache.org/solr/SolrCloud2
Hi All,
Did anyone else run multiple shards with embedded zk ensemble successfully? If
so would like some tips on any issues that
hi,
I have setup my solr installation to run with jetty and I am trying to
import an sqlite database in the solr index. I have setup a jdbc sqlite
driver:
The schema:
id
thread_title
I kickstart the import process with
"wget http://localhost:8080
On Fri, Jan 27, 2012 at 7:39 PM, mathieu lacage
wrote:
>
> It seems to work but the following command reports that only 499 documents
> were indexed (yes, there are many more documents in my database):
>
And before anyone asks:
1
499
0
2012-01-27 19:37:16
Indexing completed. Added/Updated: 499 d
Thanks for the reply guys (Cameron, David and Anderson).
I will go through the details of using multiple cores.
Thanks
Nitin
--
View this message in context:
http://lucene.472066.n3.nabble.com/Multiple-Data-Directories-and-1-SOLR-instance-tp3691644p3694412.html
Sent from the Solr - User mailin
Hi Chantal,
Thanks for your response. Isn't the DisMaxQParserPlugin the default
parser, when none is specified? I am using Solr version 3.4.
Thanks,
Dipti
On 1/23/12 3:33 AM, "Chantal Ackermann"
wrote:
>Hi Dipti,
>
>just to make sure: are you aware of
>
>http://wiki.apache.org/solr/DisMaxQParser
Hello,
this is probably a very basic question, but I haven't found an answer in my
searches.
My search engine runs fine, but I want it to return only one hit if a user
searches for a specific search string. I.e. user searches for "xyz" and,
instead of being presented hundreds of hits, is only sho
> My search engine runs fine, but I want it to return only one
> hit if a user
> searches for a specific search string. I.e. user searches
> for "xyz" and,
> instead of being presented hundreds of hits, is only shown
> one
> pre-configured result.
>
> Is there a config file somewhere to set such a
I have a project where we need to search 1B docs and still have results <
700ms. The problem is, we are using geofiltering and that is happening *
before* the queries, so we have to geofilter on the 1B docs to restrict our
set of docs first, and then do the query on a name field. But it seems that
Sure. So it is really simple. Following the Solr example for setting up two
shards and pushing some xml docs to each one and then doing a distributed
query (http://wiki.apache.org/solr/DistributedSearch), it works perfectly.
Now in my case the indices are being built outside of Solr. So basically I
No. Look in solrconfig.xml for the
and you'll see
that it's not. Appending &debugQuery=on will also
show this.
Best
Erick
On Fri, Jan 27, 2012 at 12:18 PM, Dipti Srivastava
wrote:
> Hi Chantal,
> Thanks for your response. Isn't the DisMaxQParserPlugin the default
> parser, when none is spec
Do all of the documents have unique id fields?
On Fri, Jan 27, 2012 at 10:44 AM, mathieu lacage
wrote:
> On Fri, Jan 27, 2012 at 7:39 PM, mathieu lacage
> wrote:
>
>>
>> It seems to work but the following command reports that only 499 documents
>> were indexed (yes, there are many more documents
Are there any exceptions in the Solr log? Is it possible the JSON
exporter is choking when it wants to escape gunky characters in the
final text?
On Wed, Jan 25, 2012 at 1:40 PM, Erick Erickson wrote:
> Two things:
> 1> I suspect it's your servelet container rather than Solr since your JSON
> is
I just want to verify some of the features in regards to SolrCloud
that are now on Trunk
documents added to the cluster are automatically distributed amongst
the available shards (I had seen that Yonik had ported the Murmur
hash, but I didn't see that on trunk, what is being used and where can
I l
Hi Dan,
I think this may be your problem:
> Every day we produce a new dataset of 40 GB and have to switch one for the
> othe
If you really replace an index with a new index one a day, you throw away all
the hard work the OS has been doing to cache hot parts of your index in
memory. It takes
If a query matches one or more values of a multivalued field, is it
possible to get the indexes back for WHICH values? For example, for a
document with a multivalue field having ["red", "redder", "reddest",
"yellow", "blue"] as its value, if "red" is the query, could we know that
values 0,1, and 2
Le 28 janv. 2012 à 05:17, Lance Norskog a écrit :
> Do all of the documents have unique id fields?
yes.
>
> On Fri, Jan 27, 2012 at 10:44 AM, mathieu lacage
> wrote:
>> On Fri, Jan 27, 2012 at 7:39 PM, mathieu lacage
>> wrote:
>>
>>>
>>> It seems to work but the following command reports
33 matches
Mail list logo