I have a scenario as follows:
There are 2 separate threads where each will try to update the same document
in a single index for 2 separate fields, for which we are using atomic OR
in-place updates. For e.g.
id is the unique field in the index
thread-1 will update following info:
id:1001
field-1
SOLR has a nice analysis page. You can use it to get insight what is
happening after each filter is applied at index/search time
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Problem-with-Synonyms-tp4087905p4087915.html
Sent from the Solr - User mailing
>>Is there a way to omitNorms and still be able to use {!boost b=boost} ?
OR you could let /omitNorms="false"/ as usual and have your custom
Similarity implementation with the length normalization method overridden
for using a constant value of 1.
Regards
Pravesh
--
View
ttp://wiki.apache.org/solr/CommonQueryParameters>
http://wiki.apache.org/solr/SimpleFacetParameters
<http://wiki.apache.org/solr/SimpleFacetParameters>
BTW, almost all the job sites out there (whether small/medium/big) use
SOLR/lucene to power their searches :)
Best
Pravesh
--
View thi
Update:
Also remove your range queries from the main query and specify it as a
filter query.
Best
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/TooManyClauses-maxClauseCount-is-set-to-1024-tp4056965p4056969.html
Sent from the Solr - User mailing list archive at
Just increase the value of /maxClauseCount/ in your solrconfig.xml. Keep it
large enough.
Best
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/TooManyClauses-maxClauseCount-is-set-to-1024-tp4056965p4056966.html
Sent from the Solr - User mailing list archive at
SOLR's QTime represents actual time it spent on searching, where as your c#
client response time might be the total time spent in sending HTTP request
and getting back the response(which might also include parsing the results)
.
Regards
Pravesh
--
View this message in context:
http://l
Do logstash/graylog2 do log processing/searching in real time? Or can scale
for real time need?
I guess harshadmehta is looking for real-time indexing/search.
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/need-basic-information-tp4004588p4004996.html
Sent
=true&facet.field=action
This way you have facet count for created/updated/deleted etc.
Hope this is what u r looking for.
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/need-basic-information-tp4004588p4004637.html
Sent from the Solr - User mailing list archive at Nabble.com.
Did you checked SOLR Field Collapsing/Grouping.
http://wiki.apache.org/solr/FieldCollapsing
http://wiki.apache.org/solr/FieldCollapsing
If this is what you are looking for.
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Query-during-a-query
http://lucidworks.lucidimagination.com/display/solr/Distributed+Search+with+Index+Sharding
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Query-Time-problem-on-Big-Index-Solr-3-5-tp4003660p4004630.html
Sent from the Solr - User mailing list archive at Nabble.com.
apply to other
fields which require very less analysis or have limited unique fields.
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Query-Time-problem-on-Big-Index-Solr-3-5-tp4003660p4004437.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi Dhaivat,
JMeter is a nice tool. But it all depends what sort of load are you
expecting, how complex queries are you expecting(sorting/filtering/textual
searches). You need to consider all these to benchmark.
Thanx
Pravedsh
--
View this message in context:
http://lucene.472066.n3.nabble.com
your searches/indexing performing over the time? Are there any
impact?
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/optimum-solr-core-size-tp4004251p4004424.html
Sent from the Solr - User mailing list archive at Nabble.com.
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/solr-indexing-slows-down-after-few-minutes-tp4004337p4004421.html
Sent from the Solr - User mailing list archive at Nabble.com.
We have a 48GB index size on a single shard. 20+ million documents. Recently
migrated to SOLR 3.5
But we have a cluster of SOLR servers for hosting searches. But i do see to
migrate to SOLR sharding going forward.
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3
BTW, Have you changed the MergePolicy & MergeScheduler settings also? Since
Lucene 3.x/3.5 onwards,
there have been new MergePolicy & MergeScheduler implementations available,
like TieredMergePolicy & ConcurrentMergeScheduler.
Regards
Pravesh
--
View this message in context:
h
x27;s for a
single core with multiple index directories in it).
BTW, how big your index(es) are? Total documents? total size? etc. If each
core is small(MBs/ few GBs) then you could merge few of them together.
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Dynami
lters) on both the system(1.4.1
vs 3.5) and found that 3.5 searches takes longer time than the 1.4.1(around
10-20% slower). Haven't done any load test till now
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/SOLR-3-5-Index-Optimization-not-producing-s
#x27;ve set the , true
Is it something related to the new MergePolicy being used with SOLR 3.x
onwards (I suppose it is TieredMergePolicy with 3.x version)? If yes should
i change it to the LogByteSizeMergePolicy?
Does this change requires complete rebuilt OR will do incrementally?
Regards
Pra
Hi Sujit,
The Http parameters ordering is above the SOLR level. Don't think this could
be controlled at SOLR level.
You can append all required values in a single Http param at then break at
your component level.
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabbl
e first term KENTUCKY is searched in the INST_NAME and rest terms
like TECH and PADUCAH are searched in your default search field
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Basic-SOLR-help-needed-tp3759855p375.html
Sent from the Solr - User mailing lis
be the optimal solution for doing it.
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Sorting-and-searching-on-a-field-tp3584992p3587906.html
Sent from the Solr - User mailing list archive at Nabble.com.
a particular
field.?
Since you have multiple shard infra, do the cores share the same
configurations(schema.xml/solrconfig.xml etc.)?? What error/output you are
getting for sharded query?
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-Search-Across-Mult
for the above implementation :)
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Generic-RemoveDuplicatesTokenFilter-tp3581656p3581656.html
Sent from the Solr - User mailing list archive at Nabble.com.
searching done in
milis/secs/mins?? I am trying to understand if your search could already be
performing quite good/OK.
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-improve-facet-search-tp3569910p3570048.html
Sent from the Solr - User mailing list archive
What is the type of the field on which you are getting facets (string, Text,
int, date etc.). Is it multivalued or not?
How many unique values do you have for the field?
What is your filtercache setting in your solrconfig.xml?
Regards
Pravesh
--
View this message in context:
http://lucene
Can u share more info: like what is your H/W infra, CPU, RAM, HDD??
>From where you pick the records/documents to index; RDBMS, Files, Network??
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-using-very-high-I-O-tp3567076p3569903.html
Sent from the S
Go ahead with SOLR based text search. Thats what it is meant for and does it
great.
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-or-SQL-fultext-search-tp3566654p3569894.html
Sent from the Solr - User mailing list archive at Nabble.com.
Was that field multivalued="true" earlier by any chance??? Did you rebuild
the index from scratch after changing it to multivalued="false" ???
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-sorting-issue-can-not-sort-
>>facet.limit=50
your facet.limit seems too high. Do you actually require this much?
Since there a lot of evictions from filtercache, so, increase the maxsize
value to your acceptable limit.
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com
.
Also, avoid using the q=*:* as it implicitly translates to matchalldocsquery
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/how-to-make-effective-search-with-fq-and-q-params-tp3527217p3527535.html
Sent from the Solr - User mailing list archive at Nabble.com.
I Guess,
This has nothing to do with search part. You can post process the search
results(I mean iterate through your results and sum it)
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/best-way-for-sum-of-fields-tp3477517p3486536.html
Sent from the Solr
effect(also set /omitNorms="true"/ for your fields where you need this
feature)
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/to-prevent-number-of-matching-terms-in-contributing-score-tp3486373p3486512.html
Sent from the Solr - User mailing list
Did you rebuild the index from scratch. Since this is index time factor, you
need to build complete index from scratch.
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/to-prevent-number-of-matching-terms-in-contributing-score-tp3486373p3486447.html
Sent from
>1) Is it safe to reuse a single _mgr and _client
across all 28 cores?
both are thread-safe API as per HttpClient specs. You shld go ahead with
this.
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/SolrJ-threading-http-clients-connection-manag
This approach seems fine. You might benchmark it through load test etc.
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Getting-single-documents-by-fq-on-unique-field-performance-tp3440229p3440351.html
Sent from the Solr - User mailing list archive at
Are you posting through HTTP/SOLRJ?
Your script time 'T' includes time between sending POST request -to- the
response fetched after successful response right??
Try sending in small batches like 10-20. BTW how many documents are u
indexing???
Regds
Pravesh
--
View this message
ich have term 1234567 only (facet.query
would apply to the facets,so as to which facet to be picked/shown)
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/inconsistent-results-when-faceting-on-multivalued-field-tp3438991p3440128.html
Sent from the Solr - Us
If I understood correctly, this seems you are wanting facets/hierarchical
facets.
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/hierarchical-synonym-tp344p3440090.html
Sent from the Solr - User mailing list archive at Nabble.com.
in the
item-detail page then option-1 seems better.
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/text-search-and-data-aggregation-thoughts-tp3416330p3421361.html
Sent from the Solr - User mailing list archive at Nabble.com.
Just look into your tomcat logs in more detail.specifically the logs when
tomcat loads the solr application's web context. There you might find some
clues or just post the logs snapshot here.
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/upgrading-1
Can you provide the tomcat logs full stack trace for further assistance
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Problem-while-getting-more-than-100-records-from-solr-1-4-1-tp3415443p3415619.html
Sent from the Solr - User mailing list archive at
This link might help:
http://www.lucidimagination.com/blog/2011/04/01/solr-powered-isfdb-part-8/
http://www.lucidimagination.com/blog/2011/04/01/solr-powered-isfdb-part-8/
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/upgrading-1-4-to-3-x-tp3415044p3415546
/wget /or /curl / locally from your solr server for search
with page size >100, to check it it works.
Regds
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Problem-while-getting-more-than-100-records-from-solr-1-4-1-tp3415443p3415541.html
Sent from the Solr - User
If you mean using DIH? Then you need to have a timestamp column in your DB,
which has to be updated to current timestamp, whenever you are modifying the
record in DB.
For rest just go through the http://wiki.apache.org/solr/DataImportHandler
DIH wiki here
Thanx
Pravesh
--
View this message in
ine the index which i will be making
a search in, in query time?
Then you are just duplicating the index and you might stuck-up in
maintenance issues
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/searching-documents-partially-tp3408429p3411743.html
Sent
Can you clarify following:
1) Is it that: You want to hide some documents from search when user is not
logged-in?
OR
2) Is it that: You want to hide some fields of some documents from search
when user is not logged-in?
For Point 2; one solution can be that while indexing the documents, you can
You count index the date as a text field(or use a new text field to store
date as text) and then try it on this new field
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Hierarchical-faceting-with-Date-tp3394521p3395824.html
Sent from the Solr - User mailing
You can explicitly pass /mm/ for every search, and get it in your response,
otherwise use /debugQuery=true/, it will give you all implicitly used
defaults (but you wouldn't want to use this in production)
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/is-
hat current
>user's friends (source:facebook) purchased/used/...
You can check
https://cwiki.apache.org/confluence/display/MAHOUT/Recommender+Documentation
apache mahout for this purpose. It's got recommendation engine that works
pretty well.
Thanx
Pravesh
--
View this message in c
The Wiki link that you referred is quite old and is not into active
development.
I would prefer the OS based scheduling using cron jobs. You can check below
link.
http://wiki.apache.org/solr/CollectionDistribution
http://wiki.apache.org/solr/CollectionDistribution
Thanx
Pravesh
--
View this
esire. (It's sort of reverse
engineering not supported)
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Viewing-the-complete-document-from-within-the-index-tp3288076p3292111.html
Sent from the Solr - User mailing list archive at Nabble.com.
http://lucene.472066.n3.nabble.com/how-to-update-solr-cache-when-i-delete-records-from-remote-database-td3291879.html
http://lucene.472066.n3.nabble.com/how-to-update-solr-cache-when-i-delete-records-from-remote-database-td3291879.html
--
View this message in context:
http://lucene.472066.n3.na
You would have to delete them from SOLR also, and then commit it (commit will
automatically refresh your caches).
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/how-to-update-solr-cache-when-i-delete-records-from-remote-database-tp3291879p3292074.html
Sent
(cron/scheduled-tasks).
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/what-is-scheduling-why-should-we-do-this-how-to-achieve-this-tp3287115p3292068.html
Sent from the Solr - User mailing list archive at Nabble.com.
This might be of some help:
http://wiki.apache.org/solr/SolrJmx http://wiki.apache.org/solr/SolrJmx
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-could-I-monitor-solr-cache-tp3181317p3181407.html
Sent from the Solr - User mailing list archive at
ster in load balanced manner?
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/SOLR-Shard-failover-Query-tp3178175p3181400.html
Sent from the Solr - User mailing list archive at Nabble.com.
hine
>?can a single core be presented in different shards ?
You might look into following thread:
http://lucene.472066.n3.nabble.com/difference-between-shard-and-core-in-solr-td3178214.html
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Start-parameter-m
ted across a cluster of
machines. So all shards belonging to a single core share same
schema/analysis/search requirements. You go with sharding when index is not
scalable on a single machine, or, when your index grows really big in size.
Thanx
Pravesh
--
View this message in context:
http://luce
commit would be the safest way for making sure the deleted content doesn't
show up.
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Deleted-docs-in-IndexWriter-Cache-NRT-related-tp3177877p3178179.html
Sent from the Solr - User mailing list archive at Nabble.com.
fails, or,
is this handled gracefully by SOLR?
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/SOLR-Shard-failover-Query-tp3178175p3178175.html
Sent from the Solr - User mailing list archive at Nabble.com.
You can use lucene for doing this. It provides TermEnum API to enumerate all
terms of field(s).
SOLR-1.4.+ also provides a special request handler for this purpose. Check
it if that helps
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Is-it-possible-to
You can configure analyzer for 'index-time' & for 'search-time' for each of
your field-types in schema.xml
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-do-I-specify-a-different-analyzer-at-search-time-tp3159463p3165593.html
S
>1. I assume that it's worthwhile to rely on POST method instead of GET
when issuing a search. Right? As I can see, this should work.
We do restrict users search by passing unique id's(sometimes in thousands)
in 'fq' and use POST method
Thanx
Pravesh
--
View this me
If behind proxy; then use:
ant dist ${build_files:autoproxy}
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/how-to-build-lucene-solr-espeically-if-behind-a-firewall-tp3163038p3165568.html
Sent from the Solr - User mailing list archive at Nabble.com.
You just need to allocate more heap to your JVM.
BTW are you doing any complex search while indexing is in progress, like
getting large set of documents.
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/OOM-at-solr-master-node-while-updating-document
Would you care to even index the duplicate documents? Finding duplicacy in
content fields would be not so easy as in some untokenized/keyword field.
May be you could do this filtering at indexing time before sending the
document to SOLR. Then the question comes, which one document should go(from
a
Did you do manual copy of index from Master to Slave of servers. I suppose,
it won't be copied properly.
If this is the case, then you can check the size of indexes on both servers.
Otherwise, you would've to recreate the indexes.
Thanx
Pravesh
--
View this message in context:
htt
Was your searches always slow, OR, since you did some changes at
index/config/schema level?
Is it due to 5-mins index updation? Are you warming ur searches?
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Search-is-taking-long-long-time-tp3095306p3098552.html
First commit and then try again to search.
You can also use lucene's CheckIndex tool to check & fix your index (it may
remove some corrupt segments in your index)
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Read-past-EOF-error-due-to-broken-c
evance)
>if both the fields do not contain these terms together (show as normal one
>with field1 having more relevance than field2)
You wud've to experiment with different boost values to arrive at some
benchmark.
Start with same for field-1 & field-2, then inc. for field-1 a litt
t off
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/relevant-result-for-query-with-boost-factor-on-parameters-tp3079337p3085406.html
Sent from the Solr - User mailing list archive at Nabble.com.
Yes. Then I beleive you would need multiple queries
--
View this message in context:
http://lucene.472066.n3.nabble.com/difficult-sort-tp3075563p3075802.html
Sent from the Solr - User mailing list archive at Nabble.com.
What is the type for the field's defaultquery & title in your schema.xml ?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Search-failed-even-if-it-has-the-keyword-tp3075626p3075797.html
Sent from the Solr - User mailing list archive at Nabble.com.
I'm not sure, but have looked at Collapsing feature in SOLR yet? You may have
to apply patch for 1.4.1 version, if this is what u want?
--
View this message in context:
http://lucene.472066.n3.nabble.com/difficult-sort-tp3075563p3075661.html
Sent from the Solr - User mailing list archive at Nabb
First check, in your schema.xml, which is your default search field. Also
look if you are using WordDelimiterFilterFactory in your schema.xml for the
specific field. This would tokenize your words on every capital letter, so,
for the word "DescribeYourImageWithAMovieTitle" will be broken into multi
If you are sending whole CSV in a single HTTP request using curl, why not
consider sending it in smaller chunks?
--
View this message in context:
http://lucene.472066.n3.nabble.com/SOlR-Out-of-Memory-exception-tp3074636p3075091.html
Sent from the Solr - User mailing list archive at Nabble.com.
mples). My actual
production load is much less than that(3hrs cycle is actually spans to 24
hrs on production). I will repeat this with actual load now.
Thanx all 4 ur time :)
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/High-100-CPU-usage-with-SOLR-
pe this helps
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/problem-with-the-new-IndexSearcher-when-snpainstaller-and-commit-script-happen-tp3066902p3068903.html
Sent from the Solr - User mailing list archive at Nabble.com.
ex of about 32GB with 100+ fields indexed,18 fields stored &
using an optimized index for search
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/High-100-CPU-usage-with-SOLR-1-4-1-tp3068667p3068778.html
Sent from the Solr - User mailing list archive at Nabble.com.
ibly NIOFSDirectory attributes to high CPU?
Is there a mechanism in 1.4.1 to use the SimpleFSDirectory implementation
for searching(would this require full re-index)?
Help will be appreciated :)
Thanx
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/High-100-CPU-us
>k0 --> A | C
>k1 --> A | B
>k2 --> A | B | C
>k3 --> B | C
>Now let q=k1, how do I make sure C doesn't appear as a result since it
doesn't contain any occurence of k1?
Do we bother to do that. Now that's what lucene does :)
--
View this message in context:
http://lucene.472066.n3.nabble.com/Ho
SOLR1.3+ logs only the fresh queries in the logs. If you re-run the same
query then it is served from cache, and not printed on the logs(unless
cache(s) are not warmed or sercher is reopened).
So, Otis's proposal would definitely help in doing some benchmarks &
baselining the search :)
--
View t
Since you r using expand="true" , so, every time a matching synonym entry is
found the analyzer is expanding the term with all synonyms set in the index.
This may cause the index to grow in size.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Applying-synonyms-increase-the-da
You can go ahead with the Master/Slave setup provided by SOLR. Its trivial to
setup and you also get SOLR's operational scripts for index synch'ing b/w
Master-to-Slave(s), OR the Java based replication feature.
There is no need to re-invent other architecture :)
--
View this message in context:
BTW, why r u sorting on this field?
You could also index & store this field twice. First, in its original value,
and then second, by encoding to some unique code/hash and index it and sort
on that.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Sorting-tp3017285p3019055.html
You can use DataImportHandler for your full/incremental indexing. Now NRT
indexing could vary as per business requirements (i mean delay cud be 5-mins
,10-mins,15-mins,OR, 30-mins). Then it also depends on how much volume will
be indexed incrementally.
BTW, r u having Master+Slave SOLR setup?
--
V
SOLR wiki will provide help on this. You might be interested in pure Java
based replication too. I'm not sure,whether SOLR operational will have this
feature(synch'ing only changed segments). You might need to change
configuration in searchconfig.xml
--
View this message in context:
http://lucene
>>We're using Solr to search on a Shop index and a Product index
Do you have 2 separate indexes (using distributed shard search)?? I'm sure
you are actually having only single index.
>> Currently a Shop has a field `shop_keyword` which also contains the
>> keywords of the products assigned to it.
If your index size if smaller (a few 100 MBs), you can consider the SOLR's
operational script tools provided with distribution to sync indexes from
Master to Slave servers. It will update(copies) the latest index snapshot
from Master to Slave(s). SOLR wiki provides good info on how to set them as
C
Hi,
I'm using StreamingUpdateSolrServer to post a batch of content to SOLR1.4.1.
By looking at StreamingUpdateSolrServer code, it looks it only provides the
content to be streamed in XML format only.
Can we use it to stream data in binary format?
--
View this message in context:
http://lucene
Since FieldCache is an expert level API in lucene, there is no direct control
provided by SOLR/Lucene to control its size.
--
View this message in context:
http://lucene.472066.n3.nabble.com/FieldCache-tp2987541p2989443.html
Sent from the Solr - User mailing list archive at Nabble.com.
What would be the default value for omitNorms?
--- Default value is "false"
Is general advise to ignore this and set the value explicitly?
--- Depends on your requirement. Do this on field-per-field basis. Set to
"false" on fields where you want the norms, or, set to "true" on fields
where you w
This will help:
http://cephas.net/blog/2008/03/30/how-morelikethis-works-in-lucene/
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-does-Solr-s-MoreLikeThis-component-internally-work-to-get-results-tp2938407p2988487.html
Sent from the Solr - User mailing list archive at
Do you really require multi-shards? Single core/shard will do for even
millions of documents and the search will be faster than searching on
multi-shards.
Consider multi-shard when you cannot scale-up on a single shard/machine(e.g,
CPU,RAM etc. becomes major block).
Also read through the SOLR di
Just read through:
http://www.springbyexample.org/examples/solr-client.html
http://static.springsource.org/spring-roo/reference/html/base-solr.html
--
View this message in context:
http://lucene.472066.n3.nabble.com/how-to-integrate-solr-with-spring-framework-tp2955540p2988363.html
Sent from th
For saving Memory:
1. allocate as much memory to the JVM (especially if you are using 64bit OS)
2. You can set "omitNorms=true" for your date & id fields (actually for all
fields where index-time boosting & length normalization isn't required. This
will require a full reindex)
3. Are you sorting o
I'm sure you can fix this by increasing value to some
max.
This shld apply to filter query as well
--
View this message in context:
http://lucene.472066.n3.nabble.com/Too-many-Boolean-Clause-and-Filter-Query-tp2974848p2988190.html
Sent from the Solr - User mailing list archive at Nabble.com.
tag represents to the actual SOLR document that will be posted by
the DIH. This mapping is used by the DIH to map DB-to-index document.
You can have multiple tags, as you might be pulling data from more
than 1 table.
You can only have one tag in you db-data-config.xml (remember,
the purpose of
1 - 100 of 103 matches
Mail list logo