I have used that type of location searching. But I have not used spatial
search. I wrote my logic at application end.
I have cached the location ids and their lat/lang. When queries are comming
for any location say "New Delhi" then my location searche logic at
application end calculate the distanc
As I seen the code for QueryElevationComponent ,there is no supports for
Distributed Search i.e. query elevation does not works with shards.
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Dismax-Sharding-and-Elevation-tp2247156p2247522.html
Sent from the Solr
Similar type of work I have done earlier by using spell-check component with
auto-suggest combined.
Autosuggest will provide the words starting with query term and spellcheck
returns the words similar to that.
I have combined both suggestion in single list to display
-
Grijesh
--
View this
Use this type of url for delete all data with fallowed by commit
http://localhost:8983/solr/update/?stream.body=*:*&commit=true
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Question-on-deleting-all-rows-for-an-index-tp2246726p2246948.html
Sent from the Sol
what are your Benchmarks ,
Please describe in detail your problem,
What is exactly being .
What is the way you are indexing and quering
what data you indexing,
how much data you indexing.
What are your server configurations.
How much Ram you are using
-
Grijesh
--
View this message in c
First thing is that your raw log files solr can not understand. Solr needs
data according to schema defined And also solr does not know your log file
format .
So you have to write a parser program that will parse your log files into a
existing solr writable formats .Then you can be able to index
It will not work.
I think your log files are not in solr Doc xml files.
First thing is that your log files is raw data.
you have to convert it to any of solr readable data either in solr xml DOC
or CSV format to index on solr As Gora suggested to you.
-
Grijesh
--
View this message in conte
How you parsed you log?
Which way you gone for index of log file data?
have you done any work what Gora Mohanty has suggested to you.
I am local for Delhi NCR area.
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Input-raw-log-file-tp2210043p2239505.html
Sent
what means query time index issue? Please provide more detail of your
problem.
the field type textSpell is defined in example schema.xml for Spell
Suggestion.What analysis chain you have used in your "textSpell" field.
for what purpose you are using that field ?
-
Grijesh
--
View this mess
Which type of performance issues you have index time or query time?
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/solr-search-performance-tp2239298p2239338.html
Sent from the Solr - User mailing list archive at Nabble.com.
Check their configurations they are using different Analysis .i.e their
definition is different
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/field-Type-Textsplell-tp2239237p2239275.html
Sent from the Solr - User mailing list archive at Nabble.com.
Look at Solr Function Queries they might help you
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/pruning-search-result-with-search-score-gradient-tp2233760p2233773.html
Sent from the Solr - User mailing list archive at Nabble.com.
You have to explicitly call
http://:/solr/dataimport?command=delta-import to start
delta-import.
You can setup it as cronjob.
As per my knowledge dataimport.properties file contains parameter to be used
in DtataImport xml configuration file for example database related file
contains text as
#W
Check your libraries for Tika related Jar files.Tika related files must be on
classpath of solr
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Internal-Server-Error-when-indexing-a-pdf-file-tp2214617p2226374.html
Sent from the Solr - User mailing list archive
open a new mail conversation for that
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Improving-Solr-performance-tp2210843p2211300.html
Sent from the Solr - User mailing list archive at Nabble.com.
shards are used when index size become huge and performance going down .
shards mean distributed indexes. But if you will put all shards on same
machine as multicore then it will not help too much on performance.
and also shards distributes indexes near equals in size.
There is also not enough Ra
shards are used when index size become huge and performance going down .
shards mean distributed indexes. But if you will put all shards on same
machine as multicore then it will not help too much on performance.
and also shards distributes indexes near equals in size.
There is also not enough Ra
Some questions-
1-Are all shards on same machine
2-What is your Ram Size
3-What are the size of index on each shards in GB
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Improving-Solr-performance-tp2210843p2210878.html
Sent from the Solr - User mailing list
There is a csv update handler in solr you can use it by modifying your
logfile
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Input-raw-log-file-tp2210043p2210673.html
Sent from the Solr - User mailing list archive at Nabble.com.
use as pharase it will work like "Computer / IT" for you
Here IT is a stopword when you are trying query as
category:Computer / IT
parsed as category:Computer :IT
but IT is a stopword for default search field what you have selected
so second query removed and only category:Computer is showing
add parameter fl=*,score with your query url,
It will return all fields with score of every document
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/get-score-value-tp2204353p2204458.html
Sent from the Solr - User mailing list archive at Nabble.com.
check with debugQuery=on and also with analysis.jsp ,what happening with your
query term and indexed terms.
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/forward-slash-not-working-in-my-solr-search-tp2204326p2204436.html
Sent from the Solr - User mailing list
First you have to check that your indexing process can cause of removing the
whitespace
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/forward-slash-not-working-in-my-solr-search-tp2204326p2204336.html
Sent from the Solr - User mailing list archive at Nabble.c
/usr/local/tomcat/apache-tomcat-6.0.26/bin/./solr/data/index
Thats showing that solr is getting as solr/home as
/usr/local/tomcat/apache-tomcat-6.0.26/bin/
either your solr/home settings in web.xml is not working
Or
You have the default solrconfig.xml settings ,either you
have to modify the "da
Solr is not a library like Lucene. Solr works like Webservice application
over http.
It works over http so you can easly integrate it in any application.
it takes input as http parameters and returns response in different forms
like xml, json, php, csv etc.So you can use any response type for you
/opt/solr i.e. solr/home only requires conf,data,lib directories.
why you have putted other directories under solr/home location.
The [solr/home]/data directory must required the "Write" permission ,Please
check and provide appropriate permissions to directories because solr will
try to write in
Why thinking so complex,just use result of first query as filter for your
second query
like
fq=related_id:(id1 OR id2 OR id3 )&q=q=”type:IT AND manager_12:dave”
somthing like that
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Sub-query-using-SOLR-tp2
Use parameter as Per Field Basis as
f..facet.sort=index
Solr1.4 -- the true/false values have been deprecated starting with Solr
1.4, instead use count for sorting by count, and index for sorting by index
order.
The default is true/count if facet.limit is greater than 0, false/index
otherwise.
Provide us what Analyzers/Filters are you using for Analysis of email field.
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Different-behavior-for-q-goo-com-vs-q-goo-com-in-queries-tp2168935p2183256.html
Sent from the Solr - User mailing list archive at Nabble
Second option is better. Storing Big BLOB data in index will increase index
size and it will create performance issues
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Inner-index-stock-big-BLOB-or-separate-in-a-DB-tp2166964p2167129.html
Sent from the Solr - Use
Your problem is Term Frequency, You do not want to consider term frequency
try omit Term Frequency.
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Problem-of-results-ordering-tp2139314p2150722.html
Sent from the Solr - User mailing list archive at Nabble.com.
Then what will be when we filter out only some result and want to group ,how
your index time group count will help.
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/solr-4-0-pagination-tp1812384p2124747.html
Sent from the Solr - User mailing list archive at Nabb
you can put table names in a different table and use like this
-
-
-
-
-
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/DIH-for-
You have to create again solr xml doc from modified data with same unique id
and post it to solr
You can not change any specific field of any document
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-update-solr-index-tp2038480p2038801.html
Sent from the
There is no way to update any field in solr,You have to reindex that entire
document again.
you can get that doc from index create xml with existing fields with your
updated field and post that xml to solr.
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-t
set splitWordsPart=0,splitNumberPart=0
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Regarding-WordDelimiterFactory-tp1444694p1444742.html
Sent from the Solr - User mailing list archive at Nabble.com.
yes reindexing is necessary for protwords,synanym update
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-searching-harri-finds-harry-tp1438486p1438802.html
Sent from the Solr - User mailing list archive at Nabble.com.
have u restart the solr after adding words in protwords and reindex the data?
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-searching-harri-finds-harry-tp1438486p1438735.html
Sent from the Solr - User mailing list archive at Nabble.com.
solr.xml allows you to mention the other properties as well like
instanceDir, config,schema in the cores/core tag
So , sharing the entire conf dir may not be possible , but it is
possible to share solrconfig.xml and schema.xml
U can see the detail parameters at wiki page
http://wiki.apache.org/
you need a separate solr core for that
and have to write a processor which process with your original index ,then
generate the xml data and push to the new core.That is the simple way that i
have used many times.
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.c
U have not provided much detail about analysis of that field,but I am sure
that problem because of stemming
u can see by analysis page or by debugQuery=on parameter.
To prevent stemming u have to put words in protword.txt on which u do not
need any stemming
-
Grijesh
--
View this message in
see the analysis.jsp with debug verbose and see what happens at index time
and search time during analysis with your data
Also u can use debugQuery=on for seeing what actually parsed query is.
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Phrase-search-multi
solr supports UTF8 char set so use any ofthe MappingCharFilterFactory/
UnicodeNormalizationFilterFactory/ASCIIFoldingFilterFactory for searching
and indexing of that type of data
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-enable-Unicode-Support-
Increase the Http Header size of Tomacat server.xml
and increase the maxBooleanClauses of solrconfig.xml to support the very
large queries
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/using-Solr-with-very-large-search-terms-tp1415558p1424649.html
Sent from t
1-extract the solr.war
2-edit the web.xml for setting solr/home param
3-create the solr.war
4-setup solr home directory
5-copy the solr.war to JBossAs 6 deploy directory
7-start the jboss server
--
View this message in context:
http://lucene.472066.n3.nabble.com/Deploying-Solr-1-4-1-in-JbossAs-6
As per my experience with memcache was not so good .
Finaly I have configured solr's built in cache for best perfoemance.
By memcache we were caching query,bu it solr provides already.
you can take a call after Load testing with memcache and without memcache
--
View this message in context:
htt
I have to search on multiple fields on different language at a time
--
View this message in context:
http://lucene.472066.n3.nabble.com/Spellcheck-in-multilanguage-search-tp1393357p1393431.html
Sent from the Solr - User mailing list archive at Nabble.com.
How can be spellcheck configured for multilanguage search,I have to index 17
languages in my indexes and search on them also wants to use spellcheck for
that
--
View this message in context:
http://lucene.472066.n3.nabble.com/Spellcheck-in-multilanguage-search-tp1393357p1393357.html
Sent from th
Index time synonym explore is the better way if u do not have the disk space
concern.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Implementing-synonym-NewBie-tp1379269p1392927.html
Sent from the Solr - User mailing list archive at Nabble.com.
No there is no such feature,your indexing component have to keep generation
of ids
--
View this message in context:
http://lucene.472066.n3.nabble.com/Auto-ID-for-Documents-indexed-tp1367972p1369051.html
Sent from the Solr - User mailing list archive at Nabble.com.
have u commited the data,
use *:* query to see that data is commited yet or not
--
View this message in context:
http://lucene.472066.n3.nabble.com/Search-results-after-importing-from-Dih-tp1365720p1365927.html
Sent from the Solr - User mailing list archive at Nabble.com.
DIH is not basically for deletion it is for inserting data into index.
Although it has a parameter "clean" which is by default true and it cleans
the index every time when full-import command is issued, means it create
index from scratch.
If your requirement is to delete whole index you can also
Hi Pawan,
If u r using solr1.4 or latter version then u can see terms info by using
terms request handler like
http://localhost:8080/solr/terms/?terms.fl=text&terms.sort=count
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-get-most-indexed-keyword-from-SOLR-tp124055
Using multicore is the right approach
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-for-multiple-websites-tp1173220p1219772.html
Sent from the Solr - User mailing list archive at Nabble.com.
lance, is there any Bug in solr1.4 related to sorting on date field?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Date-sorting-tp1219372p1219537.html
Sent from the Solr - User mailing list archive at Nabble.com.
provide schema.xml and solrconfig.xml to dig the problem and by which version
of solr u have indexed the data?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Date-sorting-tp1219372p1219534.html
Sent from the Solr - User mailing list archive at Nabble.com.
ramBufferSize is preferred to be 128MB more than that it does not seemes to
improve performance
--
View this message in context:
http://lucene.472066.n3.nabble.com/OutOfMemoryErrors-tp1181731p1199592.html
Sent from the Solr - User mailing list archive at Nabble.com.
U can add like this it will work I am using it
JAVA_OPTS="$JAVA_OPTS -Xms1024m -Xmx4096m "
--
View this message in context:
http://lucene.472066.n3.nabble.com/OutOfMemoryErrors-tp1181731p1183229.html
Sent from the Solr - User mailing list archive at Nabble.com.
put that line in your startup script or u can set as env var
export CATALINA_OPTS=-Xms256m -Xmx1024m;
--
View this message in context:
http://lucene.472066.n3.nabble.com/OutOfMemoryErrors-tp1181731p1182708.html
Sent from the Solr - User mailing list archive at Nabble.com.
U can set up in startup script of tomcat
--
View this message in context:
http://lucene.472066.n3.nabble.com/OutOfMemoryErrors-tp1181731p1182582.html
Sent from the Solr - User mailing list archive at Nabble.com.
increase your JVM Heap space by using params
-Xms1024m
-Xmx4096m
Like this.
--
View this message in context:
http://lucene.472066.n3.nabble.com/OutOfMemoryErrors-tp1181731p1181892.html
Sent from the Solr - User mailing list archive at Nabble.com.
Given below are the steps for auto-suggest and spellcheck in single query:
Make the change in TermComponent part in solrconfig.xml
true
termsComponent
spellcheck
Use given below query format for getting autosuggest and spellcheck
suggestion.
http
Hi ,
There is no required fields except u specify any fields to required.U can
remove or add as many fields u want.
That is an example schema which shows how feilds are configured
--
View this message in context:
http://lucene.472066.n3.nabble.com/schema-xml-tp995696p995800.html
Sent from the S
Do u have data in that field also,Solr returns field which have data only.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Dismax-query-response-field-number-tp985567p986417.html
Sent from the Solr - User mailing list archive at Nabble.com.
why using default="0" its optional remove that from field definition
--
View this message in context:
http://lucene.472066.n3.nabble.com/set-field-with-value-0-to-the-end-tp980580p986115.html
Sent from the Solr - User mailing list archive at Nabble.com.
I Think Solr does not provide any thing like that U want.
--
View this message in context:
http://lucene.472066.n3.nabble.com/nested-query-and-number-of-matched-records-tp983756p983938.html
Sent from the Solr - User mailing list archive at Nabble.com.
Integer field can be empty also, I think u have set required=true if that
remove required=true,
and u can live field without data at the time of indexing.
--
View this message in context:
http://lucene.472066.n3.nabble.com/set-field-with-value-0-to-the-end-tp980580p983728.html
Sent from the Solr
If u can do ,then don't store as 0 ,then by sorting it will come at last (if
is used SortMissingLast=true).
Either u can create a different field for shorting and for empty price don't
set anything to field.
--
View this message in context:
http://lucene.472066.n3.nabble.com/set-field-with-value
I have applied patch#236 for collapsing in solr1.3 and getting the count
after collapse.My pagination is also working fine.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Handling-Pagination-when-using-the-collapsing-feature-tp980637p980745.html
Sent from the Solr - User mai
69 matches
Mail list logo