On Wed, 2015-04-08 at 14:00 -0700, pras.venkatesh wrote:
> 1. 8 nodes, 4 shards(2 nodes per shard)
> 2. each node having about 55 GB of Data, in total there is 450 million
> documents in the collection. so the document size is not huge,
So ~120M docs/shard.
> 3. The schema has 42 fields, it gets
Thanks, I figured that might be the case (hand-editting clusterstate.json).
- Ian
On Wed, Apr 8, 2015 at 11:46 PM, ralph tice wrote:
> It looks like there's a patch available:
> https://issues.apache.org/jira/browse/SOLR-5132
>
> Currently the only way without that patch is to hand-edit
> clus
It looks like there's a patch available:
https://issues.apache.org/jira/browse/SOLR-5132
Currently the only way without that patch is to hand-edit
clusterstate.json, which is very ill advised. If you absolutely must,
it's best to stop all your Solr nodes, backup the current clusterstate
in ZK, mo
I previously created several collections with maxShardsPerNode=1 but I
would now like to change that (to "unlimited" if that is an option). Is
changing this value possible?
Cheers,
- Ian
Wups - sorry folks, I send this prematurely. After typing this out I think
I have it figured out - although SPLITSHARD ignores maxShardsPerNode,
ADDREPLICA does not. So ADDREPLICA fails because I already have too many
shards on a single node.
On Wed, Apr 8, 2015 at 11:18 PM, Ian Rose wrote:
>
On my local machine I have the following test setup:
* 2 "nodes" (JVMs)
* 1 collection named "testdrive", that was originally created with
numShards=1 and maxShardsPerNode=1.
* After a series of SPLITSHARD commands, I now have 4 shards, as follows:
testdrive_shard1_0_0_replica1 (L) Active 115
tes
Are there at least a small number of categories of users with discrete
prices, or can each user have their own price. The former is doable, the
latter is not unless the number of users is relatively small, in which case
they are equivalent to categories.
You could have a set of dynamic fields, pri
Ryan:
bq: I don't want it to issue phrase queries to that field ever
This is one of those requirements that you'd have to enforce at the
app layer. Having Solr (or Lucene) enforce a rule like this for
everyone would be terrible.
So if you're turning off TF but also saying title is "one of the
p
Matt:
How are you creating the new replica? Are you giving it an explicit
name? And especially is it the same name as one you've already
deleted?
'cause I can't really imagine why you'd be getting a ZK exception
saying the node already exists.
Shot in the dark here..
On Wed, Apr 8, 2015 at
Thanks for your thought Shawn, I don't think fq will be helpful here. The
field for which I want to turn TF off is "title", which is actually one of
the primary components of score, so I really need it in qf. I just don't
want the TF portion of the score for that field only. I don't want it to
i
On 4/8/2015 5:06 PM, Ryan Josal wrote:
> The error:
> IllegalStateException: field "foo" indexed without position data; cannot
> run PhraseQuery.
>
> It would actually be ok for us to index position data but there isn't an
> option for that without term frequencies. No TF is important for us when
Found this error which likely explains my issue with new replicas not coming
up, not sure next step. Almost looks like Zookeeper's record of a Shard's
leader is not being updated?
4/8/2015, 4:56:03 PM
ERROR
ShardLeaderElectionContext
There was a problem trying to register as the
leader:org.apa
Hey guys, it seems that omitTermFreqAndPositions is not very usable with
edismax, and I'm wondering if this is intended behavior, and how I can get
around the problem.
The setup:
define field "foo" with omitTermFreqAndPositions=true
The query:
q="ground coffee"&qf=foo bar baz
The error:
IllegalS
Erick, Anshum,
Thanks for your replies! Yes, it is replica state that I am looking at, and
this the answer I was hoping for.
I am working on a solution that involves moving some replicas to new Solr nodes
as they are made available. Before deleting the original replicas backing the
shard,
Matt:
In a word, "yes". Depending on the size of the index for that shard,
the transition from Down->Recovering->Active may be too fast to catch.
If replicating the index takes a while, though, you should at least
see the "Recovering" state, during which time there won't be any
searches forwarded
Hi Matt,
If it's the replica state that you're looking at, yes, it means that the
Replica is in sync with the leader and serving/ready to serve requests.
On Wed, Apr 8, 2015 at 2:58 PM, Matt Kuiper wrote:
> Hello,
>
> When creating a new replica, and the state is recorded as active with in
> ZK
Hello,
When creating a new replica, and the state is recorded as active with in ZK
clusterstate, does that mean that new replica has synched with the leader
replica for the particular shard?
Thanks,
Matt
Hi All,
We are trying to setup the Solr Cloud in our team and able setup multiple nodes
in one server as a cloud.
Need clarifications on the following.
Is there any good documentation, which can help us to build the Solr Cloud with
multiple physical servers?
Since the Solr Cloud is distributed
I am extremely new to SOLR and am wondering if it is possible to do something
like the following. Basically I have been tasked with researching SOLR to see
if we can replace our current searching algorithm.
We have a website with product data. Product data includes standard things
like Name, SK
I have a solr cloud instance with 8 nodes, 4 shards and facing memory leak on
the JVMs
here are the details of the instance.
1. 8 nodes, 4 shards(2 nodes per shard)
2. each node having about 55 GB of Data, in total there is 450 million
documents in the collection. so the document size is not hug
Dear all,
Hi,
As a part of my code I have to update Lucene document. For this purpose I
used writer.updateDocument() method. My problem is the update process is
not affect index until restarting Solr. Would you please tell me what part
of my code is wrong? Or what should I add in order to apply the
Specifying the default search field in the schema has been
deprecated for a while, it wasn't flexible enough.
The recommended way is to specify a "df" parameter in
your request handler defaults.
The space separates the field specification from the second term.\
Assuming that you want to search bo
See inline for a few answers:
On Wed, Apr 8, 2015 at 10:32 AM, jainam vora wrote:
> Hi,
>
> Brief:
> I am new to Solr and E commerce web apps and Java.
> i want to integrate solr in eCommerce web application (developed using Java
> on Linux).
>
> I have following queries.
>
> 1. how to setup Solr
Re,
You have to specify defautSearchField tag in Schema.xml
Regards,Andy
Le Mercredi 8 avril 2015 21h33, avinash09 a écrit
:
http://localhost:8983/solr/rna/select?q=test_name:*Uae
blow*&wt=json&rows=100
getting
{
responseHeader: {
status: 400,
QTime: 28
},
error: {
msg: "no fiel
http://localhost:8983/solr/rna/select?q=test_name:*Uae
blow*&wt=json&rows=100
getting
{
responseHeader: {
status: 400,
QTime: 28
},
error: {
msg: "no field name specified in query and no default specified via 'df'
param",
code: 400
}
}
plz help!!
--
View this message in context:
http://l
Hi,
I have installed Curl on debian linux. But when i use curl to create
collection i am getting http autentication error.
--
Thanks & Regards,
Jainam Vora
Hi,
Brief:
I am new to Solr and E commerce web apps and Java.
i want to integrate solr in eCommerce web application (developed using Java
on Linux).
I have following queries.
1. how to setup SolrCloud on Tomcat. Searched on internet but could not get
clear steps yet. Also tried some steps but no
Hello All,
I am using Master - Slave architecture setup with hundreds of cores getting
replicated between master and slave servers. I am facing very weird issue
while creating a new core.
Whenever there is a new call for a new core creation (using
CoreAdminRequest.createCore(coreName,instanceDir,
bq: After the commit, query times are unacceptable slow
First, please quantify "unacceptable". 100ms? 10,000ms? Details matter.
Second, the purpose of autowarming is exactly to smooth out the first few
searches when a new searcher is opened, are you doing any?
Third: What are your autocommit set
Well, at least it's _some_ progress ;).
Agreed, the segments hanging around is still something of a mystery
although if I really stretch I could relate them, maybe.
I believe there's clean-up logic when a core starts up to nuke cruft
in the index directory. If the cruft was created after a core
Thank you Swaraj and Jurgen for the information.
I'll just stick to the one zoo.cfg for each instance. Now I have one root
directory .../zookeeper where I create the 3 subdirectories for each of the
instance (known as zookeeper1, zookeeper2 and zookeeper3), and each of them
have their own zoo.cfg.
How much RAM do you have? Check whether your system is compute-bound or
I/O-bound? If all or most of your index doesn't fit in the system memory
available for file caching, you're asking for trouble.
Is the indexing time also unacceptably slow, or just the query time?
-- Jack Krupansky
On Wed, A
Hi,
I have a core with about 20M documents and the size on disc is about
50GB. It is running on a single EC2 instance. If the core is warmed up,
everything is running fine. The problem is the following:
We assign categories (similar to tags) to documents. Those are stored in
a multivalue string f
On 4/8/2015 6:30 AM, Chaushu, Shani wrote:
> I tried to get the SolrJ in older version, but I'm using solr-spark package
> and it fails with compilation errors probably because it uses function from
> newer versions...
> I can't find any solution...
Looking at the github repo for spark-solr, I
Hi,
I tried to get the SolrJ in older version, but I'm using solr-spark package and
it fails with compilation errors probably because it uses function from newer
versions...
I can't find any solution...
-Original Message-
From: Emre Sevinc [mailto:emre.sev...@gmail.com]
Sent: Wednesday
Hello Shani,
Are you using SolrJ? Did you try using the same version of SolrJ (e.g v.
4.4 of SolrJ, if you're using Solr 4.4)? That's what generally worked for
me.
Kind regards,
Emre Sevinç
http://www.bigindustries.be/
On Wed, Apr 8, 2015 at 1:46 PM, Chaushu, Shani
wrote:
> i'm using solr 4.
i'm using solr 4.4.
the query request works fine but when i try to add doc into solr cloud
(cloudSolrServer.request(updateRequest))
i get an error:
Exception in thread "main"
org.apache.solr.client.solrj.impl.CloudSolrServer$RouteException: Unknown type
19
at
org.apache.solr.client.so
Text search means searching of text, and special characters are not... text.
Why are you using the standard tokenizer if you are not trying to search
for standard text?
Try using the white space tokenizer, which will preserve special characters.
That said, the word delimiter filter will remove t
I have crawled a website using nutch.
When I try to index it with solr I get following error
org.apache.solr.common.SolrException: ERROR: [doc=http://xyz.htm]
unknown field 'metatag.keywords'
*unknown field 'metatag.keywords'*
I can not figure out where the error is as I have o not defined any
Hi,
I am trying to setup a SolrCloud cluster on top of Hadoop (HDP). The
upconfig and linkconfig commands were run successfully and the
configuration is now centrally manged in Zookeeper.
However, when I run the command to create a core, I am shown the following
permission denied error. This is r
To be precise: create one zoo.cfg for each of the instances. One config
file for all is a bad idea.
In each config file, use the same server.X lines, but use a unique
clientPort.
As you will also have separate data directories, I would recommend
having one root directory .../zookeeper where you c
Hi Sangeetha,
/update/extract refers to extractrequesthandler.
If you only want to index the data, you can do it with extractrequesthandler.
I dont think it requires metadata, but you need to provide literal.id to
specify which field will be unique id.
For more information :-
https://wiki.apache
Hi Zheng,
I am not sure if this command *"zkServer.cmd start zoo.cfg" * works in
windows or not, but in zkServer.cmd it calls zkEnv.cmd where "
*ZOOCFG=%ZOOCFGDIR%\zoo.cfg*" is set. So, if you want to run multiple
instances of zookeeper, change zoo.cfg to your config file and start
zookeeper.
The
Hi Swaraj,
Thanks for the answers.
From my understanding We can index,
· Using DIH from db
· Using DIH from filesystem - this is where I am concentrating on.
o For this we can use SolrJ with Tika(solr cell) from Java layer in order to
extract the content and send the data thro
Hi Pradeep,
AFAIK the mailing list doesn't allow attachments. I think pasting the
error should be enough
Best,
Andrea
On 04/08/2015 09:02 AM, Pradeep wrote:
We have installed solr-4.3.0 is our local but we are getting error.
Please find attachment. And help us to fix this error.
Thank You.
We have installed solr-4.3.0 is our local but we are getting error.
Please find attachment. And help us to fix this error.
Thank You.
Regards,
Pradeep
thanks Erick
--
View this message in context:
http://lucene.472066.n3.nabble.com/distributed-search-on-tables-tp4197456p4198285.html
Sent from the Solr - User mailing list archive at Nabble.com.
not able to search on special characters like . ,_
my query
http://localhost:8983/solr/rna/select?q=name:"UAE
B"&wt=json&fl=name&rows=100
getting result UAE_bhdgsfsdbj
but for
http://localhost:8983/solr/rna/select?q=name:"UAE_"&wt=json&fl=name&rows=100
no result found
I am using below field ty
I just want to index only certain documents and there will not be any update
happening on the indexed document.
In our existing system we already have DIH implemented which indexes
document from sql server (As you said based on last index time). In this
case the metadata is there available in dat
Thank you nutchsolruser and Shawn.
I've changed the clientPort to different port for each of the machine.
It is able to work for my another setup, in which I have 3 different
zookeeper folder, and each has its own configuration and all are using
zoo.cfg. For that setup I can start the 3 servers in
50 matches
Mail list logo