Hi List
I need your help to resolve a problem for which i had been struggling for
days.
Lets take an example of Shoes which are grouped on basis of size and Price
With first group as size and price as "7 and 7000" i have 2 documents as
below
{id:1,color:blue,item sold:10}
{id:5,price:yellow,item
Hi List,
I am unable to create a core.Unable to figure out what wrong.
I get below error.
ERROR: Failed to create collection 'XXX' due to:
org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException:Error
from server at
http://xyz.com:8983/solr:
Error CREATEing SolrCore 'docpocc_sh
Gus
You are never biased.
I explored a bit about JesterJ. Looks quite promising.
I will keep you posted on my experience to you soon.
Regards
Neo
--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
Thanks Erick
I already have gone through the link from tika example you shared.
Please look at the code in bold.
I believe still the entire contents is pushed to memory with handler object.
sorry i copied lengthy code from tika site.
Regards
Neo
*Streaming the plain text in chunks*
Sometimes, you
Thanks Shawn,
Yes I agree ERH is never suggested in production.
I am writing my custom ones.
Any pointer with this?
What exactly i am looking is a custom indexing program to compile precisely
the information
that you need and send that to Solr.
On the other hand i see the below method is very ex
Thanks Erick,
Though i saw this article in several places but never went through it
seriously.
Dont you think the below method is very exepensive
autoParser.parse(input, textHandler, metadata, context);
If the document size if bigger than it will need enough memory to hold the
document(ie Cont
Hi List,
I have a specific Requirement where i need to index below things
Meta Data of any document
Some parts from the Document that matches some keywords that i configure
The first part i am able to achieve through ERH or FilelistEntityProcessor.
I am struggling on second part.I am looking for
Hi Shawn,
Thanks for the long explanation.
Now 2 Billion limit can be overcome by using shard.
Now coming back to collection.Unless we have a logical or Business reason
we should not go for more than one collection.
Lets say i have 5 different entities and they have each 10,20,30,40 and 50
attri
Emir
I read from the link you shared that
"Shard cannot contain more than 2 billion documents since Lucene is using
integer for internal IDs."
In which java class of SOLR implimentaion repository this can be found.
Regards
Neo
--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068
Thanks every one for your beautifull explanation and valuable time.
Thanks Emir for the Nice
Link(http://www.od-bits.com/2018/01/solrelasticsearch-capacity-planning.html)
Thanks Shawn for
https://lucidworks.com/2012/07/23/sizing-hardware-in-the-abstract-why-we-dont-have-a-definitive-answer/
When
Thanks Emir
with context to DIH do we have any Resume mechanism?
Regards
Neo
--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
Hi Emir,
Thanks a lot for your reply.
so when i design a solr eco system i should start with some rough guess on
shards and increase the number of shards to make performance better.what is
the accepted/ideal Response Time.There should be a trade off between
Response time and the number of shards as
with Solrcloud What happens if indexing is partially completed and ensemble
goes down.What are the ways to Resume.In one of the scenario i am using 3 ZK
Node in ensemble.Lets say i am indexing 5 million data and i have partially
indexed the data and ZK ensemble goes down. What should be the best
Hi Team
First of all i take this opportunity to thank you all for creating a
beautiful place where people can explore ,learn and debate.
I have been on my knees for couple of days to decide on this.
When i am creating a solr cloud eco system i need to decide on number of
shards and collection.
Wh
14 matches
Mail list logo