overy, so I’m
> thinking you did a “kill -9” or similar? Were you actively indexing at the
> time? Solr _should_ manage the recovery even in that case, I’m mostly
> wondering what the sequence of events that lead up to this was…
>
> Best,
> Erick
>
> > On Feb 4, 2020, at 8
Hi all,
I have a 3 node solr cloud instance with a single collection. The solr
nodes are pointed to a 3-node zookeeper ensemble. I was doing some basic
disaster recovery testing and have encountered a problem that hasn't been
obvious to me on how to fix.
After i started back up the three solr jav
ction. Is there any way to specify a "continue on failure" mode
such that solr logs that it was unable to parse a record and why and then
continues onto the next node?
Thanks,
Joe
On Sun, Feb 2, 2020 at 4:46 PM Shawn Heisey wrote:
> On 2/2/2020 8:47 AM, Jose
k up.
>
> Finally, I mistrust just sending a large amount of data via HTTP, just
> because
> there’s not much you can do except hope it all works. If this is a
> recurring
> process I’d seriously consider writing a SolrJ program that parsed the
> csv file and sent it to Solr.
>
&
Hi all,
I have three node solr cloud cluster. The collection has a single shard. I
am importing 140 GB CSV file into solr using curl with a URL that looks
roughly like this. I am streaming the file from disk for performance
reasons.
http://localhost:8983/solr/example/update?separator=%09&stream.f
Hi all,
I am in the process of migrating a solr collection from 4 to 8. I
discovered that there was no ETL process for loading all the data into a
new collection in solr 8, so I had to build one. For technical reasons that
aren't important here, I'd prefer this tool to be a one-off.
In the future
Hi all,
I have TSV file that contains 1.2 million rows. I want to bulk import this
file into solr where each row becomes a solr document. The TSV has 24
columns. I am using the streaming API like so:
curl -v '
http://localhost:8983/solr/example/update?stream.file=/opt/solr/results.tsv&separator=%
Hi all,
I have question about the managed schema functionality. According to the
docs, "All changes to a collection’s schema require reindexing". This would
imply that if you use a managed schema and you use the schema API to update
the schema, then doing a full re-index is necessary each time.