On 3/23/2015 3:08 AM, Srinivas wrote:
> Present in my project we are using apache tika for reading metadata of the
> file,So whenever we handled large files(contained more than 100000
> characters file) tika generating the error is file contained more than
> 100000 characters, So is it possible or not handling large files by using
> tika,Please let me know.

This sounds like a Tika problem.  This is a solr mailing list.  You may
find some Tika expertise here, but this is the incorrect place for a
question about Tika.

Solr does use the Tika parser, in the contrib module for the
ExtractingRequestHandler.  I have never heard of such a limitation in
the context of the ExtractingRequestHandler, and I've heard some people
complain about OutOfMemory exceptions when they index 4 gigabyte PDF
files with our extracting handler ... so I am guessing that you are
using Tika in your own software.  If that is correct, you'll need to ask
your question on a Tika mailing list.

Thanks,
Shawn

Reply via email to