-classpath /[…]/solr-8.7.0/dist/solr-core-8.7.0.jar -Dauto=yes
> >>> -Dparams=literal.id=doc1 -Dc=gettingstarted -Ddata=files org.apa
> >>> che.solr.util.SimplePostTool example/exampledocs/solr-word.pdf
> >>> SimplePostTool version 5.0.0
> >>> Posting
solr-word.pdf (application/pdf) to [base]/extract
SimplePostTool: WARNING: Solr returned an error #404 (Not Found) for
url:
http://localhost:8983/solr/gettingstarted/update/extract?literal.id=doc1&r
esource.name=%2F[…]%2Fsolr-8.7.0%2Fexample%2Fexampledocs%2Fsolr-word.pdf
SimplePostTool: WARNING: Resp
nl,csv,pdf,doc,docx,ppt,pptx,xls,xlsx,odt,odp,ods,ott,otp,ots,rtf,htm,html,txt,log
> > POSTing file solr-word.pdf (application/pdf) to [base]/extract
> > SimplePostTool: WARNING: Solr returned an error #404 (Not Found) for
> > url:
> > http://localhost:8983/sol
l: WARNING: Solr returned an error #404 (Not Found) for
url:
http://localhost:8983/solr/gettingstarted/update/extract?literal.id=doc1&r
esource.name=%2F[…]%2Fsolr-8.7.0%2Fexample%2Fexampledocs%2Fsolr-word.pdf
SimplePostTool: WARNING: Response:
Error 404 Not Found
HTTP ERROR 404 Not Fo
Hi all,
I recently experienced some problems with the update extract handler on a Solr
7.4.0 instance. When sending a document via multipart POST update request, if a
doc parameter name contains too much chars, the POST method fails with a 500
code error and I can see the following exception
Hi all,
I recently experienced some problems with the update extract handler on a Solr
7.4.0 instance. When sending a document via multipart POST update request, if a
doc parameter name contains too much chars, the POST method fails with a 500
code error and I can see the following exception
Hi all,
I recently experienced some problems with the update extract handler on a Solr
7.4.0 instance. When sending a document via multipart POST update request, if a
doc parameter name contains too much chars, the POST method fails and I can see
the following exception in the Solr logs
Hi,
I recently experienced some problems with the update extract handler. When
sending a document via multipart POST update request, if a doc parameter name
contains too much chars, the POST method fails and I can see the following
exception in the Solr logs :
ERROR 2019-06-20T09:43:41,089
: solr-user@lucene.apache.org
Betreff: Re: update/extract override ExtractTyp
On 1/4/2017 8:12 AM, sn0...@ulysses-erp.com wrote:
> Is it possible to override the ExtractClass for a specific document?
> I would like to upload a XML Document, but this XML is not XML conform
>
> I n
testing purpose.
>
>
> The update/extract process failes every time with an 500 error.
>
> I tried to override the Content-Type with "text/plain" but get still
> the XML parse error.
If you send something to the /update handler, and don't tell Solr that
it is
Hello
Is it possible to override the ExtractClass for a specific document?
I would like to upload a XML Document, but this XML is not XML conform
I need this XML because it is part of a project where a corrupt XML is
need, for testing purpose.
The update/extract process failes every time
This question should be posted on tika mailing list. It is not related to
index or search but about parsing content of image.
On Sun, Jun 5, 2016 at 10:20 PM, SIDDHAST® Roshan
wrote:
> Hi All,
>
> we are using the application for indexing and searching text using
> solr. we refered the guide pos
Hi All,
we are using the application for indexing and searching text using
solr. we refered the guide posted
http://hortonworks.com/hadoop-tutorial/indexing-and-searching-text-within-images-with-apache-solr/
Problem: we are want to index hindi images. we want to know how to set
configuration para
Thanks Shawn, I appreciate the information.
On Wed, Apr 9, 2014 at 10:27 AM, Shawn Heisey wrote:
> On 4/9/2014 7:47 AM, Jamie Johnson wrote:
> > This is being triggered by adding the commitWithin param to
> > ContentStreamUpdateRequest (request.setCommitWithin(1);). My
> > configuration ha
On 4/9/2014 7:47 AM, Jamie Johnson wrote:
> This is being triggered by adding the commitWithin param to
> ContentStreamUpdateRequest (request.setCommitWithin(1);). My
> configuration has autoCommit max time of 15s and openSearcher set to false.
> I'm assuming that changing openSeracher to tru
at 8:43 PM, Jamie Johnson wrote:
> > Below is the log showing what I believe to be the commit
> >
> > 07-Apr-2014 23:40:55.846 INFO [catalina-exec-5]
> > org.apache.solr.update.processor.LogUpdateProcessor.finish [forums]
> > webapp=/solr path=/update/extract
> &g
-commit-in-sorlcloud/
Best,
Erick
On Mon, Apr 7, 2014 at 8:43 PM, Jamie Johnson wrote:
> Below is the log showing what I believe to be the commit
>
> 07-Apr-2014 23:40:55.846 INFO [catalina-exec-5]
> org.apache.solr.update.processor.LogUpdateProcessor.finish [forums]
> webapp=/
Below is the log showing what I believe to be the commit
07-Apr-2014 23:40:55.846 INFO [catalina-exec-5]
org.apache.solr.update.processor.LogUpdateProcessor.finish [forums]
webapp=/solr path=/update/extract
params={uprefix=attr_&literal.source_id=e4bb4bb6-96ab-4f8f-8a2a-1cf37dc1
> work when I am using the /update/extract request handler. It looks like a
> commit is happening from the logs, but the documents don't become available
> for search until I do a commit manually. Could this be some type of
> configuration issue?
7;t seem to
> work when I am using the /update/extract request handler. It looks like a
> commit is happening from the logs, but the documents don't become available
> for search until I do a commit manually. Could this be some type of
> configuration issue?
I'm running solr 4.6.0 and am noticing that commitWithin doesn't seem to
work when I am using the /update/extract request handler. It looks like a
commit is happening from the logs, but the documents don't become available
for search until I do a commit manually. Could this
HttpSolrServer allows to send multiple documents at once. But they
need to be extracted/converted on the client. However, if you know you
will be sending a lot of documents to Solr, you are better off to run
Tika locally on the client (or as a standalone network server). A lot
more performant.
I a
Who knows how to index a lot of files with ExtractingRequestHandler using a
single query?
The "#" character introduces the "fragment" portion of a URL, so
"/dev/update/extract" is not a part of the "path" of the URL. In this case
the URL "path" is "/solr/" and the server is simply complaining that there
is no code regi
ucted url in single quotes
instead of double quotes, if you're on a Unixoid platform.)
On Mon, Mar 10, 2014 at 2:51 PM, Priti Solanki wrote:
> Hi all,
>
> Following throw "The request resource is not available"
>
>
> curl "
>
> http://localhost:8080/
Hi all,
Following throw "The request resource is not available"
curl "
http://localhost:8080/solr/#/dev/update/extract?stream.file=/home/priti/$file&literal.id=document$i&commit=true
"
I don't understand what is literal.id ?? Is it mandatory. [Please share
read
gs one-by-one...
> >>
> >> Best,
> >> Erick
> >>
> >> On Sat, Sep 21, 2013 at 7:34 AM, Nutan <[hidden email]<
> http://user/SendEmail.jtp?type=node&node=4091391&i=0>>
> >> wrote:
> >>
> >> > Yes I do g
the stock config and try removing things one-by-one...
>>
>> Best,
>> Erick
>>
>> On Sat, Sep 21, 2013 at 7:34 AM, Nutan <[hidden
>> email]<http://user/SendEmail.jtp?type=node&node=4091391&i=0>>
>> wrote:
>>
>> > Yes I do g
> wrote:
>
> > Yes I do get the solr admin page.And im not using the example config
> file,I
> > have create mine own for my project as required.I have also defined
> > update/extract in solrconfig.xml.
> >
> >
> > On Tue, Sep 17, 2013 at 4:45 AM, Chris Host
d try removing things one-by-one...
Best,
Erick
On Sat, Sep 21, 2013 at 7:34 AM, Nutan wrote:
> Yes I do get the solr admin page.And im not using the example config file,I
> have create mine own for my project as required.I have also defined
> update/extract in solrconfig.xml.
>
&
Yes I do get the solr admin page.And im not using the example config file,I
have create mine own for my project as required.I have also defined
update/extract in solrconfig.xml.
On Tue, Sep 17, 2013 at 4:45 AM, Chris Hostetter-3 [via Lucene] <
ml-node+s472066n409045...@n3.nabble.com>
are
using the stock Solr 4.2 example configs, then
http://localhost:8080/solr/update/extract should not give you a 404 error.
if however you are using some other configs, it might not work unless
those configs register a handler with the path /update/extract.
Using the jetty setup provided
Is /solr/update working?
Check solrconfig to see that /update/extract is configured as in the
standard Solr example.
Does /solr/update/extract work for you using the standard Solr example?
-- Jack Krupansky
-Original Message-
From: Nutan
Sent: Sunday, September 15, 2013 2:37 AM
To
I am working on Solr 4.2 on Windows 7. I am trying to index pdf files.I
referred Solr Cookbook 4. Tomcat is using 8080 port number. I get this
error:requested url solr/update/extract not available on this server
When my curl is :
curl "http://localhost:8080/solr/update/extract?literal.id=1&a
Your email is vague in terms of what you are actually *doing* and what
behavior you are seeing.
Providing specific details like "This is my schema.xml and this is my
solrconfig.xml; when i POST this file to this URL i get this result and i
would instead like to get this result" is useful for
You can use the dynamic fields feature of Solr to map unknown field
names to types.
For example, a dynamic field named as "*_s" i.e. any field name ending
with _s can be mapped to string and so on. In your cases, if your
field names do not follow a set pattern, then you can even specify a
dynamic
hi,
while indexing document with unknown fields, its adding unknown fields in
schema but its always guessing it as string type. is it possible to specify
default field type for unknown fields to some other type, like text so that
it gets tokenized? also can we specify other properties by default l
I added to the schema.xml and now its working.
*
Thank you very much Jack. *
--
View this message in context:
http://lucene.472066.n3.nabble.com/update-extract-error-in-Solr-4-3-1-tp4079555p4079564.html
Sent from the Solr - User mailing list archive at Nabble.com.
You need a dynamic field pattern for "ignored_*" to ignore unmapped
metadata.
-- Jack Krupansky
-Original Message-
From: franagan
Sent: Monday, July 22, 2013 5:14 PM
To: solr-user@lucene.apache.org
Subject: /update/extract error
Hi all,
im testing solrcloud (version 4.3
= new CloudSolrServer("localhost:2181");
server.setDefaultCollection("tika");
ContentStreamUpdateRequest up = new
ContentStreamUpdateRequest("/update/extract");
up.addFile(new File("C:\\sample.pdf"), "applic
The Extract Request Handler invokes the classes from the extraction package.
https://svn.apache.org/repos/asf/lucene/dev/trunk/solr/contrib/extraction/src/main/java/org/apache/solr/handler/extraction/ExtractingRequestHandler.java
This is package into the apache-solr-cell jar.
Regards,
Jayendra*
Hi all,
when we handle extract request handler what class gets invoked.. I
need to know the navigation of classes when we send any files to solr.
can anybody tell me the classes or any sources where i can get the answer..
or can anyone tell me what classes get invoked when we start the
solr.
Follow-up -
This is now working (sadly I'm not sure exactly why!) but I've
successfully used curl (under windows) and the following examples to
parse content
curl
http://localhost:8080/apache-solr-1.4-dev/update/extract?extractOnly=tru
e --data-binary @curl-config.pdf -
: I see the source - but no classes or jar that seems to fit the bill.
:
: I've had problems getting ant to build from the nightly trunk. I'm of
...
: If there is an existing jar of the ExtractingRequestHandler classes that
: I might download - please point me to it.
If you are downloading a
: The \contrib and \dist directories were copied directly below the
: "webapps\apache-solr-1.4-dev" unchanged from the example.
...that doesn't sound right, they shouldn't be copied into webapps at all.
can you show a full directory structure...
: Im the catalina log I see all the "Adding spec
ll look at this today - thanks again - much appreciated.
-Original Message-
From: Grant Ingersoll [mailto:gsing...@apache.org]
Sent: Tuesday, November 03, 2009 8:12 AM
To: solr-user@lucene.apache.org
Subject: Re: Getting update/extract RequestHandler to work under Tomcat
Try making it a
Try making it a non-Lazy loaded handler. Does that help?
On Nov 2, 2009, at 4:37 PM, Glock, Thomas wrote:
Hoping someone might help with getting /update/extract
RequestHandler to
work under Tomcat.
Error 500 happens when trying to access
http://localhost:8080/apache-solr-1.4-dev/update
Hoping someone might help with getting /update/extract RequestHandler to
work under Tomcat.
Error 500 happens when trying to access
http://localhost:8080/apache-solr-1.4-dev/update/extract/ (see below)
Note /update/extract DOES work correctly under the Jetty provided
example.
I think I must
48 matches
Mail list logo