Hi,
I am new to Solr and its configuration. I need to index some pdf files and
for that reason I thought of using the extractingrequesthandler.
I use Apache tomcat and run solr from it.
I used the following command to index a pdf file.
C:\Users\vikky\Downloads\curl-7.21.0-win64-nossl>curl.e
Hello,
I had the same problem, only the first result was indexed in the sub-entity,
but was solved by setting the 'multiValued' attribute to "true" on the field
in schema.xml
--
View this message in context:
http://lucene.472066.n3.nabble.com/Sub-entities-tp2598369p2837723.html
Sent from the So
Li,
there's many ways to output data to the front-end, including solr-itas (a
velocity front-end) and the xslt. Both work almost out of the box (for /itas
you need to use the things described in contribs.
Solr can be populated, at upload time, with verbatim view code (e.g. HTML)
which, I think
Hi All!
I want to integrate Uima-solr . I followed the steps in the
readme file.I am using apache solr3.1. The jar file starts fine. But I
dont know the exact syntax in solrj to index my documents for Uima-solr
integration .Can anyone help me out rgarding this>
Thanks!
Isha Garg
I think most people are probably writing an application, at least most people
on this list. I am not aware of whether any popular CMS's somehow provide a way
to be a front-end to Solr. It seems a bit out of the mission of a CMS to me and
unlikely, but I'm not familiar with those CMSs (haven't ev
Hi all,
Does the lucene-solr git repository have a tag that marks the 3.1 release?
Context: I want to apply a patch to 3.1 and wish to start from a
well-defined point (i.e. official 3.1 release)
Executing these commands, I would have expected to see a tag marking the 3.1
release. I only see "b
H, grouping isn't *in* 1.4.1 I don't think, so how are you
trying to do this?
Have you looked at faceting?
Best
Erick
On Mon, Apr 18, 2011 at 2:41 AM, sukalyan_banerjee
wrote:
> Hi All,
> I am using apache-solr-1.4.1 and jdk1.6. I have the following scenario.
> I have 3 categories of data i
Solr is a seach *engine*. It doesn't have anything to do with the presentation.
Most users have an application layer that gets the documents from Solr
via http (in XML, JSON or other format) and then extracts the pieces to
create a web page. What you use for the application layer is totally your
ch
Right, I placed my files relative to solr_home, not in it -- but obviously
having a solr_home/lucene-libs directory didn't do me any good. :-)
- Demian
> -Original Message-
> From: Jonathan Rochkind [mailto:rochk...@jhu.edu]
> Sent: Monday, April 18, 2011 1:46 PM
> To: solr-user@lucene.a
Thanks! apache-solr-analysis-extras-3.1.jar was the missing piece that was
causing all of my trouble; I didn't see any mention of it in the documentation
-- might be worth adding!
Thanks,
Demian
> -Original Message-
> From: Robert Muir [mailto:rcm...@gmail.com]
> Sent: Monday, April 18
I'm now having the same problem but I'm not finding the problem yet.
$ bin/nutch solrindex http://localhost:8080/solr crawl/crawldb/0
crawl/linkdb crawl/segments/0/20110418100309
SolrIndexer: starting at 2011-04-18 10:03:40
java.io.IOException: Job failed!
But everything else seems to have worked
Hi All,
I am using apache-solr-1.4.1 and jdk1.6. I have the following scenario.
I have 3 categories of data indexed in solr i.e. CITIES, STATES, COUNTRY.
When I query data from SOLR I need the data from SOLR based on the following
criteria:
In a single query to Solr Engine I need data fetc
Thanks Jan. lol.
1. For example, I have a large solr database that contains 30M documents.
I want to show the datas in a web application, how should I do it? Write an
application or use a CMS like Liferay, Magnolia, or Drupal to do it.
On Mon, Apr 18, 2011 at 11:35 AM, Jan Høydahl wrote:
> Hi L
I'm sorry, you're right, I was thinking in the 2GB default value for
maxMergeMB.
*Juan*
On Mon, Apr 18, 2011 at 3:16 PM, Burton-West, Tom wrote:
> >> As far as I know, Solr will never arrive to a segment file greater than
> 2GB,
> >>so this shouldn't be a problem.
>
> Solr can easily create a fi
Hi Li,
Who are you referring to in your question, having 30M docs?
Solr is possible to integrate in tons of different ways. Perhaps if you
describe your use case and requirements, we can suggest the best way for your
particular situation. Please elaborate on what you are trying to accomplish.
Hello guys, how do you guys output the solr data into frontend? I know you
guys have 30M documents. Are you guys writing an application to do it? or
are you guys using a CMS with solr intergation? Thanks
>> As far as I know, Solr will never arrive to a segment file greater than 2GB,
>>so this shouldn't be a problem.
Solr can easily create a file size over 2GB, it just depends on how much data
you index and your particular Solr configuration, including your
ramBufferSizeMB, your mergeFactor, and
On Mon, Apr 18, 2011 at 1:31 PM, Demian Katz wrote:
> Hello,
>
> I'm interested in trying out the new ICU features in Solr 3.1. However, when
> I attempt to set up a field type using solr.ICUTokenizerFactory and/or
> solr.ICUFoldingFilterFactory, Solr refuses to start up, issuing "Error
> load
I don't think you want to put them in solr_home, I think you want to put
them in solr_home/lib/. Or did you mean that's where you put them?
On 4/18/2011 1:31 PM, Demian Katz wrote:
Hello,
I'm interested in trying out the new ICU features in Solr 3.1. However, when I attempt
to set up a fiel
Hello,
I'm interested in trying out the new ICU features in Solr 3.1. However, when I
attempt to set up a field type using solr.ICUTokenizerFactory and/or
solr.ICUFoldingFilterFactory, Solr refuses to start up, issuing "Error loading
class" exceptions.
I did see the README.txt file that menti
It's probably not accurate to say that a lot of sites were *relying* on that
feature. It's an optimization.
Getting a working patch applying to trunk is on my TODO-list within the next
couple months.
https://issues.apache.org/jira/browse/SOLR-752
"Watch" the issue to see when I get to it.
~ Dav
hi markus
i misunderstood before.
i use nutch.1.2-rc4 with solr-.4.0 trunk. You just need replace these files
apache-solr-core-4.0-SNAPSHOT.jar
apache-solr-solrj-4.0-SNAPSHOT.jar
which are in solr/dist directory with nutch 1.4.1 solrj and core.
--
View this message in context:
http://luce
There is no problem with your files. Nutch still ships SolrJ 1.4.1. If you
would be using Solr 3.1 you would get a javabin error and not a Not Found
error.
On Monday 18 April 2011 15:37:42 McGibbney, Lewis John wrote:
> Hi Ramires,
>
> I have been using Solr 1.4.1
>
> My understanding from the
Hi Ramires,
I have been using Solr 1.4.1
My understanding from the example solrconfig.xml is that jar's will be loaded
from the /lib directory. I do not have a /dist directory as I have copied the
example directory as my solr home directory therefore I have commented out
these entires in the s
The first question I'd have is whether you're somehow not committing after your
full-import command.
And have you looked at:
http://wiki.apache.org/solr/DataImportHandler#interactive?
This is a little-known feature in Solr to help with DIH.
Is it possible that your JDBC configuration is defaultin
any ideas why in this case the stats summaries are so slow ? Thank you
very much in advance for any ideas/suggestions. Johannes
2011/4/5 Johannes Goll
> Hi,
>
> thank you for making the new apache-solr-3.1 available.
>
> I have installed the version from
>
> http://apache.tradebit.com/pub//luc
Hi,
I am using a DataImportHandler to get files from the file system, if I
do the url
http://localhost:8983/solr/tester/dataimport?command=full-import it
ends up indexing 11 documents.
If I do
http://localhost:8983/solr/tester/dataimport?command=full-import&rows=817
(the number of documents I have
This is a problem of these files in nutch lib. You can easily change these
files with in solr dist directory.
apache-solr-core-1.4.0.jar
apache-solr-solrj-1.4.0.jar
--
View this message in context:
http://lucene.472066.n3.nabble.com/Indexing-from-Nutch-crawl-tp2833862p2834270.html
Sent from th
And you are really sure there's a Solr instance runnning having an update
handler at : http://localhost:8080/wombra/data/update ? Anyway, your URL is
somewhat uncommon in Solr land. It's usually something like:
http://:/solr/[]/update/
On Monday 18 April 2011 14:03:53 McGibbney, Lewis John wrot
Hi Markus,
hadoop.log from beginning of solr commands as follows
2011-04-18 11:27:05,480 INFO solr.SolrIndexer - SolrIndexer: starting at
2011-04-18 11:27:05
2011-04-18 11:27:05,562 INFO indexer.IndexerMapReduce - IndexerMapReduce:
crawldb: crawl/crawldb
2011-04-18 11:27:05,562 INFO indexer.
Hello,
I think I have found something extrange with local params and edismax. If I do
querys like :
"params":{
"hl.requireFieldMatch":"true",
"hl.fragsize":"200",
"json.wrf":"callback0",
"indent":"on",
"hl.fl":"domicilio,deno",
"wt":"json",
"hl":"true",
Hmm, ok I see the schema was wrong - I was calling the TEXT field
txt... also now I am getting results on my title search after another
restart and reindex - setting the TXT fields to be multiValued.
Thanks,
Bryan Rasmussen
On Mon, Apr 18, 2011 at 1:09 PM, bryan rasmussen
wrote:
> well basically
well basically I copied out the RSS example as I figured that would be
the closest to what I wanted to do
Can you include hadoop.log output? Likely the other commands fail as well but
don't write the exception to stdout.
On Monday 18 April 2011 12:47:41 McGibbney, Lewis John wrote:
> Hi list,
>
> I am using Nutch-1.3 branch, which I checked out today to crawl a couple of
> urls in local mode. I have
If a document contains multiple 'txt' fields, it should be marked as
'multiValued'.
But if I'm understanding well, you also tried this ? :
And for your search (MomsManual), could you give us your analyzer from the
schema.xml please ?
Ludovic.
-
Jouve
France.
--
View this message in co
Hi list,
I am using Nutch-1.3 branch, which I checked out today to crawl a couple of
urls in local mode. I have been using Solr Solr 1.4.1 within my web app but I
am running into some problems during the indexing stages. I have three commands
getting sent to Solr these are
echo "- SolrInde
hah, actually I tried with complete xpaths earlier but they weren't
working but that was because I had made a mistake in my foreach.. and
then I decided that probably the foreach and the other xpaths were
being concatenated.
however it is not absolutely correct yet, if I run
http://localhost:8983/
did you try with the comlete xpath ?
Ludovic.
-
Jouve
France.
--
View this message in context:
http://lucene.472066.n3.nabble.com/all-searches-return-0-hits-what-have-I-done-wrong-tp2833706p2833798.html
Sent from the Solr - User mailing list archive at Nabble.com.
Also if I check
solr/tester/dataimport it responds:
−
0
0
−
−
dataimporter.xml
idle
−
0
1634
0
2011-04-18 11:55:47
−
Indexing completed. Added/Updated: 0 documents. Deleted 0 documents.
2011-04-18 11:55:48
2011-04-18 11:55:48
0
0:0:0.922
−
This response format is experimental. It is
Hi,
I am starting my solr instance with the command java
-Dsolr.solr.home="./test1/solr/" -jar start.jar
where I have a solr.xml file
In the folder tester I have configurations - adapted from the rss examples
DataImporter.xml
On 18.04.11 09:23, Bill Bell wrote:
It runs delta imports faster. Normally you need to get the Pks that
changed, and then run it through query="" which is slow when you have a
lot of Ids
but the query="" only adds/updates entries. I'm not sure how to delete
entries
by running a query like "
It runs delta imports faster. Normally you need to get the Pks that
changed, and then run it through query="" which is slow when you have a
lot of Ids
It would be better if someone could just write a new fastdeltaquery="" s
that you could do it in one step and also remove the queries...
On
Hi,
when using
http://wiki.apache.org/solr/DataImportHandlerDeltaQueryViaFullImport to
periodically
run a delta-import, is it necessary to run a separate "normal"
delta-import after it to delete entries
from the index (using deletedPkQuery)?
If so, what's the point of using this method for r
43 matches
Mail list logo