we are testing the upgrade to solr 3.2. a quick test look good. solr 3.2. comes
up and we can do searches with our configs (using the "old" dismax handler,
which i have inserted in solrconfig.xml). only problem is, that i am not able
to set up user auth in jetty.
i took the same config files tha
last week we put our solr in production. it was a very smooth start. solr
really works great and without any problems so far.
its a huge improvement over our old intranet search
i wonder however whether we can increase the search performance of our solr
installation, just to make the search ex
that's not the correct lib-dir. thats the lib-dir for jetty. please check you
installation.
if you use the tgz from lucene.apache.org/solr it should look like (we have
added a few additonal jars)-
thats all of the jars in our solr-dir:
./contrib/clustering/lib/carrot2-mini-3.1.0.jar
./contrib/c
in the standard solr distribution its unter ./contrib/extraction/lib
solr 1.4.1 comes with a tika included. you want to have a newer version of
tika, right?
> -Ursprüngliche Nachricht-
> Von: Andreas Jung [mailto:li...@zopyx.com]
> Gesendet: Dienstag, 12. Oktober 2010 19:14
> An: solr-u
the second time we had the error
java.lang.OutOfMemoryError: PermGen space
and solr stopped responding.
we use the default jetty installation with jdk1.6.0_21. after the last
time i tried to set the garbage collector "right"
these are my settings:
-D64 -server -Xms892m -Xmx2048m
hi,
it's the second time i am stumble across some strange behaviour:
in my schema.xml i have defined
i can't place the PatternReplaceFilter before the WhitespaceTokenizer. i
have the schema like above, di
hi,
what is the best way to deal with searches with special chars like §
(paragraph), € (euro), @ (at in emails), % and so forth.
i think that the WordDelimiterFilters is working on such chars (on index-time
and on query-time).
the greatest problem i see is, that there can be an optional space
it is not that complicated to write an own GUI.
we are working on an integration to our intranet server...
> -Ursprüngliche Nachricht-
> Von: Peter Spam [mailto:ps...@mac.com]
> Gesendet: Donnerstag, 1. Juli 2010 03:21
> An: solr-user@lucene.apache.org
> Betreff: Re: Very basic question
hi,
i have two questions:
1) how can i set a default value on an imported field if the
field/column is missing from a SQL query
2) i had a problem with the dataimporthandler. in one database column
(WebDst) i have a string with a comma/semicolon seperated numbers, like
100,200; 300;400,
- solr.xml have to reside in the solr.home dir. you can setup this with the
java-option
-Dsolr.solr.home=
- admin is per core, so solr/CORENAME/admin will work
it is quite simple to setup.
> -Ursprüngliche Nachricht-
> Von: xdzgor [mailto:p...@alphasolutions.dk]
> Gesendet: Mittwo
did you send a commit after the last doc posted to solr?
> -Ursprüngliche Nachricht-
> Von: Scott Zhang [mailto:macromars...@gmail.com]
> Gesendet: Dienstag, 8. Juni 2010 08:30
> An: solr-user@lucene.apache.org
> Betreff: Re: Distributed Search doesn't response the result set
>
> Hi. A
> I'm not sure, whether the case you are dealing with (search for
> profi-auskunft should match profiauskunft) is appropriately
> addressed by
> the WordDelimiterFilter. What about using the
> PatternReplaceCharFilter
> at query time to eliminate all intra-word hyphens?
>
maybe it would be b
>
> Sorry Markus, I mixed up the index and query field in
> analysis.jsp. In
> fact, I meant that a search for profiauskunft matches profi-auskunft.
>
> I'm not sure, whether the case you are dealing with (search for
> profi-auskunft should match profiauskunft) is appropriately
> addressed by
i am not very sure, whether this helps me.
i see the point, that there will be problems.
but
the default-config for index is:
and for query:
with this settings i don't find "profiauskunft" when searching for
"profi-auskunft" (analyse0.jpg)
if i use "catenateWords="1""
analysis.jsp
i am wondering why a search term with hyphen doesn't match.
my search term is "prof-auskunft". in WordDelimiterFilterFactory i have
catenateWords, so my understanding is that profi-auskunft would search
for profiauskunft. when i use the analyse panel in solr admi i see that
profi-auskunft matches
ok,but is there an easy way to influence the format of json output?
eg field order, names etc. maybe i want to group the result differently or add
some infos
> -Ursprüngliche Nachricht-
> Von: Jon Baer [mailto:jonb...@gmail.com]
> Gesendet: Mittwoch, 26. Mai 2010 19:39
> An: solr-user@lu
i have also thought about an autosuggest for our intranet search.
one other solution could be:
put all the searched queries into a database and do a lookup not on the terms
indexed by solr but rather a lookup to what have been searched in the past.
we have written a small script, that takes the
sorry, the problem was sitting in front of the monitor!
it is not an error or something, i forgot that some documents didn't have all
fields filled,
so its absolutly normal, that not all fields were in the result.
markus
> -Ursprüngliche Nachricht-
> Von: markus.rietz...@rzf.fin-nrw
when i do a search, eg.
http://xxx:8983/solr/select?q=steuer&fl=score,id,__intern,title,__source,_dienststelle,_erstelldatum,__cyear,_stelle
i get a normal result, like
1.3633566
2009
0
zzz
xyz
2009020200
Presse- u. Informationsreferat
34931684
Merkblatt Vereine und Steuern
when i do a
you write:
> Our overall response (front end + SOLR) averages 0.5s to 0.7s with
> SOLR typicall taking about 100 - 300 ms.
is the 100-300ms the time your application needs to query solr and get
the response? what are the times if you query SOLR directly without
your frontend?
we are also in th
we use dismax and highlighting works fine.
the only thing we had to add to the query-url was
&hl.fl=FIELD1,FIELD2
so we had to specify which fields should be used for highlighting.
> -Ursprüngliche Nachricht-
> Von: fabritw [mailto:fabr...@gmail.com]
> Gesendet: Mittwoch, 28. A
i have send you a private mail.
markus
> -Ursprüngliche Nachricht-
> Von: Johan Cwiklinski [mailto:johan.cwiklin...@ajlsm.com]
> Gesendet: Montag, 26. April 2010 10:58
> An: solr-user@lucene.apache.org
> Betreff: "Solr 1.4 Enterprise Search Server" book examples
>
> Hello,
>
> We've r
we try to index some pdf and other documents with solr cell/tika.
so far our crawler downloads the documents and post them to solr cell.
this works, the documents get indexed, some fields are filled etc.
our crawler is written in perl. we prepare the following params to post
to solr:
my $
hello,
we want to index and search in our intranet documents.
the field "body" contains html-tags.
in our schema.xml we have a fieldType text_de (see at the end of this mail)
which uses charFilter solr.HTMLStripCharFilterFactory with index.
so this is no problem. the text is put into the index
yes i am using solr.xml, although there is only one core defined
at the moment. so reloading is only possible with a core-setup, right?
> -Ursprüngliche Nachricht-
> Von: Ahmet Arslan [mailto:iori...@yahoo.com]
> Gesendet: Samstag, 10. April 2010 15:55
> An: solr-user@lucene.apache.org
>
have you given the url like
&qf=field1^100+field2^200
i also first tried something like &qf="field1^100 field2^200" which never
worked, when i used the + sign it works.
markus
> -Ursprüngliche Nachricht-
> Von: Mark Fletcher [mailto:mark.fletcher2...@gmail.com]
> Gesendet: Sam
i am wondering how config files like synonyms.txt or stopwords.txt can
be refreshed without restarting of solr,
maybe also how changes in solrconfig.xml or schema.xml can be refreshed?
i can use a multicore setup - i just tested it with a "multicore"-setup
with one one core (core0), there i can
little question: what's the difference between a MustHavePermission and a
protected
document?
at the moment we are developing a new search for our intranet and using solr.
we also have some protected documents and implemented this kind of filter like
you.
i just think on using a true filter (fq
go to solr admin and then click on info, right in the first line you see the
solr version
> -Ursprüngliche Nachricht-
> Von: Marc Wilson [mailto:wo...@fancydressoutfitters.co.uk]
> Gesendet: Dienstag, 2. März 2010 09:55
> An: Solr
> Betreff: Solr Version
>
> Hi,
>
> This is probably a
ok,
i was looking for all types of "max" but somehow didn't saw the
maxFieldLength.
this is a global parameter, right? can this be defined on a field basis?
global would be enough at the moment.
thank you
> -Ursprüngliche Nachricht-
> Von: Shalin Shekhar Mangar [mailto:shalinman...
hi,
what is the default maximum charsize per field? i found a macChars
paramater for copyField but i don't think, that this is what i am
looking for.
we have indexed some documents via tika/solrcell. only the beginning of
these documents can be searched. where can i define the maximum size of
a do
mplete box
> on my computer (my site is built in localhost, solr is in localhost).
>
> Am I wrong ?
>
> Thanks
>
>
> Markus.Rietzler wrote:
> >
> > very simple, in the javascript you have
> >
> > $("#artist").autocomplete(
> >
very simple, in the javascript you have
$("#artist").autocomplete(
'http://localhost:8983/solr/completion/select/?wt=json', {
you have to change localhost to the correct servername
--
mit freundlichen Grüßen
Markus Rietzler -
Rechenzentrum der Finanzverwaltung NRW
0211/4572-2130
>
curl reads from local file or stdin, so you could do something like
if it only a single file from a webserver
curl http://someserver/file.html/ | curl
"http://localhost:8983/solr/update/extract?extractOnly=true"; -F na...@-
but this way no crawling, no link following etc...
--
mit freundlic
thanxs,
i know and read that page. sending additional meta-tags with the curl call is
no problem. i only thought that there might be a way to use the xml-approach
also with PDF files. i'll go the "curl"-way for that files.
--
mit freundlichen Grüßen
Markus Rietzler -
Rechenzentrum der Finanzver
hi,
we want to use SOLR as our intranet search engine.
i downloaded the nightly bild of solr 1.4. pdf extraction does via Solr
Cell/Tika. i can send the pdf via curl
to solr.
we do have a large set of meta-tags to all our intranet documents, including
PDF, PPT etc. to import html
files from our
36 matches
Mail list logo