thanks for your reply
the content of xml file is chinese
===
name="prices">
name="urls">
===
Otis Gospodnetic 写道:
Could you paste the XML you are posting?
Otis
--
Sematext -- http://semate
If you have custom XML take a look at DataImportHandler
http://wiki.apache.org/solr/DataImportHandler
On Fri, Sep 19, 2008 at 12:24 PM, Otis Gospodnetic
<[EMAIL PROTECTED]> wrote:
> The format is fixed, you can't change it -- something on the Solr end needs
> to parse that XML and expects specifi
hi, all
when i post xml files to solr, it's interrupted by this: Illegal character
how can i deal with it ?
is there any solution to ignore Illegal character in documents feeded ?
thanks
Have you actually taken a snapshot? There is more to it than just
starting rsyncd.
See http://wiki.apache.org/solr/CollectionDistribution
On Sep 18, 2008, at 12:13 PM, sunnyfr wrote:
Hi
sorry I think I've started properly rsyncd :
[EMAIL PROTECTED]:/# ./data/solr/books/bin/rsyncd-enable
Yes reloading a core can be used. I guess the proposal is a way to
update the config and schema files over the network through SOLR
rather than by the filesystem. This will make grid computing and
schema updates much faster.
On Fri, Sep 19, 2008 at 2:11 AM, Noble Paul നോബിള് नोब्ळ्
<[EMAIL PROT
You were correct, I assumed that as I did not want to break the values in the
fields into seperate tokens that I could just use "string" as the type in the
schema for these fields.
As soon as I switched to a custom field type using
solr.KeywordTokenizerFactory and reindexed it all started worki
first, u should escape some string like (code by php)
> function escapeChars($string) {
>
$string = str_replace("&", "&", $string);
$string = str_replace("<", "<", $string);
$string = str_replace(">", ">", $string);
$string = str_replace("'", "'", $string);
$string = str_replace('"', """, $str
I was finally given the go-ahead to upgrade from Solr 1.2 to 1.3 in our
environment here at work now that 1.3 is final. However I'm running into a
couple problems that I'm having trouble finding solutions to.
First, I've added a class to our Solr distribution that extends
StandardRequestHandler.
Shalin Shekhar Mangar wrote:
>
> On Fri, Sep 19, 2008 at 5:55 AM, oleg_gnatovskiy <
> [EMAIL PROTECTED]> wrote:
>
>>
>> Hello. I am using the spellcheck component
>> (https://issues.apache.org/jira/browse/SOLR-572). Since the spell checker
>> index is kept in RAM, it gets erased every time the
Save the file to disk with a name ending in ".xml", then open it in a
browser. The browser will show you a parse error, usually with the line
and column number.
You cannot ignore illegal characters. You must send legal XML.
Oddly, I answered this same question on the search_dev list yesterday.
w
On Fri, Sep 19, 2008 at 10:07 PM, oleg_gnatovskiy <
[EMAIL PROTECTED]> wrote:
>
> Is there any way to do it for an external (file-based) dictionary?
>
SpellCheckComponent always reload on the dictionary in the firstSearcher
event. This works if you are using file system based index. However with
James liu wrote:
> first, u should escape some string like (code by php)
>
>> function escapeChars($string) {
>>
> $string = str_replace("&", "&", $string);
>
> $string = str_replace("<", "<", $string);
>
> $string = str_replace(">", ">", $string);
>
> $string = str_replace("'", "'", $string);
I created one. https://issues.apache.org/jira/browse/SOLR-780
By the way you pointed out that true
would solve the problem, but that doesn't make it rebuild on startup right?
This works at rebuilding the index with every update, which is different.
Shalin Shekhar Mangar wrote:
>
> On Fri, Sep
Hello,
I tried to have the spellcheck component to write to a drive index. My
config is a s follows:
org.apache.solr.spelling.FileBasedSpellChecker
external
spellings.txt
UTF-8
true
./spellIndex
org.apache.lucene.search.spell.JaroWinklerDistance
On Fri, Sep 19, 2008 at 11:10 PM, oleg_gnatovskiy <
[EMAIL PROTECTED]> wrote:
>
> I created one. https://issues.apache.org/jira/browse/SOLR-780
>
Thanks!
> By the way you pointed out that true
> would solve the problem, but that doesn't make it rebuild on startup right?
> This works at rebuildi
Hi Oleg,
You need to specify spellCheckIndexDir to use a FSDirectory. Also, remove
the buildOnCommit because it is useless for file based spell checker.
./spellchecker
On Fri, Sep 19, 2008 at 11:23 PM, oleg_gnatovskiy <
[EMAIL PROTECTED]> wrote:
>
> Hello,
>
> I tried to have the spellcheck c
oleg_gnatovskiy wrote:
>
> Hello,
>
> I tried to have the spellcheck component to write to a drive index. My
> config is a s follows:
>
>
>name="classname">org.apache.solr.spelling.FileBasedSpellChecker
> external
> spellings.txt
> UTF-8
> true
>
>
On Sep 19, 2008, at 11:49 AM, Mark Baird wrote:
I was finally given the go-ahead to upgrade from Solr 1.2 to 1.3 in
our
environment here at work now that 1.3 is final. However I'm running
into a
couple problems that I'm having trouble finding solutions to.
First, I've added a class to our
Second, I've got some code running in the same JVM as Solr that
does some
stuff like getting the latest timestamp in the index to determine
if we need
to pull an update from our product info database, kicking off an
optimize
every night at 2:00AM, stuff like that. However I take it that
SearchComponent is the class I was missing. Looks like if I can provide an
entirely new implementation of that it will be a lot cleaner than the hack I
had been using in 1.2 over top of facets. What I'm doing is implementing
some aggregation functions like avg() and sum() that SQL has. This way
On Sep 19, 2008, at 3:22 PM, Mark Baird wrote:
SearchComponent is the class I was missing. Looks like if I can
provide an
entirely new implementation of that it will be a lot cleaner than
the hack I
had been using in 1.2 over top of facets.
thats why SearchComponets got added!
It was cr
Question -
So if I issued a dataimport?command=delta-import&entity=one,two,three
Would this also hit items w/o a delta-import like four,five,six, etc?
Im trying to set something up and I ended up with 28k+ documents which
seems more like a full import, so do I need to do something like delt
Actually how does ${deltaimporter.last_index_time} know which entity
Im specifically updating? I feel like Im missing something, can it
work like that?
Thanks.
- Jon
On Sep 19, 2008, at 4:14 PM, Jon Baer wrote:
Question -
So if I issued a dataimport?command=delta-import&entity=one,two,t
Hello
I'm doing a facet search like the following. The content field schema is
/solr/select?q=dirt
field:www.example.com&facet=true&facet.field=content&facet.limit=-1&facet.mincount=1
If I run this on a serve
Hi,
I am trying to delete a record from the index using SolrJ. When I
execute it I get a status of 0 which means success. I see that the
"cummulative_deletbyquery" count increases by 1 and also the "commit"
count increases by one. I don't see any decrease on the "numDocs" count.
When I query i
Inline below.
On Sep 17, 2008, at 6:32 PM, Andrey Shulinskiy wrote:
Hello,
First, some numbers we're expecting.
- The average size of a doc: ~100K
- The number of indexes: 1
- The query response time we're looking for: < 200 - 300ms
- The number of stored docs:
1st year: 500K - 1M
2nd
I'll shamelessly take this opportunity to plug the long-neglected SOLR-657.
Lars
19 sep 2008 kl. 23.22 skrev Grant Ingersoll:
As for HDDs, people have noted some nice speedups in Lucene using
Solid-state drives, if you can afford them.
I've seen the average response time cut in 5-10 times when switching
to SSD. 64GB SSD is starting at EUR 200 so that can be a lot cheap
> > As for HDDs, people have noted some nice speedups in Lucene using
> > Solid-state drives, if you can afford them.
>
> I've seen the average response time cut in 5-10 times when switching
> to SSD. 64GB SSD is starting at EUR 200 so that can be a lot cheaper
> to do replace the disk than
How would I handle a search for 21" or 3'. The " and ' symbols appear to get
stripped away by Lucene before passing the query off to the analyzers.
Here is my analyzer in the schema.xml:
I could certainly replace X" with X inch using regex in my custom request
handler. B
> SearchComponent is the class I was missing. Looks like if I can
provide an
> entirely new implementation of that it will be a lot cleaner than the
hack I
> had been using in 1.2 over top of facets. What I'm doing is implementing
> some aggregation functions like avg() and sum() that SQL has.
Hello,
We currently have a ton of documents that we would like to index and
make search-able. I came across solr and it seems like it offers a lot
of nice features and would suite our needs.
The documents are in similar structure to java code, blocks
representing functions, variables, comment blo
32 matches
Mail list logo