Re: How to delete documents from a SOLR cloud / balance the shards in the cloud?

2010-09-10 Thread James Liu
Stephan and all,

I am evaluating this like you are. You may want to check
http://www.tomkleinpeter.com/2008/03/17/programmers-toolbox-part-3-consistent-hashing/.
I would appreciate if others can shed some light on this, too.

Bests,
James
On Fri, Sep 10, 2010 at 6:07 AM, Stephan Raemy wrote:

> Hi solr-cloud users,
>
> I'm currently setting up a solr-cloud/zookeeper instance and so far,
> everything works out fine. I downloaded the source from the cloud branch
> yesterday and build it from source.
>
> I've got 10 shards distributed across 4 servers and a zookeeper instance.
> Searching documents with the flag "distrib=true" works out and it returns
> the expected result.
>
> But here comes the tricky question. I will add new documents every day and
> therefore, I'd like to balance my shards to keep the system speedy. The
> Wiki says that one can calculate the hash of a document id and then
> determine the corresponding shard. But IMHO, this does not take into
> account
> that the cloud may become bigger or shrink over time by adding or removing
> shards. Obviously adding has a higher priority since one wants to reduce
> the shard size to improve the response time of distributed searches.
>
> When reading through the Wikis and existing documentation, it is still
> unclear to me how to do the following operations:
> - Modify/Delete a document stored in the cloud without having to store the
>  document:shard mapping information outside of the cloud. I would expect
>  something like shard attribute on each doc in the SOLR query result
>  (activated/deactivated by a flag), so that i can query the SOLR cloud for
> a
>  doc and then delete it on the specific shard.
> - Balance a cloud when adding/removing new shards or just balance them
> after
>  many deletions.
>
> Of course there are solutions to this, but at the end, I'd love to have a
> true cloud where i do not have to worry about shard performance
> optimization.
> Hints are greatly appreciated.
>
> Cheers,
> Stephan
>


Re: response status: error 400

2009-10-22 Thread James liu
Are you sure the url is correct?


-- 
regards
j.L ( I live in Shanghai, China)


Re: Illegal character in xml file

2008-09-19 Thread James liu
first, u should escape some string like (code by php)

> function escapeChars($string) {
>
$string = str_replace("&", "&", $string);

$string = str_replace("<", "<", $string);

$string = str_replace(">", ">", $string);

$string = str_replace("'", "'", $string);

$string = str_replace('"', """, $string);


return $string;

}



second, u get xml(encode by utf-8)

third, post it by utf-8(head string "Content-Type: text/xml;charset=utf-8")


if u don't know how to, maybe u can check solr client(u can find it in
solr's wiki)

Good Luck~


On Fri, Sep 19, 2008 at 4:33 PM, 李学健 <[EMAIL PROTECTED]> wrote:

> hi, all
>
> when i post xml files to solr, it's interrupted by this: Illegal character
>
> how can i deal with it ?
>
> is there any solution to ignore Illegal character in documents feeded ?
>
>
> thanks
>



-- 
regards
j.L


Re: sole 1.3: bug in phps response writer

2008-11-17 Thread James liu
i find url not same as the others
-- 
regards
j.L


Re: Newbe! Trying to run solr-1.3.0 under tomcat. Please help

2008-11-19 Thread James liu
check procedure:
1: rm -r $tomcat/webapps/*
2: rm -r $solr/data ,,,ur index data directory
3: check xml(any xml u modified)
4: start tomcat

i had same error, but i forgot how to fix...so u can use my check procedure,
i think it will help you


i use tomcat+solr in win2003, freebsd, mac osx 10.5.5, they all work well

-- 
regards
j.L


Re: posting error in solr

2008-11-19 Thread James liu
first u sure the xml is utf-8,,and field value is utf-8,,
second u should post xml by utf-8


my advice : All encoding use utf-8...

it make my solr work well,,, i use chinese

-- 
regards
j.L


Re: Query for Distributed search -

2008-11-24 Thread James liu
Up to your solr client.

On Mon, Nov 24, 2008 at 1:24 PM, souravm <[EMAIL PROTECTED]> wrote:

> Hi,
>
> Looking for some insight on distributed search.
>
> Say I have an index distributed in 3 boxes and the index contains time and
> text data (typical log file). Each box has index for different timeline -
> say Box 1 for all Jan to April, Box 2 for May to August and Box 3 for Sep to
> Dec.
>
> Now if I try to search for a text string, will the search would happen in
> parallel in all 3 boxes or sequentially?
>
> Regards,
> Sourav
>
>  CAUTION - Disclaimer *
> This e-mail contains PRIVILEGED AND CONFIDENTIAL INFORMATION intended
> solely
> for the use of the addressee(s). If you are not the intended recipient,
> please
> notify the sender by e-mail and delete the original message. Further, you
> are not
> to copy, disclose, or distribute this e-mail or its contents to any other
> person and
> any such actions are unlawful. This e-mail may contain viruses. Infosys has
> taken
> every reasonable precaution to minimize this risk, but is not liable for
> any damage
> you may sustain as a result of any virus in this e-mail. You should carry
> out your
> own virus checks before opening the e-mail or attachment. Infosys reserves
> the
> right to monitor and review the content of all messages sent to or from
> this e-mail
> address. Messages sent to or from this e-mail address may be stored on the
> Infosys e-mail system.
> ***INFOSYS End of Disclaimer INFOSYS***
>



-- 
regards
j.L


Re: Using Chinese / How to ?

2009-06-02 Thread James liu
u means how to config solr which support chinese?

Update problem?

On Tuesday, June 2, 2009, Fer-Bj  wrote:
>
> I'm sending 3 files:
> - schema.xml
> - solrconfig.xml
> - error.txt (with the error description)
>
> I can confirm by now that this error is due to invalid characters for the
> XML format (ASCII 0 or 11).
> However, this problem now is taking a different direction: how to start
> using the CJK instead of the english!
> http://www.nabble.com/file/p23825881/error.txt error.txt
> http://www.nabble.com/file/p23825881/solrconfig.xml solrconfig.xml
> http://www.nabble.com/file/p23825881/schema.xml schema.xml
>
>
> Grant Ingersoll-6 wrote:
>>
>> Can you provide details on the errors?  I don't think we have a
>> specific how to, but I wouldn't think it would be much different from
>> 1.2
>>
>> -Grant
>> On May 31, 2009, at 10:31 PM, Fer-Bj wrote:
>>
>>>
>>> Hello,
>>>
>>> is there any "how to" already created to get me up using SOLR 1.3
>>> running
>>> for a chinese based website?
>>> Currently our site is using SOLR 1.2, and we tried to move into 1.3
>>> but we
>>> couldn't complete our reindex as it seems like 1.3 is more strict
>>> when it
>>> comes to special chars.
>>>
>>> I would appreciate any help anyone may provide on this.
>>>
>>> Thanks!!
>>> --
>>> View this message in context:
>>> http://www.nabble.com/Using-Chinese---How-to---tp23810129p23810129.html
>>> Sent from the Solr - User mailing list archive at Nabble.com.
>>>
>>
>> --
>> Grant Ingersoll
>> http://www.lucidimagination.com/
>>
>> Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids)
>> using Solr/Lucene:
>> http://www.lucidimagination.com/search
>>
>>
>>
>
> --
> View this message in context: 
> http://www.nabble.com/Using-Chinese---How-to---tp23810129p23825881.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>

-- 
regards
j.L ( I live in Shanghai, China)


Re: Solr multiple keyword search as google

2009-06-02 Thread James liu
U can find answer in tutorial or example

On Tuesday, June 2, 2009, The Spider  wrote:
>
> Hi,
>    I am using solr nightly bind for my search.
> I have to search in the location field of the table which is not my default
> search field.
> I will briefly explain my requirement below:
> I want to get the same/similar result when I give location multiple
> keywords, say  "San jose ca USA"
> or "USA ca san jose" or "CA San jose USA" (like that of google search). That
> means even if I rearranged the keywords of location I want to get proper
> results. Is there any way to do that?
> Thanks in advance
> --
> View this message in context: 
> http://www.nabble.com/Solr-multiple-keyword-search-as-google-tp23826278p23826278.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>

-- 
regards
j.L ( I live in Shanghai, China)


Re: Using Chinese / How to ?

2009-06-02 Thread James liu
1: modify ur schema.xml:
like




2: add your field:


3: add your analyzer to {solr_dir}\lib\

4: rebuild newsolr and u will find it in {solr_dir}\dist

5: follow tutorial to setup solr

6: open your browser to solr admin page, find analyzer to check analyzer, it
will tell u how to analyzer world, use which analyzer


-- 
regards
j.L ( I live in Shanghai, China)


Re: indexing Chienese langage

2009-06-04 Thread James liu
first: u not have to restart solr,,,u can use new data to replace old data
and call solr to use new search..u can find something in shell script which
with solr

two: u not have to restart solr,,,just keep id is same..example: old
id:1,title:hi, new id:1,title:welcome,,just index new data,,it will delete
old data and insert new doc,,,like replace,,but it will use more time and
resouce.

u can find indexed doc number from solr admin page.


On Fri, Jun 5, 2009 at 7:42 AM, Fer-Bj  wrote:

>
> What we usually do to reindex is:
>
> 1. stop solr
> 2. rmdir -r data  (that is to remove everything in  /opt/solr/data/
> 3. mkdir data
> 4. start solr
> 5. start reindex.   with this we're sure about not having old copies or
> index..
>
> To check the index size we do:
> cd data
> du -sh
>
>
>
> Otis Gospodnetic wrote:
> >
> >
> > I can't tell what that analyzer does, but I'm guessing it uses n-grams?
> > Maybe consider trying https://issues.apache.org/jira/browse/LUCENE-1629
> > instead?
> >
> >  Otis
> > --
> > Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
> >
> >
> >
> > - Original Message 
> >> From: Fer-Bj 
> >> To: solr-user@lucene.apache.org
> >> Sent: Thursday, June 4, 2009 2:20:03 AM
> >> Subject: Re: indexing Chienese langage
> >>
> >>
> >> We are trying SOLR 1.3 with Paoding Chinese Analyzer , and after
> >> reindexing
> >> the index size went from 1.5 Gb to 2.7 Gb.
> >>
> >> Is that some expected behavior ?
> >>
> >> Is there any switch or trick to avoid having a double + index file size?
> >>
> >> Koji Sekiguchi-2 wrote:
> >> >
> >> > CharFilter can normalize (convert) traditional chinese to simplified
> >> > chinese or vice versa,
> >> > if you define mapping.txt. Here is the sample of Chinese character
> >> > normalization:
> >> >
> >> >
> >>
> https://issues.apache.org/jira/secure/attachment/12392639/character-normalization.JPG
> >> >
> >> > See SOLR-822 for the detail:
> >> >
> >> > https://issues.apache.org/jira/browse/SOLR-822
> >> >
> >> > Koji
> >> >
> >> >
> >> > revathy arun wrote:
> >> >> Hi,
> >> >>
> >> >> When I index chinese content using chinese tokenizer and analyzer in
> >> solr
> >> >> 1.3 ,some of the chinese text files are getting indexed but others
> are
> >> >> not.
> >> >>
> >> >> Since chinese has got many different language subtypes as in standard
> >> >> chinese,simplified chinese etc which of these does the chinese
> >> tokenizer
> >> >> support and is there any method to find the type of  chiense language
> >> >> from
> >> >> the file?
> >> >>
> >> >> Rgds
> >> >>
> >> >>
> >> >
> >> >
> >> >
> >>
> >> --
> >> View this message in context:
> >>
> http://www.nabble.com/indexing-Chienese-langage-tp22033302p23864358.html
> >> Sent from the Solr - User mailing list archive at Nabble.com.
> >
> >
> >
>
> --
> View this message in context:
> http://www.nabble.com/indexing-Chienese-langage-tp22033302p23879730.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>


-- 
regards
j.L ( I live in Shanghai, China)


Re: indexing Chienese langage

2009-06-04 Thread James liu
On Mon, Feb 16, 2009 at 4:30 PM, revathy arun  wrote:

> Hi,
>
> When I index chinese content using chinese tokenizer and analyzer in solr
> 1.3 ,some of the chinese text files are getting indexed but others are not.
>

are u sure ur analyzer can do it good?

if not sure, u can use analzyer link in solr admin page to check it


>
> Since chinese has got many different language subtypes as in standard
> chinese,simplified chinese etc which of these does the chinese tokenizer
> support and is there any method to find the type of  chiense language  from
> the file?
>
> Rgds
>



-- 
regards
j.L ( I live in Shanghai, China)


Re: timeouts

2009-06-04 Thread James liu
*Collins:

*i don't know what u wanna say?

-- 
regards
j.L ( I live in Shanghai, China)


does solr support summary

2009-06-10 Thread James liu
if user use keyword to search and get summary(auto generated by
keyword)...like this

doc filed: id, text

id: 001
text:

> Open source is a development method for software that harnesses the power
> of distributed peer review and transparency of process. The promise of open
> source is better quality, higher reliability, more flexibility, lower cost,
> and an end to predatory vendor lock-in.
>
if keyword is "source",,summary is:

Open source is a development...The promise of open source is better quality
if keyword is "power ",,,summary is:
Open...harnesses the power of distributed peer review and transparency of
process...

just like google search results...

and any advice will be appreciated.

-- 
regards
j.L ( I live in Shanghai, China)


IndexMerge not found

2009-07-01 Thread James liu
i try http://wiki.apache.org/solr/MergingSolrIndexes

system: win2003, jdk 1.6

Error information:

> Caused by: java.lang.ClassNotFoundException:
> org.apache.lucene.misc.IndexMergeTo
> ol
> at java.net.URLClassLoader$1.run(Unknown Source)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(Unknown Source)
> at java.lang.ClassLoader.loadClass(Unknown Source)
> at sun.misc.Launcher$AppClassLoader.loadClass(Unknown Source)
> at java.lang.ClassLoader.loadClass(Unknown Source)
> at java.lang.ClassLoader.loadClassInternal(Unknown Source)
> Could not find the main class: org/apache/lucene/misc/IndexMergeTool.
> Program w
> ill exit.
>


-- 
regards
j.L ( I live in Shanghai, China)


Re: IndexMerge not found

2009-07-01 Thread James liu
i use lucene-core-2.9-dev.jar, lucene-misc-2.9-dev.jar

On Thu, Jul 2, 2009 at 2:02 PM, James liu  wrote:

> i try http://wiki.apache.org/solr/MergingSolrIndexes
>
> system: win2003, jdk 1.6
>
> Error information:
>
>> Caused by: java.lang.ClassNotFoundException:
>> org.apache.lucene.misc.IndexMergeTo
>> ol
>> at java.net.URLClassLoader$1.run(Unknown Source)
>> at java.security.AccessController.doPrivileged(Native Method)
>> at java.net.URLClassLoader.findClass(Unknown Source)
>> at java.lang.ClassLoader.loadClass(Unknown Source)
>> at sun.misc.Launcher$AppClassLoader.loadClass(Unknown Source)
>> at java.lang.ClassLoader.loadClass(Unknown Source)
>> at java.lang.ClassLoader.loadClassInternal(Unknown Source)
>> Could not find the main class: org/apache/lucene/misc/IndexMergeTool.
>> Program w
>> ill exit.
>>
>
>
> --
> regards
> j.L ( I live in Shanghai, China)
>



-- 
regards
j.L ( I live in Shanghai, China)


Is it problem? I use solr to search and index is made by lucene. (not EmbeddedSolrServer(wiki is old))

2009-07-02 Thread James liu
I use solr to search and index is made by lucene. (not
EmbeddedSolrServer(wiki is old))

Is it problem when i use solr to search?

which the difference between Index(made by lucene and solr)?


thks

-- 
regards
j.L ( I live in Shanghai, China)


Re: Is it problem? I use solr to search and index is made by lucene. (not EmbeddedSolrServer(wiki is old))

2009-07-02 Thread James liu
solr have much fieldtype, like: integer,long, double, sint, sfloat,
tint,tfloat,,and more.

but lucene not fieldtype,,just name and value, value only string.

so i not sure is it a problem when i use solr to search( index made by
lucene).



-- 
regards
j.L ( I live in Shanghai, China)


how to stress test solr

2010-02-03 Thread James liu
before stressing test, Should i close SolrCache?

which tool u use?

How to do stress test correctly?

Any pointers?

-- 
regards
j.L ( I live in Shanghai, China)


Re: multiple slaves on the same box

2007-07-17 Thread James liu

2007/7/18, Ryan McKinley <[EMAIL PROTECTED]>:


Xuesong Luo wrote:
> Hi, there,
> We have one master server and multiple slave servers. The multiple slave
> servers can be run either on the same box or different boxes.  For
> slaves on the same box, is there any best practice that they should use
> the same index or each should have separate indexes?
>

I'm not sure about 'best' practices, but I can tell you my experience...

We have a master and single slave on the same server using the same
index.  Since it is the same index, there really is no 'distribution'
scripts, only something that periodically calls 'commit' on the slave
index.  This is working great.



I don't know why "We have a master and single slave on the same server using
the same
index."

master which do index? or do other thing? search will use slave or master,
or first master and second to slave?


My experience: every paritition have their index, and index not same.

Master Index do backup and it in other server.




I can't think of any reason to have more then one slave server on the

same machine.  What are you trying to do?

ryan





--
regards
jl


solr index problem

2007-07-17 Thread James liu

when i index 1.7m docs and 4k-5k per doc.

OutOfMemory happen when it finish index ~1.13m docs

I just restart tomcat , delete all lock and restart do index.

No error or warning infor until it finish.


anyone know why? or have the same error?

--
regards
jl


Re: solr index problem

2007-07-18 Thread James liu

I correct it,,,i index 17M docs. not 1.7M,,,so OutOfMemory happen when it
finish index ~11.3m docs

It is new index.

i think it maybe the reason:

On 7/18/07, Otis Gospodnetic <[EMAIL PROTECTED]> wrote:

> Why?  Too small of a Java heap. :)
> Increase the size of the Java heap and lower the maxBufferedDocs number
in solrconfig.xml and then try again.

If it only happens after a lot of docs, it's probably not
maxBufferedDocs, but when a big luicene merge is triggered.




2007/7/19, Mike Klaas <[EMAIL PROTECTED]>:



On 18-Jul-07, at 2:58 PM, Yonik Seeley wrote:

> On 7/18/07, Mike Klaas <[EMAIL PROTECTED]> wrote:
>>
>> Could happen when doDeleting the pending docs too.  James: try
>> sending commit every 500k docs or so.
>
> Hmmm, right... some of the memory usage will be related to the treemap
> keeping track of deleted items, and some of it will be related to the
> reader itself (the term index, and the norms).
>
> Perhaps we should have some sort of limit on the number of deletions
> we cache so people don't have to worry about that part.

I logged this in http://issues.apache.org/jira/browse/SOLR-310

Perhaps a solrconfig.xml setting, ? Default to 100k?

I have experienced this from time to time... it is one of the reasons
I fastidiously wipe the index and restart solr before reindexing.



aha,,me too。

-Mike






--
regards
jl


why store field will be analyzed?

2007-07-31 Thread James liu
fieldset "topic" indexed='false' and stored='true'

i don't know why it will be analyzed?

now i wanna it only store not analyzed,,,how can i do?


-- 
regards
jl


what's the difference between lucene2.1 and solr1.2

2007-08-07 Thread James liu
if i wanna calc it by my method, something i should notice ?

anyone did it?



-- 
regards
jl


Re: what's the difference between lucene2.1 and solr1.2

2007-08-08 Thread James liu
aha,,,little 。

i think it is same as lucene until sometime i read something from mail list.



2007/8/8, Grant Ingersoll <[EMAIL PROTECTED]>:
>
> Hi James,
>
> Could you be a little more clear about what you are asking?  What are
> you trying to calculate?
>
> FWIW, Lucene is a search library, Solr is a search server that uses
> Lucene.
>
> Cheers,
> Grant
>
> On Aug 8, 2007, at 2:57 AM, James liu wrote:
>
> > if i wanna calc it by my method, something i should notice ?
> >
> > anyone did it?
> >
> >
> >
> > --
> > regards
> > jl
>
> --
> Grant Ingersoll
> http://lucene.grantingersoll.com
>
> Lucene Helpful Hints:
> http://wiki.apache.org/lucene-java/BasicsOfPerformance
> http://wiki.apache.org/lucene-java/LuceneFAQ
>
>
>


-- 
regards
jl


Re: Multiple indexes

2007-08-29 Thread James liu
what is the limits for Lucene and Solr.

100m, 1000m, 5000m or other number docs?

2007/8/24, Walter Underwood <[EMAIL PROTECTED]>:
>
> It should work fine to index them and search them. 13 million docs is
> not even close to the limits for Lucene and Solr. Have you had problems?
>
> wunder
>
> On 8/23/07 7:30 AM, "Jae Joo" <[EMAIL PROTECTED]> wrote:
>
> > Is there any solution to handle 13 millions document shown as below?
> > Each document is not big, but the number of ones is 13 million.
> > Any way to utilize the multiple indexes?
> >
> > Thanks,
> > Jae Joo
> >
> >
> > 
> > Unlimi-Tech Software Inc
> > ON
> > 
> > Single Location
> > 
> > Canada
> > K1C 4R1
> > Ottawa
> > Ontario
> > G2_Computer Software
> > 1447a Youville Dr
> > G_Technology & Communications
> > 203439018
> > 
> > Carleton
> > 
> > 30
> > G2173_Computer Programming Services
> > G217308D_Computer software development
> > 
>
>


-- 
regards
jl


Re: Multiple indexes

2007-08-29 Thread James liu
Does it affect with doc size?

for example 2 billion docs, 10k doc2 billion docs, but doc size is 10m.


2007/8/30, Mike Klaas <[EMAIL PROTECTED]>:
>
> 2 billion docs (signed int).
>
> On 29-Aug-07, at 6:24 PM, James liu wrote:
>
> > what is the limits for Lucene and Solr.
> >
> > 100m, 1000m, 5000m or other number docs?
> >
> > 2007/8/24, Walter Underwood <[EMAIL PROTECTED]>:
> >>
> >> It should work fine to index them and search them. 13 million docs is
> >> not even close to the limits for Lucene and Solr. Have you had
> >> problems?
> >>
> >> wunder
> >>
> >> On 8/23/07 7:30 AM, "Jae Joo" <[EMAIL PROTECTED]> wrote:
> >>
> >>> Is there any solution to handle 13 millions document shown as below?
> >>> Each document is not big, but the number of ones is 13 million.
> >>> Any way to utilize the multiple indexes?
> >>>
> >>> Thanks,
> >>> Jae Joo
> >>>
> >>>
> >>> 
> >>> Unlimi-Tech Software Inc
> >>> ON
> >>> 
> >>> Single Location
> >>> 
> >>> Canada
> >>> K1C 4R1
> >>> Ottawa
> >>> Ontario
> >>> G2_Computer Software
> >>> 1447a Youville Dr
> >>> G_Technology & Communications
> >>> 203439018
> >>> 
> >>> Carleton
> >>> 
> >>> 30
> >>> G2173_Computer Programming Services
> >>> G217308D_Computer software development
> >>> 
> >>
> >>
> >
> >
> > --
> > regards
> > jl
>
>


-- 
regards
jl


Re: Multiple indexes

2007-08-30 Thread James liu
OK...I see...thk u ,mike.

2007/8/31, Mike Klaas <[EMAIL PROTECTED]>:
>
>
> On 29-Aug-07, at 10:21 PM, James liu wrote:
>
> > Does it affect with doc size?
> >
> > for example 2 billion docs, 10k doc2 billion docs, but doc size
> > is 10m.
>
> There might be other places that have 2G limit (see lucene index
> format docs), but many things are vints and can grow larger.
>
> Of course you will hit physical limits of your machine long before
> you can achieve your hypothetical situation: that's 20,000 Tb, which
> is many, many times the size of a complete internet crawl.
>
> -Mike
> >
> > 2007/8/30, Mike Klaas <[EMAIL PROTECTED]>:
> >>
> >> 2 billion docs (signed int).
> >>
> >> On 29-Aug-07, at 6:24 PM, James liu wrote:
> >>
> >>> what is the limits for Lucene and Solr.
> >>>
> >>> 100m, 1000m, 5000m or other number docs?
> >>>
> >>> 2007/8/24, Walter Underwood <[EMAIL PROTECTED]>:
> >>>>
> >>>> It should work fine to index them and search them. 13 million
> >>>> docs is
> >>>> not even close to the limits for Lucene and Solr. Have you had
> >>>> problems?
> >>>>
> >>>> wunder
> >>>>
> >>>> On 8/23/07 7:30 AM, "Jae Joo" <[EMAIL PROTECTED]> wrote:
> >>>>
> >>>>> Is there any solution to handle 13 millions document shown as
> >>>>> below?
> >>>>> Each document is not big, but the number of ones is 13 million.
> >>>>> Any way to utilize the multiple indexes?
> >>>>>
> >>>>> Thanks,
> >>>>> Jae Joo
> >>>>>
> >>>>>
> >>>>> 
> >>>>> Unlimi-Tech Software Inc
> >>>>> ON
> >>>>> 
> >>>>> Single Location
> >>>>> 
> >>>>> Canada
> >>>>> K1C 4R1
> >>>>> Ottawa
> >>>>> Ontario
> >>>>> G2_Computer Software
> >>>>> 1447a Youville Dr
> >>>>> G_Technology & Communications
> >>>>> 203439018
> >>>>> 
> >>>>> Carleton
> >>>>> 
> >>>>> 30
> >>>>> G2173_Computer Programming Services
> >>>>> G217308D_Computer software development
> >>>>> 
> >>>>
> >>>>
> >>>
> >>>
> >>> --
> >>> regards
> >>> jl
> >>
> >>
> >
> >
> > --
> > regards
> > jl
>
>


-- 
regards
jl


How to patch

2007-09-09 Thread James liu
i wanna try patch:
https://issues.apache.org/jira/browse/SOLR-139?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel

and i download solr1.2 release

patch < SOLR-269*.pach(when in
'/tmp/apache-solr-1.2.0/src/test/org/apache/solr/update'
)

it show me

|Index: src/test/org/apache/solr/common/SolrDocumentTest.java
|===
|--- src/test/org/apache/solr/common/SolrDocumentTest.java  (revision
555897)
|+++ src/test/org/apache/solr/common/SolrDocumentTest.java  (working
copy)
--
File to patch:


i input the absolute path which file should be ...

but it not ok...

does my solr's version not right?

-- 
regards
jl


Re: How to patch

2007-09-09 Thread James liu
thks ,ryan.

2007/9/10, Ryan McKinley <[EMAIL PROTECTED]>:
>
> James liu wrote:
> > i wanna try patch:
> >
> https://issues.apache.org/jira/browse/SOLR-139?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel
> >
> > and i download solr1.2 release
> >
>
> The most recent patch is Eriks-ModifiableDocument.patch
>
> You will need to apply it to solr trunk.  (not 1.2 release)
>
> ryan
>



-- 
regards
jl


real time search

2007-09-23 Thread James liu
i wanna do it.

Maybe someone did it, if so, give me some tips.

thks

-- 
regards
jl


Re: real time search

2007-09-24 Thread James liu
it seems use somthing like ajax...

if so, it not what i wanna


2007/9/25, Matthew Runo <[EMAIL PROTECTED]>:
>
> I assume you mean something like this:
>
> http://addictedtonew.com/archives/145/wordpress-live-search-plugin/
>
> Take a look at how the search box works - is that what you mean?
>
>
> ++
> | Matthew Runo
> | Zappos Development
> | [EMAIL PROTECTED]
> | 702-943-7833
> ++
>
>
> On Sep 24, 2007, at 8:13 AM, Grant Ingersoll wrote:
>
> > Hi James,
> >
> > Can you provide more information about what you are trying to do?
> > By real time search, do you mean you want indexed documents to be
> > available immediately?  Or is a minute or two acceptable?  Do all
> > users need to see them immediately, or just the current user?
> >
> > We can better help you if you give us more details on what you are
> > trying to accomplish.
> >
> > Thanks,
> > Grant
> >
> > On Sep 23, 2007, at 10:38 AM, James liu wrote:
> >
> >> i wanna do it.
> >>
> >> Maybe someone did it, if so, give me some tips.
> >>
> >> thks
> >>
> >> --
> >> regards
> >> jl
> >
> > --
> > Grant Ingersoll
> > http://lucene.grantingersoll.com
> >
> > Lucene Helpful Hints:
> > http://wiki.apache.org/lucene-java/BasicsOfPerformance
> > http://wiki.apache.org/lucene-java/LuceneFAQ
> >
> >
>
>


-- 
regards
jl


Re: real time search

2007-09-24 Thread James liu
if new data come in and drive index it, load new search it.

if more docs, optimize time will cost much, so can't do search like real
time.

so i think new solr instance only for newest information. the docs will
be ~10K.
if it arrive 10k, it should be closed and rebuild new instance.
(if we have more memory and cpu,,,just open more instance ,not one by one.)

final, we merge result and show it to user.


that all i think, not test it.


2007/9/24, Grant Ingersoll <[EMAIL PROTECTED]>:
>
> Hi James,
>
> Can you provide more information about what you are trying to do?  By
> real time search, do you mean you want indexed documents to be
> available immediately?  Or is a minute or two acceptable?  Do all
> users need to see them immediately, or just the current user?
>
> We can better help you if you give us more details on what you are
> trying to accomplish.
>
> Thanks,
> Grant
>
> On Sep 23, 2007, at 10:38 AM, James liu wrote:
>
> > i wanna do it.
> >
> > Maybe someone did it, if so, give me some tips.
> >
> > thks
> >
> > --
> > regards
> > jl
>
> --
> Grant Ingersoll
> http://lucene.grantingersoll.com
>
> Lucene Helpful Hints:
> http://wiki.apache.org/lucene-java/BasicsOfPerformance
> http://wiki.apache.org/lucene-java/LuceneFAQ
>
>
>


-- 
regards
jl


can solr do it?

2007-09-24 Thread James liu
if use multi solr with one index, it will cache individually.

so i think can it share their cache.(they have same config)

-- 
regards
jl


anyone can send me jetty-plus

2007-09-26 Thread James liu
i can't download it from http://jetty.mortbay.org/jetty5/plus/index.html

-- 
regards
jl


Re: Availability Issues

2007-10-08 Thread James liu
* *

*i think text not need "stored='true'" unless u will show it.(it will help u
decrease index size and not affect search )*

*index and search use same box? if it is true, u should moniter search
response time when indexing.(include CPU, RAM change)*

*i have similar problem and i increase JVM size to fix it.(u can try it and
show me your response)*


2007/10/9, David Whalen <[EMAIL PROTECTED]>:
>
> Thanks for letting me know that.  Okay, here they are:
>
>
>  BEGIN SCHEMA.XML===
>
>
> 
> 
>
> 
>
> 
> 
>
> 
>
>
>
> omitNorms="true"/>
>
>
> omitNorms="true"/>
>
>
>
>
>
>
>
>
>
>
>
>
> sortMissingLast="true" omitNorms="true"/>
> sortMissingLast="true" omitNorms="true"/>
> sortMissingLast="true" omitNorms="true"/>
> sortMissingLast="true" omitNorms="true"/>
>
>
>
> omitNorms="true"/>
>
>
>
>
>
>
> positionIncrementGap="100">
>  
>
>  
>
>
>
> positionIncrementGap="100">
>  
>
>
>
> generateWordParts="1" generateNumberParts="1" catenateWords="1"
> catenateNumbers="1" catenateAll="0"/>
>
>
>
>
>  
>  
>
> ignoreCase="true" expand="true"/>
>
> generateWordParts="0" generateNumberParts="1" catenateWords="0"
> catenateNumbers="0" catenateAll="0"/>
>
>
>
>
>  
>
>
>
>
> positionIncrementGap="100" >
>  
>
> ignoreCase="true" expand="false"/>
>
> generateWordParts="0" generateNumberParts="0" catenateWords="1"
> catenateNumbers="1" catenateAll="0"/>
>
>
>
>  
>
>
> 
>
>
> 
>   
>
>   
>   
>   
>   
>   
>multiValued="true"/>
>/>
>   
>   
>/>
>   
>   
>
>   
>   
>   
>   
>   
>   
>   
>   
>   
> 
>
> 
> id
>
> 
> text
>
> 
> 
>
> 
>   
>
> 
> 
>
> 
>
>  END SCHEMA.XML===
>
>
>
>
>  BEGIN CONFIG.XML===
>
>
>
> 
> 
>
> 
>
> 
> 
> false
> 
>   
>true
>10
>1000
>2147483647
>1
>1000
>1
> 
>
> 
>
>true
>10
>1000
>2147483647
>1
>
>
>true
> 
>
> 
> 
>
>
>
>
>
>
>
>
>
> 
>
>
> 
>
>1024
>
>
>
>  class="solr.LRUCache"
>  size="512"
>  initialSize="512"
>  autowarmCount="256"/>
>
>   
>  class="solr.LRUCache"
>  size="512"
>  initialSize="512"
>  autowarmCount="256"/>
>
> 
>  class="solr.LRUCache"
>  size="512"
>  initialSize="512"
>  autowarmCount="0"/>
>
>
>false
>
>
>
>
>   
>
>   
>10
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>false
>
> 
>
>
> 
> 
>
> 
>   explicit
>
>   
>   50
>   10
>   *
>   2.1
>-->
> 
> 
>
> 
> 
>
> explicit
> 0.01
> 
>text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0 manu^1.1 cat^1.4
> 
> 
>text^0.2 features^1.1 name^1.5 manu^1.4 manu_exact^1.9
> 
> 
>ord(poplarity)^0.5 recip(rord(price),1,1000,1000)^0.3
> 
> 
>id,name,price,score
> 
> 
>2<-1 5<-2 6<90%
> 
> 100
>
> 
>
> 
> 
>
> explicit
> text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0
> 2<-1 5<-2 6<90%
> 
> incubationdate_dt:[* TO NOW/DAY-1MONTH]^2.2
>
>
>
>  inStock:true
>
>
>
>  cat
>  manu_exact
>  price:[* TO 500]
>  price:[500 TO *]
>
> 
>
> 
>
> 
>inStock:true
> 
> 
>text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0 manu^1.1 cat^1.4
> 
> 
>2<-1 5<-2 6<90%
> 
> 
>
> 
>
> 
> 
>5
> 
>
> 
> 
>solr
>solrconfig.xml schema.xml admin-extra.html
> 
>
>
> qt=dismax&q=solr&start=3&fq=id:[* TO *]&fq=cat:[* TO
> *]
>
>
> 
>
> 
>
>
>
>  END CONFIG.XML===
>
>
>
>
>
> > -Original Message-
> > From: Chris Hostetter [mailto:[EMAIL PROTECTED]
> > Sent: Monday, October 08, 2007 4:56 PM
> > To: solr-user
> > Subject: RE: Availability Issues
> >
> > : I've attached our schema/config files.  They are pretty much
> > : out-of-the-box values, except for our index.
> >
> > FYI: the mailing list strips most attachemnts ... the best
> > thing to do is just inline them in your mail.
> >
> > Quick question: do you have autoCommit turned on in your
> > solrconfig.xml?
> >
> > Second question: do you have autowarming on your caches?
> >
> >
> >
> > -Hoss
> >
> >
> >
>



-- 
regards
jl


index become bigger and the only way seems to add hardware, another way?

2007-10-09 Thread James liu
i just wanna know is it exist which can decrease index size,,not by
increasing hardware or optimizing lucene params.

-- 
regards
jl


Re: index become bigger and the only way seems to add hardware, another way?

2007-10-09 Thread James liu
i did what u said, but it not fix my problem.

so i wanna try to find another way.




2007/10/10, Otis Gospodnetic <[EMAIL PROTECTED]>:
>
> Here are some ways:
>
>
>
> Index less data, store fewer fields and less data, compress fields,
> change Lucene's the term index interval (default 128; increasing it
> will make your index a little bit smaller, but will slow down
> queries)... But in general, the more your index the more hw you'll
> need.  I saw 1TB disks for ~$300 USD the other day.  You are in China
> and this stuff is even cheaper there.
>
> Otis
>
> --
>
> Lucene - Solr - Nutch - Consulting -- http://sematext.com/
>
>
>
>
> - Original Message 
> From: James liu <[EMAIL PROTECTED]>
> To: solr-user@lucene.apache.org
> Sent: Tuesday, October 9, 2007 11:15:56 PM
> Subject: index become bigger and the only way seems to add hardware,
> another way?
>
>
> i just wanna know is it exist which can decrease index size,,not by
> increasing hardware or optimizing lucene params.
>
> --
> regards
> jl
>
>
>
>


-- 
regards
jl


random lost segments* files,,

2007-10-14 Thread James liu
i have 40 instances,,,one instance lost segments* file(happen after commit
and optimize)

anyone have similar problem?

can i fix this problem?

can i recovery this instance data?



-- 
regards
jl


Re: random lost segments* files,,

2007-10-14 Thread James liu
i find it happen when it do commit.

i use solr 1.2 release.

i use crontab to do index work.


2007/10/15, James liu <[EMAIL PROTECTED]>:
>
> i have 40 instances,,,one instance lost segments* file(happen after commit
> and optimize)
>
> anyone have similar problem?
>
> can i fix this problem?
>
> can i recovery this instance data?
>
>
>
> --
> regards
> jl
>



-- 
regards
jl


Re: SOLR 1.3 Release?

2007-10-25 Thread James liu
where i can read 1.3 new features?

2007/10/26, Venkatraman S <[EMAIL PROTECTED]>:
>
> On 10/26/07, Mike Klaas <[EMAIL PROTECTED]> wrote:
> >
> > If we did a 1.2.x, it shoud (imo) contain no new features, only
> > important bugfixes.
>
>
> I have been having a look at the trunk for quite sometime now, and must
> say
> that its changing pretty fast. Having an  interim release now will require
> more work with the comitters having to put an extra effort to spruce up
> things. I would personally suggest 1-jan-2008 as the next Solr release
> which
> would help in sprucing up the code and also releasing the bunch of new
> features that have been lying in the trunk to be used in
> *the-next-stable-release*.
>
> -Venkat
>
> --
> Blog @ http://blizzardzblogs.blogspot.com
>



-- 
regards
jl


Re: get all query not working

2007-10-31 Thread James liu
if u rebuild solr , safe method is rm -r *tomcat*/webapps/*.

2007/11/1, Chris Hostetter <[EMAIL PROTECTED]>:
>
>
> : Is there an easy to find out which version of solr is running.  I
> installed
> : solr 1.2 and set up an instance using Tomcat. It was successful before.
>
> FYI: starting a while back, the "Info" page (registry.jsp) of the admin
> interface gives you specifics on the Solr and Lucene versions in use.
>
>
> -Hoss
>
>


-- 
regards
jl


how to use PHP AND PHPS?

2007-11-05 Thread James liu
i find they all return string

http://localhost:8080/solr/select/?q=solr&version=2.2&start=0&rows=10&indent=on&wt=php
';
  var_dump(file_get_contents($url);
?>


-- 
regards
jl


Re: how to use PHP AND PHPS?

2007-11-05 Thread James liu
i know it...but u try it,,u will find simlar question.

On 11/5/07, Robert Young <[EMAIL PROTECTED]> wrote:
>
> I would imagine you have to unserialize
>
> On 11/5/07, James liu <[EMAIL PROTECTED]> wrote:
> > i find they all return string
> >
> >  >   $url = '
> >
> http://localhost:8080/solr/select/?q=solr&version=2.2&start=0&rows=10&indent=on&wt=php
> > ';
> >   var_dump(file_get_contents($url);
> > ?>
> >
> >
> > --
> > regards
> > jl
> >
>



-- 
regards
jl


Re: how to use PHP AND PHPS?

2007-11-05 Thread James liu
first: i m sure i enable php and phps in my solrconfig.xml
two: i can't get answer.

*phps:
*http://localhost:8080/solr1/select/?q=2&version=2.2&start=0&rows=10&indent=on&wt=phps
';
$a = file_get_contents($url);
echo 'before unserialize';
var_dump($a);
$a = unserialize($a);
echo 'after unserialize...';
var_dump($a);
?>*

*


On 11/6/07, Stu Hood <[EMAIL PROTECTED]> wrote:
>
> Did you enable the PHP serialized response writer in your solrconfig.xml?
> It is not enabled by default.
>
> Thanks,
> Stu
>
>
> -Original Message-
> From: James liu <[EMAIL PROTECTED]>
> Sent: Monday, November 5, 2007 9:03am
> To: solr-user@lucene.apache.org
> Subject: Re: how to use PHP AND PHPS?
>
> i know it...but u try it,,u will find simlar question.
>
> On 11/5/07, Robert Young <[EMAIL PROTECTED]> wrote:
> >
> > I would imagine you have to unserialize
> >
> > On 11/5/07, James liu <[EMAIL PROTECTED]> wrote:
> > > i find they all return string
> > >
> > >  > >   $url = '
> > >
> >
> http://localhost:8080/solr/select/?q=solr&version=2.2&start=0&rows=10&indent=on&wt=php
> > > ';
> > >   var_dump(file_get_contents($url);
> > > ?>
> > >
> > >
> > > --
> > > regards
> > > jl
> > >
> >
>
>
>
> --
> regards
> jl
>
>
>


-- 
regards
jl


Re: how to use PHP AND PHPS?

2007-11-06 Thread James liu
first var_dump result(part not all):

string(50506)
> "a:2:{s:14:"responseHeader";a:3:{s:6:"status";i:0;s:5:"QTime";i:2906;s:6:"params";a:6:{s:6:"indent";s:2:"on";s:5:"start";s:1:"0";s:1:"q";s:1:"2";s:2:"wt";s:4:"phps";s:4:"rows";s:2:"10";s:7:"version";s:3:"
> 2.2";}}
>

two var_dump result:
bool(false)


On Nov 6, 2007 10:36 PM, Dave Lewis <[EMAIL PROTECTED]> wrote:
> What are the results of the two var_dumps?
>
> dave
>
>
> On Nov 5, 2007, at 10:06 PM, James liu wrote:
>
> > first: i m sure i enable php and phps in my solrconfig.xml
> > two: i can't get answer.
> >
> > *phps:
> > * > $url = '
> > http://localhost:8080/solr1/select/?
> > q=2&version=2.2&start=0&rows=10&indent=on&wt=phps
> > ';
> > $a = file_get_contents($url);
> > echo 'before unserialize';
> > var_dump($a);
> > $a = unserialize($a);
> > echo 'after unserialize...';
> > var_dump($a);
> > ?>*
> >
> > *
> >
> >
> > On 11/6/07, Stu Hood <[EMAIL PROTECTED]> wrote:
> >>
> >> Did you enable the PHP serialized response writer in your
> >> solrconfig.xml?
> >> It is not enabled by default.
> >>
> >> Thanks,
> >> Stu
> >>
> >>
> >> -Original Message-
> >> From: James liu <[EMAIL PROTECTED]>
> >> Sent: Monday, November 5, 2007 9:03am
> >> To: solr-user@lucene.apache.org
> >> Subject: Re: how to use PHP AND PHPS?
> >>
> >> i know it...but u try it,,u will find simlar question.
> >>
> >> On 11/5/07, Robert Young <[EMAIL PROTECTED]> wrote:
> >>>
> >>> I would imagine you have to unserialize
> >>>
> >>> On 11/5/07, James liu <[EMAIL PROTECTED]> wrote:
> >>>> i find they all return string
> >>>>
> >>>>  >>>>   $url = '
> >>>>
> >>>
> >> http://localhost:8080/solr/select/?
> >> q=solr&version=2.2&start=0&rows=10&indent=on&wt=php
> >>>> ';
> >>>>   var_dump(file_get_contents($url);
> >>>> ?>
> >>>>
> >>>>
> >>>> --
> >>>> regards
> >>>> jl
> >>>>
> >>>
> >>
> >>
> >>
> >> --
> >> regards
> >> jl
> >>
> >>
> >>
> >
> >
> > --
> > regards
> > jl
>
>



-- 
regards
jl


Re: how to use PHP AND PHPS?

2007-11-06 Thread James liu
afternoon,,i will update svn...and try the newest...



On Nov 7, 2007 11:23 AM, Dave Lewis <[EMAIL PROTECTED]> wrote:

>
> On Nov 6, 2007, at 8:10 PM, James liu wrote:
>
> > first var_dump result(part not all):
> >
> > string(50506)
> >> "a:2:{s:14:"responseHeader";a:3:{s:6:"status";i:0;s:5:"QTime";i:
> >> 2906;s:6:"params";a:6:{s:6:"indent";s:2:"on";s:5:"start";s:1:"0";s:
> >> 1:"q";s:1:"2";s:2:"wt";s:4:"phps";s:4:"rows";s:2:"10";s:
> >> 7:"version";s:3:"
> >> 2.2";}}
> >>
> >
>
> This is exactly correct.
>
> > two var_dump result:
> > bool(false)
> >
>
> So, unserializing is failing.  Are you running from the trunk or from
> a nightly?  There was a bug a couple of weeks ago that sent back
> faulty serialized data.  It's fixed now.  It's possible this is your
> issue.
>
> dave
>
>
> >
> > On Nov 6, 2007 10:36 PM, Dave Lewis <[EMAIL PROTECTED]> wrote:
> >> What are the results of the two var_dumps?
> >>
> >> dave
> >>
> >>
> >> On Nov 5, 2007, at 10:06 PM, James liu wrote:
> >>
> >>> first: i m sure i enable php and phps in my solrconfig.xml
> >>> two: i can't get answer.
> >>>
> >>> *phps:
> >>> * >>> $url = '
> >>> http://localhost:8080/solr1/select/?
> >>> q=2&version=2.2&start=0&rows=10&indent=on&wt=phps
> >>> ';
> >>> $a = file_get_contents($url);
> >>> echo 'before unserialize';
> >>> var_dump($a);
> >>> $a = unserialize($a);
> >>> echo 'after unserialize...';
> >>> var_dump($a);
> >>> ?>*
> >>>
> >>> *
> >>>
> >>>
> >>> On 11/6/07, Stu Hood <[EMAIL PROTECTED]> wrote:
> >>>>
> >>>> Did you enable the PHP serialized response writer in your
> >>>> solrconfig.xml?
> >>>> It is not enabled by default.
> >>>>
> >>>> Thanks,
> >>>> Stu
> >>>>
> >>>>
> >>>> -Original Message-
> >>>> From: James liu <[EMAIL PROTECTED]>
> >>>> Sent: Monday, November 5, 2007 9:03am
> >>>> To: solr-user@lucene.apache.org
> >>>> Subject: Re: how to use PHP AND PHPS?
> >>>>
> >>>> i know it...but u try it,,u will find simlar question.
> >>>>
> >>>> On 11/5/07, Robert Young <[EMAIL PROTECTED]> wrote:
> >>>>>
> >>>>> I would imagine you have to unserialize
> >>>>>
> >>>>> On 11/5/07, James liu <[EMAIL PROTECTED]> wrote:
> >>>>>> i find they all return string
> >>>>>>
> >>>>>>  >>>>>>   $url = '
> >>>>>>
> >>>>>
> >>>> http://localhost:8080/solr/select/?
> >>>> q=solr&version=2.2&start=0&rows=10&indent=on&wt=php
> >>>>>> ';
> >>>>>>   var_dump(file_get_contents($url);
> >>>>>> ?>
> >>>>>>
> >>>>>>
> >>>>>> --
> >>>>>> regards
> >>>>>> jl
> >>>>>>
> >>>>>
> >>>>
> >>>>
> >>>>
> >>>> --
> >>>> regards
> >>>> jl
> >>>>
> >>>>
> >>>>
> >>>
> >>>
> >>> --
> >>> regards
> >>> jl
> >>
> >>
> >
> >
> >
> > --
> > regards
> > jl
>
>


-- 
regards
jl


Re: how to use PHP AND PHPS?

2007-11-06 Thread James liu
same answer.

On Nov 7, 2007 11:41 AM, James liu <[EMAIL PROTECTED]> wrote:

> afternoon,,i will update svn...and try the newest...
>
>
>
>
> On Nov 7, 2007 11:23 AM, Dave Lewis <[EMAIL PROTECTED]> wrote:
>
> >
> > On Nov 6, 2007, at 8:10 PM, James liu wrote:
> >
> > > first var_dump result(part not all):
> > >
> > > string(50506)
> > >> "a:2:{s:14:"responseHeader";a:3:{s:6:"status";i:0;s:5:"QTime";i:
> > >> 2906;s:6:"params";a:6:{s:6:"indent";s:2:"on";s:5:"start";s:1:"0";s:
> > >> 1:"q";s:1:"2";s:2:"wt";s:4:"phps";s:4:"rows";s:2:"10";s:
> > >> 7:"version";s:3:"
> > >> 2.2";}}
> > >>
> > >
> >
> > This is exactly correct.
> >
> > > two var_dump result:
> > > bool(false)
> > >
> >
> > So, unserializing is failing.  Are you running from the trunk or from
> > a nightly?  There was a bug a couple of weeks ago that sent back
> > faulty serialized data.  It's fixed now.  It's possible this is your
> > issue.
> >
> > dave
> >
> >
> > >
> > > On Nov 6, 2007 10:36 PM, Dave Lewis <[EMAIL PROTECTED]> wrote:
> > >> What are the results of the two var_dumps?
> > >>
> > >> dave
> > >>
> > >>
> > >> On Nov 5, 2007, at 10:06 PM, James liu wrote:
> > >>
> > >>> first: i m sure i enable php and phps in my solrconfig.xml
> > >>> two: i can't get answer.
> > >>>
> > >>> *phps:
> > >>> * > >>> $url = '
> > >>> http://localhost:8080/solr1/select/?
> > >>> q=2&version= 2.2&start=0&rows=10&indent=on&wt=phps
> > >>> ';
> > >>> $a = file_get_contents($url);
> > >>> echo 'before unserialize';
> > >>> var_dump($a);
> > >>> $a = unserialize($a);
> > >>> echo 'after unserialize...';
> > >>> var_dump($a);
> > >>> ?>*
> > >>>
> > >>> *
> > >>>
> > >>>
> > >>> On 11/6/07, Stu Hood <[EMAIL PROTECTED]> wrote:
> > >>>>
> > >>>> Did you enable the PHP serialized response writer in your
> > >>>> solrconfig.xml?
> > >>>> It is not enabled by default.
> > >>>>
> > >>>> Thanks,
> > >>>> Stu
> > >>>>
> > >>>>
> > >>>> -Original Message-
> > >>>> From: James liu <[EMAIL PROTECTED]>
> > >>>> Sent: Monday, November 5, 2007 9:03am
> > >>>> To: solr-user@lucene.apache.org
> > >>>> Subject: Re: how to use PHP AND PHPS?
> > >>>>
> > >>>> i know it...but u try it,,u will find simlar question.
> > >>>>
> > >>>> On 11/5/07, Robert Young < [EMAIL PROTECTED]> wrote:
> > >>>>>
> > >>>>> I would imagine you have to unserialize
> > >>>>>
> > >>>>> On 11/5/07, James liu < [EMAIL PROTECTED]> wrote:
> > >>>>>> i find they all return string
> > >>>>>>
> > >>>>>>  > >>>>>>   $url = '
> > >>>>>>
> > >>>>>
> > >>>> http://localhost:8080/solr/select/?
> > >>>> q=solr&version=2.2&start=0&rows=10&indent=on&wt=php
> > >>>>>> ';
> > >>>>>>   var_dump(file_get_contents($url);
> > >>>>>> ?>
> > >>>>>>
> > >>>>>>
> > >>>>>> --
> > >>>>>> regards
> > >>>>>> jl
> > >>>>>>
> > >>>>>
> > >>>>
> > >>>>
> > >>>>
> > >>>> --
> > >>>> regards
> > >>>> jl
> > >>>>
> > >>>>
> > >>>>
> > >>>
> > >>>
> > >>> --
> > >>> regards
> > >>> jl
> > >>
> > >>
> > >
> > >
> > >
> > > --
> > > regards
> > > jl
> >
> >
>
>
> --
> regards
> jl




-- 
regards
jl


Re: how to use PHP AND PHPS?

2007-11-06 Thread James liu
i just decrease answer information...and u will see my result(full, not
part)

*before unserialize*
> string(433)
> "a:2:{s:14:"responseHeader";a:3:{s:6:"status";i:0;s:5:"QTime";i:0;s:6:"params";a:7:{s:2:"fl";s:5:"Title";s:6:"indent";s:2:"on";s:5:"start";s:1:"0";s:1:"q";s:1:"2";s:2:"wt";s:4:"phps";s:4:"rows";a:2:{i:0;s:1:"2";i:1;s:2:"10";}s:7:"version";s:3:"
> 2.2";}}s:8:"response";a:3:{s:8:"numFound";i:28;s:5:"start";i:0;s:4:"docs";a:2:{i:0;a:1:{s:5:"Title";d:诺基亚N-Gage基本数据;}i:1;a:1:{s:5:"Title";d:索尼爱立信P908基本数据;"
>
> *after unserialize...*
> bool(false)
>

and i write serialize test code..

 $ar = array (
> array('id' => 123, 'Title'=> "中文测试"),
> array('id' => 123, 'Title'=> "中国上海"),
> );
>
> echo serialize($ar);
>
> ?>
>


and result is :

>
> a:2:{i:0;a:2:{s:2:"id";i:123;s:5:"Title";s:12:"中文测试";}i:1;a:2:{s:2:"id";i:123;s:5:"Title";s:12:"中国上海";}}
>



*php* result is:

string(369) "array( 'responseHeader'=>array( 'status'=>0, 'QTime'=>0,
'params'=>array( 'fl'=>'Title', 'indent'=>'on', 'start'=>'0', 'q'=>'2',
'wt'=>'php', 'rows'=>array('2', '10'), 'version'=>'2.2')),
'response'=>array('numFound'=>28,'start'=>0,'docs'=>array( array(
'Title'=>诺基亚N-Gage基本数据), array( 'Title'=>索尼爱立信P908基本数据)) ))"

it is string, so i can't read it correctly by php.


-- 
regards
jl


Re: how to use PHP AND PHPS?

2007-11-07 Thread James liu
hmm

i find error,,,that is my error not about php and phps ..

i use old config to testso config have a problem..

that is Title i use double as its type...it should use text.


On Nov 8, 2007 10:29 AM, James liu <[EMAIL PROTECTED]> wrote:

>  php now is ok..
>
> but phps failed
>
> mycode:
>
> >  > $url = 
> > 'http://localhost:8080/solr1/select/?q=2&version=2.2&rows=2&fl=Title&start=0&rows=10&indent=on&wt=phps
> > ';
> > $a = file_get_contents($url);
> > //eval('$solrResults = ' .$serializedSolrResults . ';');
> > echo 'before unserialize';
> > var_dump($a);
> > echo '';
> > $a = unserialize($a);
> > echo 'after unserialize...';
> > var_dump($a);
> > ?>
> >
>
> and result:
>
> > *before unserialize*
> > string(434)
> > "a:2:{s:14:"responseHeader";a:3:{s:6:"status";i:0;s:5:"QTime";i:32;s:6:"params";a:7:{s:2:"fl";s:5:"Title";s:6:"indent";s:2:"on";s:5:"start";s:1:"0";s:1:"q";s:1:"2";s:2:"wt";s:4:"phps";s:4:"rows";a:2:{i:0;s:1:"2";i:1;s:2:"10";}s:7:"version";s:3:"
> > 2.2";}}s:8:"response";a:3:{s:8:"numFound";i:28;s:5:"start";i:0;s:4:"docs";a:2:{i:0;a:1:{s:5:"Title";d:诺基亚N-Gage基本数据;}i:1;a:1:{s:5:"Title";d:索尼爱立信P908基本数据;"
> >
> > *after unserialize...*
> > bool(false)
> >
>
>
> On Nov 7, 2007 9:30 PM, Dave Lewis <[EMAIL PROTECTED]> wrote:
>
> >
> > On Nov 7, 2007, at 2:04 AM, James liu wrote:
> >
> > > i just decrease answer information...and u will see my result(full,
> > > not
> > > part)
> > >
> > > *before unserialize*
> > >> string(433)
> > >> "a:2:{s:14:"responseHeader";a:3:{s:6:"status";i:0;s:5:"QTime";i:
> > >> 0;s:6:"params";a:7:{s:2:"fl";s:5:"Title";s:6:"indent";s:2:"on";s:
> > >> 5:"start";s:1:"0";s:1:"q";s:1:"2";s:2:"wt";s:4:"phps";s:4:"rows";a:
> > >> 2:{i:0;s:1:"2";i:1;s:2:"10";}s:7:"version";s:3:"
> > >> 2.2";}}s:8:"response";a:3:{s:8:"numFound";i:28;s:5:"start";i:0;s:
> > >> 4:"docs";a:2:{i:0;a:1:{s:5:"Title";d:诺基亚N-Gage基本数据;}i:1;a:1:
> > >> {s:5:"Title";d:索尼爱立信P908基本数据;"
> > >>
> > >> *after unserialize...*
> > >> bool(false)
> > >>
> > >
> > > and i write serialize test code..
> > >
> > >  > >> $ar = array (
> > >> array('id' => 123, 'Title'=> "中文测试"),
> > >> array('id' => 123, 'Title'=> "中国上海"),
> > >> );
> > >>
> > >> echo serialize($ar);
> > >>
> > >> ?>
> > >>
> > >
> > >
> > > and result is :
> > >
> > >>
> > >> a:2:{i:0;a:2:{s:2:"id";i:123;s:5:"Title";s:12:"中文测试";}i:1;a:2:
> > >> {s:2:"id";i:123;s:5:"Title";s:12:"中国上海";}}
> > >>
> > >
> > >
> > >
> > > *php* result is:
> > >
> > > string(369) "array( 'responseHeader'=>array( 'status'=>0, 'QTime'=>0,
> > > 'params'=>array( 'fl'=>'Title', 'indent'=>'on', 'start'=>'0',
> > > 'q'=>'2',
> > > 'wt'=>'php', 'rows'=>array('2', '10'), 'version'=>'2.2')),
> > > 'response'=>array('numFound'=>28,'start'=>0,'docs'=>array( array(
> > > 'Title'=>诺基亚N-Gage基本数据), array( 'Title'=>索尼爱立信P908基本数
> > > 据)) ))"
> > >
> > > it is string, so i can't read it correctly by php.
> > >
> > >
> >
> >
> > This part (after string(369)) is exactly what it you should be seeing
> > if you use the php handler, and it's what you get after you
> > unserialize when using phps.
> >
> > You can access your search results as:
> >
> > $solrResults['response']['docs'];
> >
> > In your example above, that would be:
> >
> > array( array('Title'=>诺基亚N-Gage基本数据), array( 'Title'=>索尼爱立信
> > P908基本数据))
> >
> > When using the php handler, you must do something like this:
> >
> > eval('$solrResults = ' .$serializedSolrResults . ';');
> >
> > Then, as above, you can access $solrResults['response']['docs'].
> >
> > To sum up, if you use phps, you must unserialize the results.  If you
> > use php, you must eval the results (including some sugar to get a
> > variable set to that value).
> >
> >
> > dave
> >
> >
> >
>
>
> --
> regards
> jl




-- 
regards
jl


Re: how to use PHP AND PHPS?

2007-11-07 Thread James liu
 php now is ok..

but phps failed

mycode:

>  $url = '
> http://localhost:8080/solr1/select/?q=2&version=2.2&rows=2&fl=Title&start=0&rows=10&indent=on&wt=phps
> ';
> $a = file_get_contents($url);
> //eval('$solrResults = ' .$serializedSolrResults . ';');
> echo 'before unserialize';
> var_dump($a);
> echo '';
> $a = unserialize($a);
> echo 'after unserialize...';
> var_dump($a);
> ?>
>

and result:

> *before unserialize*
> string(434)
> "a:2:{s:14:"responseHeader";a:3:{s:6:"status";i:0;s:5:"QTime";i:32;s:6:"params";a:7:{s:2:"fl";s:5:"Title";s:6:"indent";s:2:"on";s:5:"start";s:1:"0";s:1:"q";s:1:"2";s:2:"wt";s:4:"phps";s:4:"rows";a:2:{i:0;s:1:"2";i:1;s:2:"10";}s:7:"version";s:3:"
> 2.2";}}s:8:"response";a:3:{s:8:"numFound";i:28;s:5:"start";i:0;s:4:"docs";a:2:{i:0;a:1:{s:5:"Title";d:诺基亚N-Gage基本数据;}i:1;a:1:{s:5:"Title";d:索尼爱立信P908基本数据;"
>
> *after unserialize...*
> bool(false)
>


On Nov 7, 2007 9:30 PM, Dave Lewis <[EMAIL PROTECTED]> wrote:

>
> On Nov 7, 2007, at 2:04 AM, James liu wrote:
>
> > i just decrease answer information...and u will see my result(full,
> > not
> > part)
> >
> > *before unserialize*
> >> string(433)
> >> "a:2:{s:14:"responseHeader";a:3:{s:6:"status";i:0;s:5:"QTime";i:
> >> 0;s:6:"params";a:7:{s:2:"fl";s:5:"Title";s:6:"indent";s:2:"on";s:
> >> 5:"start";s:1:"0";s:1:"q";s:1:"2";s:2:"wt";s:4:"phps";s:4:"rows";a:
> >> 2:{i:0;s:1:"2";i:1;s:2:"10";}s:7:"version";s:3:"
> >> 2.2";}}s:8:"response";a:3:{s:8:"numFound";i:28;s:5:"start";i:0;s:
> >> 4:"docs";a:2:{i:0;a:1:{s:5:"Title";d:诺基亚N-Gage基本数据;}i:1;a:1:
> >> {s:5:"Title";d:索尼爱立信P908基本数据;"
> >>
> >> *after unserialize...*
> >> bool(false)
> >>
> >
> > and i write serialize test code..
> >
> >  >> $ar = array (
> >> array('id' => 123, 'Title'=> "中文测试"),
> >> array('id' => 123, 'Title'=> "中国上海"),
> >> );
> >>
> >> echo serialize($ar);
> >>
> >> ?>
> >>
> >
> >
> > and result is :
> >
> >>
> >> a:2:{i:0;a:2:{s:2:"id";i:123;s:5:"Title";s:12:"中文测试";}i:1;a:2:
> >> {s:2:"id";i:123;s:5:"Title";s:12:"中国上海";}}
> >>
> >
> >
> >
> > *php* result is:
> >
> > string(369) "array( 'responseHeader'=>array( 'status'=>0, 'QTime'=>0,
> > 'params'=>array( 'fl'=>'Title', 'indent'=>'on', 'start'=>'0',
> > 'q'=>'2',
> > 'wt'=>'php', 'rows'=>array('2', '10'), 'version'=>'2.2')),
> > 'response'=>array('numFound'=>28,'start'=>0,'docs'=>array( array(
> > 'Title'=>诺基亚N-Gage基本数据), array( 'Title'=>索尼爱立信P908基本数
> > 据)) ))"
> >
> > it is string, so i can't read it correctly by php.
> >
> >
>
>
> This part (after string(369)) is exactly what it you should be seeing
> if you use the php handler, and it's what you get after you
> unserialize when using phps.
>
> You can access your search results as:
>
> $solrResults['response']['docs'];
>
> In your example above, that would be:
>
> array( array('Title'=>诺基亚N-Gage基本数据), array( 'Title'=>索尼爱立信
> P908基本数据))
>
> When using the php handler, you must do something like this:
>
> eval('$solrResults = ' .$serializedSolrResults . ';');
>
> Then, as above, you can access $solrResults['response']['docs'].
>
> To sum up, if you use phps, you must unserialize the results.  If you
> use php, you must eval the results (including some sugar to get a
> variable set to that value).
>
>
> dave
>
>
>


-- 
regards
jl


Re: how to use PHP AND PHPS?

2007-11-07 Thread James liu
Thks everybody who give me help.

especial Dave, thk u.

On Nov 8, 2007 11:21 AM, James liu <[EMAIL PROTECTED]> wrote:

> hmm
>
> i find error,,,that is my error not about php and phps ..
>
> i use old config to testso config have a problem..
>
> that is Title i use double as its type...it should use text.
>
>
>
> On Nov 8, 2007 10:29 AM, James liu <[EMAIL PROTECTED]> wrote:
>
> >  php now is ok..
> >
> > but phps failed
> >
> > mycode:
> >
> > >  > > $url = 
> > > 'http://localhost:8080/solr1/select/?q=2&version=2.2&rows=2&fl=Title&start=0&rows=10&indent=on&wt=phps
> > > ';
> > > $a = file_get_contents($url);
> > > //eval('$solrResults = ' .$serializedSolrResults . ';');
> > > echo 'before unserialize';
> > > var_dump($a);
> > > echo '';
> > > $a = unserialize($a);
> > > echo 'after unserialize...';
> > > var_dump($a);
> > > ?>
> > >
> >
> > and result:
> >
> > > *before unserialize *
> > > string(434)
> > > "a:2:{s:14:"responseHeader";a:3:{s:6:"status";i:0;s:5:"QTime";i:32;s:6:"params";a:7:{s:2:"fl";s:5:"Title";s:6:"indent";s:2:"on";s:5:"start";s:1:"0";s:1:"q";s:1:"2";s:2:"wt";s:4:"phps";s:4:"rows";a:2:{i:0;s:1:"2";i:1;s:2:"10";}s:7:"version";s:3:"
> > > 2.2";}}s:8:"response";a:3:{s:8:"numFound";i:28;s:5:"start";i:0;s:4:"docs";a:2:{i:0;a:1:{s:5:"Title";d:诺基亚N-Gage基本数据;}i:1;a:1:{s:5:"Title";d:索尼爱立信P908基本数据;"
> > >
> > > *after unserialize...*
> > > bool(false)
> > >
> >
> >
> > On Nov 7, 2007 9:30 PM, Dave Lewis < [EMAIL PROTECTED]> wrote:
> >
> > >
> > > On Nov 7, 2007, at 2:04 AM, James liu wrote:
> > >
> > > > i just decrease answer information...and u will see my result(full,
> > > > not
> > > > part)
> > > >
> > > > *before unserialize*
> > > >> string(433)
> > > >> "a:2:{s:14:"responseHeader";a:3:{s:6:"status";i:0;s:5:"QTime";i:
> > > >> 0;s:6:"params";a:7:{s:2:"fl";s:5:"Title";s:6:"indent";s:2:"on";s:
> > > >> 5:"start";s:1:"0";s:1:"q";s:1:"2";s:2:"wt";s:4:"phps";s:4:"rows";a:
> > > >> 2:{i:0;s:1:"2";i:1;s:2:"10";}s:7:"version";s:3:"
> > > >> 2.2";}}s:8:"response";a:3:{s:8:"numFound";i:28;s:5:"start";i:0;s:
> > > >> 4:"docs";a:2:{i:0;a:1:{s:5:"Title";d:诺基亚N-Gage基本数据;}i:1;a:1:
> > > >> {s:5:"Title";d:索尼爱立信P908基本数据;"
> > > >>
> > > >> *after unserialize...*
> > > >> bool(false)
> > > >>
> > > >
> > > > and i write serialize test code..
> > > >
> > > >  > > >> $ar = array (
> > > >> array('id' => 123, 'Title'=> "中文测试"),
> > > >> array('id' => 123, 'Title'=> "中国上海"),
> > > >> );
> > > >>
> > > >> echo serialize($ar);
> > > >>
> > > >> ?>
> > > >>
> > > >
> > > >
> > > > and result is :
> > > >
> > > >>
> > > >> a:2:{i:0;a:2:{s:2:"id";i:123;s:5:"Title";s:12:"中文测试";}i:1;a:2:
> > > >> {s:2:"id";i:123;s:5:"Title";s:12:"中国上海";}}
> > > >>
> > > >
> > > >
> > > >
> > > > *php* result is:
> > > >
> > > > string(369) "array( 'responseHeader'=>array( 'status'=>0,
> > > 'QTime'=>0,
> > > > 'params'=>array( 'fl'=>'Title', 'indent'=>'on', 'start'=>'0',
> > > > 'q'=>'2',
> > > > 'wt'=>'php', 'rows'=>array('2', '10'), 'version'=>'2.2')),
> > > > 'response'=>array('numFound'=>28,'start'=>0,'docs'=>array( array(
> > > > 'Title'=>诺基亚N-Gage基本数据), array( 'Title'=>索尼爱立信P908基本数
> > > > 据)) ))"
> > > >
> > > > it is string, so i can't read it correctly by php.
> > > >
> > > >
> > >
> > >
> > > This part (after string(369)) is exactly what it you should be seeing
> > > if you use the php handler, and it's what you get after you
> > > unserialize when using phps.
> > >
> > > You can access your search results as:
> > >
> > > $solrResults['response']['docs'];
> > >
> > > In your example above, that would be:
> > >
> > > array( array('Title'=>诺基亚N-Gage基本数据), array( 'Title'=>索尼爱立信
> > > P908基本数据))
> > >
> > > When using the php handler, you must do something like this:
> > >
> > > eval('$solrResults = ' .$serializedSolrResults . ';');
> > >
> > > Then, as above, you can access $solrResults['response']['docs'].
> > >
> > > To sum up, if you use phps, you must unserialize the results.  If you
> > > use php, you must eval the results (including some sugar to get a
> > > variable set to that value).
> > >
> > >
> > > dave
> > >
> > >
> > >
> >
> >
> > --
> > regards
> > jl
>
>
>
>
> --
> regards
> jl




-- 
regards
jl


Re: Does SOLR supports multiple instances within the same webapplication?

2007-11-12 Thread James liu
if I understand correct,,u just do it like that:(i use php)

$data1 = getDataFromInstance1($url);
$data2 = getDataFromInstance2($url);

it just have multi solr Instance. and getData from the distance.


On Nov 12, 2007 11:15 PM, Dilip.TS <[EMAIL PROTECTED]> wrote:

> Hello,
>
>  Does SOLR supports multiple instances within the same web application? If
> so how is this achieved?
>
>  Thanks in advance.
>
> Regards,
> Dilip TS
>
>


-- 
regards
jl


Re: 2Gb process on 32 bits

2007-11-12 Thread James liu
if u use tomcat,,,it default port: 8080 and other default port.

so u just use other tomcat which use 8181 and other port...(i remember u
should modify three port(one tomcat) )

I used to have four tomcat in One SERVER.

On Nov 9, 2007 7:39 AM, Isart Montane <[EMAIL PROTECTED]> wrote:

> Hi all,
>
> i'm experiencing some trouble when i'm trying to lauch solr with more
> than 1.6GB. My server is a FC5 with 8GB RAM but when I start solr like
> this
>
> java -Xmx2000m -jar start.jar
>
> i get the following errors:
>
> Error occurred during initialization of VM
> Could not reserve enough space for object heap
> Could not create the Java virtual machine.
>
> I've tried to start a virtual machine like this
>
> java -Xmx2000m -version
>
> but i get the same errors.
>
> I've read there's a kernel limitation for a 32 bits architecture of 2Gb
> per process, and i just wanna know if anybody knows an alternative to
> get a new 64bits server.
>
> Thanks
> Isart
>



-- 
regards
jl


Re: CJK Analyzers for Solr

2007-11-26 Thread James liu
I don't think NGram is good method for Chinese.

CJKAnalyzer of Lucene is 2-Gram.

Eswar K:
  if it is chinese analyzer,,i recommend hylanda(www.hylanda.com),,,it is
the best chinese analyzer and it not free.
  if u wanna free chinese analyzer, maybe u can try je-analyzer. it have
some problem when using it.



On Nov 27, 2007 5:56 AM, Otis Gospodnetic <[EMAIL PROTECTED]>
wrote:

> Eswar,
>
> We've uses the NGram stuff that exists in Lucene's contrib/analyzers
> instead of CJK.  Doesn't that allow you to do everything that the Chinese
> and CJK analyzers do?  It's been a few months since I've looked at Chinese
> and CJK Analzyers, so I could be off.
>
> Otis
>
> --
> Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
>
> - Original Message 
> From: Eswar K <[EMAIL PROTECTED]>
> To: solr-user@lucene.apache.org
> Sent: Monday, November 26, 2007 8:30:52 AM
> Subject: CJK Analyzers for Solr
>
> Hi,
>
> Does Solr come with Language analyzers for CJK? If not, can you please
> direct me to some good CJK analyzers?
>
> Regards,
> Eswar
>
>
>
>


-- 
regards
jl


Re: CJK Analyzers for Solr

2007-11-26 Thread James liu
if ur analyzer is standard, u can try use tokenize.(u can find the answer
from analyzer source code and schema.xml)


On Nov 27, 2007 9:39 AM, zx zhang <[EMAIL PROTECTED]> wrote:

> lance,
>
> The following is a instance schema fieldtype using solr1.2 and CJK
> package.
> And it works. As you said, CJK does parse cjk string in a bi-gram way,
> just
> like turning 'C1C2C3C4' into 'C1C2 C2C3 C3C4'.
>
> More to the point, it is worthwhile to mention that the index expand
> beyond
> tolerance to use cjk package, and it will take a long time to index
> document. For most enterprise applications, I think, it need a more
> effective string parser.
>
>
> 
>  
> 
>
>
>
> On 11/27/07, Norskog, Lance <[EMAIL PROTECTED]> wrote:
> >
> > I notice this is in the future tense. Is the CJKTokenizer available yet?
> > From what I can see, the CJK code should be a Filter instead anyway.
> > Also, the ChineseFilter and CJKTokenizer do two different things.
> >
> > CJKTokenizer turns C1C2C3C4 into 'C1C2 C2C3 C3C4'. ChineseFilter (from
> > 2001) turns C1C2 into 'C1 C2'. I hope someone who speaks Mandarin or
> > Cantonese understands what this should do.
> >
> > Lance
> >
> > -Original Message-
> > From: Eswar K [mailto:[EMAIL PROTECTED]
> > Sent: Monday, November 26, 2007 10:28 AM
> > To: solr-user@lucene.apache.org
> > Subject: Re: CJK Analyzers for Solr
> >
> > Hoss,
> >
> > Thanks a lot. Will look into it.
> >
> > Regards,
> > Eswar
> >
> > On Nov 26, 2007 11:55 PM, Chris Hostetter <[EMAIL PROTECTED]>
> > wrote:
> >
> > >
> > > : Does Solr come with Language analyzers for CJK? If not, can you
> > > please
> > > : direct me to some good CJK analyzers?
> > >
> > > Lucene has a CJKTokenizer and CJKAnalyzer in the contrib/analyzers
> > jar.
> > > they can be used in Solr.  both have been included in Solr for a while
> >
> > > now, so you can specify CJKAnalyzer in your schema with Solr 1.2, but
> > > starting with Solr 1.3 a Factory for the Tokenizer will also be
> > > included so it can be used in a more complex analysis chain defined in
> > the schema.
> > >
> > >
> > >
> > > -Hoss
> > >
> > >
> >
>



-- 
regards
jl


Re: CJK Analyzers for Solr

2007-11-26 Thread James liu
i not use HYLANDA analyzer.

i use je-analyzer and indexing at least 18m docs.

i m sorry i only use chinese analyzer.


On Nov 27, 2007 10:01 AM, Eswar K <[EMAIL PROTECTED]> wrote:

> What is the performance of these CJK analyzers (one in lucene and hylanda
> )?
> We would potentially be indexing millions of documents.
>
> James,
>
> We would have a look at hylanda too. What abt japanese and korean
> analyzers,
> any recommendations?
>
> - Eswar
>
> On Nov 27, 2007 7:21 AM, James liu <[EMAIL PROTECTED]> wrote:
>
> > I don't think NGram is good method for Chinese.
> >
> > CJKAnalyzer of Lucene is 2-Gram.
> >
> > Eswar K:
> >  if it is chinese analyzer,,i recommend hylanda(www.hylanda.com),,,it is
> > the best chinese analyzer and it not free.
> >  if u wanna free chinese analyzer, maybe u can try je-analyzer. it have
> > some problem when using it.
> >
> >
> >
> > On Nov 27, 2007 5:56 AM, Otis Gospodnetic <[EMAIL PROTECTED]>
> > wrote:
> >
> > > Eswar,
> > >
> > > We've uses the NGram stuff that exists in Lucene's contrib/analyzers
> > > instead of CJK.  Doesn't that allow you to do everything that the
> > Chinese
> > > and CJK analyzers do?  It's been a few months since I've looked at
> > Chinese
> > > and CJK Analzyers, so I could be off.
> > >
> > > Otis
> > >
> > > --
> > > Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
> > >
> > > - Original Message 
> > > From: Eswar K <[EMAIL PROTECTED]>
> > > To: solr-user@lucene.apache.org
> > > Sent: Monday, November 26, 2007 8:30:52 AM
> > > Subject: CJK Analyzers for Solr
> > >
> > > Hi,
> > >
> > > Does Solr come with Language analyzers for CJK? If not, can you please
> > > direct me to some good CJK analyzers?
> > >
> > > Regards,
> > > Eswar
> > >
> > >
> > >
> > >
> >
> >
> > --
> > regards
> > jl
> >
>



-- 
regards
jl


Re: CJK Analyzers for Solr

2007-12-03 Thread James liu
it seems good.

On Dec 3, 2007 1:01 AM, Ken Krugler <[EMAIL PROTECTED]> wrote:

> >Wunder - are you aware of any free dictionaries
> >for either C or J or K?  When I dealt with this
> >in the past, I looked for something free, but
> >found only commercial dictionaries.
>
> I would use data files from:
>
> http://ftp.monash.edu.au/pub/nihongo/00INDEX.html
>
> -- Ken
>
>
> >Sematext -- http://sematext.com/ -- Lucene -
> >Solr - Nutch - Original Message  From:
> >Walter Underwood <[EMAIL PROTECTED]> To:
> >solr-user@lucene.apache.org Sent: Wednesday,
> >November 28, 2007 5:43:32 PM Subject: Re: CJK
> >Analyzers for Solr With Ultraseek, we switched
> >to a dictionary-based segmenter for Chinese
> >because the N-gram highlighting wasn't
> >acceptable to our Chinese customers. I guess it
> >is something to check for each application.
> >wunder On 11/27/07 10:46 PM, "Otis Gospodnetic"
> ><[EMAIL PROTECTED]> wrote: > For what
> >it's worth I worked on indexing and searching a
> >*massive* pile of > data, a good portion of
> >which was in CJ and some K.  The n-gram approach
> >was > used for all 3 languages and the quality
> >of search results, including > highlighting was
> >evaluated and okay-ed by native speakers of
> >these languages. > > Otis > -- > Sematext --
> >http://sematext.com/ -- Lucene - Solr -
> >Nutch > > - Original Message  > From:
> >Walter Underwood <[EMAIL PROTECTED]> > To:
> >solr-user@lucene.apache.org > Sent: Tuesday,
> >November 27, 2007 2:41:38 PM > Subject: Re: CJK
> >Analyzers for Solr > > Dictionaries are
> >surprisingly expensive to build and maintain
> >and > bi-gram is surprisingly effective for
> >Chinese. See this paper: > >
> >http://citeseer.ist.psu.edu/kwok97comparing.html > >
> >I expect that n-gram indexing would be less
> >effective for Japanese > because it is an
> >inflected language. Korean is even harder. It
> >might > work to break Korean into the phonetic
> >subparts and use n-gram on > those. > > You
> >should not do term highlighting with any of the
> >n-gram methods. > The relevance can be very
> >good, but the highlighting just looks dumb. > >
> >wunder > > On 11/27/07 8:54 AM, "Eswar K"
> ><[EMAIL PROTECTED]> wrote: > >> Is there any
> >specific reason why the CJK analyzers in Solr
> >were >  chosen to be >> n-gram based instead of
> >it being a morphological analyzer which is >
> >kind of >> implemented in Google as it
> >considered to be more effective than the >
> >n-gram >> ones? >> >> Regards, >>
> >Eswar >> >> >> >> On Nov 27, 2007 7:57 AM, Eswar
> >K <[EMAIL PROTECTED]> wrote: >> >>> thanks
> >james... >>> >>> How much time does it take to
> >index 18m docs? >>> >>> - Eswar >>> >>> >>> On
> >Nov 27, 2007 7:43 AM, James liu
> ><[EMAIL PROTECTED] > wrote: >>> >>>> i not
> >use HYLANDA analyzer. >>>> >>>> i use
> >je-analyzer and indexing at least 18m
> >docs. >>>> >>>> i m sorry i only use chinese
> >analyzer. >>>> >>>> >>>> On Nov 27, 2007 10:01
> >AM, Eswar K <[EMAIL PROTECTED]>
> >wrote: >>>> >>>>> What is the performance of
> >these CJK analyzers (one in lucene and >>>>
> >hylanda >>>>> )? >>>>> We would potentially be
> >indexing millions of documents. >>>>> >>>>>
> >James, >>>>> >>>>> We would have a look at
> >hylanda too. What abt japanese and korean >>>>>
> >analyzers, >>>>> any
> >recommendations? >>>>> >>>>> - Eswar >>>>> >>>>>
> >On Nov 27, 2007 7:21 AM, James liu
> ><[EMAIL PROTECTED]> >  wrote: >>>>> >>>>>>
> >I don't think NGram is good method for
> >Chinese. >>>>>> >>>>>> CJKAnalyzer of Lucene is
> >2-Gram. >>>>>> >>>>>> Eswar K: >>>>>>  if it is
> >chinese analyzer,,i recommend >
> >hylandaÅiwww.hylanda.comÅj,,,it >>>> is >>>>>>
> >the 

i can't start http://localhost:8983/solr/admin/

2006-08-26 Thread James liu

i use winxp java1.5.

i started "[EMAIL PROTECTED]:~/tmp/solr/solr-nightly/example$ java -jar
start.jar", and everything is ok.

but i wanna open http://localhost:8983/solr/admin/,,,firefox display nothing
and gcc shows long error and warning.

my java setting is:
JAVA_HOME:c:\program files\java\jre1.5.0_08
classpath: %JAVA_HOME%\lib\tools.jar;%JAVA_HOME%\lib\dt.jar


and i try

[EMAIL PROTECTED]:~/tmp/solr/solr-nightly/example/exampledocs$ sh post.sh 
solr.xml

http://localhost:8983/solr/select/?stylesheet=&q=solr&version=2.1&start=0&rows=10&indent=on

that is ok.

anyone can tell me?


Re: i can't start http://localhost:8983/solr/admin/

2006-08-26 Thread James liu

sorry.spell error. i just mean i find error information in cygwin.

like that:

"Generated servlet error:
2006-8-27 org.apache.jasper.compiler.Compiler generateClass

Generated servlet error:
??: Env: Compile:
javaFileName=/C:/DOCUME~1/JL/LOCALS~1/Temp/Jetty__8983__solr//org/apache/jsp/admin\index_jsp.java





2006/8/27, WHIRLYCOTT <[EMAIL PROTECTED]>:


What is the error that you are getting?  And why do you think that
it's coming from gcc?
phil.

On Aug 26, 2006, at 10:26 PM, James liu wrote:

> but i wanna open http://localhost:8983/solr/admin/,,,firefox
> display nothing
> and gcc shows long error and warning.


--
Whirlycott
Philip Jacob
[EMAIL PROTECTED]
http://www.whirlycott.com/phil/





Re: i can't start http://localhost:8983/solr/admin/

2006-08-26 Thread James liu

... thk u.

2006/8/27, WHIRLYCOTT <[EMAIL PROTECTED]>:


I see.  You need a JDK, not a JRE.

phil.

On Aug 26, 2006, at 10:40 PM, James liu wrote:

> sorry.spell error. i just mean i find error information in cygwin.
>
> like that:
>
> "Generated servlet error:
> 2006-8-27 org.apache.jasper.compiler.Compiler generateClass
>
> Generated servlet error:
> ??: Env: Compile:
> javaFileName=/C:/DOCUME~1/JL/LOCALS~1/Temp/Jetty__8983__solr//org/
> apache/jsp/admin\index_jsp.java
>
>
>
>
>
> 2006/8/27, WHIRLYCOTT <[EMAIL PROTECTED]>:
>>
>> What is the error that you are getting?  And why do you think that
>> it's coming from gcc?
>> phil.
>>
>> On Aug 26, 2006, at 10:26 PM, James liu wrote:
>>
>> > but i wanna open http://localhost:8983/solr/admin/,,,firefox
>> > display nothing
>> > and gcc shows long error and warning.
>>
>>
>> --
>> Whirlycott
>> Philip Jacob
>> [EMAIL PROTECTED]
>> http://www.whirlycott.com/phil/
>>
>>
>>


--
Whirlycott
Philip Jacob
[EMAIL PROTECTED]
http://www.whirlycott.com/phil/





about analyzer and index

2006-08-27 Thread James liu

lucene have ChineseAnalyzer and CJKAnalyzer,,,so i can search chinese
keyword with it.

solr have it? if not, how can i add it?


if i use php+mysql build  data.xml,,,use post.sh data.xml? it is the only
way to index?


i remember i must use same analyzer to index and search when i use lucene2.0
,,,

what is solr analyzer? and how support user defined?(if it not support
chinese)


Re: Simple Faceted Searching out of the box

2006-09-09 Thread James liu

Good. Thk u,Hoss.

2006/9/9, Tim Archambault <[EMAIL PROTECTED]>:


Hoss,

What is "faceted browsing"? Maybe an example of a site interface that is
using it would be good. Dumb question, I know.


On 9/8/06, Chris Hostetter <[EMAIL PROTECTED]> wrote:
>
>
> Hey everybody, I just wanted to officially announce that as of the
> solr-2006-09-08.zip nightly build, Solr supports some simple Faceted
> Searching options right out of the box.
>
> Both the StandardRequestHandler and DisMaxRequestHandler now support
some
> query params for specifying simple queries to use as facet constraints,
or
> fields in your index you wish to use as facets - generating a constraint
> count for each term in the field.  All of these params can be configured
> as "defaults" when registering the RequestHandler in your solrconfig.xml
>
> Information on what the new facet parameters are, how to use them, and
> what types of resultsthey generate can be found in the wiki...
>
> http://wiki.apache.org/solr/SimpleFacetParameters
> http://wiki.apache.org/solr/StandardRequestHandler
> http://wiki.apache.org/solr/DisMaxRequestHandler
>
> ...as allways: feedback, comments, suggestions and general discussion is
> strongly encouraged :)
>
>
> -Hoss
>
>




Re: Got it working! And some questions

2006-09-09 Thread James liu

- Is the solr php in the wiki working out of the box for anyone?
show your php.ini. did you performance your php?




2006/9/10, Brian Lucas <[EMAIL PROTECTED]>:


Hi Michael,

I apologize for the lack of testing on the SolPHP.  I had to "strip" it
down
significantly to turn it into a general class that would be usable and the
version up there has not been extensively tested yet (I'm almost ready to
get back to that and "revise" it), plus much of my coding is done in Rails
at the moment.  However...

If you have a new version, could you send it over my way or just upload it
to the wiki?  I'd like to take a look at the changes and throw your
revised
version up there or integrate both versions into a cleaner revision of the
version already there.

With respect to batch queries, it's already designed to do that (that's
why
you see "array($array)" in the example, because it accepts an array of
updates) but I'd definitely like to see how you revised it.

Thanks,
Brian


-Original Message-
From: Michael Imbeault [mailto:[EMAIL PROTECTED]
Sent: Saturday, September 09, 2006 12:30 PM
To: solr-user@lucene.apache.org
Subject: Got it working! And some questions

First of all, in reference to
http://www.mail-archive.com/solr-user@lucene.apache.org/msg00808.html ,
I got it working! The problem(s) was coming from solPHP; the
implementation in the wiki isn't really working, to be honest, at least
for me. I had to modify it significantly at multiple places to get it
working. Tomcat 5.5, WAMP and Windows XP.

The main problem was that addIndex was sending 1 doc at a time to solr;
it would cause a problem after a few thousand docs because i was running
out of resources. I modified solr_update.php to handle batch queries,
and i'm now sending batches of 1000 docs at a time. Great indexing speed.

Had a slight problem with the curl function of solr_update.php; the
custom HTTP header wasn't recognized; I now use curl_setopt($ch,
CURLOPT_POST, 1); curl_setopt($ch, CURLOPT_POSTFIELDS, $post_string); -
much simpler, and now everything works!

Up so far I indexed 15.000.000 documents (my whole collection,
basically) and the performance i'm getting is INCREDIBLE (sub 100ms
query time without warmup and no optimization at all on a 7 gigs index -
and with the cache, it gets stupid fast)! Seriously, Solr amaze me every
time I use it. I increased HashDocSet Maxsize to 75000, will continue to
optimize this value - it helped a great deal. I will try disMaxHandler
soon too; right now the standard one is great. And I will index with a
better stopword file; the default one could really use improvements.

Some questions (couldn't find the answer in the docs):

- Is the solr php in the wiki working out of the box for anyone? Else we
could modify the wiki...

- What is the loadFactor variable of HashDocSet? Should I optimize it too?

- What's the units on the size value of the caches? Megs, number of
queries, kilobytes? Not described anywhere.

- Any way to programatically change the OR/AND preference of the query
parser? I set it to AND by default for user queries, but i'd like to set
it to OR for some server-side queries I must do (find related articles,
order by score).

- Whats the difference between the 2 commits type? Blocking and
non-blocking. Didn't see any differences at all, tried both.

- Every time I do an  command, I get the following in my
catalina logs - should I do anything about it?

9-Sep-2006 2:24:40 PM org.apache.solr.core.SolrException log
SEVERE: Exception during commit/optimize:java.io.EOFException: no more
data available - expected end tag  to close start tag
 from line 1, parser stopped on START_TAG seen ...
@1:10

- Any benefits of setting the allowed memory for Tomcat higher? Right
now im allocating 384 megs.

Can't wait to try the new Faceted Queries... seriously, solr is really,
really awesome up so far. Thanks for all your work, and sorry for all
the questions!

--
Michael Imbeault
CHUL Research Center (CHUQ)
2705 boul. Laurier
Ste-Foy, QC, Canada, G1V 4G2
Tel: (418) 654-2705, Fax: (418) 654-2212




i wanna setup solr project,,but something i don't know

2006-09-16 Thread James liu

if i use Jetty , how i make my
project? for example, make project directory and sub directory which include
filesi don't find from wiki.

where i place

  -

  solrconfig.xml - SolrConfigXml
  -

  schema.xml - SchemaXml 




i don't know how to use tomcat setup my project. who can tell me ?


i wanna use php+mysql,,,and i find solrphp.


does solr know classpath

2006-09-16 Thread James liu

i set classpath where i put lucene-analyzers-2.0.0.jar...i can use it.

but solr not find it..

where i should put it in?


solrphp+ jetty,,,anyone can success?

2006-09-16 Thread James liu

i use xml which i write by myself to sendUpdate,,,but error show like:
"string(1536) "ERROR:unknown field 'pname'
org.xmlpull.v1.XmlPullParserException: expected START_TAG or END_TAG not
END_DOCUMENT (position: END_DOCUMENT seen ...\r\n... @20:7) at
org.xmlpull.mxp1.MXParser.nextTag(MXParser.java:1083) at
org.apache.solr.core.SolrCore.update(SolrCore.java:681) at
org.apache.solr.servlet.SolrUpdateServlet.doPost(SolrUpdateServlet.java:52)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:616) at
javax.servlet.http.HttpServlet.service(HttpServlet.java:689) at
org.mortbay.jetty.servlet.ServletHolder.handle(ServletHolder.java:428) at
org.mortbay.jetty.servlet.WebApplicationHandler.dispatch(
WebApplicationHandler.java:473) at
org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:568) at
org.mortbay.http.HttpContext.handle(HttpContext.java:1530) at
org.mortbay.jetty.servlet.WebApplicationContext.handle(
WebApplicationContext.java:633) at org.mortbay.http.HttpContext.handle(
HttpContext.java:1482) at org.mortbay.http.HttpServer.service(
HttpServer.java:909) at org.mortbay.http.HttpConnection.service(
HttpConnection.java:820) at org.mortbay.http.HttpConnection.handleNext(
HttpConnection.java:986) at org.mortbay.http.HttpConnection.handle(
HttpConnection.java:837) at org.mortbay.http.SocketListener.handleConnection
(SocketListener.java:245) at org.mortbay.util.ThreadedServer.handle(
ThreadedServer.java:357) at org.mortbay.util.ThreadPool$PoolThread.run(
ThreadPool.java:534) "


my xml:

$dom_string = '
 
   a
   a
   a

 
 
   b
   b
   b

 
 
   c
   c
   c

 
';


Re: does solr know classpath

2006-09-16 Thread James liu

i use jetty,,,and put it into solr-nightly\lib with
lucene-snowball-nightly.jar and lucene-core-nightly.jar...


2006/9/17, Simon Willnauer <[EMAIL PROTECTED]>:


/solrwebapp/WEB-INF/lib

to point out one solution

best regards simon

On 9/16/06, James liu <[EMAIL PROTECTED]> wrote:
> i set classpath where i put lucene-analyzers-2.0.0.jar...i can use it.
>
> but solr not find it..
>
> where i should put it in?
>
>



Re: i wanna setup solr project,,but something i don't know

2006-09-16 Thread James liu

u r smart boy.  i get the answer. thk u reply

2006/9/17, Chris Hostetter <[EMAIL PROTECTED]>:



: project? for example, make project directory and sub directory which
include
: filesi don't find from wiki.
:
: where i place
:solrconfig.xml - SolrConfigXml<
http://wiki.apache.org/solr/SolrConfigXml>
:schema.xml - SchemaXml 

based on your followup posts, i think you've figured out the answers to
this question, but in case i'm wrong, plese try the tutorial and see if
that helps clear some things up for you...

http://incubator.apache.org/solr/tutorial.html


-Hoss




how to check index's data

2006-09-16 Thread James liu

lucene's index can be checked by lukeall.jar;

how to check solr's index?


Re: how to check index's data

2006-09-16 Thread James liu

i think you misunderstand.

i wanna check indexed data,,,not how to index.

because my testxml seems to be indexed ,,,but i can't find it by
http://localhost:8080/solr/admin

2006/9/17, Mike Klaas <[EMAIL PROTECTED]>:


On 9/16/06, James liu <[EMAIL PROTECTED]> wrote:
> lucene's index can be checked by lukeall.jar;
>
> how to check solr's index?

Solr's index is a lucene index.  It's location is specified by the
 directive in solrconfig.xml.

-Mike



Re: how to check index's data

2006-09-16 Thread James liu

1: my lucene index will have _1.cfs,deletable,segmentsand solr just only
segments,,,am i wrong? luke can't open solr's index,and can open lucene's
index.

2:   i m sure commit after adddocs.



2006/9/17, Chris Hostetter <[EMAIL PROTECTED]>:


:
: i think you misunderstand.
:
: i wanna check indexed data,,,not how to index.

I think mike's point is that a solr index *is* a lucene index, so you can
point luke at it just like you're use to if you've used lucene.

: because my testxml seems to be indexed ,,,but i can't find it by
: http://localhost:8080/solr/admin

off the top of my head: did you remember to do a commit after you added
the docs?

: > > lucene's index can be checked by lukeall.jar;
: > >
: > > how to check solr's index?
: >
: > Solr's index is a lucene index.  It's location is specified by the
: >  directive in solrconfig.xml.



-Hoss




Re: how to check index's data

2006-09-16 Thread James liu

i just modify addIndex

if i index data,i can see "updated"...



function addIndex($resultarray) {
$dom_string = '
 
 111
   jl
   jldepart
   jlcompany
 
 
 112
   ben
   bendepart
   bencompany
 
 
 113
   gf
   gfdepart
   gfcompany
 
';
   if ($dom_string) {
   $ok = $this->sendUpdate($dom_string);
   echo 'updated';
   } else {
   echo "Error with xml document";
   print_r($resultarray);
   }
   return $ok;
   }


Re: how to check index's data

2006-09-17 Thread James liu

I find solrphp is problem.

i use post.sh,,it can index data and i can use luke


tomcat install

2006-09-17 Thread James liu

tomcat 5.5 win2003,,i can find http://localhost:8080,,it is ok.

i copy solr-nightly.war to "c:\tomcat5.5\webapps\solr.war"

i start tomcat,,it auto generate solr directory in
"c:\tomcat5.5\webapps\solr"

http://localhost:8080/solr/admin/,,,it show me error information:

HTTP Status 500 -

type Exception report

message

description The server encountered an internal error () that prevented it
from fulfilling this request.

exception
org.apache.jasper.JasperException: Exception in JSP: /admin/_info.jsp:8

5:
6: <%@ page import="java.util.Date" %>
7:
8: 
9: 
10: <[EMAIL PROTECTED] file="header.jsp" %>
11:


and i don't know where i should put example/solr


if i wanna add lucene-analyzers-2.0.0.jar,,,where i put it?


thks.


Re: tomcat install

2006-09-17 Thread James liu

thk Nick.

i do you tell me and i can see admin page.

but when i click search ,,,error information:

java.lang.NullPointerException
at org.apache.solr.search.SolrQueryParser.(SolrQueryParser.java:37)
at org.apache.solr.search.QueryParsing.parseQuery(QueryParsing.java:47)
at org.apache.solr.request.StandardRequestHandler.handleRequest(
StandardRequestHandler.java:94)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:586)
at org.apache.solr.servlet.SolrServlet.doGet(SolrServlet.java:91)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:689)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:802)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(
ApplicationFilterChain.java:252)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(
ApplicationFilterChain.java:173)
at org.apache.catalina.core.StandardWrapperValve.invoke(
StandardWrapperValve.java:213)
at org.apache.catalina.core.StandardContextValve.invoke(
StandardContextValve.java:178)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java
:126)
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java
:105)
at org.apache.catalina.core.StandardEngineValve.invoke(
StandardEngineValve.java:107)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java
:148)
at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java
:869)
at
org.apache.coyote.http11.Http11BaseProtocol$Http11ConnectionHandler.processConnection
(Http11BaseProtocol.java:664)
at org.apache.tomcat.util.net.PoolTcpEndpoint.processSocket(
PoolTcpEndpoint.java:527)
at org.apache.tomcat.util.net.LeaderFollowerWorkerThread.runIt(
LeaderFollowerWorkerThread.java:80)
at org.apache.tomcat.util.threads.ThreadPool$ControlRunnable.run(
ThreadPool.java:684)
at java.lang.Thread.run(Thread.java:595)


it seems i should do something..



2006/9/18, Nick Snels <[EMAIL PROTECTED]>:


Hi James,

you should put the directory solr (which you find under /example/solr)
directly under c:\tomcat5.5\, so you get c:\tomcat5.5\solr\bin en
c:\tomcat5.5\solr\conf . The error you are getting is that tomcat cann't
find the solr configuration directory. Hope you get it running.

Kind regards,

Nick

On 9/17/06, James liu <[EMAIL PROTECTED]> wrote:
>
> tomcat 5.5 win2003,,i can find http://localhost:8080,,it is ok.
>
> i copy solr-nightly.war to "c:\tomcat5.5\webapps\solr.war"
>
> i start tomcat,,it auto generate solr directory in
> "c:\tomcat5.5\webapps\solr"
>
> http://localhost:8080/solr/admin/,,,it show me error information:
>
> HTTP Status 500 -
>
> type Exception report
>
> message
>
> description The server encountered an internal error () that prevented
it
> from fulfilling this request.
>
> exception
> org.apache.jasper.JasperException: Exception in JSP: /admin/_info.jsp:8
>
> 5:
> 6: <%@ page import="java.util.Date" %>
> 7:
> 8: 
> 9: 
> 10: <[EMAIL PROTECTED] file="header.jsp" %>
> 11:
>
>
> and i don't know where i should put example/solr
>
>
> if i wanna add lucene-analyzers-2.0.0.jar,,,where i put it?
>
>
> thks.
>
>




Re: tomcat install

2006-09-18 Thread James liu

Hi Nick,

It is very funny. when i reboot my pc,it is ok and i do nothing.

my new question is how to add lucene-analyzers-2.0.0.jar to tomcat or jetty.

i add useful classes to solr.war which exist
"C:\cygwin\tmp\solr-nightly\example\webapps\solr.war",,,but it is not
effect...

do u know how to solve it?


Regards,

JL

2006/9/18, Nick Snels <[EMAIL PROTECTED]>:


Hi James,

the problem is most likely a xml error in either schema.xml or
solrconfig.xml. Go through your Tomcat logs, if it is an xml error you
should find the line where the xml parsing went wrong.

Kind regards,

Nick

On 9/18/06, James liu <[EMAIL PROTECTED]> wrote:
>
> thk Nick.
>
> i do you tell me and i can see admin page.
>
> but when i click search ,,,error information:
>
> java.lang.NullPointerException
> at org.apache.solr.search.SolrQueryParser.(SolrQueryParser.java
:37)
> at org.apache.solr.search.QueryParsing.parseQuery(QueryParsing.java:47)
> at org.apache.solr.request.StandardRequestHandler.handleRequest(
> StandardRequestHandler.java:94)
> at org.apache.solr.core.SolrCore.execute(SolrCore.java:586)
> at org.apache.solr.servlet.SolrServlet.doGet(SolrServlet.java:91)
> at javax.servlet.http.HttpServlet.service(HttpServlet.java:689)
> at javax.servlet.http.HttpServlet.service(HttpServlet.java:802)
> at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(
> ApplicationFilterChain.java:252)
> at org.apache.catalina.core.ApplicationFilterChain.doFilter(
> ApplicationFilterChain.java:173)
> at org.apache.catalina.core.StandardWrapperValve.invoke(
> StandardWrapperValve.java:213)
> at org.apache.catalina.core.StandardContextValve.invoke(
> StandardContextValve.java:178)
> at org.apache.catalina.core.StandardHostValve.invoke(
> StandardHostValve.java
> :126)
> at org.apache.catalina.valves.ErrorReportValve.invoke(
> ErrorReportValve.java
> :105)
> at org.apache.catalina.core.StandardEngineValve.invoke(
> StandardEngineValve.java:107)
> at org.apache.catalina.connector.CoyoteAdapter.service(
CoyoteAdapter.java
> :148)
> at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java
> :869)
> at
>
>
org.apache.coyote.http11.Http11BaseProtocol$Http11ConnectionHandler.processConnection
> (Http11BaseProtocol.java:664)
> at org.apache.tomcat.util.net.PoolTcpEndpoint.processSocket(
> PoolTcpEndpoint.java:527)
> at org.apache.tomcat.util.net.LeaderFollowerWorkerThread.runIt(
> LeaderFollowerWorkerThread.java:80)
> at org.apache.tomcat.util.threads.ThreadPool$ControlRunnable.run(
> ThreadPool.java:684)
> at java.lang.Thread.run(Thread.java:595)
>




about sortMissingLast and sortMissingFirst

2006-09-18 Thread James liu

i look expression about them in schema.xml.

but i m not clear.

My
understand: if sortMissingLast is true do last in descending sort. for example:
 field name : pname,  if sortMissingLast is true, pname will be sorted desc
and we can't find result with pname, only other field information.

am i wrong?



--
regards
jl


no example to CollectionDistribution?

2006-09-18 Thread James liu

not find example in wiki.


anyone know?

--
regards
jl


Re: no example to CollectionDistribution?

2006-09-18 Thread James liu

maybe i should get cron through cygwin..

my system is win2003,not unix.

today i try ./snappuller,,,but it seems wrong and i set master port,
directory,snap directory

tomorrow i will try again.



2006/9/18, Chris Hostetter <[EMAIL PROTECTED]>:



: Subject: no example to CollectionDistribution?
:
: not find example in wiki.
:
:
: anyone know?

you really need to be more specific James:  what kinds of examples are you
looking for? ... the CollectionDistribution describes in depth how the
replication/distribution works, and has examples of the only things that
need to be croned: snapcleaner, snappuller, and snapinstaller.


-Hoss





--
regards
jl


Re: tomcat install

2006-09-18 Thread James liu

thank u, with your step and add junit,  it is ok.

you can analyzer your language?

i modify schema:
   
 
   
   
   
   
   
   
   

but nothing changed.



2006/9/19, Nick Snels <[EMAIL PROTECTED]>:


Hi James,

I also needed the DutchAnalyzer from Lucene in my Solr project. I did it
the
following way. Which is probably the hard way, because my Java knowledge
isn't that great.

1. I unzipped the solr-nightly build
2. I downloaded the latest code from lucene, preferrably from svn :
http://svn.apache.org/viewvc/lucene/java/ and all necessary analyzers from
the lucene sandbox
3. I put it into c:\solr-nightly\src\java\org\apache\lucene
4. I installed ant (unzip it and add ANT_HOME to your path)
5. than open a DOS prompt and go to c:\solr-nightly and run 'ant dist',
this
makes a new solr-1.0.war file in c:\solr-nightly\dist. That war file
contains also the lucene code along with your analyzers

This is how I did it, don't know if this is the right or the easiest way
to
do it.

Kind regards,

Nick


On 9/18/06, James liu <[EMAIL PROTECTED]> wrote:
>
> Hi Nick,
>
> It is very funny. when i reboot my pc,it is ok and i do nothing.
>
> my new question is how to add lucene-analyzers-2.0.0.jar to tomcat or
> jetty.
>
> i add useful classes to solr.war which exist
> "C:\cygwin\tmp\solr-nightly\example\webapps\solr.war",,,but it is not
> effect...
>
> do u know how to solve it?
>
>
> Regards,
>
> JL
>
> 2006/9/18, Nick Snels <[EMAIL PROTECTED]>:
> >
> > Hi James,
> >
> > the problem is most likely a xml error in either schema.xml or
> > solrconfig.xml. Go through your Tomcat logs, if it is an xml error you
> > should find the line where the xml parsing went wrong.
> >
> > Kind regards,
> >
> > Nick
> >
> > On 9/18/06, James liu <[EMAIL PROTECTED]> wrote:
> > >
> > > thk Nick.
> > >
> > > i do you tell me and i can see admin page.
> > >
> > > but when i click search ,,,error information:
> > >
> > > java.lang.NullPointerException
> > > at org.apache.solr.search.SolrQueryParser.(
SolrQueryParser.java
> > :37)
> > > at org.apache.solr.search.QueryParsing.parseQuery(QueryParsing.java
> :47)
> > > at org.apache.solr.request.StandardRequestHandler.handleRequest(
> > > StandardRequestHandler.java:94)
> > > at org.apache.solr.core.SolrCore.execute(SolrCore.java:586)
> > > at org.apache.solr.servlet.SolrServlet.doGet(SolrServlet.java:91)
> > > at javax.servlet.http.HttpServlet.service(HttpServlet.java:689)
> > > at javax.servlet.http.HttpServlet.service(HttpServlet.java:802)
> > > at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(
> > > ApplicationFilterChain.java:252)
> > > at org.apache.catalina.core.ApplicationFilterChain.doFilter(
> > > ApplicationFilterChain.java:173)
> > > at org.apache.catalina.core.StandardWrapperValve.invoke(
> > > StandardWrapperValve.java:213)
> > > at org.apache.catalina.core.StandardContextValve.invoke(
> > > StandardContextValve.java:178)
> > > at org.apache.catalina.core.StandardHostValve.invoke(
> > > StandardHostValve.java
> > > :126)
> > > at org.apache.catalina.valves.ErrorReportValve.invoke(
> > > ErrorReportValve.java
> > > :105)
> > > at org.apache.catalina.core.StandardEngineValve.invoke(
> > > StandardEngineValve.java:107)
> > > at org.apache.catalina.connector.CoyoteAdapter.service(
> > CoyoteAdapter.java
> > > :148)
> > > at org.apache.coyote.http11.Http11Processor.process(
> Http11Processor.java
> > > :869)
> > > at
> > >
> > >
> >
>
org.apache.coyote.http11.Http11BaseProtocol$Http11ConnectionHandler.processConnection
> > > (Http11BaseProtocol.java:664)
> > > at org.apache.tomcat.util.net.PoolTcpEndpoint.processSocket(
> > > PoolTcpEndpoint.java:527)
> > > at org.apache.tomcat.util.net.LeaderFollowerWorkerThread.runIt(
> > > LeaderFollowerWorkerThread.java:80)
> > > at org.apache.tomcat.util.threads.ThreadPool$ControlRunnable.run(
> > > ThreadPool.java:684)
> > > at java.lang.Thread.run(Thread.java:595)
> > >
> >
> >
>
>





--
regards
jl


Re: tomcat install

2006-09-19 Thread James liu
today not Ok。i check source of cjk: CJKAnalyzer.java and CJKTokenizer.java (these from lucene 2.0 source code)and your code,,,i write CJKJLFilterFactory.java and CJKJLTokenizerFactory.javaant is ok.i
 copy new solr.war to tomcat's webappsand modify schema.xml use admin page, i use http://localhost:8484/solr/admin/analysis.jsp?highlight=on
 to check word analyzeit show me *()*)&*^&*, oh my god. i m failure.i see org.apache.lucene.analysis.nl's code, i find something difference. like your Tokenizer sames as StandardTokenizer,,i have to use myself..
 thk u very much, no your code, i think i maybe give up. i only use delphi and php, no java and unix before i meet lucene.i use lucene well and i think i can use solr well.thk u again.my msn: 
[EMAIL PROTECTED],,,maybe we can be friend.


wana use CJKAnalyzer

2006-09-19 Thread James liu
My step to support CJK...:1:add lucene-analyzers-2.0.0.jar to "C:\cygwin\tmp\solr-nightly\lib"2:use cmd, "cd C:\cygwin\tmp\solr-nightly","ant dist"3:copy "C:\cygwin\tmp\solr-nightly\dist\solr- 
1.0.war" to "C:\cygwin\tmp\solr-nightly\example\webapps\solr.war"4:modify schema(conf/schema.conf), like yours,,just "" 
5:delete solr/data/index;6:start jetty,java -jar start.jar7:no error.8: http://localhost:8983/solr/admin,,,i click analyzer link,,,and try test analyzer chinese word,but nothing happend. 
9: i use  xml.php to add index(english is well),it show me ok10: i try lukeall.jar to see solr's index data. but it show me like my attachements.xml.php maybe error althrough no error show.
i write jl.xml to example/exampledocsuse cygwin: sh post.sh jl.xml,no error。and i use lukeall.jar to see,,nothing changed.i failure.maybe someone can give me some advice to solve it.
-- regardsjl 


  
  	111
ÐÕÃûÊÇÁõƽ
  
  
  	112
ÐÕÃûÊÇСÍõ
  
  
  	113
ÀÏÆÅ²»ÔÚ¼Ò
  


Re: no example to CollectionDistribution?

2006-09-19 Thread James liu

i see,thk u.

2006/9/20, Chris Hostetter <[EMAIL PROTECTED]>:



: maybe i should get cron through cygwin..
:
: my system is win2003,not unix.
:
: today i try ./snappuller,,,but it seems wrong and i set master port,
: directory,snap directory

The CollectionDistribution scripts may not work well on windows -- many of
them require hardlinks which may or may-not be supported by windows
orcygwin (i've heard differet things) ... snappuller in particular
requires that you have rsync running.



-Hoss





--
regards
jl


Re: tomcat install

2006-09-19 Thread James liu

i'd like to hear "I would start by trying to use the CJKAnalyzer as is with
the syntax,described above."

if need tester, call me.




2006/9/20, Chris Hostetter <[EMAIL PROTECTED]>:



: I have went through my archives and I have found that people also have
used
: something similar to:
:
: 
:  
: 

Correct.  If you want to use a Lucene analyzer "as is" all you need to do
is specify the class name.  if you wnat to make an analyzer on the fly
from a tokenizer and some tokenfilters -- you need factories for each.

I would start by trying to use the CJKAnalyzer as is with the syntax
described above.  once you get that working, then look at what it takes to
write factories for the tokeinizer so you can mix/match it with other
token filters.


-Hoss





--
regards
jl


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu

2006/9/20, Yonik Seeley <[EMAIL PROTECTED]>:


On 9/20/06, James liu <[EMAIL PROTECTED]> wrote:
> My step to support CJK...:
> 1:add lucene-analyzers-2.0.0.jar to
> "C:\cygwin\tmp\solr-nightly\lib"
> 2:use cmd, "cd C:\cygwin\tmp\solr-nightly","ant dist"
> 3:copy "C:\cygwin\tmp\solr-nightly\dist\solr- 1.0.war" to
> "C:\cygwin\tmp\solr-nightly\example\webapps\solr.war"
>
> 4:modify schema(conf/schema.conf), like yours,,just " class="org.apache.lucene.analysis.cjk.CJKAnalyzer"/>"
> 5:delete solr/data/index;
> 6:start jetty,java -jar start.jar
> 7:no error.
> 8: http://localhost:8983/solr/admin,,,i click analyzer
> link,,,and try test analyzer chinese word,but nothing happend.

When you say nothing happened, do you mean the analyzer didn't change
the text at all, or you didn't see any output at all?  Did you type
some text into the input fields?  Does it work for you with english
text?



i think it shows clear. step 8: i use admin page 's analyzer, analyzer
didn't change the text at all, and nothing output. im sure i type chinese
word into the input fields.
it work with english text.



9: i use xml.php to add index(english is well),it show me ok
> 10: i try lukeall.jar to see solr's index data. but it show me like my
> attachements.

Please be explicit on what the problem is... not many people on this
list can look at CJK and see what is wrong.


yes i know. im sorry not be explicit .

 Do you mean that the

analyzer isn't breaking up your text into words?

-Yonik



i means i follow these step, it break my text into words.but
i can't know these words..

you find index data from my attachements. its name is solr.jpg and lucene
breaking well, its name is lucene.jpg


--
regards
jl


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu

i m java newer. so i print these steps.

solr tutorial i test is ok.

anything you wanna know, mail me.


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu
attachements:  schema.xml2006/9/20, James liu <[EMAIL PROTECTED]>:
i m java newer. so i  print these steps.solr tutorial i test is ok.anything you wanna know, mail me.-- regardsjl 


  

  


  
  


  
  id
  content


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu

to mike:
" Are you testing the same field to which you are adding the analyzer?
I noticed in another mail that you added this to the "text_lu" field
type--the solr example uses "text", as I recall."

now my schema.xml:


 
 111
   姓名是刘平
 
 
 112
   姓名是小王
 
 
 113
   老婆不在家
 


but i m failed..


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu

Hoss, thk for ur help

2006/9/21, Chris Hostetter <[EMAIL PROTECTED]>:




: 6:start jetty,java -jar start.jar
: 7:no error.
: 8: http://localhost:8983/solr/admin,,,i click analyzer link,,,and try
: test analyzer chinese word,but nothing happend.

...i don't know much about non latin characters but i tried making the
same changes you did, and asked a coworker who speaks/types chinese to try
outthe Analyziz page, and he said it worked fine for him.

one comment he had was that it only works if your www browser is
configured to use utf-8 or to auto-select the character encoding (in which
case it uses utf-8 because that's what the HTML page itself specifies as
the encoding).  if you browser is explicitly configured to use Simplified
Chinese (or, i assume, Traditional Chinese) as the encoding, then it won't
work (the page he got looks like it might be what you are seeing: no data
returned under the form, as if you had provided no input)



i tried to avoid browser and use post.sh (in example/exampledocs),,i put
jl.xml


 
 111
   姓名是刘平
 
 
 112
   姓名是小王
 
 
 113
   老婆不在家
 


under cygwin,i use sh post.sh jl.xml.

so i think it is not www browser's problem.

which difference? system: i use win2003,,java i use
"C:\Sun\AppServer\jdk"tutorial i try is ok. cygwin is install from
internet.

test by jetty(it include solr) and tomcat 5.5


could you zip your code to me,,i try it.

if failure,,i think only envirement make problem.

can i contact with the chinese coworker?

Traditional Chinese <> Simplified Chinese..

and i use Simplified Chinese。


--
regards
jl


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu

sorry,,it is wrong...

my schema.xml



 
   
 
   
   
 
 
   
   
 
 id
 content




在06-9-21,James liu <[EMAIL PROTECTED]> 写道:


to mike:

" Are you testing the same field to which you are adding the analyzer?
I noticed in another mail that you added this to the "text_lu" field
type--the solr example uses "text", as I recall."

now my schema.xml:


  
  111
姓名是刘平
  
  
  112
姓名是小王
  
  
  113
老婆不在家
  


but i m failed..





--
regards
jl


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu

i use lukeall.jar to check indexdata.

u can find picture from http://www.flickr.com/photos/[EMAIL PROTECTED]/?saved=1


solr.jpg is i use lukeall.jar to check solr's index data.

lucene.jpg is  i use lukeall.jar to check lucene's index data.


now i use lucene is ok.


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu

i dont know it is import i add junit

when i use ant dist,,,it show me error information : not found junit,,,so i
download and add it.

Is it problem about CJKAnalyzer?


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu

i recompile it.

when i ant dist...cmd shows some api is old and uncheck...

Is it problem?


my java version you can find
http://www.flickr.com/photos/[EMAIL PROTECTED]/?saved=1



2006/9/21, James liu <[EMAIL PROTECTED]>:


i dont know it is import i add junit

when i use ant dist,,,it show me error information : not found junit,,,so
i download and add it.

Is it problem about CJKAnalyzer?






--
regards
jl


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu

thk Hoss,Nick Snels,Koji,Mike and  everybody who helped me and wanna help
me..

i can use solr with Chinese Word.


Re: wana use CJKAnalyzer

2006-09-20 Thread James liu

i just wanna say: no your help,maybe i will give up.thk u again.

http://www.flickr.com/photos/[EMAIL PROTECTED]/248815068/

2006/9/21, James liu <[EMAIL PROTECTED]>:


thk Hoss,Nick Snels,Koji,Mike and  everybody who helped me and wanna help
me..

i can use solr with Chinese Word.









--
regards
jl


Re: wana use CJKAnalyzer

2006-09-21 Thread James liu

2006/9/22, Chris Hostetter <[EMAIL PROTECTED]>:



: i just wanna say: no your help,maybe i will give up.thk u again.
:
: http://www.flickr.com/photos/[EMAIL PROTECTED]/248815068/

: > thk Hoss,Nick Snels,Koji,Mike and  everybody who helped me and wanna
help
: > me..
: >
: > i can use solr with Chinese Word.

I'm sorry, i'm really confused now ... it seems like you got things
working, but you also say "maybe i will give up" ... ?




Express problem. It is maybe i would have given up...in fact i not give
up.my english is poor.
Thk Yonik.

1) if you did get things working, what was the root of your problem, was

it the utf-8 issue when using the forms in your browser or adding docs?



Yes,it working. the root of my problem is xml muse be encoded by utf-8.
if use php,it not about www browser. just notice that
curl header information must be utf-8.
if use post.sh,xml muse be encoded by utf-8.(my editplus default encode
style is ansi)


2) if things aren't working right, what is the current state of things?

... from the picture "solr_chinese" on your flicker page, Luke seems to be
showing you Chinese characters in a Solr index ... are they not being
tokenized properly or something?




-Hoss





--
regards
jl


Re: wana use CJKAnalyzer

2006-09-22 Thread James liu
2006/9/23, Walter Underwood <[EMAIL PROTECTED]>:
On 9/21/06 5:37 PM, "James liu" <[EMAIL PROTECTED]> wrote:> Yes,it working. the root of my problem is xml muse be encoded by utf-8.> if use php,it not about www browser. just notice that
> curl header information must be utf-8.> if use post.sh,xml muse be encoded by utf-8.(my editplus default encode> style is ansi)This might be a Solr bug. Solr should be able to accept XML in any
of the required encodings (ASCII, Latin 1, UTF-8, and UTF-16).Getting XML content types exactly right is tricky, see RFC 3023.What curl command line was used?No sepcial curl command i 
use.just solr-nightly/example/exampledocs post.sh.but my jl.xml encoded  utf-8(i use editplus, i tried to use  xml encoding utf 8, but it is not effect).solrphp i use curl "$header=array("Content-Type: text/xml;charset=utf-8");curl_setopt($ch, CURLOPT_HTTPHEADER, $header);", this is php.

What encoding is the XML?Can you give a sample XML file?see attachments, anything you need mail me. 
wunder--Walter UnderwoodSearch Guru, Netflix-- regardsjl 

  
  	111
姓名是刘平
  
  
  	112
姓名是小王
  
  
  	113
老婆不在家
  
  



Re: wana use CJKAnalyzer

2006-09-25 Thread James liu

2006/9/25, Walter Underwood <[EMAIL PROTECTED]>:


This document has two problems. First, the document is not well-formed
XML.
Open it  in Firefox and you will see this error:

   XML Parsing Error: mismatched tag. Expected: .
   Location: file:///Users/wunderwood/Desktop/jl.xml
   Line Number 15, Column 3:

After I fix that, it still is not legal UTF-8.



Im sorry that it have more , because i test more data in
solr. In order to transfter attachements, i reduced jl.xml and not check.
so, you find this problem.
yes, it is not legal utf-8.
utf-8 encoding i mean that is file encoding mode.
when you create new xml by using editplus, and save it, it appears window
that have a selection encoding mode.(u can find it with attachements)
That is jl.xml,Index it by post.sh.

if you use "script language", like solrphp(my solrphp not from solr's wiki)
that i modified. you must send your xml with encoding utf-8.
for instance, i try send my.xml to http://localhost:8983/solr/update-< this
url's head information should have ""Content-Type: text/xml;charset=utf-8"";
Solr work well after with head information.


Does Solr report parsing errors? It really should. Maybe a 400 Bad Request

response with a text/plain body showing the error message.



after i fixed "more 


On 9/22/06 6:24 PM, "James liu" <[EMAIL PROTECTED]> wrote:
>
> 2006/9/23, Walter Underwood <[EMAIL PROTECTED]>:
>> On 9/21/06 5:37 PM, "James liu" <[EMAIL PROTECTED]> wrote:
>>
>>> > Yes,it working. the root of my problem is xml muse be encoded by
utf-8.
>>> > if use php,it not about www browser. just notice that
>>> > curl header information must be utf-8.
>>> > if use post.sh,xml muse be encoded by utf-8.(my editplus default
encode
>>> > style is ansi)
>>
>> This might be a Solr bug. Solr should be able to accept XML in any
>> of the required encodings (ASCII, Latin 1, UTF-8, and UTF-16).
>> Getting XML content types exactly right is tricky, see RFC 3023.
>>
>> What curl command line was used?
>
> No sepcial curl command i use.just solr-nightly/example/exampledocs
post.sh.
> but my jl.xml encoded  utf-8(i use editplus, i tried to use  xml
encoding utf
> 8, but it is not effect).
> solrphp i use curl "$header=array("Content-Type:
> text/xml;charset=utf-8");curl_setopt($ch, CURLOPT_HTTPHEADER,
$header);", this
> is php.
>
>> What encoding is the XML?
>>
>> Can you give a sample XML file?
>
> see attachments, anything you need mail me.
>
>> wunder
>> --
>> Walter Underwood
>> Search Guru, Netflix
>>
>
>







--
regards
jl


SolrDistribution is tested by FreeBSD?

2006-12-21 Thread James liu

i use fb 6.1

when i use
# sh rsyncd-enable

it show me:
cd: can't cd to rsyncd-enable/..
.: Can't open /usr/local/www/data/solr/example/solr/bin/bin/scripts-util: No
such file or directory




--
regards
jl


  1   2   3   >