3 problems:
a- he wanted to read it locally.
b- crawling the open web is imperfect.
c- /browse needs to get at the files with the same URL as the uploader.
a and b- Try downloading the whole thing with 'wget'. It has a 'make
links point to the downloaded files' option. Wget is great.
I have do
That's why I think this could be a nice joint project with Apacha Infra.
They provide Moin export, we build a way to index it with Solr for local
usage. Start with our own - Solr - project , then sell it to others once it
has been dog-fooded enough. Instant increased Solr exposure to all Apache
pro
I have permission to provide an export. Right now I'm thinking of it
being a one off dump, without the user dir. If someone wants to research
how to make moin automate it, I at least promise to listen.
Upayavira
On Tue, Jan 1, 2013, at 08:10 AM, Alexandre Rafalovitch wrote:
> That's why I think t
Can you please clarify your answer?
You said to try to bootstap:
> > On the solr cloud page, admittedly down the page a ways, is the line
> > below.
> > Does that apply?
> > Best
> > Erick
> >
> > # try bootstrapping all the conf dirs in solr.xml
> > java -classpath example/solr-webapp/WEB-INF/lib/
I had a partial success with executing a wget as follows:
wget --recursive --page-requisites --no-clobber --html-extension
--convert-links --restrict-file-names=windows --domains
wiki.apache.org http://wiki.apache.org/solr/ -w 10 -l 5
configuring a web server to serve that location and then ind
I was reading a solr wiki located at
http://wiki.apache.org/solr/NearRealtimeSearch
It says all commitWithin are now soft commits.
can any one explain what does it means?
Does It means commitWithin will not cause a hard commit?
Moreover that wiki itself is insufficient,as feature is NRT.
can a
Hmmm... quite some time ago I switched from ASCIIFoldingFilterFactory to
MappingCharFilterFactory, because I was told (by who I can't recall) that
the latter was "better/preferred". Is there any particular reason to favor
one over the other?
So, if one is preferred over the other, the questio
hi everyone,
I have a system written in .net that I'm using lucene .net.
I'm thinking of changing that system to be based on solr instead of lucene.
My situation is that I have a lot of code that was written in .net, that I
don't want to waste because moving to solr makes me to change code to java
On 1/1/2013 1:03 PM, dafna wrote:
hi everyone,
I have a system written in .net that I'm using lucene .net.
I'm thinking of changing that system to be based on solr instead of lucene.
My situation is that I have a lot of code that was written in .net, that I
don't want to waste because moving to
thanks shawn for your answer.
I know that all of the analyzers, tokenizers, filters usally are on server
side.
That is exactly my question, if I will take them outside the engine, and do
it outside, is it still work the same ?
--
View this message in context:
http://lucene.472066.n3.nabble.com
On 1/1/2013 1:46 PM, dafna wrote:
thanks shawn for your answer.
I know that all of the analyzers, tokenizers, filters usally are on server
side.
That is exactly my question, if I will take them outside the engine, and do
it outside, is it still work the same ?
Solr can use a lucene index if th
The code that I don't want to rewrite is the analyzers.
I have written many analyzers.
I know the benefits of solr, I just want to spend as less time as I can to
see the system in solr.
I don't think you understood my idea:
1. The data handling will be just in lucene .net.
2. The index itself will
Solr is written in Java. It can't as it is make use of .Net code. You'd
need to find a way to convert your code from .Net to Java. Or you'd need
to write adapters in Java that can interface to your .Net code. One way
or another, you're gonna have to work out how to get the two languages
to work tog
If all of your analyzers are on the indexing side, you can look into
preanalyzed field type and just push already processed tokens into Solr.
Not sure about query time though.
Regards,
Alex
On 1 Jan 2013 16:13, "dafna" wrote:
> The code that I don't want to rewrite is the analyzers.
> I have
Hi, Upayavira
I know multiple segments are not problem.
But I always optimize index on master server before replicate.
So just single segment file is on master.
File lists of the master server directory are below.
Additionally, segments_1 and segments_2 on slave server are deleted by hand.
[root@s
15 matches
Mail list logo