Hi All,
I am very new to solr as well as java too.
I require to use solrj for indexing also require to index automatically once
in 24 hour.
I wrote java code for indexing now I want to do further coding for automatic
process.
Could you suggest or give me sample code for automatic index process..
p
Hi Jonty,
what is your specific problem?
You could use a cronjob or the Java-lib called quartz to automate this task.
Or did you mean replication?
Regards,
Peter.
> Hi All,
>
> I am very new to solr as well as java too.
> I require to use solrj for indexing also require to index automatically on
Hi Peter,
actually I want the index process should start automatically. right now
I am doing mannually.
same thing I want to start indexing when less load on server i.e. late
night. So setting auto will fix my
problem..
On Wednesday 02 June 2010 02:00 PM, Peter Karich wrote:
Hi Jonty,
what
Hi Peter,
actually I want the index process should start automatically. right now I am
doing mannually.
same thing I want to start indexing when less load on server i.e. late
night. So setting auto will fix my
problem..
On Wed, Jun 2, 2010 at 2:00 PM, Peter Karich wrote:
> Hi Jonty,
>
> what is
Hi,
I have solr 1.4. In schema i have a field called "title" of type "text"
Now problem is, when I search for "Test_Title" it brings all documents with
titles like "Test-Title", "Test_Title", "Test,Title, "Test Title",
"Test.Title"
What to do to avoid this?
"Test_Title" should only return documen
What analyzer you are using to index and search? Check out schema.xml. You
are currently using analyzer which breaks the words. If you don't want to
break then you need to use .
Regards
Aditya
www.findbestopensource.com
On Wed, Jun 2, 2010 at 2:41 PM, M.Rizwan wrote:
> Hi,
>
> I have solr 1.4.
my batchSize is -1 and the load ist to big for us. why i should increase it ?
what is a normal serverload ? our server is a fast server. 4 cores 3 GB Ram
but we dont want a serverload from over 2 when index a starts.
--
View this message in context:
http://lucene.472066.n3.nabble.com/DIH-
You need to do schedule your task. Check out schedulers available in all
programming languages.
http://www.findbestopensource.com/tagged/job-scheduler
Regards
Aditya
www.findbestopensource.com
On Wed, Jun 2, 2010 at 2:39 PM, Jonty Rhods wrote:
> Hi Peter,
>
> actually I want the index process
Those aren't in the default parameters. They are config for the SearchHandler
itself.
On Jun 1, 2010, at 9:00 PM, Lance Norskog wrote:
> In the "/spell" declaration in the example solrconfig.xml, we find
> these lines among the default parameters:
>
>
> spellcheck
>
>
> How does
On Jun 1, 2010, at 9:54 PM, Blargy wrote:
>
> We have around 5 million items in our index and each item has a description
> located on a separate physical database. These item descriptions vary in
> size and for the most part are quite large. Currently we are only indexing
> items and not their
On 2010-06-02 12:42, Grant Ingersoll wrote:
>
> On Jun 1, 2010, at 9:54 PM, Blargy wrote:
>
>>
>> We have around 5 million items in our index and each item has a description
>> located on a separate physical database. These item descriptions vary in
>> size and for the most part are quite large.
On Jun 2, 2010, at 6:53 AM, Andrzej Bialecki wrote:
> On 2010-06-02 12:42, Grant Ingersoll wrote:
>>
>> On Jun 1, 2010, at 9:54 PM, Blargy wrote:
>>
>>>
>>> We have around 5 million items in our index and each item has a description
>>> located on a separate physical database. These item descr
On 2010-06-02 13:12, Grant Ingersoll wrote:
>
> On Jun 2, 2010, at 6:53 AM, Andrzej Bialecki wrote:
>
>> On 2010-06-02 12:42, Grant Ingersoll wrote:
>>>
>>> On Jun 1, 2010, at 9:54 PM, Blargy wrote:
>>>
We have around 5 million items in our index and each item has a description
loc
Hi,
I want to hit the query given below :
?q=*:*&facet=true&facet.date=pub&facet.date.start=2000-01-01T00:00:00Z&facet.date.end=2010-01-01T00:00:00Z&facet.date.gap=%2B1YEAR
using SolrJ. I am browsing the net but not getting any clues about how
should I approach it. How can SolJ API be used to c
Hi Ninad,
SolrQuery q = new SolrQuery();
q.setQuery("*:*");
q.setFacet(true);
q.set("facet.data", "pub");
q.set("facet.date.start", "2000-01-01T00:00:00Z")
... etc.
basically you can completely build your entire query with the 'raw' set (and
add) methods.
The specific methods are just helpers.
S
Thanks Greet-Jan. Din't know about this trick. [?]
On Wed, Jun 2, 2010 at 5:39 PM, Geert-Jan Brits wrote:
> Hi Ninad,
>
> SolrQuery q = new SolrQuery();
> q.setQuery("*:*");
> q.setFacet(true);
> q.set("facet.data", "pub");
> q.set("facet.date.start", "2000-01-01T00:00:00Z")
> ... etc.
>
> basic
Hallo Users...
I have a Problem...
In my SolR, i have a lot of multiValued, dynamicFields and now i must print
ther Fields in php..
But i dont know how...
In schema.xml:
output from Solr:
A201005311740560002.xml
NO
A201005311740560002
2010-05-31 17:40:56
−
Q:\DatenI
You CAN easily turn spellchecking on or off, or set the spellcheck dictionary,
in request parameters. So there's really no need, that I can think of, to try
to actually add or remove the spellcheck component in request parameters; you
could just leave it turned off in your default parameters,
Hello.
Our Server is a 8-Core Server with 12 GB RAM.
Solr is running with 4 Cores.
55 Tomcat 5.5 processes are running. ist this normal ???
htop show me a list of these processes of the server. and tomcat have about
55.
every process using:
/usr/share/java/commons-daemon.jar:/usr/share/tom
You probably should try the php or phps response writer - it'll likely
make your PHP integration easier.
Erik
On Jun 2, 2010, at 9:50 AM, Jörg Agatz wrote:
Hallo Users...
I have a Problem...
In my SolR, i have a lot of multiValued, dynamicFields and now i
must print
ther Fields i
My guess would be that commons-daemon is somehow thinking that Tomcat has gone
down and started up multiple copies... You only need one Tomcat process for
your 4 core Solr instance! You may have many other WAR applications hosted in
Tomcat, I know a lot of places would have 1 tomcat per depl
Is your server Linux?
In this case this is very normal.. any java application spawns many
new processes on linux... it's not exactly bound to threads
unfortunately.
And, of course, they all refer to the same invocation path.
paul
Le 02-juin-10 à 15:59, stockii a écrit :
Hello.
Our Ser
yes i done.. but i dont know how i get the information out of the big
Array...
Al fields like P_VIP_ADR_*
Glad to hear someone looking at Solr not just as web enabled search engine, but
as a simpler/more powerful interface to Lucene!
When you download the source code, look at the Chapter 8 "Crawler" project,
specifically "Indexer.java", it demonstrates how to index into both a
traditional separa
yes, its a Linux... Debian System.
when i running a import. only 2-3 tomcat processes are running. the other
doing nothing ... thats what is strange for me .. ^^
--
View this message in context:
http://lucene.472066.n3.nabble.com/Many-Tomcat-Processes-on-Server-tp864732p864804.html
Sent from th
You'd need to search explanations for this at generic java forums.
It's the same with any java process on Linux.
In the Unix family Solaris and MacOSX do it better, fortunately and is
probably due to the very old time where the Linux java was a
translation of the Solaris java with the special
Am 02.06.2010 16:13, schrieb Paul Libbrecht:
> Is your server Linux?
> In this case this is very normal.. any java application spawns many new
> processes on linux... it's not exactly bound to threads unfortunately.
Uh, no. New threads in Java typically don't spawn new processes on OS level.
I ne
Am 02.06.2010 16:15, schrieb Jörg Agatz:
> yes i done.. but i dont know how i get the information out of the big
> Array...
They're simply the keys of a single response array.
Maybe he was looking at the output from top or htop?
-Original Message-
From: Michael Kuhlmann [mailto:michael.kuhlm...@zalando.de]
Sent: Wednesday, June 02, 2010 10:29 AM
To: solr-user@lucene.apache.org
Subject: Re: Many Tomcat Processes on Server ?!?!?
Am 02.06.2010 16:13, schrieb Paul
oha... "ps aux" shows only 3 processes from tomcat55.
but why show htop 55 ? close the garbage collector these not ?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Many-Tomcat-Processes-on-Server-tp864732p864849.html
Sent from the Solr - User mailing list archive at Nabble.
This is impressive, I had this in any Linux I've been using: SuSE,
Ubuntu, Debian, Mandrake, ...
Maybe there's some modern JDK with a modern Linux where it doesn't
happen?
It surely is not one process per thread though.
paul
Le 02-juin-10 à 16:29, Michael Kuhlmann a écrit :
Am 02.06.2010
i don't understand what you mean!
Am 02.06.2010 16:39, schrieb Paul Libbrecht:
> This is impressive, I had this in any Linux I've been using: SuSE,
> Ubuntu, Debian, Mandrake, ...
> Maybe there's some modern JDK with a modern Linux where it doesn't happen?
> It surely is not one process per thread though.
I'm not a linux thread ex
Am 02.06.2010 16:42, schrieb Jörg Agatz:
> i don't understand what you mean!
>
Then you should ask more precisely.
all the process in in htop show, have a own PID. so thats are no threads ?
i restart my tomcat via " /etc/init.d/tomcat restart "
do you think that after ervery resart the processes arent closed ?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Many-Tomcat-Processes-on-S
Try shutting tomcat down instead of restarting. If processes remain, then I'd
say further investigation is warranted. If no processes remain, then I think
it's safe to disregard unless you notice any problems.
-Original Message-
From: stockii [mailto:st...@shopgate.com]
Sent: Wednesday,
Le 02-juin-10 à 16:57, stockii a écrit :
all the process in in htop show, have a own PID. so thats are no
threads ?
No, you can't say that.
In general it is sufficient for the "mother process" to be killed but
it can take several attempts.
i restart my tomcat via " /etc/init.d/tomcat r
Thanks Yonik.
I guess the confusing thing is if the lucene query parser (for nested
querries) does backslash escaping, and the LocalParams also does
backslash escaping when you have a nested query with local params,
with quotes at both places... the inner scope needs... double escaping?
Solr is used to manage lists of indexes.
We have a database containing documents of different types.
Each document type is defined by a list of properties and we want to associate
some of these properties with lists of indexes to help users during query.
For example:
The property contains a text
okay you are right. thats all threads and no processes ...
but so many ? :D hehe
so when all the "processes" are threads i think its okay so ?! i can ignore
this ... XD
--
View this message in context:
http://lucene.472066.n3.nabble.com/Many-Tomcat-Processes-on-Server-tp864732p865008.html
Sent
As a data point, I routinely see clients index 5M items on normal
> hardware in approx. 1 hour (give or take 30 minutes).
Our master solr machine is running 64-bit RHEL 5.4 on dedicated machine with
4 cores and 16G ram so I think we are good on the hardware. Our DB is MySQL
version 5.0.67 (exa
Andrzej Bialecki wrote:
>
> On 2010-06-02 12:42, Grant Ingersoll wrote:
>>
>> On Jun 1, 2010, at 9:54 PM, Blargy wrote:
>>
>>>
>>> We have around 5 million items in our index and each item has a
>>> description
>>> located on a separate physical database. These item descriptions vary in
>>> si
As a data point, I routinely see clients index 5M items on normal hardware
in approx. 1 hour (give or take 30 minutes).
Also wanted to add that our main entity (item) consists of 5 sub-entities
(ie, joins). 2 of those 5 are fairly small so I am using
CachedSqlEntityProcessor for them but the ot
I see. It's still a little confusing to me but I'm fine as long as
this is the expected behavior. I also tried the "example" index
with data that come with the solr distribution and observe the
same behavior - only String fields are displayed. So Lucene is
sharing _some_ types with Solr but not all
One thing that might help indexing speed - create a *single* SQL query
to grab all the data you need without using DIH's sub-entities, at
least the non-cached ones.
Erik
On Jun 2, 2010, at 12:21 PM, Blargy wrote:
As a data point, I routinely see clients index 5M items on normal
: I see. It's still a little confusing to me but I'm fine as long as
: this is the expected behavior. I also tried the "example" index
: with data that come with the solr distribution and observe the
: same behavior - only String fields are displayed. So Lucene is
: sharing _some_ types with Solr
: In the "/spell" declaration in the example solrconfig.xml, we find
: these lines among the default parameters:
as grant pointed out: these aren't in the default params
: How does one supply such an array of strings in HTTP parameters? Does
: Solr have a parsing option for this?
in general, ig
: Is it possible to combine index and file spellcheck dictionaries?
off the top of my head -- i don't think so. however you could add special
docs to your index, which only contain the "spell" field you use to build
your spellcheck index, based on the contents of your dictionary file.
-Hoss
> One thing that might help indexing speed - create a *single* SQL query
> to grab all the data you need without using DIH's sub-entities, at
> least the non-cached ones.
>
Not sure how much that would help. As I mentioned that without the item
description import the full process takes 4 h
: Obviously I could implement this in userland (like like mincount for
: that matter), but I wonder if anyone else see's use in being able to
: define that a facet must match a minimum percentage of all documents in
: the result set, rather than a hardcoded value? The idea being that while
: I
thx for your reply!
On 02.06.2010, at 20:27, Chris Hostetter wrote:
> feel free to file a feature request -- truthfully this is kind of a hard
> problem to solve in userland, you'd either have to do two queries (the
> first to get the numFound, the second with facet.mincount set as an
> intege
How long does it take to do a grab of all the data via SQL? I found by
denormalizing the data into a lookup table meant that I was able to
index about 300k rows of similar data size with dih regex spilting on
some fields in about 8mins I know it's not quite the scale bit with
batching...
: When I query for a word say Tiger woods, and sort results by score... i do
: notice that the results are mixed up i.e first 5 results match Tiger woods
: the next 2 match either tiger/tigers or wood/woods
: the next 2 after that i notice again match tiger woods.
:
: How do i make sure that when
I've got a situation where I'm looking to build an auto-suggest where any
term entered will lead to suggestions. For example, if I type "wine" I want
to see suggestions like this:
french *wine* classes
*wine* book discounts
burgundy *wine*
etc.
I've tried some tricks with shingles, but the only
I'm painfully new to Solr so please be gentle if my suggestion is terrible!
Could you use highlighting to do this? Take the first n results from a query
and show their highlights, customizing the highlights to show the desired
number of words.
Just a thought.
Patrick
-Original Message
I was interested in the same thing and stumbled upon this article:
http://www.mattweber.org/2009/05/02/solr-autosuggest-with-termscomponent
-and-jquery/
I haven't followed through, but it looked promising to me.
Tim
-Original Message-
From: Jay Hill [mailto:jayallenh...@gmail.com]
Sent
Hi,
I installed Solr Server on my machine and able to access with localhost. I
tried accessing from a different machine with IP Address but not able to access
it. What do I need to do to be able to access the Solr instance from any
machine within the network?
Thanks,
Murali
details... detailseverybody let's say details !
Which app server are you using ?
What is the error message that you get when trying to access solr admin from
another machine ?
On Wed, Jun 2, 2010 at 9:39 PM, Bondiga, Murali <
murali.krishna.bond...@hmhpub.com> wrote:
> Hi,
>
> I installed
Thank you so much for the reply.
I am using Jetty which comes with Solr installation.
http://localhost:8983/solr/
The above URL works fine.
The below URL does not work:
http://177.44.9.119:8983/solr/
-Original Message-
From: Abdelhamid ABID [mailto:aeh.a...@gmail.com]
Sent: Wednes
When you access from another machine what message error do you get ?
Check your remote access with Telnet to see if the server respond
On Wed, Jun 2, 2010 at 10:26 PM, Bondiga, Murali <
murali.krishna.bond...@hmhpub.com> wrote:
> Thank you so much for the reply.
>
> I am using Jetty which comes
Thank you Chris. I'm clear now. I'll give Luke's latest version a try
when it's out.
On Wed, Jun 2, 2010 at 9:47 AM, Chris Hostetter
wrote:
>
> : I see. It's still a little confusing to me but I'm fine as long as
> : this is the expected behavior. I also tried the "example" index
> : with data th
Hi,
Can I restrict the facet search within the result count?
Example: A total of 100 documents were fetched for a given query x, and
facet worked in these 100 documents. I want that facet should work only on
first 10 documents fetched from query x.
Regards,
Sushan Rungta
: I ran through some more failure scenarios (scenarios and results below). The
: concerning ones in my deployment are when data does not get updated, but the
: DIH's .properties file does. I could only simulate that scenario when I ran
: out of disk space (all all disk space issues behaved consist
Hi,
I'm new to SOLR and have some basic questions that hopefully steer me in the
right direction.
- I want my search to "auto" spell check - that is if someone types
"restarant" I'd like the system to automatically search for restaurant.
I've seen the SpellCheckComponent but that doesn't seem to
Wait! You're fetching records from one database and then doing lookups
against another DB? That makes this a completely different problem.
The DIH does not to my knowledge have the ability to "pool" these
queries. That is, it will not build a batch of 1000 keys from
datasource1 and then do a query
Well, I hope to have around 5 million datasets/documents within 1 year, so this
is good info. BUT if I DO have that many, then the market I am aiming at will
end giving me 100 times more than than within 2 years.
Are there good references/books on using Solr/Lucen/(linux/nginx) for 500
million
When adding data continuously, that data is available after committing and is
indexed, right?
If so, how often is reindexing do some good?
Dennis Gearon
Signature Warning
EARTH has a Right To Life,
otherwise we all die.
Read 'Hot, Flat, and Crowded'
Laugh at http://www.yert.
That's promising!!! That's how I have been desigining my project. It must be
all the joins that are causing the problems for him?
Dennis Gearon
Signature Warning
EARTH has a Right To Life,
otherwise we all die.
Read 'Hot, Flat, and Crowded'
Laugh at http://www.yert.com/film.ph
Lance Norskog-2 wrote:
>
> Wait! You're fetching records from one database and then doing lookups
> against another DB? That makes this a completely different problem.
>
> The DIH does not to my knowledge have the ability to "pool" these
> queries. That is, it will not build a batch of 1000 key
Erik Hatcher-4 wrote:
>
> One thing that might help indexing speed - create a *single* SQL query
> to grab all the data you need without using DIH's sub-entities, at
> least the non-cached ones.
>
> Erik
>
> On Jun 2, 2010, at 12:21 PM, Blargy wrote:
>
>>
>>
>> As a data point, I ro
Would dumping the databases to a local file help at all?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Importing-large-datasets-tp863447p866538.html
Sent from the Solr - User mailing list archive at Nabble.com.
Ah! Thank you.
On Wed, Jun 2, 2010 at 9:52 AM, Chris Hostetter
wrote:
>
> : In the "/spell" declaration in the example solrconfig.xml, we find
> : these lines among the default parameters:
>
> as grant pointed out: these aren't in the default params
>
> : How does one supply such an array of stri
Hi,
I'm trying to use the field collapsing feature.
For that I need to take a checkout of the trunk and apply the patch
available at https://issues.apache.org/jira/browse/SOLR-236
When I take a checkout and run the example-DIH, I get following error in
browser on doing dataimport?command=full-i
Thanks for you advice. I did as you said and i still cannot search my
content.
One thing i notice here i can search for only the words within first 100
rows or maybe bigger than this not sure but not all. So is it the limitation
of the index it self? When I create another sample content with only
Modify all settings in solrconfig.xml and try again, by
default solr will only index the first 1 fields.
Best Regards,
Yandong
-Original Message-
From: Mint o_O! [mailto:mint@gmail.com]
Sent: 2010年6月3日 13:58
To: solr-user@lucene.apache.org
Subject: Re: Solr Search problem; cann
On 3 Jun 2010, at 02:58, Dennis Gearon wrote:
When adding data continuously, that data is available after
committing and is indexed, right?
Yes
If so, how often is reindexing do some good?
You should only need to reindex if the data changes or you change your
schema. The DIH in solr 1.4
On 3 Jun 2010, at 02:51, Dennis Gearon wrote:
Well, I hope to have around 5 million datasets/documents within 1
year, so this is good info. BUT if I DO have that many, then the
market I am aiming at will end giving me 100 times more than than
within 2 years.
Are there good references/b
Hi Yandong,
You are right. It works!!!
You are the best.
Thanks,
Mint
2010/6/3 Zero Yao
> Modify all settings in solrconfig.xml and try again, by
> default solr will only index the first 1 fields.
>
> Best Regards,
> Yandong
>
> -Original Message-
> From: Mint o_O! [mailto:mint.
On 3 Jun 2010, at 03:51, Blargy wrote:
Would dumping the databases to a local file help at all?
I would suspect not especally with the size of your data. But it would
be good to know how long that takes i.e. Creating a SQL script that
just pulls that data out how long does that take?
79 matches
Mail list logo