Re: indexing a xml file

2011-09-24 Thread GR
i think the xml to be indexed has to follow a certain schema, defined  
in schema.xml under conf directory. maybe, your solr.xml is not doing  
that


Sent from my iPhone

On 24 Sep 2011, at 18:15, ahmad ajiloo  wrote:


hello
Solr Tutorial page explains about index a xml file. but when I try  
to index

a xml file with this command:
~/Desktop/apache-solr-3.3.0/example/exampledocs$ java -jar post.jar  
solr.xml

I get this error:
SimplePostTool: FATAL: Solr returned an error #400 ERROR:unknown field
'name'

can anyone help me?
thanks


Re: OS Cache - Solr

2011-10-19 Thread GR

i wonder how do you manage 200 instances

On 20 Oct 2011, at 09:21, Sujatha Arun  wrote:


Yes 200 Individual Solr Instances not solr cores.

We get an avg response time of below 1 sec.

The number of documents is  not many most of the isntances ,some of  
the

instnaces have about 5 lac documents on average.

Regards
Sujahta

On Thu, Oct 20, 2011 at 3:35 AM, Jaeger, Jay - DOT >wrote:


200 instances of what?  The Solr application with lucene, etc. per  
usual?

Solr cores? ???

Either way, 200 seems to be very very very many: unusually so.  Why  
so

many?

If you have 200 instances of Solr in a 20 GB JVM, that would only  
be 100MB

per Solr instance.

If you have 200 instances of Solr all accessing the same physical  
disk, the
results are not likely to be satisfactory - the disk head will go  
nuts

trying to handle all of the requests.

JRJ

-Original Message-
From: Sujatha Arun [mailto:suja.a...@gmail.com]
Sent: Wednesday, October 19, 2011 12:25 AM
To: solr-user@lucene.apache.org; Otis Gospodnetic
Subject: Re: OS Cache - Solr

Thanks ,Otis,

This is our Solr Cache  Allocation.We have the same Cache  
allocation for

all
our *200+ instances* in the single Server.Is this too high?

*Query Result Cache*:LRU Cache(maxSize=16384, initialSize=4096,
autowarmCount=1024, )

*Document Cache *:LRU Cache(maxSize=16384, initialSize=16384)


*Filter Cache* LRU Cache(maxSize=16384, initialSize=4096,
autowarmCount=4096, )

Regards
Sujatha

On Wed, Oct 19, 2011 at 4:05 AM, Otis Gospodnetic <
otis_gospodne...@yahoo.com> wrote:

Maybe your Solr Document cache is big and that's consuming a big  
part of

that JVM heap?
If you want to be able to run with a smaller heap, consider making  
your

caches smaller.

Otis

Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch
Lucene ecosystem search :: http://search-lucene.com/




From: Sujatha Arun 
To: solr-user@lucene.apache.org
Sent: Tuesday, October 18, 2011 12:53 AM
Subject: Re: OS Cache - Solr

Hello Jan,

Thanks for your response and  clarification.

We are monitoring the JVM cache utilization and we are currently  
using

about
18 GB of the 20 GB assigned to JVM. Out total index size being  
abt 14GB


Regards
Sujatha

On Tue, Oct 18, 2011 at 1:19 AM, Jan Høydahl t.com>

wrote:



Hi Sujatha,

Are you sure you need 20Gb for Tomcat? Have you profiled using

JConsole

or

similar? Try with 15Gb and see how it goes. The reason why this is
beneficial is that you WANT your OS to have available memory for  
disk
caching. If you have 17Gb free after starting Solr, your OS will  
be

able

to

cache all index files in memory and you get very high search

performance.
With your current settings, there is only 12Gb free for both  
caching

the
index and for your MySql activities.  Chances are that when you  
backup

MySql, the cached part of your Solr index gets flushed from disk

caches

and

need to be re-cached later.

How to interpret memory stats vary between OSes, and seing 163Mb  
free

may

simply mean that your OS has used most RAM for various caches and

paging,
but will flush it once an application asks for more memory. Have  
you

seen

http://wiki.apache.org/solr/SolrPerformanceFactors ?

You should also slim down your index maximally by setting  
stored=false

and

indexed=false wherever possible. I would also upgrade to a more

current

Solr

version.

--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
Solr Training - www.solrtraining.com

On 17. okt. 2011, at 19:51, Sujatha Arun wrote:


Hello

I am trying to understand the  OS cache utilization of Solr .Our

server

has
several solr instances on a server .The total combined Index  
size of

all
instances is abt 14 Gb and the size of the maximum single Index  
is

abt

2.5

GB .

Our Server has Quad processor with 32 GB RAM .Out of which 20  
GB has

been
assigned to  JVM. We are running solr1.3  on tomcat 5.5 and  
Java 1.6


Our current Statistics indicate that  solr uses 18-19 GB of 20 GB

RAM
assigned to JVM .However the  Free physical seems to remain  
constant

as

below.
Free physical memory = 163 Mb
Total physical memory = 32,232 Mb,

The server also serves as a backup server for Mysql where the

application

DB
is backed up and restored .During this activity we see that lot  
of

queries

that nearly take even 10+ minutes to execute .But other wise
maximum query time is less than  1-2 secs

The physical memory that is free seems to be constant . Why is  
this

constant
and how this will be used between the  Mysql backup and solr  
while
backup activity is  happening How much free physical memory  
should

be

available to OS given out stats.?

Any pointers would be helpful.

Regards
Sujatha