Re: Problem with linux
Well, an awful lot of people use it successfully, so the first place I'd look is my classpath and be sure you don't have unexpected jars in there. Perhaps ones from a different version of Lucene, perhaps ones pointing to "interesting" places in your JVM. Perhaps. Second, have you tried it on any other linux boxes? Is it possible that the one you're using is hosed? Without more information though, it's hard to say anything. You haven't included the version of Lucene you're using (which is suggested by your code snippet, perhaps the Lucene user's list would be a better forum?). Nor the version of your JVM. Nor your linux version. Nor Best Erick On Fri, Jul 9, 2010 at 7:49 PM, sarfaraz masood < sarfarazmasood2...@yahoo.com> wrote: > I have problems when i execute my prog on linux having this following piece > of code. > { > > Document d; > Analyzer analyzer = new PorterStemAnalyzer(); > System.out.println("1"); > > > Directory index = FSDirectory.open(new File("index1")); > System.out.println("2"); > > IndexWriter w = new IndexWriter(index, analyzer, true, > IndexWriter.MaxFieldLength.UNLIMITED ) ; // MY PROG HANGS UP HERE > System.out.println("3"); > . > . > . > } > > > Strangely this exact prog runs well on windows. It simply hangs > up(doesnt halt) while creating the IndexWriter object in linux. The account > via > which im logged in has sufficient rights for the concerned folder. > > > -Sarfaraz > > > >
Database connections during data import
Hi, We are indexing a large amount of data into Solr from a MS-SQL database (don't ask!). There are approximately 4 million records, and a total database size of the order of 20GB. There is also a need for incremental updates, but these are only a few % of the total. After some trials-and-error, things are working great. Indexing is a little slow as per our original expectations, but this is probably to be expected, given that: * There are a fair number of queries per record indexed into Solr * Only one database server is in use at the moment, and this could well be a bottle-neck (please see below). * The index has many fields, and we are also storing everything in this phase, so that we can recover data directly from the Solr index. * Transformers are used pretty liberally * Finally, we are no longer so concerned about the indexing speed of a single Solr instance, as thanks to the possibility of merging indexes, we can simply throw more hardware at the problem. (Incidentally, a big thank-you to everyone who has contributed to Solr. The above work was way easier than we had feared.) As a complete indexing takes about 20h, sometimes the process gets interrupted due to a loss of the database connection. I can tell that that a loss of connection is the problem from the Solr Tomcat logs, but it is difficult to tell whether it is the database dropping connections (the database server is at 60-70% CPU utilisation, but close to being maxed out at 4GB, and I am told that MS-SQL/the OS cannot handle more RAM), or a network glitch. What happens is that the logs report a reconnection, but the number of processed records reported by the DataImportHandler at /solr/dataimport?command=full-import stops incrementing, even several hours after the reconnection. Is there any way to recover from a reconnection, and continue DataImportHandler indexing at the point where the process left off? Regards, Gora P.S. Incidentally, would there be any interest in a GDataRequestHandler for Solr queries, and a GDataResponseWriter? We wrote one in the interests of trying to adhere to a de-facto standard, and can consider contributing these, after further testing, and cleanup.
Re: Using symlinks to alias cores
Chris Hostetter-3 wrote: > > a cleaner way to deal with this would be do use something like > RewriteRule -- either in your appserver (if it supports a feature like > that) or in a proxy sitting in front of Solr. > I think we'll go with this -- seems like the most bulletproof way. Cheers, Andrew. -- View this message in context: http://lucene.472066.n3.nabble.com/Using-symlinks-to-alias-cores-tp942567p956394.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Problem with linux
well thanx a lot Mr Uwe.. This problem solved by making an object of SimpleFSDirectory & passing it to Directory. Directory index = FSDirectory.open(new File("index1"),sfs); Well Erick i m using Lucene2.9.2. --- On Sat, 10/7/10, Erick Erickson wrote: From: Erick Erickson Subject: Re: Problem with linux To: solr-user@lucene.apache.org Date: Saturday, 10 July, 2010, 6:32 PM Well, an awful lot of people use it successfully, so the first place I'd look is my classpath and be sure you don't have unexpected jars in there. Perhaps ones from a different version of Lucene, perhaps ones pointing to "interesting" places in your JVM. Perhaps. Second, have you tried it on any other linux boxes? Is it possible that the one you're using is hosed? Without more information though, it's hard to say anything. You haven't included the version of Lucene you're using (which is suggested by your code snippet, perhaps the Lucene user's list would be a better forum?). Nor the version of your JVM. Nor your linux version. Nor Best Erick On Fri, Jul 9, 2010 at 7:49 PM, sarfaraz masood < sarfarazmasood2...@yahoo.com> wrote: > I have problems when i execute my prog on linux having this following piece > of code. > { > > Document d; > Analyzer analyzer = new PorterStemAnalyzer(); > System.out.println("1"); > > > Directory index = FSDirectory.open(new File("index1")); > System.out.println("2"); > > IndexWriter w = new IndexWriter(index, analyzer, true, > IndexWriter.MaxFieldLength.UNLIMITED ) ; // MY PROG HANGS UP HERE > System.out.println("3"); > . > . > . > } > > > Strangely this exact prog runs well on windows. It simply hangs > up(doesnt halt) while creating the IndexWriter object in linux. The account > via > which im logged in has sufficient rights for the concerned folder. > > > -Sarfaraz > > > >