Null pointer exception in spell checker at addchecker method

2013-12-06 Thread sweety
Im trying to use spell check component. My *schema* is:(i have included only fields necessary for spell check not the entire schema) My *solrconfig* is: text direct contents solr.DirectSolrSpellChecker internal 0.8 1 1 5 3 0.01 wordbreak solr.Wor

Re: Null pointer exception in spell checker at addchecker method

2013-12-09 Thread sweety
yes, it worked. And i got the reason for the error. Thanks a lot. -- View this message in context: http://lucene.472066.n3.nabble.com/Null-pointer-exception-in-spell-checker-at-addchecker-method-tp4105489p4105636.html Sent from the Solr - User mailing list archive at Nabble.com.

Java heap space:out of memory

2013-12-10 Thread sweety
I just indexed 10 doc of total 15mb.For some queries it works fine but, for some queries i get this error: java.lang.OutOfMemoryError: Java heap space java.lang.RuntimeException: java.lang.OutOfMemoryError: Java heap space at org.apache.solr.servlet.SolrDispatchFilter.sendError(SolrDispatchFilt

Re: Java heap space:out of memory

2013-12-10 Thread sweety
4gb ram. I m running on Windows 7,with Tomcat as webserver. -- View this message in context: http://lucene.472066.n3.nabble.com/Java-heap-space-out-of-memory-tp4105903p4105929.html Sent from the Solr - User mailing list archive at Nabble.com.

Re: Java heap space:out of memory

2013-12-10 Thread sweety
sorry but i dont know how to check that? -- View this message in context: http://lucene.472066.n3.nabble.com/Java-heap-space-out-of-memory-tp4105903p4105947.html Sent from the Solr - User mailing list archive at Nabble.com.

Re: Java heap space:out of memory

2013-12-10 Thread sweety
okay thanks, here it is: max heap size : 63.56MB(it is howing 37.2% usage though) How to increase that size?? -- View this message in context: http://lucene.472066.n3.nabble.com/Java-heap-space-out-of-memory-tp4105903p4105952.html Sent from the Solr - User mailing list archive at Nabble.com.

Re: Java heap space:out of memory

2013-12-10 Thread sweety
I have set : JAVA_OPTS as value: -Xms1024M-Xmx1024M But the dashboard still shows 64M,but now the usage is only 18% How could that be? yesterday it was 87%. -- View this message in context: http://lucene.472066.n3.nabble.com/Java-heap-space-out-of-memory-tp4105903p4106069.html Sent from the S

Re: Java heap space:out of memory

2013-12-10 Thread sweety
yes,i did put the space,as in the image -- View this message in context: http://lucene.472066.n3.nabble.com/Java-heap-space-out-of-memory-tp4105903p4106077.html Sent from the Solr - User mailing list archive at Nabble.com.

Re: Java heap space:out of memory

2013-12-10 Thread sweety
You were right the changes made in JAVA_OPTs didn't show increase in the heap size, I made changes in the UI of tomcat Initial pool memory : 512 MB Maximum pool memory : 1024 MB Now the heap size has increased. Thanks you all for your suggestions,it really saved my time. -- View this message i

indexing .docx using solrj

2013-12-21 Thread sweety
i am trying to index .docx file using solrj, i referred this link: http://wiki.apache.org/solr/ContentStreamUpdateRequestExample My code is : import java.io.File; import java.io.IOException; import org.apache.solr.client.solrj.SolrServer; import org.apache.solr.client.solrj.SolrServerException;

Re: indexing .docx using solrj

2013-12-21 Thread sweety
I have added that jar,in the build path. but the same error,i get. Why is eclipse not recognising that jar?? Logs also show this, Caused by: java.lang.NoClassDefFoundError: org/apache/xml/serialize/BaseMarkupSerializer at org.apache.solr.handler.extraction.ExtractingRequestHandler.newLoad

Re: indexing .docx using solrj

2013-12-21 Thread sweety
Jar is already there in the lib folder of solr home. -- View this message in context: http://lucene.472066.n3.nabble.com/indexing-docx-using-solrj-tp4107737p4107748.html Sent from the Solr - User mailing list archive at Nabble.com.

Re: program termination in solrj

2013-12-21 Thread sweety
Before and after running client,stats remain same only, class:org.apache.solr.update.DirectUpdateHandler2 version:1.0 description:Update handler that efficiently directly updates the on-disk main lucene index src:$URL: https:/​/​svn.apache.org/​repos/​asf/​lucene/​dev/​branches/​branch_4x/​solr/​c

Re: indexing .docx using solrj

2013-12-21 Thread sweety
solr: 4.2 tomcat: 7.0 jdk1.7.0.45 i have created solr home in c:\solr as in java options: -Dsolr.solr.home=C:\solr c:solr/lib contains: tika jars, actually i pasted all the jars from the solr 4.2 dist,contrib folders in c:solr/lib tomcat/lib contains: all the jars when installed. -- View t

Re: program termination in solrj

2013-12-21 Thread sweety
also my default search handler has no dismax. explicit 20 * contents 2.1 -- View this message in context: http://lucene.472066.n3.nabble.com/program-termination-in-solrj-tp4107706p4107753.html Sent from the Solr - User mailing li

Re: program termination in solrj

2013-12-21 Thread sweety
okay, i did a mistake, i did not refresh the stats,so the stats after running java program: commits:1 autocommits:0 soft autocommits:0 optimizes:0 rollbacks:0 expungeDeletes:0 docsPending:0 adds:0 deletesById:0 deletesByQuery:0 errors:0 cumulative_adds:1 cumulative_deletesById:0 cumulative_deletes

Re: indexing .docx using solrj

2013-12-21 Thread sweety
It is working now,i just restarted computer. But i dont still get the reason for the error. Thank you though,for your efforts. -- View this message in context: http://lucene.472066.n3.nabble.com/indexing-docx-using-solrj-tp4107737p4107755.html Sent from the Solr - User mailing list archive at N

Re: indexing .docx using solrj

2013-12-21 Thread sweety
yes,i copied all jars from contrib/extraction to solr/lib. It is not getting the poi jar now, as mentioned in above post of mine, new error it shows now. -- View this message in context: http://lucene.472066.n3.nabble.com/indexing-docx-using-solrj-tp4107737p4107758.html Sent from the Solr - Use

to index byte array

2014-01-01 Thread sweety
I am converting .doc and .docx files to byte array in c#, now I need to index this byte array of doc files. Is it possible in solr to index byte array of files?? -- View this message in context: http://lucene.472066.n3.nabble.com/to-index-byte-array-tp4108999.html Sent from the Solr - User mail

Re: to index byte array

2014-01-01 Thread sweety
For indexing .docx files using tika, requires file system path, but i dont want to give the path. I read in DIH faq's that by using transformer the output can be converted from byte to string. -- View this message in context: http://lucene.472066.n3.nabble.com/to-index-byte-array-tp4108999p41

Re: to index byte array

2014-01-01 Thread sweety
For indexing .docx files using tika, requires file system path, but i dont want to give the path. I read in DIH faq's that by using transformer the output can be converted from byte to string. -- View this message in context: http://lucene.472066.n3.nabble.com/to-index-byte-array-tp4108999p41

Re: to index byte array

2014-01-01 Thread sweety
If you consider a client-server architecture, the documents will sent in binary format to server, now for solr this binary format will be the source to index, so i need to index byte array. Also if store this byte-array into db and then index in solr, then will the contents of document be searchabl

using extract handler: data not extracted

2014-01-11 Thread sweety
I need to index rich text documents, this is* solrconfig.xml for extract handler*: true ignored_ true My *schema.xml* is: But after *indexing using this curl*: curl "http://localhost:8080/solr/document/update/extract?literal.id=12&commit=true"; -F"myfile=Coding.pdf" when queried as

Re: using extract handler: data not extracted

2014-01-11 Thread sweety
Sorry, that my question was not clear. Initially when indexed pdf files it showed the data within this pdf in the contents field.as follows:(this is output for initially indexed documents) Cloud ctured As tale in size as well as complexity. We need a cloud based system that will solve this problem

Re: using extract handler: data not extracted

2014-01-11 Thread sweety
I set the level of extract handler to finest, now the logs are : INFO: [document] webapp=/solr path=/update/extract params={commit=true&literal.id=12&debug=true} {add=[12 (1456944038966984704)],commit=} 0 2631 Jan 11, 2014 7:51:57 PM org.apache.solr.servlet.SolrDispatchFilter handleAdminRequest INF

Re: using extract handler: data not extracted

2014-01-11 Thread sweety
how set finest for tika package?? -- View this message in context: http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110888.html Sent from the Solr - User mailing list archive at Nabble.com.

Re: using extract handler: data not extracted

2014-01-11 Thread sweety
the logging screen does not show tika package, also i searched on net, it requires log4j and slf4j jars, is it true?? Do i need to do the configurations for package level log? -- View this message in context: http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850

Re: using extract handler: data not extracted

2014-01-11 Thread sweety
this is the output i get when indexed through* solrj*, i followed the link you suggested. i tried indexing .doc file. 400 17 org.apache.solr.search.SyntaxError: Cannot parse 'id:C:\solr\document\src\new_index_doc\document_1.doc': Encountered " ":" ": "" at line 1, column 4. Was expecting one o

Re: using extract handler: data not extracted

2014-01-12 Thread sweety
ya right all 3 points are right. Let me solve the 1 first, there is some errror in tika level indexing, for that i need to debug at tika level right?? but how to do that?? Solr admin does not show package wise logging. -- View this message in context: http://lucene.472066.n3.nabble.com/using-e

Re: using extract handler: data not extracted

2014-01-12 Thread sweety
through command line(>java -jar tika-app-1.4.jar -v C:Cloud.docx) apache tika is able to parse .docx files, so can i use this tika-app-1.4.jar in solr?? how to do that?? -- View this message in context: http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110

Re: using extract handler: data not extracted

2014-01-12 Thread sweety
Sorry for the mistake. im using solr 4.2, it has tika-1.3. So now, java -jar tika-app-1.3.jar -v C:Coding.pdf , parses pdf document without error or msg. Also, java -jar tika-app-1.4.jar* -t *C:Cloud.docx, shows the entire document. Which means there is no problem in tika right?? -- View this

Re: using extract handler: data not extracted

2014-01-12 Thread sweety
Sorry for the mistake. im using solr 4.2, it has tika-1.3. So now, java -jar tika-app-1.3.jar -v C:\Coding.pdf , parses pdf document without error or msg. Also, java -jar tika-app-1.3.jar -t C:\Coding.pdf, shows the entire document. Which means there is no problem in tika right?? -- View t

Re: using extract handler: data not extracted

2014-01-12 Thread sweety
Sorry for the mistake. im using solr 4.2, it has tika-1.3. So now, java -jar tika-app-1.3.jar -v C:\Coding.pdf , parses pdf document without error or msg. Also, java -jar tika-app-1.3.jar -t C:\Coding.pdf, shows the entire document. Which means there is no problem in tika right?? -- View t

Re: using extract handler: data not extracted

2014-01-12 Thread sweety
I am working on Windows 7 -- View this message in context: http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110993.html Sent from the Solr - User mailing list archive at Nabble.com.

Solrcloud: no registered leader found and new searcher error

2014-02-17 Thread sweety
I have configured solrcloud as follows, Solr.xml: I have added all the required config for solrcloud, referred this : http://wiki.apache.org/solr/SolrCloud#Required_Config I am adding data to core:document. Now w

Re: Solrcloud: no registered leader found and new searcher error

2014-02-17 Thread sweety
How do i get them running? -- View this message in context: http://lucene.472066.n3.nabble.com/Solrcloud-no-registered-leader-found-and-new-searcher-error-tp4117724p4117830.html Sent from the Solr - User mailing list archive at Nabble.com.

to reduce indexing time

2014-03-05 Thread sweety
Before indexing , this was the memory layout, System Memory : 63.2% ,2.21 gb JVM Memory : 8.3% , 81.60mb of 981.38mb I have indexed 700 documents of total size 12MB. Following are the results i get : Qtime: 8122, System time : 00:00:12.7318648 System Memory : 65.4% ,2.29 gb JVM Memory : 15.3% ,

Re: to reduce indexing time

2014-03-05 Thread sweety
Now i have batch indexed, with batch of 250 documents.These were the results. After 7,000 documents, Qtime: 46894, System time : 00:00:55.9384892 JVM memory : 249.02mb, 24.8% This shows quite a reduction in timing. After 70,000 documents, Qtime: 480435, System time : 00:09:29.5206727 System memor

Re: to reduce indexing time

2014-03-05 Thread sweety
I will surely read about JVM Garbage collection. Thanks a lot, all of you. But, is the time required for my indexing good enough? I dont know about the ideal timings. I think that my indexing is taking more time. -- View this message in context: http://lucene.472066.n3.nabble.com/to-reduce-ind

no such field error:smaller big block size details while indexing doc files

2013-10-07 Thread sweety
Im trying to index .doc,.docx,pdf files, im using this url: curl "http://localhost:8080/solr/document/update/extract?literal.id=12&commit=true"; -F"myfile=@complex.doc" This is the error I get: Oct 07, 2013 5:02:18 PM org.apache.solr.common.SolrException log SEVERE: null:java.lang.RuntimeException

Re: no such field error:smaller big block size details while indexing doc files

2013-10-08 Thread sweety
This my new schema.xml:         id I still get the same error. From: Erick Erickson [via Lucene] To: sweety Sent: Tuesday, October 8, 2013 7:16 AM Subject: Re: no such field error:smaller big block size details while indexing doc files

Re: no such field error:smaller big block size details while indexing doc files

2013-10-09 Thread sweety
You. On Wednesday, October 9, 2013 12:54 PM, sweety shinde wrote: I will try using solrJ. Now I tried indexing .docx files and I get some different error,logs are: SEVERE: null:java.lang.RuntimeException: java.lang.VerifyError: (class: org/apache/poi/extractor/ExtractorFactory, method

Re: no such field error:smaller big block size details while indexing doc files

2013-10-09 Thread sweety
s in there, it's a relatively simple SolrJ program, here's a sample: http://searchhub.org/2012/02/14/indexing-with-solrj/ Best, Erick On Tue, Oct 8, 2013 at 4:15 PM, sweety <[hidden email]> wrote: > This my new schema.xml: > > > multiValued="false&