Thanks Kevin, this worked for me. On Mon, Oct 3, 2016 at 11:48 AM, Kevin Risden <compuwizard...@gmail.com> wrote:
> You need to have the hadoop pieces on the classpath. Like core-site.xml and > hdfs-site.xml. There is an hdfs classpath command that would help but it > may have too many pieces. You may just need core-site and hdfs-site so you > don't get conflicting jars. > > Something like this may work for you: > > java -cp > "$(hdfs classpath):./server/solr-webapp/webapp/WEB-INF/lib/*:./server/lib/ > ext/*:/hadoop/hadoop-client/lib/servlet-api-2.5.jar" > -ea:org.apache.lucene... org.apache.solr.index.hdfs.CheckHdfsIndex > hdfs://<NODE_NAME>:8020/apps/solr/data/ExampleCollection/ > core_node1/data/index > > Kevin Risden > > On Mon, Oct 3, 2016 at 1:38 PM, Rishabh Patel < > rishabh.mahendra.pa...@gmail.com> wrote: > > > Hello, > > > > My SolrCloud 5.5 installation has Kerberos enabled. The CheckHdfsIndex > test > > fails to run. However, without Kerberos, I am able to run the test with > no > > issues. > > > > I ran the following command: > > > > java -cp > > "./server/solr-webapp/webapp/WEB-INF/lib/*:./server/lib/ > > ext/*:/hadoop/hadoop-client/lib/servlet-api-2.5.jar" > > -ea:org.apache.lucene... org.apache.solr.index.hdfs.CheckHdfsIndex > > hdfs://<NODE_NAME>:8020/apps/solr/data/ExampleCollection/ > > core_node1/data/index > > > > The error is: > > > > ERROR: could not open hdfs directory " > > hdfs://<NODE_NAME>:8020/apps/solr/data/ExampleCollection/ > > core_node1/data/index > > "; > > exiting org.apache.hadoop.ipc.RemoteException(org.apache. > hadoop.security. > > AccessControlException): > > SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] > > Does this error message imply that the test cannot run with Kerberos > > enabled? > > > > For reference, I followed this blog > > http://yonik.com/solr-5-5/ > > > > -- > > Regards, > > *Rishabh Patel* > > > -- Sincerely, *Rishabh Patel*