Thanks Kevin, this worked for me.

On Mon, Oct 3, 2016 at 11:48 AM, Kevin Risden <compuwizard...@gmail.com>
wrote:

> You need to have the hadoop pieces on the classpath. Like core-site.xml and
> hdfs-site.xml. There is an hdfs classpath command that would help but it
> may have too many pieces. You may just need core-site and hdfs-site so you
> don't get conflicting jars.
>
> Something like this may work for you:
>
> java -cp
> "$(hdfs classpath):./server/solr-webapp/webapp/WEB-INF/lib/*:./server/lib/
> ext/*:/hadoop/hadoop-client/lib/servlet-api-2.5.jar"
> -ea:org.apache.lucene... org.apache.solr.index.hdfs.CheckHdfsIndex
> hdfs://<NODE_NAME>:8020/apps/solr/data/ExampleCollection/
> core_node1/data/index
>
> Kevin Risden
>
> On Mon, Oct 3, 2016 at 1:38 PM, Rishabh Patel <
> rishabh.mahendra.pa...@gmail.com> wrote:
>
> > Hello,
> >
> > My SolrCloud 5.5 installation has Kerberos enabled. The CheckHdfsIndex
> test
> > fails to run. However, without Kerberos, I am able to run the test with
> no
> > issues.
> >
> > I ran the following command:
> >
> > java -cp
> > "./server/solr-webapp/webapp/WEB-INF/lib/*:./server/lib/
> > ext/*:/hadoop/hadoop-client/lib/servlet-api-2.5.jar"
> > -ea:org.apache.lucene... org.apache.solr.index.hdfs.CheckHdfsIndex
> > hdfs://<NODE_NAME>:8020/apps/solr/data/ExampleCollection/
> > core_node1/data/index
> >
> > The error is:
> >
> > ERROR: could not open hdfs directory "
> > hdfs://<NODE_NAME>:8020/apps/solr/data/ExampleCollection/
> > core_node1/data/index
> > ";
> > exiting org.apache.hadoop.ipc.RemoteException(org.apache.
> hadoop.security.
> > AccessControlException):
> > SIMPLE authentication is not enabled.  Available:[TOKEN, KERBEROS]
> > Does this error message imply that the test cannot run with Kerberos
> > enabled?
> >
> > For reference, I followed this blog
> > http://yonik.com/solr-5-5/
> >
> > --
> > Regards,
> > *Rishabh Patel*
> >
>



-- 
Sincerely,
*Rishabh Patel*

Reply via email to