I'll try it out. Thanks
Abhi On Fri, Mar 23, 2018, 6:22 PM Rick Leir <rl...@leirtech.com> wrote: > Abhi > Check your lib directives. > > https://lucene.apache.org/solr/guide/6_6/lib-directives-in-solrconfig.html#lib-directives-in-solrconfig > > I suspect your jars are not in a lib dir mentioned in solrconfig.xml > Cheers -- Rick > > On March 23, 2018 11:12:17 AM EDT, Abhi Basu <9000r...@gmail.com> wrote: > >MS Azure does not support Solr 4.9 on HDI, so I am posting here. I > >would > >like to write index collection data to HDFS (hosted on ADL). > > > >Note: I am able to get to ADL from hadoop fs command like, so hadoop is > >configured correctly to get to ADL: > >hadoop fs -ls adl:// > > > >This is what I have done so far: > >1. Copied all required jars to sol ext lib folder: > >sudo cp -f /usr/hdp/current/hadoop-client/*.jar > >/usr/hdp/current/solr/example/lib/ext > >sudo cp -f /usr/hdp/current/hadoop-client/lib/*.jar > >/usr/hdp/current/solr/example/lib/ext > >sudo cp -f /usr/hdp/current/hadoop-hdfs-client/*.jar > >/usr/hdp/current/solr/example/lib/ext > >sudo cp -f /usr/hdp/current/hadoop-hdfs-client/lib/*.jar > >/usr/hdp/current/solr/example/lib/ext > >sudo cp -f > >/usr/hdp/current/storm-client/contrib/storm-hbase/storm-hbase*.jar > >/usr/hdp/current/solr/example/lib/ext > >sudo cp -f /usr/hdp/current/phoenix-client/lib/phoenix*.jar > >/usr/hdp/current/solr/example/lib/ext > >sudo cp -f /usr/hdp/current/hbase-client/lib/hbase*.jar > >/usr/hdp/current/solr/example/lib/ext > > > >This includes the Azure active data lake jars also. > > > >2. Edited my solr-config.xml file for my collection: > > > ><dataDir>${solr.core.name}/data/</dataDir> > > > ><directoryFactory name="DirectoryFactory" > >class="solr.HdfsDirectoryFactory"> > ><str > >name="solr.hdfs.home">adl:// > esodevdleus2.azuredatalakestore.net/clusters/esohadoopdeveus2/solr/</str> > > <str name="solr.hdfs.confdir">/usr/hdp/2.6.2.25-1/hadoop/conf</str> > ><str > > >name="solr.hdfs.blockcache.global">${solr.hdfs.blockcache.global:true}</str> > > <bool name="solr.hdfs.blockcache.enabled">true</bool> > > <int name="solr.hdfs.blockcache.slab.count">1</int> > > <bool name="solr.hdfs.blockcache.direct.memory.allocation">true</bool> > > <int name="solr.hdfs.blockcache.blocksperbank">16384</int> > > <bool name="solr.hdfs.blockcache.read.enabled">true</bool> > > <bool name="solr.hdfs.nrtcachingdirectory.enable">true</bool> > > <int name="solr.hdfs.nrtcachingdirectory.maxmergesizemb">16</int> > ></directoryFactory> > > > > > >When this collection is deployed to solr, I see this error message: > > > ><response> > ><lst name="responseHeader"> > ><int name="status">0</int> > ><int name="QTime">2189</int></lst> > ><lst name="failure"> > > ><str>org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Error > >CREATEing SolrCore 'ems-collection_shard2_replica2': > >Unable to create core: ems-collection_shard2_replica2 Caused by: Class > >org.apache.hadoop.fs.adl.HdiAdlFileSystem not > > >found</str><str>org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Error > >CREATEing SolrCore 'ems-collection_shard2_replica1': Unable to create > >core: ems-collection_shard2_replica1 Caused by: Class > >org.apache.hadoop.fs.adl.HdiAdlFileSystem not > > >found</str><str>org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Error > >CREATEing SolrCore 'ems-collection_shard1_replica1': Unable to create > >core: ems-collection_shard1_replica1 Caused by: Class > >org.apache.hadoop.fs.adl.HdiAdlFileSystem not > > >found</str><str>org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Error > >CREATEing SolrCore 'ems-collection_shard1_replica2': Unable to create > >core: ems-collection_shard1_replica2 Caused by: Class > >org.apache.hadoop.fs.adl.HdiAdlFileSystem not found</str> > ></lst> > ></response> > > > > > >Has anyone done this and can help me out? > > > >Thanks, > > > >Abhi > > > > > >-- > >Abhi Basu > > -- > Sorry for being brief. Alternate email is rickleir at yahoo dot com