I am trying to setup Solr with HDFS following this wiki https://cwiki.apache.org/confluence/display/solr/Running+Solr+on+HDFS
My Setup: *********** VMWare: Cloudera Quick Start VM 4.4.0-1 default setup (only hdfs1, hive1,hue1,mapreduce1 and zookeeper1 is running) http://www.cloudera.com/content/support/en/downloads/download-components/download-products.html?productID=F6mO278Rvo SolrCloud: Mac OS 10.7.5 -> -Running Solr 4.6 with maven jetty plugin in eclipse outside from HDFS (Cloudera VM) , so it is accessing HDFS as remote service External zookeeper 3 nodes Java 1.6, Jett Container 8.1 Collection with 1 shard and 1 replica ************ But I am getting below error "Problem creating directory:" I have created this directory manually in hdfs. Do I need to setup some special user permission in Solr?. or do I need to always run solr instance in HDFS (Data Node)? [cloudera@localhost ~]$ sudo -u hdfs hadoop fs -mkdir /solr-hdfs Directory permisson in HDFS: solr-hdfs rwxr-xr-x hdfs supergroup Startup Log: 2014-01-01 20:21:57.433:INFO:oejs.Server:jetty-8.1.7.v20120910 2014-01-01 20:21:59.710:INFO:omjp.MavenWebInfConfiguration:Adding overlay: file:/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs/target/tmp/solr-4_6_0_war/ 2014-01-01 20:22:02.249:INFO:oejpw.PlusConfiguration:No Transaction manager found - if your webapp requires one, please configure one. 2014-01-01 20:22:03.368:INFO:oejsh.ContextHandler:started o.m.j.p.JettyWebAppContext{/solr,[file:/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs/src/main/webapp/, file:/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs/target/tmp/solr-4_6_0_war/]},file:/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs/src/main/webapp/ 2014-01-01 20:22:03.369:INFO:oejsh.ContextHandler:started o.m.j.p.JettyWebAppContext{/solr,[file:/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs/src/main/webapp/, file:/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs/target/tmp/solr-4_6_0_war/]},file:/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs/src/main/webapp/ 2014-01-01 20:22:03.369:INFO:oejsh.ContextHandler:started o.m.j.p.JettyWebAppContext{/solr,[file:/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs/src/main/webapp/, file:/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs/target/tmp/solr-4_6_0_war/]},file:/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs/src/main/webapp/ 0 [main] INFO org.apache.solr.servlet.SolrDispatchFilter – SolrDispatchFilter.init() 29 [main] INFO org.apache.solr.core.SolrResourceLoader – No /solr/home in JNDI 30 [main] INFO org.apache.solr.core.SolrResourceLoader – using system property solr.solr.home: /Users/gpatwa/opensource/solr-hdfs-home 32 [main] INFO org.apache.solr.core.SolrResourceLoader – new SolrResourceLoader for directory: '/Users/gpatwa/opensource/solr-hdfs-home/' 220 [main] INFO org.apache.solr.core.ConfigSolr – Loading container configuration from /Users/gpatwa/opensource/solr-hdfs-home/solr.xml 348 [main] INFO org.apache.solr.core.ConfigSolrXml – Config-defined core root directory: 358 [main] INFO org.apache.solr.core.CoreContainer – New CoreContainer 445620464 359 [main] INFO org.apache.solr.core.CoreContainer – Loading cores into CoreContainer [instanceDir=/Users/gpatwa/opensource/solr-hdfs-home/] 374 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting socketTimeout to: 120000 375 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting urlScheme to: http:// 375 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting connTimeout to: 15000 375 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting maxConnectionsPerHost to: 20 375 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting corePoolSize to: 0 376 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting maximumPoolSize to: 2147483647 376 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting maxThreadIdleTime to: 5 376 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting sizeOfQueue to: -1 378 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting fairnessPolicy to: false 645 [main] INFO org.apache.solr.logging.LogWatcher – SLF4J impl is org.slf4j.impl.Log4jLoggerFactory 646 [main] INFO org.apache.solr.logging.LogWatcher – Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)] 647 [main] INFO org.apache.solr.core.ZkContainer – Zookeeper client=localhost:2181/search/catalog 653 [main] INFO org.apache.solr.cloud.ZkController – zkHost includes chroot 762 [main] INFO org.apache.solr.common.cloud.ConnectionManager – Waiting for client to connect to ZooKeeper 5781 [main-EventThread] INFO org.apache.solr.common.cloud.ConnectionManager – Watcher org.apache.solr.common.cloud.ConnectionManager@25630eb6name:ZooKeeperConnection Watcher:localhost:2181 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None 5783 [main] INFO org.apache.solr.common.cloud.ConnectionManager – Client is connected to ZooKeeper 5792 [main] INFO org.apache.solr.common.cloud.ConnectionManager – Waiting for client to connect to ZooKeeper 5827 [main-EventThread] INFO org.apache.solr.common.cloud.ConnectionManager – Watcher org.apache.solr.common.cloud.ConnectionManager@52f5bad0name:ZooKeeperConnection Watcher:localhost:2181/search/catalog got event WatchedEvent state:SyncConnected type:None path:null path:null type:None 5827 [main] INFO org.apache.solr.common.cloud.ConnectionManager – Client is connected to ZooKeeper 5852 [main] INFO org.apache.solr.common.cloud.ZkStateReader – Updating cluster state from ZooKeeper... 6877 [main] INFO org.apache.solr.cloud.ZkController – Register node as live in ZooKeeper:/live_nodes/127.0.0.1:8983_solr 6880 [main] INFO org.apache.solr.common.cloud.SolrZkClient – makePath: /live_nodes/127.0.0.1:8983_solr 6885 [main-EventThread] INFO org.apache.solr.common.cloud.ZkStateReader – Updating live nodes... (1) 6892 [main] INFO org.apache.solr.common.cloud.SolrZkClient – makePath: /overseer_elect/leader 6896 [main] INFO org.apache.solr.cloud.Overseer – Overseer (id=90988393900081217-127.0.0.1:8983_solr-n_0000000017) starting 6913 [Overseer-90988393900081217-127.0.0.1:8983_solr-n_0000000017] INFO org.apache.solr.cloud.OverseerCollectionProcessor – Process current queue of collection creations 6917 [Thread-5] INFO org.apache.solr.cloud.Overseer – Starting to work on the main queue 6946 [main] INFO org.apache.solr.core.CoresLocator – Looking for core definitions underneath /Users/gpatwa/opensource/solr-hdfs-home 6979 [main] INFO org.apache.solr.core.CoresLocator – Found core event_shard1_replica1 in /Users/gpatwa/opensource/solr-hdfs-home/event_shard1_replica1/ 6980 [main] INFO org.apache.solr.core.CoresLocator – Found 1 core definitions 6981 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.cloud.ZkController – publishing core=event_shard1_replica1 state=down 6984 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.cloud.ZkController – waiting to find shard id in clusterstate for event_shard1_replica1 6984 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.CoreContainer – Creating SolrCore 'event_shard1_replica1' using instanceDir: /Users/gpatwa/opensource/solr-hdfs-home/event_shard1_replica1 6984 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.cloud.ZkController – Check for collection zkNode:event 6985 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.cloud.ZkController – Collection zkNode exists 6986 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.cloud.ZkController – Load collection config from:/collections/event 6987 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.SolrResourceLoader – new SolrResourceLoader for directory: '/Users/gpatwa/opensource/solr-hdfs-home/event_shard1_replica1/' 7036 [coreLoadExecutor-4-thread-1] WARN org.apache.solr.core.Config – You should not use LUCENE_CURRENT as luceneMatchVersion property: if you use this setting, and then Solr upgrades to a newer release of Lucene, sizable changes may happen. If precise back compatibility is important then you should instead explicitly specify an actual Lucene version. 7172 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.SolrConfig – Using Lucene MatchVersion: LUCENE_CURRENT 7283 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.Config – Loaded SolrConfig: solrconfig.xml 7292 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.schema.IndexSchema – Reading Solr Schema from schema.xml 7354 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.schema.IndexSchema – [event_shard1_replica1] Schema name=event-hdfs 7686 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.schema.IndexSchema – default search field in schema is searchKeywords_en_US 7688 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.schema.IndexSchema – query parser default operator is OR 7691 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.schema.IndexSchema – unique key field: id 7826 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.SolrCore – solr.HdfsDirectoryFactory 7836 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.HdfsDirectoryFactory – Solr Kerberos Authentication disabled 7836 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.SolrCore – [event_shard1_replica1] Opening new SolrCore at /Users/gpatwa/opensource/solr-hdfs-home/event_shard1_replica1/, dataDir=hdfs://10.249.132.29:8020/solr-hdfs/ 7838 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.JmxMonitoredMap – No JMX servers found, not exposing Solr information with JMX. 7845 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.HdfsDirectoryFactory – creating directory factory for path hdfs://10.249.132.29:8020/solr-hdfs 7857 [coreLoadExecutor-4-thread-1] INFO org.apache.hadoop.metrics.jvm.JvmMetrics – Initializing JVM Metrics with processName=blockcache, sessionId=1388636531350 2014-01-01 20:22:11.488 java[62306:10b03] Unable to load realm info from SCDynamicStore 8001 [coreLoadExecutor-4-thread-1] WARN org.apache.hadoop.util.NativeCodeLoader – Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 8426 [Thread-5] INFO org.apache.solr.common.cloud.ZkStateReader – Updating cloud state from ZooKeeper... 8428 [Thread-5] INFO org.apache.solr.cloud.Overseer – Update state numShards=1 message={ "operation":"state", "state":"down", "base_url":"http://127.0.0.1:8983/solr", "core":"event_shard1_replica1", "roles":null, "node_name":"127.0.0.1:8983_solr", "shard":"shard1", "shard_range":null, "shard_state":"active", "shard_parent":null, "collection":"event", "numShards":"1", "core_node_name":"core_node1"} 8450 [main-EventThread] INFO org.apache.solr.common.cloud.ZkStateReader – A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1) 8513 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.SolrCore – [event_shard1_replica1] CLOSING SolrCore org.apache.solr.core.SolrCore@5035135a 8513 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.update.SolrCoreState – Closing SolrCoreState 8513 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.update.DefaultSolrCoreState – SolrCoreState ref count has reached 0 - closing IndexWriter 8514 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.SolrCore – [event_shard1_replica1] Closing main searcher on request. 8514 [coreLoadExecutor-4-thread-1] INFO org.apache.solr.core.CachingDirectoryFactory – Closing HdfsDirectoryFactory - 0 directories currently being tracked 8516 [coreLoadExecutor-4-thread-1] ERROR org.apache.solr.core.CoreContainer – Unable to create core: event_shard1_replica1 org.apache.solr.common.SolrException: Problem creating directory: hdfs:// 10.249.132.29:8020/solr-hdfs at org.apache.solr.core.SolrCore.<init>(SolrCore.java:834) at org.apache.solr.core.SolrCore.<init>(SolrCore.java:625) at org.apache.solr.core.ZkContainer.createFromZk(ZkContainer.java:256) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:590) at org.apache.solr.core.CoreContainer$1.call(CoreContainer.java:271) at org.apache.solr.core.CoreContainer$1.call(CoreContainer.java:263) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask( ThreadPoolExecutor.java:895) at java.util.concurrent.ThreadPoolExecutor$Worker.run( ThreadPoolExecutor.java:918) at java.lang.Thread.run(Thread.java:680) Caused by: java.lang.RuntimeException: Problem creating directory: hdfs:// 10.249.132.29:8020/solr-hdfs at org.apache.solr.store.hdfs.HdfsDirectory.<init>(HdfsDirectory.java:68) at org.apache.solr.core.HdfsDirectoryFactory.create( HdfsDirectoryFactory.java:154) at org.apache.solr.core.CachingDirectoryFactory.get( CachingDirectoryFactory.java:350) at org.apache.solr.core.SolrCore.getNewIndexDir(SolrCore.java:251) at org.apache.solr.core.SolrCore.initIndex(SolrCore.java:465) at org.apache.solr.core.SolrCore.<init>(SolrCore.java:755) ... 13 more Caused by: java.io.IOException: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: Protocol message contained an invalid tag (zero).; Host Details : local host is: "LM-SFA-00713958/192.168.1.66"; destination host is: "10.249.132.29":8020; at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:764) at org.apache.hadoop.ipc.Client.call(Client.java:1351) at org.apache.hadoop.ipc.Client.call(Client.java:1300) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke( ProtobufRpcEngine.java:206) at com.sun.proxy.$Proxy22.getFileInfo(Unknown Source) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke( NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke( DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod( RetryInvocationHandler.java:186) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke( RetryInvocationHandler.java:102) at com.sun.proxy.$Proxy22.getFileInfo(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo( ClientNamenodeProtocolTranslatorPB.java:651) at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1679) at org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall( DistributedFileSystem.java:1106) at org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall( DistributedFileSystem.java:1102) at org.apache.hadoop.fs.FileSystemLinkResolver.resolve( FileSystemLinkResolver.java:81) at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus( DistributedFileSystem.java:1102) at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1397) at org.apache.solr.store.hdfs.HdfsDirectory.<init>(HdfsDirectory.java:63) ... 18 more Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol message contained an invalid tag (zero). at com.google.protobuf.InvalidProtocolBufferException.invalidTag( InvalidProtocolBufferException.java:89) at com.google.protobuf.CodedInputStream.readTag(CodedInputStream.java:108) at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.<init>( RpcHeaderProtos.java:1398) at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.<init>( RpcHeaderProtos.java:1362) at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto$1.parsePartialFrom( RpcHeaderProtos.java:1492) at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto$1.parsePartialFrom( RpcHeaderProtos.java:1487) at com.google.protobuf.AbstractParser.parsePartialFrom( AbstractParser.java:200) at com.google.protobuf.AbstractParser.parsePartialDelimitedFrom( AbstractParser.java:241) at com.google.protobuf.AbstractParser.parseDelimitedFrom( AbstractParser.java:253) at com.google.protobuf.AbstractParser.parseDelimitedFrom( AbstractParser.java:259) at com.google.protobuf.AbstractParser.parseDelimitedFrom( AbstractParser.java:49) at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom( RpcHeaderProtos.java:2364) at org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse( Client.java:996) at org.apache.hadoop.ipc.Client$Connection.run(Client.java:891) 8519 [coreLoadExecutor-4-thread-1] ERROR org.apache.solr.core.CoreContainer – null:org.apache.solr.common.SolrException: Unable to create core: event_shard1_replica1 at org.apache.solr.core.CoreContainer.recordAndThrow(CoreContainer.java:977) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:601) at org.apache.solr.core.CoreContainer$1.call(CoreContainer.java:271) at org.apache.solr.core.CoreContainer$1.call(CoreContainer.java:263) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask( ThreadPoolExecutor.java:895) at java.util.concurrent.ThreadPoolExecutor$Worker.run( ThreadPoolExecutor.java:918) at java.lang.Thread.run(Thread.java:680) Caused by: org.apache.solr.common.SolrException: Problem creating directory: hdfs://10.249.132.29:8020/solr-hdfs at org.apache.solr.core.SolrCore.<init>(SolrCore.java:834) at org.apache.solr.core.SolrCore.<init>(SolrCore.java:625) at org.apache.solr.core.ZkContainer.createFromZk(ZkContainer.java:256) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:590) ... 10 more Caused by: java.lang.RuntimeException: Problem creating directory: hdfs:// 10.249.132.29:8020/solr-hdfs at org.apache.solr.store.hdfs.HdfsDirectory.<init>(HdfsDirectory.java:68) at org.apache.solr.core.HdfsDirectoryFactory.create( HdfsDirectoryFactory.java:154) at org.apache.solr.core.CachingDirectoryFactory.get( CachingDirectoryFactory.java:350) at org.apache.solr.core.SolrCore.getNewIndexDir(SolrCore.java:251) at org.apache.solr.core.SolrCore.initIndex(SolrCore.java:465) at org.apache.solr.core.SolrCore.<init>(SolrCore.java:755) ... 13 more Caused by: java.io.IOException: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: Protocol message contained an invalid tag (zero).; Host Details : local host is: "LM-SFA-00713958/192.168.1.66"; destination host is: "10.249.132.29":8020; at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:764) at org.apache.hadoop.ipc.Client.call(Client.java:1351) at org.apache.hadoop.ipc.Client.call(Client.java:1300) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke( ProtobufRpcEngine.java:206) at com.sun.proxy.$Proxy22.getFileInfo(Unknown Source) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke( NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke( DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod( RetryInvocationHandler.java:186) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke( RetryInvocationHandler.java:102) at com.sun.proxy.$Proxy22.getFileInfo(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo( ClientNamenodeProtocolTranslatorPB.java:651) at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1679) at org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall( DistributedFileSystem.java:1106) at org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall( DistributedFileSystem.java:1102) at org.apache.hadoop.fs.FileSystemLinkResolver.resolve( FileSystemLinkResolver.java:81) at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus( DistributedFileSystem.java:1102) at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1397) at org.apache.solr.store.hdfs.HdfsDirectory.<init>(HdfsDirectory.java:63) ... 18 more Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol message contained an invalid tag (zero). at com.google.protobuf.InvalidProtocolBufferException.invalidTag( InvalidProtocolBufferException.java:89) at com.google.protobuf.CodedInputStream.readTag(CodedInputStream.java:108) at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.<init>( RpcHeaderProtos.java:1398) at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.<init>( RpcHeaderProtos.java:1362) at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto$1.parsePartialFrom( RpcHeaderProtos.java:1492) at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto$1.parsePartialFrom( RpcHeaderProtos.java:1487) at com.google.protobuf.AbstractParser.parsePartialFrom( AbstractParser.java:200) at com.google.protobuf.AbstractParser.parsePartialDelimitedFrom( AbstractParser.java:241) at com.google.protobuf.AbstractParser.parseDelimitedFrom( AbstractParser.java:253) at com.google.protobuf.AbstractParser.parseDelimitedFrom( AbstractParser.java:259) at com.google.protobuf.AbstractParser.parseDelimitedFrom( AbstractParser.java:49) at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom( RpcHeaderProtos.java:2364) at org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse( Client.java:996) at org.apache.hadoop.ipc.Client$Connection.run(Client.java:891) 8520 [main] INFO org.apache.solr.servlet.SolrDispatchFilter – user.dir=/Users/gpatwa/workspaces/workspace_pb_search_platform_fr/solr-hdfs 8521 [main] INFO org.apache.solr.servlet.SolrDispatchFilter – SolrDispatchFilter.init() done 2014-01-01 20:22:12.344:INFO:oejs.AbstractConnector:Started SelectChannelConnector@127.0.0.1:8983 [INFO] Started Jetty Server