Hello Team, I come to resolution of this issue by allowing the iproute table entry for the specific ports used for namenode as well as datanode. now hdfs is running and cluster is running.
thanks a lot for the suggestion. now i have another issue of interface as i am running console view of RHEL is their any way i can connect to webinterface by url so that interface and jobs details are visible.?? On 24 August 2017 at 19:18, surendra lilhore <[email protected]> wrote: > Hi Atul, > > > > Please can you share the datanode exception logs ?. Check if namenode and > datanode hostname mapping is proper or not in /etc/hosts. > > > > Put operation is failing because datanode’s are not connected to the > namenode. > > > > -Surendra > > > > *From:* Atul Rajan [mailto:[email protected]] > *Sent:* 24 August 2017 09:32 > *To:* [email protected] > *Subject:* Data streamer java exception > > > > Hello Team, > > > > I am setting up a hadoop 3.0 alpha cluster of 4 nodes in RHEL 7.2 > everything is set namenode datanode resource manager node manager but data > node is not able to connect to namenode i am getting retrying logs in > datanode. > > Also when copying files from local to hdfs data streamer java exceptions > are being thrown. > > > > can you please help me out here. > > Thanks and Regards > > Atul Rajan > > > > -Sent from my iPhone > -- *Best Regards* *Atul Rajan*
