What configuration do you want me to check? Each of the three nodes can
access each other via password-less SSH, can ping each other's IP.

Thanks
Bhushan Pathak

Thanks
Bhushan Pathak

On Wed, May 17, 2017 at 10:11 PM, Sidharth Kumar <
[email protected]> wrote:

> Hi,
>
> The error you mentioned below " 'Name or service not known'" means
> servers not able to communicate to each other. Check network configurations.
>
> Sidharth
> Mob: +91 8197555599
> LinkedIn: www.linkedin.com/in/sidharthkumar2792
>
> On 17-May-2017 12:13 PM, "Bhushan Pathak" <[email protected]>
> wrote:
>
> Apologies for the delayed reply, was away due to some personal issues.
>
> I tried the telnet command as well, but no luck. I get the response that
> 'Name or service not known'
>
> Thanks
> Bhushan Pathak
>
> Thanks
> Bhushan Pathak
>
> On Wed, May 3, 2017 at 7:48 AM, Sidharth Kumar <
> [email protected]> wrote:
>
>> Can you check if the ports are opened by running telnet command.
>> Run below command from source machine to destination machine and check if
>> this help
>>
>> $telnet <IP address> <port number>
>> Ex: $telnet 192.168.1.60 9000
>>
>>
>> Let's Hadooping....!
>>
>> Bests
>> Sidharth
>> Mob: +91 8197555599
>> LinkedIn: www.linkedin.com/in/sidharthkumar2792
>>
>> On 28-Apr-2017 10:32 AM, "Bhushan Pathak" <[email protected]>
>> wrote:
>>
>>> Hello All,
>>>
>>> 1. The slave & master can ping each other as well as use passwordless SSH
>>> 2. The actual IP starts with 10.x.x.x, I have put in the config file as
>>> I cannot share  the actual IP
>>> 3. The namenode is formatted. I executed the 'hdfs namenode -format'
>>> again just to rule out the possibility
>>> 4. I did not configure anything in the master file. I don;t think Hadoop
>>> 2.7.3 has a master file to be configured
>>> 5. The netstat command [sudo netstat -tulpn | grep '51150' ] does not
>>> give any output.
>>>
>>> Even if I change  the port number to a different one, say 52220, 50000,
>>> I still get the same error.
>>>
>>> Thanks
>>> Bhushan Pathak
>>>
>>> Thanks
>>> Bhushan Pathak
>>>
>>> On Fri, Apr 28, 2017 at 7:52 AM, Lei Cao <[email protected]>
>>> wrote:
>>>
>>>> Hi Mr. Bhushan,
>>>>
>>>> Have you tried to format namenode?
>>>> Here's the command:
>>>> hdfs namenode -format
>>>>
>>>> I've encountered such problem as namenode cannot be started. This
>>>> command line easily fixed my problem.
>>>>
>>>> Hope this can help you.
>>>>
>>>> Sincerely,
>>>> Lei Cao
>>>>
>>>>
>>>> On Apr 27, 2017, at 12:09, Brahma Reddy Battula <
>>>> [email protected]> wrote:
>>>>
>>>> *Please check “hostname –i” .*
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> *1)      **What’s configured in the “master” file.(you shared only
>>>> slave file).?*
>>>>
>>>>
>>>>
>>>> *2)      **Can you able to “ping master”?*
>>>>
>>>>
>>>>
>>>> *3)      **Can you configure like this check once..?*
>>>>
>>>> *                1.1.1.1 master*
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Regards
>>>>
>>>> Brahma Reddy Battula
>>>>
>>>>
>>>>
>>>> *From:* Bhushan Pathak [mailto:[email protected]
>>>> <[email protected]>]
>>>> *Sent:* 27 April 2017 18:16
>>>> *To:* Brahma Reddy Battula
>>>> *Cc:* [email protected]
>>>> *Subject:* Re: Hadoop 2.7.3 cluster namenode not starting
>>>>
>>>>
>>>>
>>>> Some additional info -
>>>>
>>>> OS: CentOS 7
>>>>
>>>> RAM: 8GB
>>>>
>>>>
>>>>
>>>> Thanks
>>>>
>>>> Bhushan Pathak
>>>>
>>>>
>>>> Thanks
>>>>
>>>> Bhushan Pathak
>>>>
>>>>
>>>>
>>>> On Thu, Apr 27, 2017 at 3:34 PM, Bhushan Pathak <
>>>> [email protected]> wrote:
>>>>
>>>> Yes, I'm running the command on the master node.
>>>>
>>>>
>>>>
>>>> Attached are the config files & the hosts file. I have updated the IP
>>>> address only as per company policy, so that original IP addresses are not
>>>> shared.
>>>>
>>>>
>>>>
>>>> The same config files & hosts file exist on all 3 nodes.
>>>>
>>>>
>>>>
>>>> Thanks
>>>>
>>>> Bhushan Pathak
>>>>
>>>>
>>>> Thanks
>>>>
>>>> Bhushan Pathak
>>>>
>>>>
>>>>
>>>> On Thu, Apr 27, 2017 at 3:02 PM, Brahma Reddy Battula <
>>>> [email protected]> wrote:
>>>>
>>>> Are you sure that you are starting in same machine (master)..?
>>>>
>>>>
>>>>
>>>> Please share “/etc/hosts” and configuration files..
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Regards
>>>>
>>>> Brahma Reddy Battula
>>>>
>>>>
>>>>
>>>> *From:* Bhushan Pathak [mailto:[email protected]]
>>>> *Sent:* 27 April 2017 17:18
>>>> *To:* [email protected]
>>>> *Subject:* Fwd: Hadoop 2.7.3 cluster namenode not starting
>>>>
>>>>
>>>>
>>>> Hello
>>>>
>>>>
>>>>
>>>> I have a 3-node cluster where I have installed hadoop 2.7.3. I have
>>>> updated core-site.xml, mapred-site.xml, slaves, hdfs-site.xml,
>>>> yarn-site.xml, hadoop-env.sh files with basic settings on all 3 nodes.
>>>>
>>>>
>>>>
>>>> When I execute start-dfs.sh on the master node, the namenode does not
>>>> start. The logs contain the following error -
>>>>
>>>> 2017-04-27 14:17:57,166 ERROR 
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode:
>>>> Failed to start namenode.
>>>>
>>>> java.net.BindException: Problem binding to [master:51150]
>>>> java.net.BindException: Cannot assign requested address; For more details
>>>> see:  http://wiki.apache.org/hadoop/BindException
>>>>
>>>>         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>>>> Method)
>>>>
>>>>         at sun.reflect.NativeConstructorAccessorImpl.newInstance(Native
>>>> ConstructorAccessorImpl.java:62)
>>>>
>>>>         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(De
>>>> legatingConstructorAccessorImpl.java:45)
>>>>
>>>>         at java.lang.reflect.Constructor.newInstance(Constructor.java:4
>>>> 23)
>>>>
>>>>         at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java
>>>> :792)
>>>>
>>>>         at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:7
>>>> 21)
>>>>
>>>>         at org.apache.hadoop.ipc.Server.bind(Server.java:425)
>>>>
>>>>         at org.apache.hadoop.ipc.Server$Listener.<init>(Server.java:574
>>>> )
>>>>
>>>>         at org.apache.hadoop.ipc.Server.<init>(Server.java:2215)
>>>>
>>>>         at org.apache.hadoop.ipc.RPC$Server.<init>(RPC.java:951)
>>>>
>>>>         at org.apache.hadoop.ipc.ProtobufRpcEngine$Server.<init>(Protob
>>>> ufRpcEngine.java:534)
>>>>
>>>>         at org.apache.hadoop.ipc.ProtobufRpcEngine.getServer(ProtobufRp
>>>> cEngine.java:509)
>>>>
>>>>         at org.apache.hadoop.ipc.RPC$Builder.build(RPC.java:796)
>>>>
>>>>         at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.<in
>>>> it>(NameNodeRpcServer.java:345)
>>>>
>>>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.createRpcSer
>>>> ver(NameNode.java:674)
>>>>
>>>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(N
>>>> ameNode.java:647)
>>>>
>>>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameN
>>>> ode.java:812)
>>>>
>>>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameN
>>>> ode.java:796)
>>>>
>>>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNo
>>>> de(NameNode.java:1493)
>>>>
>>>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNod
>>>> e.java:1559)
>>>>
>>>> Caused by: java.net.BindException: Cannot assign requested address
>>>>
>>>>         at sun.nio.ch.Net.bind0(Native Method)
>>>>
>>>>         at sun.nio.ch.Net.bind(Net.java:433)
>>>>
>>>>         at sun.nio.ch.Net.bind(Net.java:425)
>>>>
>>>>         at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelI
>>>> mpl.java:223)
>>>>
>>>>         at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java
>>>> :74)
>>>>
>>>>         at org.apache.hadoop.ipc.Server.bind(Server.java:408)
>>>>
>>>>         ... 13 more
>>>>
>>>> 2017-04-27 14:17:57,171 INFO org.apache.hadoop.util.ExitUtil: Exiting
>>>> with status 1
>>>>
>>>> 2017-04-27 14:17:57,176 INFO 
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode:
>>>> SHUTDOWN_MSG:
>>>>
>>>> /************************************************************
>>>>
>>>> SHUTDOWN_MSG: Shutting down NameNode at master/1.1.1.1
>>>>
>>>> ************************************************************/
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> I have changed the port number multiple times, every time I get the
>>>> same error. How do I get past this?
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Thanks
>>>>
>>>> Bhushan Pathak
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>
>
>

Reply via email to