Thanks guys, I figured out what my issue was.I did not setup the ssh key
correctly, it was for my user but I started the service as root.
Now it is working except none of the namenodes are transitioning to active on
startup, and the datanodes are not starting automatically (I think because no
namenode is active).
I can start everything manually though.
On Tuesday, October 3, 2023 at 11:03:33 AM PDT, Susheel Kumar Gadalay
<[email protected]> wrote:
Why you have set this again in hdfs-site.xml at the end.
<property> <name>dfs.namenode.rpc-address</name> <value>nn1:8020</value>
</property>
Remove this and start name node again.
Regards Susheel Kumar On Tue, 3 Oct 2023, 10:09 pm Harry Jamison,
<[email protected]> wrote:
OK here is where I am at now.
When I start the namenodes, they work, but they are all in standby mode.When I
start my first datanode it seems to kill one of the namenodes (the active one I
assume)
I am getting 2 different warnings in the namenode
[2023-10-03 09:03:52,162] WARN Unable to initialize FileSignerSecretProvider,
falling back to use random secrets. Reason: Could not read signature secret
file: /root/hadoop-http-auth-signature-secret
(org.apache.hadoop.security.authentication.server.AuthenticationFilter)
[2023-10-03 09:03:52,350] WARN Only one image storage directory
(dfs.namenode.name.dir) configured. Beware of data loss due to lack of
redundant storage directories!
(org.apache.hadoop.hdfs.server.namenode.FSNamesystem)
I am using a journal node, so I am not clear if I am supposed to have multiple
dfs.namenode.name.dir directoriesI thought each namenode has 1 directory.
Susheel Kumar Gadalay said that my shared.edits.dir Is wrong, but I am not
clear how it is wrongFrom here mine looks
righthttps://hadoop.apache.org/docs/r3.3.6/hadoop-project-dist/hadoop-hdfs/HDFSHighAvailabilityWithQJM.html
This is what is in the logs right before the namenode dies[2023-10-03
09:01:22,054] INFO Listener at vmnode3:8020
(org.apache.hadoop.ipc.Server)[2023-10-03 09:01:22,054] INFO Starting Socket
Reader #1 for port 8020 (org.apache.hadoop.ipc.Server)[2023-10-03 09:01:22,097]
INFO Registered FSNamesystemState, ReplicatedBlocksState and ECBlockGroupsState
MBeans. (org.apache.hadoop.hdfs.server.namenode.FSNamesystem)[2023-10-03
09:01:22,119] INFO Number of blocks under construction: 0
(org.apache.hadoop.hdfs.server.namenode.LeaseManager)[2023-10-03 09:01:22,122]
INFO Initialized the Default Decommission and Maintenance monitor
(org.apache.hadoop.hdfs.server.blockmanagement.DatanodeAdminDefaultMonitor)[2023-10-03
09:01:22,131] INFO STATE* Leaving safe mode after 0 secs
(org.apache.hadoop.hdfs.StateChange)[2023-10-03 09:01:22,131] INFO STATE*
Network topology has 0 racks and 0 datanodes
(org.apache.hadoop.hdfs.StateChange)[2023-10-03 09:01:22,131] INFO STATE*
UnderReplicatedBlocks has 0 blocks
(org.apache.hadoop.hdfs.StateChange)[2023-10-03 09:01:22,130] INFO Start
MarkedDeleteBlockScrubber thread
(org.apache.hadoop.hdfs.server.blockmanagement.BlockManager)[2023-10-03
09:01:22,158] INFO IPC Server Responder: starting
(org.apache.hadoop.ipc.Server)[2023-10-03 09:01:22,159] INFO IPC Server
listener on 8020: starting (org.apache.hadoop.ipc.Server)[2023-10-03
09:01:22,165] INFO NameNode RPC up at: vmnode3/192.168.1.103:8020
(org.apache.hadoop.hdfs.server.namenode.NameNode)[2023-10-03 09:01:22,166] INFO
Starting services required for standby state
(org.apache.hadoop.hdfs.server.namenode.FSNamesystem)[2023-10-03 09:01:22,168]
INFO Will roll logs on active node every 120 seconds.
(org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer)[2023-10-03
09:01:22,171] INFO Starting standby checkpoint thread...Checkpointing active NN
to possible NNs: [http://vmnode1:9870, http://vmnode2:9870]Serving checkpoints
at http://vmnode3:9870
(org.apache.hadoop.hdfs.server.namenode.ha.StandbyCheckpointer)real-time
non-blocking time (microseconds, -R) unlimitedcore file size
(blocks, -c) 0data seg size (kbytes, -d) unlimitedscheduling
priority (-e) 0file size (blocks, -f)
unlimitedpending signals (-i) 15187max locked memory
(kbytes, -l) 8192max memory size (kbytes, -m) unlimitedopen
files (-n) 1024pipe size (512 bytes,
-p) 8POSIX message queues (bytes, -q) 819200real-time priority
(-r) 0stack size (kbytes, -s) 8192cpu time
(seconds, -t) unlimitedmax user processes (-u)
15187virtual memory (kbytes, -v) unlimitedfile locks
(-x) unlimited
On Tuesday, October 3, 2023 at 03:54:23 AM PDT, Liming Cui
<[email protected]> wrote:
Harry,
Great question.I would say the same configurations in core-site.xml and
hdfs-site.xml will be overwriting each other in some way.
Glad you found the root cause.
Keep going.
On Tue, Oct 3, 2023 at 10:27 AM Harry Jamison <[email protected]> wrote:
Liming
After looking at my config, I think that maybe my problem is because my
fs.defaultFS is inconsistent between hdfs-site.xml and core-site.xmlWhat does
hdfs-site.xml vs core-site.xml do why is the same setting in 2 different
places?Or do I just have it there mistakenly?
this is what I have in hdfs-site.xml
<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet type="text/xsl"
href="configuration.xsl"?><configuration> <property>
<name>fs.defaultFS</name> <value>hdfs://mycluster</value> </property>
<property> <name>ha.zookeeper.quorum</name>
<value>nn1:2181,nn2:2181,nn3:2181</value> </property>
<property> <name>dfs.nameservices</name> <value>mycluster</value>
</property>
<property> <name>dfs.ha.namenodes.mycluster</name>
<value>nn1,nn2,nn3</value> </property>
<property> <name>dfs.namenode.rpc-address.mycluster.nn1</name>
<value>nn1:8020</value> </property> <property>
<name>dfs.namenode.rpc-address.mycluster.nn2</name> <value>nn2:8020</value>
</property> <property> <name>dfs.namenode.rpc-address.mycluster.nn3</name>
<value>nn3:8020</value> </property>
<property> <name>dfs.namenode.http-address.mycluster.nn1</name>
<value>nn1:9870</value> </property> <property>
<name>dfs.namenode.http-address.mycluster.nn2</name> <value>nn2:9870</value>
</property> <property>
<name>dfs.namenode.http-address.mycluster.nn3</name> <value>nn3:9870</value>
</property>
<property> <name>dfs.namenode.shared.edits.dir</name>
<value>qjournal://nn1:8485;nn2:8485;nn3:8485/mycluster</value> </property>
<property> <name>dfs.client.failover.proxy.provider.mycluster</name>
<value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
</property>
<property> <name>dfs.ha.fencing.methods</name> <value>sshfence</value>
</property>
<property> <name>dfs.ha.fencing.ssh.private-key-files</name>
<value>/home/harry/.ssh/id_rsa</value> </property>
<property> <name>dfs.namenode.name.dir</name>
<value>file:/hadoop/data/hdfs/namenode</value> </property> <property>
<name>dfs.datanode.data.dir</name>
<value>file:/hadoop/data/hdfs/datanode</value> </property> <property>
<name>dfs.journalnode.edits.dir</name>
<value>/hadoop/data/hdfs/journalnode</value> </property> <property>
<name>dfs.namenode.rpc-address</name> <value>nn1:8020</value> </property>
<property> <name>dfs.ha.nn.not-become-active-in-safemode</name>
<value>true</value> </property>
</configuration>
In core-site.xml I have this
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property overrides in this file. -->
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://nn1:8020</value>
</property>
</configuration>
On Tuesday, October 3, 2023 at 12:54:26 AM PDT, Liming Cui
<[email protected]> wrote:
Can you show us the configuration files? Maybe I can help you with some
suggestions.
On Tue, Oct 3, 2023 at 9:05 AM Harry Jamison <[email protected]>
wrote:
I am trying to setup a HA HDFS cluster, and I am running into a problem
I am not sure what I am doing wrong, I thought I followed the HA namenode
guide, but it is not working.
Apache Hadoop 3.3.6 – HDFS High Availability
I have 2 namenodes and 3 journal nodes, and 3 zookeeper nodes.
After some period of time I see the following and my namenode and journal node
die.I am not sure where the problem is, or how to diagnose what I am doing
wrong here. And the logging here does not make sense to me.
NamenodeServing checkpoints at http://nn1:9870
(org.apache.hadoop.hdfs.server.namenode.ha.StandbyCheckpointer)
real-time non-blocking time (microseconds, -R) unlimited
core file size (blocks, -c) 0
data seg size (kbytes, -d) unlimited
scheduling priority (-e) 0
file size (blocks, -f) unlimited
pending signals (-i) 15187
max locked memory (kbytes, -l) 8192
max memory size (kbytes, -m) unlimited
open files (-n) 1024
pipe size (512 bytes, -p) 8
POSIX message queues (bytes, -q) 819200
real-time priority (-r) 0
stack size (kbytes, -s) 8192
cpu time (seconds, -t) unlimited
max user processes (-u) 15187
virtual memory (kbytes, -v) unlimited
file locks (-x) unlimited
[2023-10-02 23:53:46,693] ERROR RECEIVED SIGNAL 15: SIGTERM
(org.apache.hadoop.hdfs.server.namenode.NameNode)
[2023-10-02 23:53:46,701] INFO SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at nn1/192.168.1.159
************************************************************/
(org.apache.hadoop.hdfs.server.namenode.NameNode)
JournalNode[2023-10-02 23:54:19,162] WARN Journal at nn1/192.168.1.159:8485 has
no edit logs (org.apache.hadoop.hdfs.qjournal.server.JournalNodeSyncer)
real-time non-blocking time (microseconds, -R) unlimited
core file size (blocks, -c) 0
data seg size (kbytes, -d) unlimited
scheduling priority (-e) 0
file size (blocks, -f) unlimited
pending signals (-i) 15187
max locked memory (kbytes, -l) 8192
max memory size (kbytes, -m) unlimited
open files (-n) 1024
pipe size (512 bytes, -p) 8
POSIX message queues (bytes, -q) 819200
real-time priority (-r) 0
stack size (kbytes, -s) 8192
cpu time (seconds, -t) unlimited
max user processes (-u) 15187
virtual memory (kbytes, -v) unlimited
file locks (-x) unlimited
--
Best
Liming
--
Best
Liming