下面记录下hadoop完全分布式安装的过程,其中hadoop使用的版本是apache下的,不是cdh,linux版本为centos6。
完全分布式示意图
下面在三台节点上安装hadoop完全分布式,其中一个服务器节点上将有多个hadoop相关的节点,最后是压缩到三台的安装效果,正常来说至少13个服务节点。
(1)zookeeper用于管理namenode,用于故障转移主备切换,其中zookeeper通过failoverController进程来进行namenode主备切换。
(2)namenode主备之间通过journalNode来进行通信,进行数据同步。
(3)resourceManager也会有两个,一个挂了另外一个顶上。
(4)datanode上储存数据,MR计算有数据本地化策略,nodeManager一般和datanode在一起。
以上是最后安装的节点分布图,下面开始安装部署。
前置准备
前置准备包括关闭linux防火墙、修改主机名、ip映射、配置jdk和免密登录,可参考,其中这里使用的主机名分别为hadoop01、hadoop02和hadoop03。ip映射需修改/etc/hosts文件,添加三台ip和节点名的映射关系。以上操作三台都需要准备好,容易出现问题的就是免密登录,下面记录一下。
hadoop01节点生成公私钥对,并分发公钥到hadoop02和hadoop03,还包括自己hadoop01。
# ssh-keygen命令完成公私钥生成,这里省略了,具体参考上面博文记录
# 拷贝公钥到hadoop02
[root@hadoop01 ~]# ssh-copy-id root@hadoop02
# 直接回车
root@hadoop02's password:
Now try logging into the machine, with "ssh 'root@hadoop02'", and check in:
.ssh/authorized_keys
to make sure we haven't added extra keys that you weren't expecting.
You have new mail in /var/spool/mail/root
# 拷贝公钥到hadoop03
[root@hadoop01 ~]# ssh-copy-id root@hadoop03
# 直接回车
root@hadoop03's password:
Now try logging into the machine, with "ssh 'root@hadoop03'", and check in:
.ssh/authorized_keys
to make sure we haven't added extra keys that you weren't expecting.
# 拷贝公钥到hadoop01自己
[root@hadoop01 ~]# ssh-copy-id root@hadoop01
完成这一步,所有节点~/.ssh目录下会在authorized_keys文件里保存hadoop01上的公钥。
# 第二个节点保存hadoop01上的公钥,第三个节点也是
[root@hadoop02 ~/.ssh]# ll
total 8
-rw-------. 1 root root 395 Dec 28 23:11 authorized_keys
hadoop02和hadoop03节点都进行同样的操作后,就完成免密登录,查看authorized_keys文件,发现已经获取到三台的公钥。
# 任何一个节点上,都有三台节点的公钥信息
[root@hadoop03 ~/.ssh]# cat authorized_keys
ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEAwcYd1kp2gM0dszWemOTSYy7MpGAXcg9DQEDUxcr8x0oz3Wcm8iZs8i6EZj0GQtUAExKX8Y/6havVAfF6DF6HiszXB4P7CjJUVwrLJhNt1YtRj7e0gPe6/HMIsOkRzrDGQMoD3L5b5wmOP+0KBTTFWDNTNoIZlbttYLk5Ir0bW0VtfLlJhU2FJSYkcmqKAtZXKvsqclx1oQG1aKDQV6OIJA72/YVw0HvhFk82UPEZBoW9HYrmJcn6QOOH1AtR5eC9V2omrpXXsHrx71iXxDQesCvOlew6IOB3z6F2Zlll3ReNCjhP9RCr7ILJpfF3VKFbmwl8jnGJiEH56qMfwVesNQ== root@hadoop01
ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEApyq6QsZvkHPkrOtTgmtsI365ZI8dV3DVgAs4ANWmPEaLb2j3GpJ8fLzml77HvX4oYO3fXIa9Nyg2oNgLGa3ipvmGCig58fR8P2wSGjQY0oHGowl5+Mt2BNRu6O+u7mBSIIn8Q2BR+K0AotfSGgQrw7U9fRUlA38dOU9qrWSuFNE+A8VgdzpLtV6Z5DgPJjzCbPZRCnHIex1CSg18X5Wg0jE1pbiumWvx2J9od36JheosEiJPljVTnbSYeCJnFJyyUsFeZq7nTwklifqeaY2RRUGSB72vrPXxLLaHT3tnwqV/DmcdIfIebLFsIUbzNeLQP03EmISLsJfLafYTCi0N/Q== root@hadoop02
ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEA3E+tThIDFezbq3qvPFFJDDGzKTSJBuRDl5dR1LdY5r38ODcEw7uslJMqiryXwlgSvLkzwEAQIE7FUiLMlOnmLJ+aiwdTHheEC1M9cPAYQZ6QMMkZp8Y6YQ9QeWqvMSfbtRPPpFrndEjaXML0+dJUh8Srgyzchu7zp1sSPu5wZXGuIbcTha8UNV7c0VjIQAimjQIdZSqA4zmTa2rt6MIVEwlGsq/RhzXnnaXopMLAyk97bVK8XWpWgIUJ/DSiKZBSDAa8usxaqTF8qw1AvxGf52Xuye5EV2WvpU5LeOaBj/XxmCnrtRRkqYrtvObEXvDRq1MG9wU90LFHeH68PRCrCQ== root@hadoop03
可以使用'ssh 节点名'来进行进行免密登录测试,第一次提示可能需要输入密码,这样会将登录信息保存在ssh目录下known_hosts中,后续不再提示需要密码,另外本次设置不需要密码。
安装zookeeper
三台节点安装zookeeper,参考博文。
安装完全分布式hadoop
完成以上配置后,就可以开始进入完全分布式hadoop的安装和部署,按照如下步骤傻瓜式安装就能完成。先在hadoop01节点上安装配置hadoop,然后将安装包分发到另外两台。
解压hadoop
解压apache版本的hadoop安装包。
[root@hadoop01 /home/software]# tar -zxvf hadoop-2.7.1.tar.gz
修改hadoop-env.sh
hadoop-env.sh是跟环境相关的,配置JAVA_HOME和HADOOP_CONF_DIR,JAVA_HOME就是前面配置好的jdk,后者就是指定hadoop配置文件的目录。
# 配置项
export JAVA_HOME=/home/software/jdk1.8.0_181
export HADOOP_CONF_DIR=/home/software/hadoop-2.7.1/etc/hadoop
# 保存退出,source使其重新生效
[root@hadoop01 /home/software/hadoop-2.7.1/etc/hadoop]# source hadoop-env.sh
修改core-site.xml
使用zookeeper来管理hadoop节点,ns就是注册在zookeeper上节点的名字。
<configuration>
<!--为集群起别名,以确定向Zookeeper注册的节点的名字-->
<property>
<name>fs.defaultFS</name>
<value>hdfs://ns</value>
</property>
<!--指定Hadoop数据临时存放目录-->
<property>
<name>hadoop.tmp.dir</name>
<value>/home/software/hadoop-2.7.1/tmp</value>
</property>
<!--指定zookeeper的连接地址-->
<property>
<name>ha.zookeeper.quorum</name>
<value>hadoop01:2181,hadoop02:2181,hadoop03:2181</value>
</property>
</configuration>
修改hdfs-site.xml
配置hadoop高可用相关内容,配置两个namenode,指定节点位置,需要journalnode的支持,同样在里面配置。
<!--注意要和core-site.xml中的名称保持一致-->
<property>
<name>dfs.nameservices</name>
<value>ns</value>
</property>
<!--ns集群下有两个namenode,分别为nn1, nn2-->
<property>
<name>dfs.ha.namenodes.ns</name>
<value>nn1,nn2</value>
</property>
<!--nn1的RPC通信-->
<property>
<name>dfs.namenode.rpc-address.ns.nn1</name>
<value>hadoop01:9000</value>
</property>
<!--nn1的http通信-->
<property>
<name>dfs.namenode.http-address.ns.nn1</name>
<value>hadoop01:50070</value>
</property>
<!-- nn2的RPC通信地址 -->
<property>
<name>dfs.namenode.rpc-address.ns.nn2</name>
<value>hadoop02:9000</value>
</property>
<!-- nn2的http通信地址 -->
<property>
<name>dfs.namenode.http-address.ns.nn2</name>
<value>hadoop02:50070</value>
</property>
<!--指定namenode的元数据在JournalNode上存放的位置,这样,namenode2可以从journalnode集群里的指定位置上获取信息,达到热备效果-->
<property>
<name>dfs.namenode.shared.edits.dir</name>
<value>qjournal://hadoop01:8485;hadoop02:8485;hadoop03:8485/ns</value>
</property>
<!-- 指定JournalNode在本地磁盘存放数据的位置 -->
<property>
<name>dfs.journalnode.edits.dir</name>
<value>/home/software/hadoop-2.7.1/tmp/journal</value>
</property>
<!-- 开启NameNode故障时自动切换 -->
<property>
<name>dfs.ha.automatic-failover.enabled</name>
<value>true</value>
</property>
<!-- 配置失败自动切换实现方式 -->
<property>
<name>dfs.client.failover.proxy.provider.ns</name>
<value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
</property>
<!-- 配置隔离机制 -->
<property>
<name>dfs.ha.fencing.methods</name>
<value>sshfence</value>
</property>
<!-- 使用隔离机制时需要ssh免登陆 -->
<property>
<name>dfs.ha.fencing.ssh.private-key-files</name>
<value>/root/.ssh/id_rsa</value>
</property>
<!--配置namenode存放元数据的目录,可以不配置,如果不配置则默认放到hadoop.tmp.dir下-->
<property>
<name>dfs.namenode.name.dir</name>
<value>file:///home/software/hadoop-2.7.1/tmp/hdfs/name</value>
</property>
<!--配置datanode存放元数据的目录,可以不配置,如果不配置则默认放到hadoop.tmp.dir下-->
<property>
<name>dfs.datanode.data.dir</name>
<value>file:///home/software/hadoop-2.7.1/tmp/hdfs/data</value>
</property>
<!--配置复本数量-->
<property>
<name>dfs.replication</name>
<value>3</value>
</property>
<!--设置用户的操作权限,false表示关闭权限验证,任何用户都可以操作-->
<property>
<name>dfs.permissions</name>
<value>false</value>
</property>
修改mapred-site.xml
将MapReduce配置基于yarn来运行。
<!--mapreduce运行基于yarn,也可以调整为local-->
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
修改yarn-site.xml
主要配置resourcemanager的高可用,在节点1和节点3上配置,同样两个rm交给zookeeper管理。
<!--配置yarn的高可用-->
<property>
<name>yarn.resourcemanager.ha.enabled</name>
<value>true</value>
</property>
<!--指定两个resourcemaneger的名称-->
<property>
<name>yarn.resourcemanager.ha.rm-ids</name>
<value>rm1,rm2</value>
</property>
<!--配置rm1的主机-->
<property>
<name>yarn.resourcemanager.hostname.rm1</name>
<value>hadoop01</value>
</property>
<!--配置rm2的主机-->
<property>
<name>yarn.resourcemanager.hostname.rm2</name>
<value>hadoop03</value>
</property>
<!--开启yarn恢复机制-->
<property>
<name>yarn.resourcemanager.recovery.enabled</name>
<value>true</value>
</property>
<!--执行rm恢复机制实现类-->
<property>
<name>yarn.resourcemanager.store.class</name>
<value>org.apache.hadoop.yarn.server.resourcemanager.recovery.ZKRMStateStore</value>
</property>
<!--配置zookeeper的地址-->
<property>
<name>yarn.resourcemanager.zk-address</name>
<value>hadoop01:2181,hadoop02:2181,hadoop03:2181</value>
</property>
<!--执行yarn集群的别名-->
<property>
<name>yarn.resourcemanager.cluster-id</name>
<value>ns-yarn</value>
</property>
<!-- 指定nodemanager启动时加载server的方式为shuffle server -->
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<!-- 指定resourcemanager地址 -->
<property>
<name>yarn.resourcemanager.hostname</name>
<value>hadoop03</value>
</property>
修改slaves文件
# 添加三个节点信息
hadoop01
hadoop02
hadoop03
远程拷贝
将hadoop01上配置好的hadoop安装目录远程拷贝到hadoop02和hadoop03。
# $PWD就是/home/software,也可以写绝对路径
[root@hadoop01 /home/software]# scp -r hadoop2.7.1/ root@hadoop02:$PWD
[root@hadoop01 /home/software]# scp -r haoop2.7.1/ root@hadoop03:$PWD
三节点配置hadoop环境变量
三个节点都需要配置hadoop环境变量,此外还包括前面准备阶段的java环境变量。配置完后需要source /etc/profile一下让环境变量生效。
export HADOOP_HOME=/home/software/hadoop-2.7.1
export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin
启动zookeeper
三个节点均需要启动zookeeper,并记得启动后查看状态是否为一主两从。
# 顺序启动后hadoop02上的zk就是leader,其他均为follower
[root@hadoop02 /home/software/zookeeper-3.4.8/bin]# ./zkServer.sh start
ZooKeeper JMX enabled by default
Using config: /home/software/zookeeper-3.4.8/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
[root@hadoop02 /home/software/zookeeper-3.4.8/bin]# ./zkServer.sh status
ZooKeeper JMX enabled by default
Using config: /home/software/zookeeper-3.4.8/bin/../conf/zoo.cfg
Mode: leader
第一个节点格式化zookeeper
使用hdfs zkfc -formatZK命令,这个操作将会在zk上注册节点。
# 格式化zookeeper
[root@hadoop01 /home/software]# hdfs zkfc -formatZK
19/12/29 01:16:20 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
19/12/29 01:16:20 INFO tools.DFSZKFailoverController: Failover controller configured for NameNode NameNode at hadoop01/192.168.200.140:9000
19/12/29 01:16:21 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.6-1569965, built on 02/20/2014 09:09 GMT
19/12/29 01:16:21 INFO zookeeper.ZooKeeper: Client environment:host.name=hadoop01
19/12/29 01:16:21 INFO zookeeper.ZooKeeper: Client environment:java.version=1.8.0_181
19/12/29 01:16:21 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation
...省略
19/12/29 01:16:21 INFO zookeeper.ZooKeeper: Initiating client connection, connectString=hadoop01:2181,hadoop02:2181,hadoop03:2181 sessionTimeout=5000 watcher=org.apache.hadoop.ha.ActiveStandbyElector$WatcherWithClientRef@436813f3
19/12/29 01:16:21 INFO zookeeper.ClientCnxn: Opening socket connection to server hadoop02/192.168.200.150:2181. Will not attempt to authenticate using SASL (unknown error)
19/12/29 01:16:21 INFO zookeeper.ClientCnxn: Socket connection established to hadoop02/192.168.200.150:2181, initiating session
19/12/29 01:16:21 INFO zookeeper.ClientCnxn: Session establishment complete on server hadoop02/192.168.200.150:2181, sessionid = 0x26f4d75770d0000, negotiated timeout = 5000
19/12/29 01:16:21 INFO ha.ActiveStandbyElector: Session connected.
# 成功创建节点ns,ns就是前面配置的集群别名
19/12/29 01:16:21 INFO ha.ActiveStandbyElector: Successfully created /hadoop-ha/ns in ZK.
19/12/29 01:16:21 INFO zookeeper.ZooKeeper: Session: 0x26f4d75770d0000 closed
19/12/29 01:16:21 INFO zookeeper.ClientCnxn: EventThread shut down
You have new mail in /var/spool/mail/root
三个节点启动journalNode
三个节点均需要启动,命令在sbin目录下。
[root@hadoop01 /home/software/hadoop-2.7.1/sbin]# ./hadoop-daemon.sh start journalnode
第一个节点格式化namenode
注意只在第一个节点上使用如下命令格式化namenode,第二个节点还需要格式化namenode,使用的命令不一样,需要注意。
[root@hadoop01 /home/software/hadoop-2.7.1/sbin]# hadoop namenode -format
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.
19/12/29 01:22:26 INFO namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG: host = hadoop01/192.168.200.140
STARTUP_MSG: args = [-format]
STARTUP_MSG: version = 2.7.1
...省略
STARTUP_MSG: build = https://git-wip-us.apache.org/repos/asf/hadoop.git -r 15ecc87ccf4a0228f35af08fc56de536e6ce657a; compiled by 'jenkins' on 2015-06-29T06:04Z
STARTUP_MSG: java = 1.8.0_181
************************************************************/
19/12/29 01:22:26 INFO namenode.NameNode: registered UNIX signal handlers for [TERM, HUP, INT]
19/12/29 01:22:26 INFO namenode.NameNode: createNameNode [-format]
19/12/29 01:22:26 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Formatting using clusterid: CID-0deff869-4530-4fc5-a9e9-10babaaa7d48
19/12/29 01:22:26 INFO namenode.FSNamesystem: No KeyProvider found.
19/12/29 01:22:26 INFO namenode.FSNamesystem: fsLock is fair:true
19/12/29 01:22:26 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit=1000
19/12/29 01:22:26 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true
19/12/29 01:22:26 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
19/12/29 01:22:26 INFO blockmanagement.BlockManager: The block deletion will start around 2019 Dec 29 01:22:26
19/12/29 01:22:26 INFO util.GSet: Computing capacity for map BlocksMap
19/12/29 01:22:26 INFO util.GSet: VM type = 64-bit
19/12/29 01:22:26 INFO util.GSet: 2.0% max memory 966.7 MB = 19.3 MB
19/12/29 01:22:26 INFO util.GSet: capacity = 2^21 = 2097152 entries
19/12/29 01:22:26 INFO blockmanagement.BlockManager: dfs.block.access.token.enable=false
19/12/29 01:22:26 INFO blockmanagement.BlockManager: defaultReplication = 3
19/12/29 01:22:26 INFO blockmanagement.BlockManager: maxReplication = 512
19/12/29 01:22:26 INFO blockmanagement.BlockManager: minReplication = 1
19/12/29 01:22:26 INFO blockmanagement.BlockManager: maxReplicationStreams = 2
19/12/29 01:22:26 INFO blockmanagement.BlockManager: shouldCheckForEnoughRacks = false
19/12/29 01:22:26 INFO blockmanagement.BlockManager: replicationRecheckInterval = 3000
19/12/29 01:22:26 INFO blockmanagement.BlockManager: encryptDataTransfer = false
19/12/29 01:22:26 INFO blockmanagement.BlockManager: maxNumBlocksToLog = 1000
19/12/29 01:22:26 INFO namenode.FSNamesystem: fsOwner = root (auth:SIMPLE)
19/12/29 01:22:26 INFO namenode.FSNamesystem: supergroup = supergroup
19/12/29 01:22:26 INFO namenode.FSNamesystem: isPermissionEnabled = false
19/12/29 01:22:26 INFO namenode.FSNamesystem: Determined nameservice ID: ns
19/12/29 01:22:26 INFO namenode.FSNamesystem: HA Enabled: true
19/12/29 01:22:26 INFO namenode.FSNamesystem: Append Enabled: true
19/12/29 01:22:27 INFO util.GSet: Computing capacity for map INodeMap
19/12/29 01:22:27 INFO util.GSet: VM type = 64-bit
19/12/29 01:22:27 INFO util.GSet: 1.0% max memory 966.7 MB = 9.7 MB
19/12/29 01:22:27 INFO util.GSet: capacity = 2^20 = 1048576 entries
19/12/29 01:22:27 INFO namenode.FSDirectory: ACLs enabled? false
19/12/29 01:22:27 INFO namenode.FSDirectory: XAttrs enabled? true
19/12/29 01:22:27 INFO namenode.FSDirectory: Maximum size of an xattr: 16384
19/12/29 01:22:27 INFO namenode.NameNode: Caching file names occuring more than 10 times
19/12/29 01:22:27 INFO util.GSet: Computing capacity for map cachedBlocks
19/12/29 01:22:27 INFO util.GSet: VM type = 64-bit
19/12/29 01:22:27 INFO util.GSet: 0.25% max memory 966.7 MB = 2.4 MB
19/12/29 01:22:27 INFO util.GSet: capacity = 2^18 = 262144 entries
19/12/29 01:22:27 INFO namenode.FSNamesystem: dfs.namenode.safemode.threshold-pct = 0.9990000128746033
19/12/29 01:22:27 INFO namenode.FSNamesystem: dfs.namenode.safemode.min.datanodes = 0
19/12/29 01:22:27 INFO namenode.FSNamesystem: dfs.namenode.safemode.extension = 30000
19/12/29 01:22:27 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10
19/12/29 01:22:27 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10
19/12/29 01:22:27 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25
19/12/29 01:22:27 INFO namenode.FSNamesystem: Retry cache on namenode is enabled
19/12/29 01:22:27 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
19/12/29 01:22:27 INFO util.GSet: Computing capacity for map NameNodeRetryCache
19/12/29 01:22:27 INFO util.GSet: VM type = 64-bit
19/12/29 01:22:27 INFO util.GSet: 0.029999999329447746% max memory 966.7 MB = 297.0 KB
19/12/29 01:22:27 INFO util.GSet: capacity = 2^15 = 32768 entries
19/12/29 01:22:28 INFO namenode.FSImage: Allocated new BlockPoolId: BP-880978136-192.168.200.140-1577553748138
# 格式化成功
19/12/29 01:22:28 INFO common.Storage: Storage directory /home/software/hadoop-2.7.1/tmp/hdfs/name has been successfully formatted.
19/12/29 01:22:28 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0
19/12/29 01:22:28 INFO util.ExitUtil: Exiting with status 0
19/12/29 01:22:28 INFO namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at hadoop01/192.168.200.140
************************************************************/
You have new mail in /var/spool/mail/root
第一个节点启动namenode
[root@hadoop01 /home/software/hadoop-2.7.1/sbin]# ./hadoop-daemon start namenode
第二个节点格式化namenode
第二个节点也格式化namenode,其为standby状态。
[root@hadoop02 /home/software/hadoop-2.7.1/sbin]# hdfs namenode -bootstrapStandby
19/12/29 01:26:42 INFO namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG: host = hadoop02/192.168.200.150
STARTUP_MSG: args = [-bootstrapStandby]
STARTUP_MSG: version = 2.7.1
...省略
STARTUP_MSG: build = https://git-wip-us.apache.org/repos/asf/hadoop.git -r 15ecc87ccf4a0228f35af08fc56de536e6ce657a; compiled by 'jenkins' on 2015-06-29T06:04Z
STARTUP_MSG: java = 1.8.0_181
************************************************************/
19/12/29 01:26:42 INFO namenode.NameNode: registered UNIX signal handlers for [TERM, HUP, INT]
19/12/29 01:26:42 INFO namenode.NameNode: createNameNode [-bootstrapStandby]
19/12/29 01:26:43 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
=====================================================
About to bootstrap Standby ID nn2 from:
Nameservice ID: ns
Other Namenode ID: nn1
Other NN's HTTP address: http://hadoop01:50070
Other NN's IPC address: hadoop01/192.168.200.140:9000
Namespace ID: 2053051572
Block pool ID: BP-880978136-192.168.200.140-1577553748138
Cluster ID: CID-0deff869-4530-4fc5-a9e9-10babaaa7d48
Layout version: -63
isUpgradeFinalized: true
=====================================================
# 第二个nn格式化成功
19/12/29 01:26:43 INFO common.Storage: Storage directory /home/software/hadoop-2.7.1/tmp/hdfs/name has been successfully formatted.
19/12/29 01:26:44 INFO namenode.TransferFsImage: Opening connection to http://hadoop01:50070/imagetransfer?getimage=1&txid=0&storageInfo=-63:2053051572:0:CID-0deff869-4530-4fc5-a9e9-10babaaa7d48
19/12/29 01:26:44 INFO namenode.TransferFsImage: Image Transfer timeout configured to 60000 milliseconds
19/12/29 01:26:44 INFO namenode.TransferFsImage: Transfer took 0.00s at 0.00 KB/s
19/12/29 01:26:44 INFO namenode.TransferFsImage: Downloaded file fsimage.ckpt_0000000000000000000 size 351 bytes.
19/12/29 01:26:44 INFO util.ExitUtil: Exiting with status 0
19/12/29 01:26:44 INFO namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at hadoop02/192.168.200.150
************************************************************/
第二个节点启动namenode
[root@hadoop02 /home/software/hadoop-2.7.1/sbin]# ./hadoop-daemon.sh start namenode
三个节点启动datanode
[root@hadoop01 /home/software/hadoop-2.7.1/sbin]# ./hadoop-daemon.sh start datanode
第一个节点和第二个节点启动故障转移进程
启动故障转移进程后,web界面能查看到节点1上的hadoop为active,节点2上为standby。
[root@hadoop01 /home/software/hadoop-2.7.1/sbin]# ./hadoop-daemon.sh start zkfc
[root@hadoop02 /home/software/hadoop-2.7.1/sbin]# ./hadoop-daemon.sh start zkfc
WEB界面查看效果,现在是节点2在对外提供服务,状态为active。
第三个节点启动yarn
第三个节点前面配置为resourcemanager的主节点,启动。
[root@hadoop03 /home/software/hadoop-2.7.1/sbin]# ./start-yarn.sh
starting yarn daemons
starting resourcemanager, logging to /home/software/hadoop-2.7.1/logs/yarn-root-resourcemanager-hadoop03.out
hadoop01: starting nodemanager, logging to /home/software/hadoop-2.7.1/logs/yarn-root-nodemanager-hadoop01.out
hadoop02: starting nodemanager, logging to /home/software/hadoop-2.7.1/logs/yarn-root-nodemanager-hadoop02.out
hadoop03: starting nodemanager, logging to /home/software/hadoop-2.7.1/logs/yarn-root-nodemanager-hadoop03.out
第一个节点单独启动resouceManager
[root@hadoop01 /home/software/hadoop-2.7.1/sbin]# ./yarn-daemon.sh start resourcemanager
检查最后节点信息
最后检查节点,如果三个节点信息如下,说明安装ok。
节点1
[root@hadoop01 /home/software/hadoop-2.7.1/sbin]# jps
2833 Jps
2756 ResourceManager
2548 NodeManager
2293 DataNode
1510 QuorumPeerMain
1879 JournalNode
2425 DFSZKFailoverController
2074 NameNode
节点2
[root@hadoop02 /home/software/hadoop-2.7.1/sbin]# jps
2388 DFSZKFailoverController
1526 QuorumPeerMain
2246 DataNode
1895 JournalNode
2106 NameNode
2668 Jps
2495 NodeManager
节点3
[root@hadoop03 /home/software/hadoop-2.7.1/sbin]# jps
2209 ResourceManager
1859 JournalNode
2311 NodeManager
1544 QuorumPeerMain
2041 DataNode
2655 Jps
关闭节点
在节点1上使用stop-all.sh来关闭节点。
[root@hadoop01 /home/software/hadoop-2.7.1/sbin]# stop-all.sh
This script is Deprecated. Instead use stop-dfs.sh and stop-yarn.sh
19/12/29 01:38:46 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Stopping namenodes on [hadoop01 hadoop02]
hadoop02: stopping namenode
hadoop01: stopping namenode
hadoop02: stopping datanode
hadoop01: stopping datanode
hadoop03: stopping datanode
Stopping journal nodes [hadoop01 hadoop02 hadoop03]
hadoop02: stopping journalnode
hadoop01: stopping journalnode
hadoop03: stopping journalnode
19/12/29 01:39:06 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Stopping ZK Failover Controllers on NN hosts [hadoop01 hadoop02]
hadoop02: stopping zkfc
hadoop01: stopping zkfc
stopping yarn daemons
stopping resourcemanager
hadoop01: stopping nodemanager
hadoop03: stopping nodemanager
hadoop02: stopping nodemanager
no proxyserver to stop
关闭后三个节点的信息,zk的进程还保留,还有节点3上的rm进程保留。
# 节点1
[root@hadoop01 /home/software/hadoop-2.7.1/sbin]# jps
1510 QuorumPeerMain
3448 Jps
# 节点2
[root@hadoop02 /home/software/hadoop-2.7.1/sbin]# jps
1526 QuorumPeerMain
2905 Jps
# 节点3
[root@hadoop03 /home/software/hadoop-2.7.1/sbin]# jps
2817 Jps
2209 ResourceManager
1544 QuorumPeerMain
开启所有节点
在节点1上使用start-all.sh,可以将关闭的节点全部重新启动,无需单个命令启动。
[root@hadoop01 /home/software/hadoop-2.7.1/sbin]# start-all.sh
This script is Deprecated. Instead use start-dfs.sh and start-yarn.sh
19/12/29 01:43:12 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Starting namenodes on [hadoop01 hadoop02]
hadoop02: starting namenode, logging to /home/software/hadoop-2.7.1/logs/hadoop-root-namenode-hadoop02.out
hadoop01: starting namenode, logging to /home/software/hadoop-2.7.1/logs/hadoop-root-namenode-hadoop01.out
hadoop02: starting datanode, logging to /home/software/hadoop-2.7.1/logs/hadoop-root-datanode-hadoop02.out
hadoop01: starting datanode, logging to /home/software/hadoop-2.7.1/logs/hadoop-root-datanode-hadoop01.out
hadoop03: starting datanode, logging to /home/software/hadoop-2.7.1/logs/hadoop-root-datanode-hadoop03.out
Starting journal nodes [hadoop01 hadoop02 hadoop03]
hadoop03: starting journalnode, logging to /home/software/hadoop-2.7.1/logs/hadoop-root-journalnode-hadoop03.out
hadoop02: starting journalnode, logging to /home/software/hadoop-2.7.1/logs/hadoop-root-journalnode-hadoop02.out
hadoop01: starting journalnode, logging to /home/software/hadoop-2.7.1/logs/hadoop-root-journalnode-hadoop01.out
19/12/29 01:43:31 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Starting ZK Failover Controllers on NN hosts [hadoop01 hadoop02]
hadoop01: starting zkfc, logging to /home/software/hadoop-2.7.1/logs/hadoop-root-zkfc-hadoop01.out
hadoop02: starting zkfc, logging to /home/software/hadoop-2.7.1/logs/hadoop-root-zkfc-hadoop02.out
starting yarn daemons
starting resourcemanager, logging to /home/software/hadoop-2.7.1/logs/yarn-root-resourcemanager-hadoop01.out
hadoop01: starting nodemanager, logging to /home/software/hadoop-2.7.1/logs/yarn-root-nodemanager-hadoop01.out
hadoop02: starting nodemanager, logging to /home/software/hadoop-2.7.1/logs/yarn-root-nodemanager-hadoop02.out
hadoop03: starting nodemanager, logging to /home/software/hadoop-2.7.1/logs/yarn-root-nodemanager-hadoop03.out
这样所有关闭节点又重新启动,以上就是完全分布式hadoop集群在三个节点上的安装过程,尝试启动后上传一个文件到集群,发现上传到了active的节点里,standby的节点提示不可以访问。
[root@hadoop01 /home]# hadoop fs -put book.txt /
19/12/29 11:32:30 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
You have new mail in /var/spool/mail/root
# 上传到了节点2
[root@hadoop01 /home]# hadoop fs -ls /
19/12/29 11:32:39 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Found 1 items
-rw-r--r-- 3 root supergroup 18 2019-12-29 11:32 /book.txt
standby的节点提示不可以访问,数据可以在active的节点上查看到。
以上就是安装和初步使用过程,记录下后面使用。