一、准备工作:
1、环境:CentOS6.4-64bit JDK1.7-64bit
基于虚拟机拷贝4个虚拟机,一个作为Master,另外三个作为Slave,在这拷贝出来的四台虚拟机上分别执行下面的脚本初始化网卡eth0设备:
/install/initNetwork.sh
2、配置集群网络
A、Master机器:
1 #配置主机名 2 hostname master01 3 vi /etc/sysconfig/network 4 NETWORKING=yes 5 HOSTNAME=master01
1 #配置网络地址 2 vi /etc/sysconfig/network-scripts/ifcfg-eth0 3 DEVICE=eth0 4 TYPE=Ethernet 5 ONBOOT=yes 6 NM_CONTROLLED=yes 7 BOOTPROTO=static 8 IPADDR=192.168.37.154 9 NETMASK=255.255.255.0 10 GATEWAY=192.168.37.2 11 DNS1=192.168.37.2 12 DNS2=8.8.8.8
1 #配置本地IP地址解析 2 [root@master01 ~]# cat /etc/hosts 3 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 5 192.168.37.154 master01 6 192.168.37.156 slave01 7 192.168.37.157 slave02 8 192.168.37.158 slave03
1 #重新启动服务器 2 reboot
B、Slave机器:
1 #配置主机名 2 hostname slave01 3 vi /etc/sysconfig/network 4 NETWORKING=yes 5 HOSTNAME=slave01
1 #配置网络地址 2 vi /etc/sysconfig/network-scripts/ifcfg-eth0 3 DEVICE=eth0 4 TYPE=Ethernet 5 ONBOOT=yes 6 NM_CONTROLLED=yes 7 BOOTPROTO=static 8 IPADDR=192.168.37.156 9 NETMASK=255.255.255.0 10 GATEWAY=192.168.37.2 11 DNS1=192.168.37.2 12 DNS2=8.8.8.8
1 #配置本地IP地址解析 2 [root@master01 ~]# cat /etc/hosts 3 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 5 192.168.37.154 master01 6 192.168.37.156 slave01 7 192.168.37.157 slave02 8 192.168.37.158 slave03
1 #重新启动服务器 2 reboot
注意:
另外两台Slave虚拟机的主机名和IP地址分别为:
slave02 192.168.37.157
slave03 192.168.37.158
集群中每台虚拟机的网络配置完成之后就可以使用SecureCRT或XShell连接上去操作了
3、关闭所有虚拟机的防火墙和Selinux
1 [root@master01 ~]# service iptables stop 2 [root@master01 software]# chkconfig iptables --level 35 off 3 [root@master01 ~]# setenforce 0 4 [root@master01 ~]# getenforce 5 Permissive 6 [root@master01 ~]# grep -Ev "^#|^$" /etc/sysconfig/selinux 7 SELINUX=disabled 8 SELINUXTYPE=targeted
注意:
如果使用上述方法SELINUX还是为enforcing,则可以使用以下方法解决:
修改/etc/selinux/config 文件
将SELINUX=enforcing改为SELINUX=disabled
重启机器即可
4、配置ntp服务
A、Master机器:
1 #校正Master机器的时间 2 [root@master01 ~]# date -s "2017-07-16 16:17:50" 3 2017年 07月 16日 星期日 16:17:50 CST 4 [root@master01 ~]# clock -w 5 [root@master01 ~]# date '+%F %T %A' 6 2017-07-16 16:20:31 星期日
1 #检查是否安装ntp组件,若未安装则执行yum install -y ntp安装它 2 [root@master01 ~]# rpm -qa|grep ntp 3 fontpackages-filesystem-1.41-1.1.el6.noarch 4 ntp-4.2.4p8-3.el6.centos.x86_64 5 ntpdate-4.2.4p8-3.el6.centos.x86_64
1 #配置时间服务 2 [root@master01 ~]# grep -Ev "^#|^$" /etc/ntp.conf 3 #server 0.centos.pool.ntp.org 4 #server 1.centos.pool.ntp.org 5 #server 2.centos.pool.ntp.org 6 driftfile /var/lib/ntp/drift 7 restrict default kod nomodify notrap nopeer noquery 8 restrict -6 default kod nomodify notrap nopeer noquery 9 restrict 127.0.0.1 10 restrict -6 ::1 11 server 127.127.1.0 12 fudge 192.168.37.154 stratum 0 13 includefile /etc/ntp/crypto/pw 14 keys /etc/ntp/keys
1 #启动时间服务 2 service ntpd start
1 #查看ntp端口123是否启动 2 [root@master01 ~]# lsof -i:123 3 COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME 4 ntpd 1694 ntp 16u IPv4 14631 0t0 UDP *:ntp 5 ntpd 1694 ntp 17u IPv6 14632 0t0 UDP *:ntp 6 ntpd 1694 ntp 18u IPv6 14636 0t0 UDP localhost:ntp 7 ntpd 1694 ntp 19u IPv6 14637 0t0 UDP [fe80::20c:29ff:feaf:886]:ntp 8 ntpd 1694 ntp 20u IPv4 14638 0t0 UDP localhost:ntp 9 ntpd 1694 ntp 21u IPv4 14639 0t0 UDP 192.168.37.154:ntp
1 #查看时间服务器配置 2 [root@master01 ~]# ntpq -p 3 remote refid st t when poll reach delay offset jitter 4 ============================================================================== 5 *LOCAL(0) .LOCL. 5 l 30 64 377 0.000 0.000 0.000
B、Slave机器:
1 #首先校正Slave机器的时间 2 [root@slave01 ~]# date -s "2017-07-16 16:17:50" 3 2017年 07月 16日 星期日 16:17:50 CST 4 [root@slave01 ~]# clock -w 5 [root@slave01 ~]# date '+%F %T %A' 6 2017-07-16 16:20:31 星期日
1 #手动同步一次时间 2 [root@slave01 ~]# ntpdate 192.168.37.154 3 16 Jul 16:39:18 ntpdate[1834]: step time server 192.168.37.154 offset 23.818568 sec
1 #修改配置文件 2 [root@slave01 ~]# grep -Ev "^#|^$" /etc/ntp.conf 3 #server 0.centos.pool.ntp.org 4 #server 1.centos.pool.ntp.org 5 #server 2.centos.pool.ntp.org 6 driftfile /var/lib/ntp/drift 7 restrict default kod nomodify notrap nopeer noquery 8 restrict -6 default kod nomodify notrap nopeer noquery 9 restrict 127.0.0.1 10 restrict -6 ::1 11 server 192.168.37.154 12 fudge 192.168.37.156 stratum 10 13 includefile /etc/ntp/crypto/pw 14 keys /etc/ntp/keys
1 #启动时间服务 2 service ntpd start
1 #查看ntp端口123是否启动 2 [root@master01 ~]# lsof -i:123 3 COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME 4 ntpd 1694 ntp 16u IPv4 14631 0t0 UDP *:ntp 5 ntpd 1694 ntp 17u IPv6 14632 0t0 UDP *:ntp 6 ntpd 1694 ntp 18u IPv6 14636 0t0 UDP localhost:ntp 7 ntpd 1694 ntp 19u IPv6 14637 0t0 UDP [fe80::20c:29ff:feaf:886]:ntp 8 ntpd 1694 ntp 20u IPv4 14638 0t0 UDP localhost:ntp 9 ntpd 1694 ntp 21u IPv4 14639 0t0 UDP 192.168.37.154:ntp
1 #查看客户端的配置 2 [root@slave01 ~]# ntpq -p 3 remote refid st t when poll reach delay offset jitter 4 ============================================================================== 5 192.168.37.154 LOCAL(0) 6 u 44 64 7 0.756 -0.646 0.122
1 #查看Slave端的同步状态 2 [root@slave01 ~]# ntpstat 3 unsynchronised 4 time server re-starting 5 polling server every 64 s
注意:参照slave01节点的时间服务配置将另外两台Slave节点配置好
5、配置Hadoop用户的SSH免密码登录
从官方给出的建议来看只需要Master到各个Slave节点之间实现单向免密码登录即可
A、Master机器
1 #切换到Hadoop用户下 2 [root@master01 software]# su -l hadoop
1 #生成无密码的公私密匙 2 [hadoop@master01 ~]$ ssh-keygen -t rsa -P '' 3 Generating public/private rsa key pair. 4 Enter file in which to save the key (/home/hadoop/.ssh/id_rsa): 5 Created directory '/home/hadoop/.ssh'. 6 Your identification has been saved in /home/hadoop/.ssh/id_rsa. 7 Your public key has been saved in /home/hadoop/.ssh/id_rsa.pub. 8 The key fingerprint is: 9 3a:3c:57:36:31:67:f4:d8:2f:df:f9:88:47:45:c1:26 hadoop@master01 10 The key's randomart image is: 11 +--[ RSA 2048]----+ 12 | . ...| 13 | . E o.| 14 | o + =. | 15 | = ..| 16 | S + ...| 17 | . . o . .oo| 18 | = . . .o| 19 | + ....| 20 | ... .| 21 +-----------------+
1 #首先拷贝公匙给自己实现无密码登录自己(使用ssh-copy-id拷贝公匙可以直接将其追加到authorized_keys文件且自动设置authorized_keys文件权限为600、.ssh文件夹权限为700) 2 [hadoop@master01 ~]$ ssh-copy-id master01 3 The authenticity of host 'master01 (192.168.37.154)' can't be established. 4 RSA key fingerprint is 82:72:60:05:6d:dc:3e:bf:f7:aa:2d:f5:08:c1:59:3a. 5 Are you sure you want to continue connecting (yes/no)? yes 6 Warning: Permanently added 'master01,192.168.37.154' (RSA) to the list of known hosts. 7 hadoop@master01's password: 8 Now try logging into the machine, with "ssh 'master01'", and check in: 9 10 .ssh/authorized_keys 11 12 to make sure we haven't added extra keys that you weren't expecting.
1 #测试无密码登录自己 2 [hadoop@master01 ~]$ ssh master01 3 [hadoop@master01 ~]$ exit 4 logout 5 Connection to master01 closed.
1 #拷贝公匙到第1台Slave 2 [hadoop@master01 ~]$ ssh-copy-id slave01 3 The authenticity of host 'slave01 (192.168.37.156)' can't be established. 4 RSA key fingerprint is 82:72:60:05:6d:dc:3e:bf:f7:aa:2d:f5:08:c1:59:3a. 5 Are you sure you want to continue connecting (yes/no)? yes 6 Warning: Permanently added 'slave01,192.168.37.156' (RSA) to the list of known hosts. 7 hadoop@slave01's password: 8 Now try logging into the machine, with "ssh 'slave01'", and check in: 9 10 .ssh/authorized_keys 11 12 to make sure we haven't added extra keys that you weren't expecting.
1 #测试免密码登录第1台Slave 2 [hadoop@master01 ~]$ ssh slave01
1 #回退到当前环境 2 [hadoop@slave01 ~]$ exit 3 logout 4 Connection to slave01 closed.
注意:参照上面的操作继续将生成公匙拷贝到另外两台Slave机器上
B、Slave机器(双向免密码通信需要配置各个Slave节点):
#可以不用配置(略)
二、搭建Hadoop集群:
1、Master机器:
A、上传并解压
1 [hadoop@master01 install]# ls 2 hadoop-2.7.3.tar.gz initNetwork.sh 3 [hadoop@master01 install]$ tar -zxvf hadoop-2.7.3.tar.gz -C /software/ 4 [hadoop@master01 install]$ ls /software/ 5 hadoop-2.7.3 jdk1.7.0_79 6 [hadoop@master01 install]$ rm -rf hadoop-2.7.3.tar.gz
B、配置环境变量
1 [hadoop@master01 install]$ su -lc "vi /etc/profile" 2 密码:123456 3 JAVA_HOME=/software/jdk1.7.0_79 4 HADOOP_HOME=/software/hadoop-2.7.3 5 PATH=$PATH:$JAVA_HOME/bin:$JAVA_HOME/lib:$HADOOP_HOME/bin:$HADOOP_HOME/sbin 6 export PATH JAVA_HOME HADOOP_HOME 7 [hadoop@master01 hadoop-2.7.3]$ source /etc/profile
C、配置Hadoop环境
1 [hadoop@master01 hadoop-2.7.3]$ cd /software/hadoop-2.7.3/etc/hadoop/ 2 #配置hadoop-env.sh(注意必须配置,即便/etc/profile中配置了也要配置) 3 [hadoop@master01 hadoop]$ vi hadoop-env.sh 4 export JAVA_HOME=/software/jdk1.7.0_79
1 #配置core-site.xml 2 [hadoop@master01 hadoop]$ vi core-site.xml 3 <configuration> 4 <property> 5 <name>fs.defaultFS</name> 6 <value>hdfs://master01:9000</value> 7 </property> 8 <property> 9 <name>hadoop.tmp.dir</name> 10 <value>/software/hadoop-2.7.3/work</value> 11 </property> 12 </configuration>
1 #配置hdfs-site.xml 2 [hadoop@master01 hadoop]$ vi hdfs-site.xml 3 <configuration> 4 <property> 5 <name>dfs.replication</name> 6 <value>3</value> 7 </property> 8 </configuration>
1 #配置mapred-site.xml 2 [hadoop@master01 hadoop]$ mv mapred-site.xml.template mapred-site.xml 3 [hadoop@master01 hadoop]$ vi mapred-site.xml 4 <configuration> 5 <property> 6 <name>mapreduce.framework.name</name> 7 <value>yarn</value> 8 </property> 9 </configuration>
1 #配置yarn-site.xml 2 <configuration> 3 <property> 4 <name>yarn.resourcemanager.hostname</name> 5 <value>master01</value> 6 </property> 7 <property> 8 <name>yarn.nodemanager.aux-services</name> 9 <value>mapreduce_shuffle</value> 10 </property> 11 </configuration>
1 #配置slaves 2 [hadoop@master01 hadoop]$ vi slaves 3 slave01 4 slave02 5 slave03
2、Slave机器:
1 #将Master机器上Hadoop的安装目录拷贝到各个Slave节点上 2 [hadoop@master01 software]$ cd /software/ 3 [hadoop@master01 software]$ scp -r hadoop-2.7.3 slave01:/software/ 4 [hadoop@master01 software]$ scp -r hadoop-2.7.3 slave02:/software/ 5 [hadoop@master01 software]$ scp -r hadoop-2.7.3 slave03:/software/
1 #将Master机器上/etc/profile文件拷贝到各个Slave节点上 2 [root@master01 ~]# scp /etc/profile slave01:/etc/ 3 [root@master01 ~]# scp /etc/profile slave02:/etc/ 4 [root@master01 ~]# scp /etc/profile slave03:/etc/ 5 #在各个Slave节点上执行source指令使配置生效 6 [hadoop@slave01 ~]$ source /etc/profile
3、启动HDFS集群并测试
1 #查看hadoop的版本 2 [hadoop@master01 software]$ hadoop version 3 Hadoop 2.7.3 4 Subversion https://git-wip-us.apache.org/repos/asf/hadoop.git -r baa91f7c6bc9cb92be5982de4719c1c8af91ccff 5 Compiled by root on 2016-08-18T01:41Z 6 Compiled with protoc 2.5.0 7 From source with checksum 2e4ce5f957ea4db193bce3734ff29ff4 8 This command was run using /software/hadoop-2.7.3/share/hadoop/common/hadoop-common-2.7.3.jar
1 #格式化文件系统 2 [hadoop@master01 software]$ hdfs namenode -format 3 注意:当出现下面的退出状态值为0时标识格式化成功 4 INFO util.ExitUtil: Exiting with status 0
1 #启动HDFS集群 2 [hadoop@master01 hadoop]$ start-dfs.sh 3 Starting namenodes on [master01] 4 master01: starting namenode, logging to /software/hadoop-2.7.3/logs/hadoop-hadoop-namenode-master01.out 5 slave01: starting datanode, logging to /software/hadoop-2.7.3/logs/hadoop-hadoop-datanode-slave01.out 6 slave02: starting datanode, logging to /software/hadoop-2.7.3/logs/hadoop-hadoop-datanode-slave02.out 7 slave03: starting datanode, logging to /software/hadoop-2.7.3/logs/hadoop-hadoop-datanode-slave03.out 8 Starting secondary namenodes [0.0.0.0] 9 0.0.0.0: starting secondarynamenode, logging to /software/hadoop-2.7.3/logs/hadoop-hadoop-secondarynamenode-master01.out 10 [hadoop@master01 hadoop]$ 11 [hadoop@master01 hadoop]$
1 #检查进程NN进程 2 [hadoop@master01 hadoop]$ jps 3 5640 NameNode 4 5847 SecondaryNameNode 5 5974 Jps
1 #分别在各个从节点上检查DN进程 2 [hadoop@slave01 software]$ jps 3 3032 DataNode 4 3111 Jps 5 [hadoop@slave02 software]$ jps 6 3181 Jps 7 3102 DataNode 8 [hadoop@slave03 software]$ jps 9 3119 Jps 10 3040 DataNode
#HDFS分布式文件系统简单命令行操作
#查看分布式系统根目录
[hadoop@master01 hadoop]$ hdfs dfs -ls /
#创建单级目录
[hadoop@master01 hadoop]$ hdfs dfs -mkdir /bnyw
1 #创建多级目录 2 [hadoop@master01 hadoop]$ hdfs dfs -mkdir -p /test/lyf 3 [hadoop@master01 hadoop]$ hdfs dfs -ls / 4 Found 2 items 5 drwxr-xr-x - hadoop supergroup 0 2017-07-16 23:22 /bnyw 6 drwxr-xr-x - hadoop supergroup 0 2017-07-16 23:24 /test
1 #删除单级目录、多级目录及其文件 2 [hadoop@master01 hadoop]$ hdfs dfs -rm -r /test/lyf 3 17/07/16 23:28:25 INFO fs.TrashPolicyDefault: Namenode trash configuration: Deletion interval = 0 minutes, Emptier interval = 0 minutes. 4 Deleted /test/lyf
1 #上传文件 2 [hadoop@master01 install]$ ls /install/ 3 initNetwork.sh test.sh 4 [hadoop@master01 install]$ hdfs dfs -put initNetwork.sh /test/ 5 [hadoop@master01 install]$ hdfs dfs -ls /test/ 6 Found 1 items 7 -rw-r--r-- 3 hadoop supergroup 492 2017-07-16 23:33 /test/initNetwork.sh
1 #查看文件内容 2 [hadoop@master01 install]$ hdfs dfs -cat /test/initNetwork.sh 3 #!/bin/sh 4 declare -a checkEth0=`ifconfig -a|awk '/eth0/{print $1}'` 5 declare -a eth0Device='/etc/sysconfig/network-scripts/ifcfg-eth0' 6 declare -a deviceFile='/etc/udev/rules.d/70-persistent-net.rules' 7 [ -z $checkEth0 ] && { 8 [ -f $deviceFile ] && printf "">$deviceFile 9 [ -f $eth0Device ] && sed -ri -e "/HWADDR/d" -e "/UUID/d" $eth0Device 10 echo "init eth0 device finished....." 11 echo "reboot current unix system....." 12 sleep 1 13 reboot 14 } || echo "eth0 already exists,no require init..."
1 #下载文件 2 [hadoop@master01 install]$ mkdir -p test && cd test 3 [hadoop@master01 test]$ pwd 4 /install/test 5 [hadoop@master01 test]$ hdfs dfs -get /test/initNetwork.sh /install/test/ 6 [hadoop@master01 test]$ ls /install/test/ 7 initNetwork.sh
1 #删除文件 2 [hadoop@master01 test]$ hdfs dfs -rm -r /test/initNetwork.sh 3 17/07/16 23:40:24 INFO fs.TrashPolicyDefault: Namenode trash configuration: Deletion interval = 0 minutes, Emptier interval = 0 minutes. 4 Deleted /test/initNetwork.sh 5 [hadoop@master01 test]$ hdfs dfs -ls /test/
1 #拷贝文件 2 [hadoop@CloudDeskTop hive]$ hdfs dfs -mkdir -p /test/test1 3 [hadoop@CloudDeskTop hive]$ hdfs dfs -mkdir -p /test/test2 4 [hadoop@CloudDeskTop hive]$ hdfs dfs -put user02 /test/test1 5 [hadoop@CloudDeskTop hive]$ hdfs dfs -ls /test/test1 6 Found 1 items 7 -rw-r--r-- 3 hadoop supergroup 53 2017-07-30 03:27 /test/test1/user02 8 [hadoop@CloudDeskTop hive]$ hdfs dfs -cp /test/test1/user02 /test/test2 9 [hadoop@CloudDeskTop hive]$ hdfs dfs -ls /test/test2 10 Found 1 items 11 -rw-r--r-- 3 hadoop supergroup 53 2017-07-30 03:29 /test/test2/user02 12 [hadoop@CloudDeskTop hive]$ hdfs dfs -cat /test/test2/user02 13 20 chenhualiang 90 14 21 chaochunhong 80 15 22 guaxixi 100
#重命名或移动文件
1 #重命名操作 2 [hadoop@CloudDeskTop hive]$ hdfs dfs -mv /test/test2/user02 /test/test2/user03 3 [hadoop@CloudDeskTop hive]$ hdfs dfs -ls /test/test2/ 4 Found 1 items 5 -rw-r--r-- 3 hadoop supergroup 53 2017-07-30 03:29 /test/test2/user03 6 [hadoop@CloudDeskTop hive]$ hdfs dfs -cat /test/test2/user03 7 20 chenhualiang 90 8 21 chaochunhong 80 9 22 guaxixi 100
1 #移动文件操作 2 [hadoop@CloudDeskTop hive]$ hdfs dfs -mv /test/test2/user03 /test/test1 3 [hadoop@CloudDeskTop hive]$ hdfs dfs -ls /test/test2/ 4 [hadoop@CloudDeskTop hive]$ hdfs dfs -ls /test/test1/ 5 Found 2 items 6 -rw-r--r-- 3 hadoop supergroup 53 2017-07-30 03:27 /test/test1/user02 7 -rw-r--r-- 3 hadoop supergroup 53 2017-07-30 03:29 /test/test1/user03
1 #查看HDFS分布式系统的整体运行工况 2 [hadoop@master01 test]$ hdfs dfsadmin -report 3 Configured Capacity: 128831840256 (119.98 GB) 4 Present Capacity: 114848382976 (106.96 GB) 5 DFS Remaining: 114848260096 (106.96 GB) 6 DFS Used: 122880 (120 KB) 7 DFS Used%: 0.00% 8 Under replicated blocks: 0 9 Blocks with corrupt replicas: 0 10 Missing blocks: 0 11 Missing blocks (with replication factor 1): 0 12 13 ------------------------------------------------- 14 Live datanodes (3): 15 16 Name: 192.168.37.158:50010 (slave03) 17 Hostname: slave03 18 Decommission Status : Normal 19 Configured Capacity: 42943946752 (39.99 GB) 20 DFS Used: 40960 (40 KB) 21 Non DFS Used: 4660379648 (4.34 GB) 22 DFS Remaining: 38283526144 (35.65 GB) 23 DFS Used%: 0.00% 24 DFS Remaining%: 89.15% 25 Configured Cache Capacity: 0 (0 B) 26 Cache Used: 0 (0 B) 27 Cache Remaining: 0 (0 B) 28 Cache Used%: 100.00% 29 Cache Remaining%: 0.00% 30 Xceivers: 1 31 Last contact: Sun Jul 16 23:49:36 CST 2017 32 33 34 Name: 192.168.37.156:50010 (slave01) 35 Hostname: slave01 36 Decommission Status : Normal 37 Configured Capacity: 42943946752 (39.99 GB) 38 DFS Used: 40960 (40 KB) 39 Non DFS Used: 4662685696 (4.34 GB) 40 DFS Remaining: 38281220096 (35.65 GB) 41 DFS Used%: 0.00% 42 DFS Remaining%: 89.14% 43 Configured Cache Capacity: 0 (0 B) 44 Cache Used: 0 (0 B) 45 Cache Remaining: 0 (0 B) 46 Cache Used%: 100.00% 47 Cache Remaining%: 0.00% 48 Xceivers: 1 49 Last contact: Sun Jul 16 23:49:39 CST 2017 50 51 52 Name: 192.168.37.157:50010 (slave02) 53 Hostname: slave02 54 Decommission Status : Normal 55 Configured Capacity: 42943946752 (39.99 GB) 56 DFS Used: 40960 (40 KB) 57 Non DFS Used: 4660391936 (4.34 GB) 58 DFS Remaining: 38283513856 (35.65 GB) 59 DFS Used%: 0.00% 60 DFS Remaining%: 89.15% 61 Configured Cache Capacity: 0 (0 B) 62 Cache Used: 0 (0 B) 63 Cache Remaining: 0 (0 B) 64 Cache Used%: 100.00% 65 Cache Remaining%: 0.00% 66 Xceivers: 1 67 Last contact: Sun Jul 16 23:49:38 CST 2017
4、启动Yarn集群并测试
1 #启动yarn集群 2 [hadoop@master01 test]$ start-yarn.sh 3 starting yarn daemons 4 starting resourcemanager, logging to /software/hadoop-2.7.3/logs/yarn-hadoop-resourcemanager-master01.out 5 slave02: starting nodemanager, logging to /software/hadoop-2.7.3/logs/yarn-hadoop-nodemanager-slave02.out 6 slave01: starting nodemanager, logging to /software/hadoop-2.7.3/logs/yarn-hadoop-nodemanager-slave01.out 7 slave03: starting nodemanager, logging to /software/hadoop-2.7.3/logs/yarn-hadoop-nodemanager-slave03.out
1 #在Master节点上查看RM进程 2 [hadoop@master01 test]$ jps 3 7297 Jps 4 5640 NameNode 5 5847 SecondaryNameNode 6 7032 ResourceManager
1 #在其它Slave节点上查看NM进程 2 [hadoop@slave01 software]$ jps 3 3032 DataNode 4 3178 NodeManager 5 3291 Jps 6 [hadoop@slave02 software]$ jps 7 3250 NodeManager 8 3102 DataNode 9 3363 Jps 10 [hadoop@slave03 software]$ jps 11 3299 Jps 12 3040 DataNode 13 3186 NodeManager
5、WebPortal端监控
1 #查看HDFS集群状态信息 2 http://192.168.37.154:50070
1 #查看Yarn集群状态信息 2 http://192.168.37.154:8088
1 #停止Yarn集群 2 [hadoop@master01 test]$ stop-yarn.sh 3 stopping yarn daemons 4 stopping resourcemanager 5 slave02: stopping nodemanager 6 slave01: stopping nodemanager 7 slave03: stopping nodemanager 8 no proxyserver to stop
1 #停止HDFS集群 2 [hadoop@master01 test]$ stop-dfs.sh 3 Stopping namenodes on [master01] 4 master01: stopping namenode 5 slave03: stopping datanode 6 slave02: stopping datanode 7 slave01: stopping datanode 8 Stopping secondary namenodes [0.0.0.0] 9 0.0.0.0: stopping secondarynamenode 10 [hadoop@master01 test]$ jps 11 7972 Jps 12 [hadoop@slave01 software]$ jps 13 3432 Jps