1.基本信息
| 版本 | 2.7.3 |
|---|---|
| 安装机器 | 三台机器 |
| 账号 | hadoop |
| 源路径 | /opt/software/hadoop-2.7.3.tar.gz |
| 目标路径 | /opt/hadoop -> /opt/hadoop-2.7.3 |
| 依赖关系 | 无 |
2.安装过程
1).切换到hadoop账户,通过tar -zxvf命令将hadoop解压缩至目的安装目录:
[root@bgs-5p173-wangwenting opt]# su hadoop[hadoop@bgs-5p173-wangwenting opt]$ cd /opt/software[hadoop@bgs-5p173-wangwenting software]$ tar -zxvf hadoop-${version}.tar.gz -C /opt[hadoop@bgs-5p173-wangwenting software]$ cd /opt[hadoop@bgs-5p173-wangwenting opt]$ ln -s /opt/hadoop-${version} /opt/hadoop |
2).创建tmpdir目录:
[hadoop@bgs-5p173-wangwenting opt]$ cd /opt/hadoop[hadoop@bgs-5p173-wangwenting hadoop]$ mkdir -p tmpdir |
3).配置hadoop-env.sh文件:
[hadoop@bgs-5p173-wangwenting hadoop]$ cd /opt/hadoop/etc/hadoop/[hadoop@bgs-5p173-wangwenting hadoop]$ mkdir -p /opt/hadoop/pids[hadoop@bgs-5p173-wangwenting hadoop]$ vi hadoop-env.sh在hadoop-env.sh文件中添加如下配置:export JAVA_HOME=/opt/javaexport HADOOP_PID_DIR=/opt/hadoop/pids |
4.配置mapred-env.sh文件:
[hadoop@bgs-5p173-wangwenting hadoop]$ cd /opt/hadoop/etc/hadoop/[hadoop@bgs-5p173-wangwenting hadoop]$ vim mapred-env.sh在mapred-env.sh文件中添加如下配置:export JAVA_HOME=/opt/java |
5.配置core-site.xml文件
[hadoop@bgs-5p173-wangwenting hadoop]$ cd /opt/hadoop/etc/hadoop/[hadoop@bgs-5p173-wangwenting hadoop]$ vim core-site.xml在core-site.xml文件中添加如下配置:<configuration><property>//namenode的临时工作目录 <name>hadoop.tmp.dir</name> <value>/opt/hadoop/tmpdir</value> </property><property>//hdfs的入口,告诉namenode在那个机器上,端口号是什么。 <name>fs.defaultFS</name> <value>hdfs://bgs-5p173-wangwenting:9000</value> </property> <property> <name>io.file.buffer.size</name> <value>131072</value> </property> <property> <name>fs.trash.interval</name> <value>1440</value> </property></configuration> |
6.配置hdfs-site.xml文件
[hadoop@bgs-5p173-wangwenting hadoop]$ cd /opt/hadoop/etc/hadoop/[hadoop@bgs-5p173-wangwenting hadoop]$ vim hdfs-site.xml在hdfs-site.xml文件中添加如下配置:<configuration><property>//副本数量,一般是小于等于datanode的数量, <name>dfs.replication</name> <value>2</value> </property> <property> <name>dfs.namenode.name.dir</name> <value>file:/opt/hadoop/tmpdir/dfs/name</value> </property> <property> <name>dfs.datanode.name.dir</name> <value>file:/opt/hadoop/tmpdir/dfs/data</value> </property> <property> <name>dfs.webhdfs.enabled</name> <value>true</value> </property><property> <name>dfs.secondary.http.address</name> <value>bgs-5p173-wangwenting:50090</value> </property></configuration> |
7.配置mapred-site.xml文件
[hadoop@bgs-5p173-wangwenting hadoop]$ cd /opt/hadoop/etc/hadoop/[hadoop@bgs-5p173-wangwenting hadoop]$ vi mapred-site.xml在mapred-site.xml文件中添加如下配置:<configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> <property> <name>mapreduce.jobhistory.address</name> <value>bgs-5p173-wangwenting:10020</value> </property> <property> <name>mapreduce.jobhistory.webapp.address</name> <value>bgs-5p173-wangwenting:19888</value> </property></configuration> |
8.配置yarn-site.xml文件:
[hadoop@bgs-5p173-wangwenting hadoop]$ cd /opt/hadoop/etc/hadoop/[hadoop@bgs-5p173-wangwenting hadoop]$ vim yarn-site.xml在yarn-site.xml文件中添加如下配置:<configuration> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <property> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> </property> <property> <name>yarn.resourcemanager.scheduler.address</name> <value>bgs-5p173-wangwenting:8030</value> </property> <property> <name>yarn.resourcemanager.resource-tracker.address</name> <value>bgs-5p173-wangwenting:8031</value> </property> <property> <name>yarn.resourcemanager.address</name> <value>bgs-5p173-wangwenting:8032</value> </property> <property> <name>yarn.resourcemanager.admin.address</name> <value>bgs-5p173-wangwenting:8033</value> </property> <property> <name>yarn.resourcemanager.webapp.address</name> <value>bgs-5p173-wangwenting:8088</value> </property><!-- Site specific YARN configuration properties --></configuration> |
9.配置hadoop运行的环境变量
[hadoop@bgs-5p173-wangwenting hadoop]$ vim /etc/profileexport HADOOP_HOME=/opt/hadoopexport PATH=$HADOOP_HOME/bin:$PATH配置成功后,执行source /etc/profile使配置生效[hadoop@bgs-5p173-wangwenting hadoop]$ source /etc/profile |
10.修改slaves文件:
[hadoop@bgs-5p173-wangwenting hadoop]$ cd /opt/hadoop/etc/hadoop[hadoop@bgs-5p173-wangwenting hadoop]$ vim slaves在slaves文件中添加//datanode的节点的位置bgs-5p173-wangwentingbgs-5p174-wangwentingbgs-5p175-wangwenting |
11.在bgs-5p173-wangwenting上复制hadoop-2.7.3到hadoop@bgs-5p174-wangwenting和hadoop@bgs-5p174-wangwenting机器并按照步骤9修改环境变量并执行以下操作:
[hadoop@bgs-5p173-wangwenting hadoop]$ scp -r /opt/hadoop-${version} hadoop@bgs-5p174-wangwenting:/opt/[hadoop@bgs-5p173-wangwenting hadoop]$ ln -s /opt/hadoop-${version} /opt/hadoop[hadoop@bgs-5p173-wangwenting hadoop]$ scp -r /opt/hadoop-${version} hadoop@bgs-5p175-wangwenting:/opt/[hadoop@bgs-5p173-wangwenting hadoop]$ ln -s /opt/hadoop-${version} /opt/hadoop |
12.格式化namenode,启动hadoop,并启动jobhistory服务:
[hadoop@bgs-5p173-wangwenting hadoop]$ hadoop namenode -format[hadoop@bgs-5p173-wangwenting hadoop]$ ${HADOOP_HOME}/sbin/start-all.sh[hadoop@bgs-5p173-wangwenting hadoop]$ ${HADOOP_HOME}/sbin/mr-jobhistory-daemon.sh start historyserverstart-all.sh包含dfs和yarn两个模块的启动,分别为start-dfs.sh 、 start-yarn.sh,所以dfs和yarn可以单独启动。注意:如果datanode没有启动起来,看看是不是tmpdir中有之前的脏数据,删除这个目录其他两台机器也要删除。 |
13.检查每台机器的服务,hadoop@bgs-5p173-wangwenting、hadoop@bgs-5p174-wangwenting、hadoop@bgs-5p175-wangwenting三台机器上分别输入jps:
[hadoop@bgs-5p173-wangwenting ~]$ jps24429 Jps22898 ResourceManager23229 NodeManager24383 JobHistoryServer22722 SecondaryNameNode22488 NameNode6945 DataNode[ahdoop@bgs-5p174-wangwenting ~]$ jps7650 DataNode7788 NodeManager8018 Jps[hadoop@bgs-5p175-wangwenting ~]$ jps28407 Jps28038 DataNode28178 NodeManager如果三台机器正常输出上述内容,则表示hadoop集群的服务正常工作。 |
访问hadoop的服务页面:在浏览器中输入如下地址
http://bgs-5p173-wangwenting:8088
http://bgs-5p173-wangwenting:50070
http://bgs-5p173-wangwenting:19888
如能正常反应,则表明上述三类服务正常运行.
