zoukankan      html  css  js  c++  java
  • hadoop3.0.0 分布式集群安装过程

    1. 环境

    主机名  ip  角色

    hdp01  192.168.184.61  ResourceManager/NameNode/SecondaryNameNode

    hdp02  192.168.184.62  NodeManager/DataNode

    hdp03  192.168.184.63  NodeManager/DataNode

    (1个namenode 2个datanode)

    2.准备工作

    • windows10(物理机系统)
    • VMware12 workstation(虚拟机软件)
    • centos7.0(虚拟机系统)
    • hadoop3.0.0
    • jdk1.8
    • SecureCRT

    3.配置IP/hostname及SSH免密码登录/hosts

    3.1 vi /etc/sysconfig/network-scripts/ifcfg-eth0(删掉UUID  HWADDR)

    rm -rf  /etc/udev/rules.d/70-persistent-net.rules(删掉eth0,将eth1改为eth0)

    ---修改虚拟网卡,删掉旧的

    service network restart(重启网关)

    3.2 vi /etc/sysconfig/network

    HOSTNAME = hdp-server01(192.168.184.61)

    HOSTNAME = hdp-server02(192.168.184.62)

    HOSTNAME = hdp-server03(192.168.184.63)

    3.3 ssh-keygen 

    ssh-copy-id hdp01

    ssh-copy-id hdp02

    ssh-copy-id hdp03

    3.4 vi /etc/hosts

    192.168.184.61 master

    192.168.184.62 node1

    192.168.184.63 node2

    4.关闭防火墙

    service iptables stop(关闭防火墙)
    chkconfig iptables --list(检查防火墙启动状态)
    chkconfig iptables off(关闭防火墙启动)
    chkconfig iptables status(查看防火墙状态)

    5.安装JDK

    tar -zxvf ....tar.gz -C ./app/

    vi /etc/profile

    export JAVA_HOME=/usr/local/jdk1.8.0_11

    export PATH=$PATH:$JAVA_HOME/bin

    source /etc/profile

    验证:

    echo $JAVA_HOME
    java -version

    6.配置hadoop

    6.1 hadoop-env.sh

    export JAVA_HOME=/usr/local/jdk1.8.0_11
    export HDFS_NAMENODE_USER=root
    export HDFS_DATANODE_USER=root
    export HDFS_JOURNALNODE_USER=root
    export YARN_RESOURCEMANAGER_USER=root
    export YARN_NODEMANAGER_USER=root
    export HDFS_SECONDARYNAMENODE_USER=root

    6.2 core-site.xml

    <configuration>
    <property>
    <name>fs.defaultFS</name>
    <value>hdfs://hdp01:9000</value>
    </property>

    <property>
    <name>hadoop.tmp.dir</name>
    <value>/usr/local/hadoop/tmp</value>
    </property>
    </configuration>

    6.3 hdfs-site.xml

    <configuration>
    <property>
    <name>dfs.replication</name>
    <value>2</value>
    </property>
    <property>
    <name>dfs.namenode.name.dir</name>
    <value>/usr/local/hadoop/hdfs/name</value>
    </property>
    <property>
    <name>dfs.datanode.data.dir</name>
    <value>/usr/local/hadoop/hdfs/data</value>
    </property>
    <property>
    <name>dfs.namenode.secondary.http-address</name>
    <value>hdp01:9001</value>
    </property>
    <property>
    <name>dfs.http.address</name>
    <value>0.0.0.0:50070</value>
    </property>
    </configuration>

    6.4 mapred-site.xml

    <configuration>
    <property>
    <name>mapred.job.tracker.http.address</name>
    <value>0.0.0.0:50030</value>
    </property>
    <property>
    <name>mapred.task.tracker.http.address</name>
    <value>0.0.0.0:50060</value>
    </property>

    <property>
    <name>mapreduce.framework.name</name>
    <value>yarn</value>
    </property>

    <property>
    <name>mapreduce.application.classpath</name>
    <value>
    /usr/local/hadoop/etc/hadoop,
    /usr/local/hadoop/share/hadoop/common/*,
    /local/hadoop/share/hadoop/common/lib/*,
    /usr/local/hadoop/share/hadoop/hdfs/*,
    /usr/local/hadoop/share/hadoop/hdfs/lib/*,
    /usr/local/hadoop/share/hadoop/mapreduce/*,
    /usr/local/hadoop/share/hadoop/mapreduce/lib/*,
    /usr/local/hadoop/share/hadoop/yarn/*,
    /usr/local/hadoop/share/hadoop/yarn/lib/*
    </value>
    </property>
    </configuration>

    6.5 workers

    hdp02
    hdp03

    6.6 yarn-site.xml

    <configuration>

    <!-- Site specific YARN configuration properties -->

    <property>
    <name>yarn.nodemanager.aux-services</name>
    <value>mapreduce_shuffle</value>
    </property>
    <property>
    <name>yarn.resourcemanager.address</name>
    <value>hdp01:8032</value>
    </property>
    <property>
    <name>yarn.resourcemanager.scheduler.address</name>
    <value>hdp01:8030</value>
    </property>
    <property>
    <name>yarn.resourcemanager.resource-tracker.address</name>
    <value>hdp01:8031</value>
    </property>
    <property>
    <name>yarn.resourcemanager.admin.address</name>
    <value>hdp01:8033</value>
    </property>
    <property>
    <name>yarn.resourcemanager.webapp.address</name>
    <value>hdp01:8088</value>
    </property>

    </configuration>

    6.7 配置环境变量

    vi /etc/profile

    export JAVA_HOME=/usr/local/jdk1.8.0_11
    export HADOOP_HOME=/usr/local/hadoop

    export PATH=$PATH:$JAVA_HOME/bin

    export PATH=$PATH:$HADOOP_HOME/bin

    export PATH=$PATH:$HADOOP_HOME/sbin

    7.启动hadoop守护进程及检查进程启动情况

    格式化:

    hadoop namenode -format

    start-dfs.sh

    start-yarn.sh

    ui:http://master:50070

    yarn:http://master:8088

    正常启动节点情况:

    8598 ResourceManager
    8343 SecondaryNameNode
    8077 NameNode

    5654 Jps
    4759 DataNode
    4877 NodeManager

    4503 Jps
    3578 DataNode
    3695 NodeManager

    小case 验证:

    hadoop fs -put ~/xxx.log hdfs://hdp01:9000/data/(hdfs文件路径)

    hadoop jar /usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-3.0.3.jar wordcount /data/wordcount /output/

    时间不同步:date -s "2018-08-13 17:05:08"

    查看启动进程 
    /usr/local/jdk/jdk1.8.0_141/bin/jps

    查看集群状态 
    hadoop dfsadmin -report

    =====================================================

    删除hadoop文件:

    hadoop fs -rm -r hdfs://hdp01:9000/*

  • 相关阅读:
    Linux性能监测:CPU篇
    Linux性能监测:监测目的与工具介绍
    敏捷开发学习笔记
    Sonar+Hudson+Maven构建系列之三:安装Hudson
    Sonar+Hudson+Maven构建系列之二:迁移Sonar
    Sonar+Hudson+Maven构建系列之一:安装Sonar
    nohup之no hang up, kill, ps -ef, ps aux, grep
    Sonar相关资料
    自动、手动同步FishEye, JIRA的联系人信息
    SOAPFaultException
  • 原文地址:https://www.cnblogs.com/geek-sharing/p/9468907.html
Copyright © 2011-2022 走看看