zoukankan      html  css  js  c++  java
  • #Hadoop集群部署

    #

    1. 配置JAVA环境(vi/etc/profile.d/java.sh)

    [root@master ~]# yum install java-1.8.0-openjdk-devel

     

    [root@master ~]# vi /etc/profile.d/java.sh
    添加JAVA环境变量
    export JAVA_HOME=/usr

    或者编译安装。

    export JAVA_HOME=/usr/java/jdk1.8.0_151
    export JRE_HOME=$JAVA_HOME/jre
    export CLASSPATH=$JAVA_HOME/lib:$JAVA_HOME/lib/tools.jar
    export PATH=$PATH:$JAVA_HOME/bin:$JAVA_HOME/jre/bin

    node节点也需要配置JAVA环境。

    2.创建hadoop用户

    [root@master hadoop]# useradd hadoop
    [root@master hadoop]#echo 'hadoop'|passwd -stdin hadoop

    node2,node3,node4都需要创建。

    3. 配置hosts(/etc/hosts)

    127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
    ::1         localhost localhost.localdomain localhost6 localhost6.localdomain6
    192.168.15.128 master192.168.15.129 node2
    192.168.15.130  node3
    192.168.15.131 node4

    node2,node3,node4都需要配置。

     

    4. 配置master免密登录node节点

    [root@master ~]# su - hadoop
    [hadoop@master ~]$ ssh-keygen -t rsa -P ''
    [hadoop@master ~]$ for i in 2 3 4;do ssh-copy-id -i .ssh/id_rsa_pub hadoop@node${i};done

     

    5.  解压hadoop安装包

    [root@master ~]# tar -zxvf hadoop-2.8.2.tar.gz -C /usr/local/
    [root@master ~]# cd /usr/local
    [root@master ~]# ln -sv hadoop-2.8.2 hadoop

     

    6. 创建数据目录和日志目录

    [root@master ~]# mkdir -pv /data/hadoop/hdfs/{name,sname,data}
    [root@master ~]# chown -R hadoop:hadoop /data
    创建日志目录
    [root@master ~]# mkdir /usr/local/hadoop/logs
    [root@master ~]# chown -R hadoop:hadoop /usr/local/hadoop/
    [root@master ~]# chmod g+x /usr/local/hadoop/logs

     

    7. 配置hadoop配置文件(/usr/local/hadoop/etc/hadoop/)下面的配置Node节点也需要同样的配置,master配置完,

    可以通过命令[hadoop@master ~]$ scp /usr/local/hadoop/etc/hadoop/*  hadoop@node2:/usr/local/hadoop/etc/hadoop/

    core-site.xml

    <configuration>
            <property>
                    <name>fs.defaultFS</name>
                    <value>hdfs://master:8082</value>
            </property>
    </configuration>

    hdfs-site.xml

    <configuration>
            <property>
                    <name>dfs.namenode.name.dir</name>
                    <value>file:///data/hadoop/hdfs/name</value>
            </property>
            <property>
                    <name>dfs.datanode.data.dir</name>
                    <value>file:///data/hadoop/hdfs/data</value>
            </property>
            <property>
                    <name>dfs.replication</name>
                    <value>2</value>
            </property>
            <property>
                    <name>fs.checkpoint.dir</name>
                    <value>file:///data/hadoop/hdfs/sname</value>
            </property>
            <property>
                    <name>fs.checkpoint.edits.dir</name>
                    <value>file:///data/hadoop/hdfs/sname</value>
            </property>
    </configuration>

     yarn-site.xml

    <configuration>
    
    <!-- Site specific YARN configuration properties -->
     <property>
             <name>yarn.resourcemanager.address</name>
            <value>master:8032</value>
    </property>
    <property>
            <name>yarn.resourcemanager.scheduler.address</name>
            <value>master:8030</value>
    </property>
    <property>
            <name>yarn.resourcemanager.resource-tracker.address</name>
            <value>master:8031</value>
    </property>
    <property>
            <name>yarn.resourcemanager.admin.address</name>
            <value>master:8033</value>
    </property>
    <property>
            <name>yarn.resourcemanager.webapp.address</name>
            <value>master:8088</value>
    </property>
    <property>
            <name>yarn.nodemanager.aux-services</name>
            <value>mapreduce_shuffle</value>
    </property>
    <property>
            <name>yarn.nodemanager.auxservices.mapreduce_shuffle.class</name>
            <value>org.apache.hadoop.mapred.ShuffleHandler</value>
    </property>
    <property>
            <name>yarn.resourcemanager.scheduler.class</name>
            <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>
    </property>
    </configuration>

    mapred-site.xml

    <configuration>
            <property>
                    <name>mapreduce.framework.name</name>
                    <value>yarn</value>
            </property>
    </configuration>

     slaves

    node2
    node3
    node4

    hadoop-env.sh(需要添加JAVA环境变量)

    export JAVA_HOME=/usr/java/jdk1.8.0_151


    8. 格式化HDFS

    [root@master ~]# su - hadoop
    [hadoop@master ~]$ hdfs namenode -format

    输出结果为 17/11/13 23:13:47 INFO common.Storage: Storage directory /data/hadoop/hdfs/name has been successfully formatted.
    说明格式化成功了。

    9. 启动dfs

    [hadoop@master ~]$ start-dfs.sh
    Starting namenodes on [master]
    master: starting namenode, logging to /bdapps/hadoop/logs/hadoop-hadoop-namenode-master.out
    node2: starting datanode, logging to /bdapps/hadoop-2.8.2/logs/hadoop-hadoop-datanode-node2.out
    node3: starting datanode, logging to /bdapps/hadoop-2.8.2/logs/hadoop-hadoop-datanode-node3.out
    node4: starting datanode, logging to /bdapps/hadoop-2.8.2/logs/hadoop-hadoop-datanode-node4.out
    Starting secondary namenodes [0.0.0.0]
    0.0.0.0: starting secondarynamenode, logging to /bdapps/hadoop/logs/hadoop-hadoop-secondarynamenode-master.out


    10. 上传一个文件(/etc/fstab)

    [hadoop@master ~]$ hdfs dfs -mkdir /test
    [hadoop@master ~]$ hdfs dfs -put /etc/fstab /test
    [hadoop@master ~]$ hdfs dfs -ls /test
    Found 1 items
    -rw-r--r--   2 hadoop supergroup        779 2017-11-14 00:03 /test/fstab
    [hadoop@master ~]$ hdfs dfs -cat /test/fstab
    
    #
    # /etc/fstab
    # Created by anaconda on Mon Nov 13 08:43:43 2017
    #
    # Accessible filesystems, by reference, are maintained under '/dev/disk'
    # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
    #
    /dev/mapper/VolGroup-lv_root /                       ext4    defaults        1 1
    UUID=98ed038f-f133-4809-bad8-c4139402b11b /boot                   ext4    defaults        1 2
    /dev/mapper/VolGroup-lv_swap swap                    swap    defaults        0 0
    tmpfs                   /dev/shm                tmpfs   defaults        0 0
    devpts                  /dev/pts                devpts  gid=5,mode=620  0 0
    sysfs                   /sys                    sysfs   defaults        0 0
    proc                    /proc                   proc    defaults        0 0
    [hadoop@master ~]$ 


    11. 启动yarn

    [hadoop@master ~]$ start-yarn.sh 
    starting yarn daemons
    starting resourcemanager, logging to /bdapps/hadoop/logs/yarn-hadoop-resourcemanager-master.out
    node4: starting nodemanager, logging to /bdapps/hadoop/logs/yarn-hadoop-nodemanager-node4.out
    node3: starting nodemanager, logging to /bdapps/hadoop/logs/yarn-hadoop-nodemanager-node3.out
    node2: starting nodemanager, logging to /bdapps/hadoop/logs/yarn-hadoop-nodemanager-node2.out


    可以通过命令jps查看java进程

    Master节点

    [hadoop@master ~]$ jps
    7600 ResourceManager
    7106 NameNode
    8483 Jps
    7304 SecondaryNameNode
    


    Node节点

    [root@node4 ~]# jps
    2802 Jps
    2678 NodeManager
    2553 DataNode


    12. 访问

    hadoop总览
    http://192.168.15.128:50070

    查看hadoop的应用
    http://192.168.15.128:8088

    #

  • 相关阅读:
    清北学堂模拟赛d5t6 cube
    清北学堂模拟赛d5t5 exLCS
    清北学堂模拟赛d5t4 套路
    清北学堂模拟赛d3t2 b
    清北学堂模拟赛d4t2 b
    Android(java)学习笔记196:ContentProvider使用之内容观察者01
    Android(java)学习笔记195:ContentProvider使用之添加数据到联系人(掌握)
    Android(java)学习笔记194:ContentProvider使用之获得系统联系人信息02(掌握)
    Android(java)学习笔记193:ContentProvider使用之获得系统联系人信息01
    Android(java)学习笔记192:ContentProvider使用之虚拟短信
  • 原文地址:https://www.cnblogs.com/lwsup/p/7838420.html
Copyright © 2011-2022 走看看