zoukankan      html  css  js  c++  java
  • hadoop伪分布式

    Hadoop集群搭建笔记

    1、安装虚拟机

    VMware workstation

    CentOS 镜像

    安装Linux虚拟机:(在Win7上)

    1)安装VMwareWorkstations(可修改配置)

    2)添加CentOS镜像(启动安装 ->配置网络)

    网络配置:NAT模式

    网络重启:service network restart

    关闭系统:shutdown -h now

    2、远程连接

    Xshell5

    Xftp5

    3、在Linux上搭建Hadoop集群

    下载软件

    1. Jdk:http://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.html

    2. Hadoop包:http://apache.fayea.com/hadoop/common/hadoop-2.7.2/

    步骤:

    1、修改/etc/hosts

    2、配置JDK

    安装:rpm -ivh jdk-8u101-linux-x64.rpm

    vi /etc/profile
    JAVA_HOME
    source /etc/profile
    java -version

    3、配置SSH(免密码登录)

    ssh-keygen -t rsa
    cat xxx.pub >> authorized_keys
    chmod 644 authorized_keys
    ssh IP/HOSTNAME

     

    配置环境JAVA_HOME, HADOOP_HOME

    JAVA_HOME=/usr/java/default

    HADOOP_HOME=/software/hadoop-2.8.0

    PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$PATH


    Linux修改/etc/profile配置错误command is not found自救方法

    # export PATH=/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin:/root/bin    

    执行完以上的命令后,你可以重新执行 source /etc/profile命令来让你的配置文件生效

    如果不幸的是你的profile配置文件没有备份,在直接使用vi命令无法使用的情况下,你可以在vi前面加上命令的地址

    #/bin/vi /etc/profile



    虚拟机:/bin/vi 修改完profile ,直接重启虚拟机也可以

    4、安装及配置Hadoop

    tar zxf hadoop-2.7.2.tar.gz
    cd /opt/hadoop-2.7.2/etc/hadoop/

    core-site.xml

    <property>
        <name>fs.default.name</name>
        <value>hdfs://bigdata:9000</value>
      </property>
     
     <property>
        <name>hadoop.tmp.dir</name>
        <value>/software/hadoop-2.8.0/current/tmp</value>
      </property>
     <property>
        <name>fs.trash.interval</name>
        <value>4320</value>
      </property>

    mkdir -p /software/hadoop-2.8.0/current/tmp

    hdfs-site.xml

    <property>
       <name>dfs.namenode.name.dir</name>
       <value>/software/hadoop-2.8.0/current/dfs/name</value>
     </property>
     <property>
       <name>dfs.datanode.data.dir</name>
       <value>/software/hadoop-2.8.0/current/data</value>
     </property>
     <property>
       <name>dfs.replication</name>
       <value>1</value>
     </property>
     <property>
       <name>dfs.webhdfs.enabled</name>
       <value>true</value>
     </property>
     <property>
       <name>dfs.permissions.superusergroup</name>
       <value>staff</value>
     </property>
     <property>
       <name>dfs.permissions.enabled</name>
       <value>false</value>
     </property>


    [root@bigdata hadoop]# mkdir -p /software/hadoop-2.8.0/current/dfs/name
    [root@bigdata hadoop]# mkdir -p /software/hadoop-2.8.0/current/data

    yarn-site.xml

    <property>
       <name>yarn.resourcemanager.hostname</name>
       <value>bigdata</value>
     </property>
     <property>
       <name>yarn.nodemanager.aux-services</name>
       <value>mapreduce_shuffle</value>
     </property>
     <property>
       <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
       <value>org.apache.hadoop.mapred.ShuffleHandler</value>
     </property>
     <property>
       <name>yarn.resourcemanager.address</name>
       <value>bigdata:18040</value>
     </property>
    <property>
       <name>yarn.resourcemanager.scheduler.address</name>
       <value>bigdata:18030</value>
     </property>
     <property>
       <name>yarn.resourcemanager.resource-tracker.address</name>
       <value>bigdata:18025</value>
     </property> <property>
       <name>yarn.resourcemanager.admin.address</name>
       <value>bigdata:18141</value>
     </property>
    <property>
       <name>yarn.resourcemanager.webapp.address</name>
       <value>bigdata:18088</value>
     </property>
    <property>
       <name>yarn.log-aggregation-enable</name>
       <value>true</value>
     </property>
    <property>
       <name>yarn.log-aggregation.retain-seconds</name>
       <value>86400</value>
     </property>
    <property>
       <name>yarn.log-aggregation.retain-check-interval-seconds</name>
       <value>86400</value>
     </property>
    <property>
       <name>yarn.nodemanager.remote-app-log-dir</name>
       <value>/tmp/logs</value>
     </property>
    <property>
       <name>yarn.nodemanager.remote-app-log-dir-suffix</name>
       <value>logs</value>
     </property>

    mapred-site.xml

    <property>
      <name>mapreduce.framework.name</name>
      <value>yarn</value>
    </property>
    <property>
      <name>mapreduce.jobtracker.http.address</name>
      <value>bigdata:50030</value>
    </property>
    <property>
      <name>mapreduce.jobhisotry.address</name>
      <value>bigdata:10020</value>
    </property>
    <property>
      <name>mapreduce.jobhistory.webapp.address</name>
      <value>bigdata:19888</value>
    </property>
    <property>
      <name>mapreduce.jobhistory.done-dir</name>
      <value>/jobhistory/done</value>
    </property>
    <property>
      <name>mapreduce.intermediate-done-dir</name>
      <value>/jobhisotry/done_intermediate</value>
    </property>
    <property>
      <name>mapreduce.job.ubertask.enable</name>
      <value>true</value>
    </property>

    slaves

    bigdata

    hadoop-env.sh

    JAVA_HOME

    5、格式化HDFS

    hdfs namenode -format

    ‘16/09/0403:07:30 INFO common.Storage: Storage directory /opt/hadoop-2.7.2/current/dfs/namehas been successfully formatted.

     

    6、启动Hadoop集群

    /opt/hadoop-2.7.2/sbin/start-all.sh

     

    7、验证Hadoop集群

    1)jps

    2)

    关闭防火墙 或者 在防火墙的规则中开放这些端口

    hdfs   http://bigdata:50070

    yarn  http://bigdata:18088

  • 相关阅读:
    MVC4学习-View(0)
    javascript与常用正则表达式
    uhfreader&rfid标签测试
    WebClient文件上传很方便哈
    NAudio的简单用法
    与wmi交互,调非托管代码,单元测试遇到的一些问题
    我在这里骑美团单车被交警罚了50元,这个地方不能骑共享单车大家留意了
    发邮件,美化table格式
    学习jwt的简单使用
    学习redis的基本使用
  • 原文地址:https://www.cnblogs.com/Bighua123/p/7684301.html
Copyright © 2011-2022 走看看