zoukankan      html  css  js  c++  java
  • hadoop 随笔1

    sed -i 's/#auth required pam_wheel.so/auth required pam_wheel.so/g' '/etc/pam.d/su'
    cp /etc/login.defs /etc/login.defs_bak
    echo "SU_WHEEL_ONLY yes" >> /etc/login.defs
    gpasswd -a hadoop wheel





    cat /etc/hosts
    192.168.80.166 nn1.hadoop
    192.168.80.167 nn2.hadoop
    192.168.80.168 s1.hadoop
    192.168.80.169 s2.hadoop
    192.168.80.170 s3.hadoop

    scp /etc/hosts root@192.168.80.167:/etc/
    scp /etc/hosts root@192.168.80.168:/etc/
    scp /etc/hosts root@192.168.80.169:/etc/
    scp /etc/hosts root@192.168.80.170:/etc/

    yum install -y ntp
    /usr/sbin/ntpdate ntp1.aliyun.com

    su - hadoop
    rm -rf .ssh
    mkdir .ssh
    chmod 700 ./.ssh
    ssh-keygen -t rsa


    mv id_rsa.pub id_rsa.pubs2
    mv id_rsa.pub id_rsa.pubs1
    mv id_rsa.pub id_rsa.pubnn1
    mv id_rsa.pub id_rsa.pubnn2

    nn1.hadoop
    touch authorized_keys
    chmod 600 authorized_keys

    scp id_rsa.pubnn1 hadoop@nn2.hadoop:$PWD
    scp id_rsa.pubnn1 hadoop@s1.hadoop:$PWD
    scp id_rsa.pubnn1 hadoop@s2.hadoop:$PWD

    nn2.hadoop
    cd ~/.ssh
    cat id_rsa.pubnn1 >> authorized_keys

    sudo chown -R hadoop:hadoop /home/hadoop

    zkServer.sh start
    zkServer.sh status
    jps


    yum -y install svn ncurses-devel gcc* lzo-devel zlib-devel autoconf automake libtool cmake openssl-devel bzip2



    export JAVA_HOME=/usr/local/jdk1.8.0_201
    export JRE_HOME=$JAVA_HOME/jre
    export PATH=$PATH:$JAVA_HOME/bin
    export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar


    #!/bin/bash
    #文件名:ssh_all.sh
    RUN_HOME=$(cd "$(dirname "$0")"; echo "${PWD}")

    NOW_LIST=(`cat ${RUN_HOME}/ips`)

    SSH_USER="hadoop"
    for i in ${NOW_LIST[@]}; do
        f_cmd="ssh $SSH_USER@$i "$*""
        echo $f_cmd
        if eval $f_cmd; then
            echo "OK"
        else
            echo "FAIL"
        fi
    done




    #!/bin/bash
    #文件名:ssh_root.sh
    RUN_HOME=$(cd "$(dirname "$0")"; echo "${PWD}")

    NOW_LIST=(`cat ${RUN_HOME}/ips`)

    SSH_USER="hadoop"
    for i in ${NOW_LIST[@]}; do
        f_cmd="ssh $SSH_USER@i ~/exe.sh "$*""
        echo $f_cmd
        if eval $f_cmd; then
            echo "OK"
        else
            echo "FAIL"
        fi
    done



    #文件名exe.sh
    cmd=$*

    su - <<EOF
    $cmd

    EOF



    #!/bin/bash
    RUN_HOME=$(cd "(dirname "$0")"; echo "${PWD}")

    NOW_LIST=(`cat ${UN_HOME}/ips`)

    SSH_USER="hadoop"
    for i in ${NOW_LIST[@]}; do
        f_cmd="scp $1 $SSH_USER@i:$2"
        echo $f_cmd
        if eval $f_cmd; then
            echo "ok"
        else
            echo "FAIL"
        fi
    done


    chown -R hadoop:hadoop /usr/local/hadoop-2.7.3
    chmod -R 770 /usr/local/hadoop-2.7.3
    ln -s /usr/local/hadoop-2.7.3 /usr/local/hadoop
    chown -R hadoop:hadoop /usr/local/hadoop



    export JAVA_HOME=/usr/local/jdk1.8.0_201
    export JRE_HOME=$JAVA_HOME/jre
    export PATH=$PATH:$JAVA_HOME/bin
    export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar

    export HADOOP_HOME=/usr/local/hadoop
    export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:
    export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
    export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib:$HADOOP_COMMON_LIB_NATIVE_DIR"
    #Zookeeper
    export ZOOKEEPER_HOME=/home/hadoop/apps/zookeeper-3.4.14
    #export PATH=$PATH:$ZOOKEEPER_HOME/bin


    #set Hadoop_compile
    export MAVEN_HOME=/usr/local/apache-maven-3.3.9
    export FINDBUGS_HOME=/usr/local/findbugs-3.0.1
    export PROTOBUF_HOME=/usr/local/protobuf-2.5.0
    export ANT_HOME=/usr/local/apache-ant-1.9.14
    export PATH=$PATH:$ZOOKEEPER_HOME/bin:PATH=$JAVA_HOME/bin:$MAVEN_HOME/bin:$FINDBUGS_HOME/bin:$ANT_HOME/bin
    export MAVEN_OPTS="-Xmx2g -XX:MaxMetaspaceSize=512M -XX:ReservedCodeCacheSize=512m"





    cp -a /etc/yum.repos.d/CentOS-Base.repo /etc/yum.repos.d/CentOS-Base.repo.bak
    wget -O /etc/yum.repos.d/CentOS-Base.repo https://repo.huaweicloud.com/repository/conf/CentOS-7-reg.repo
    yum clean all
    yum makecache


    yum install -y openssh-server vim gcc gcc-c++ glibc-headers bzip2-devel lzo-devel curl wget openssh-clients zlib-devel autoconf automake cmake libtool openssl-devel fuse-devel snappy-devel telnet unzip zip net-tools.x86_64 firewalld systemd



    dataDir=/data/zkdata
    logDir=/data/log/zklog

    sudo mkdir -p /data/zkdata

    sudo mkdir -p /data/log/zklog

    sudo chown -R hadoop:hadoop /data
    sudo chown -R hadoop:hadoop /usr/local/zookeeper


    文件在末尾添加
    #set Hadoop Path
    export JAVA_HOME=/usr/local/jdk1.8.0_201
    export JRE_HOME=$JAVA_HOME/jre
    export PATH=$PATH:$JAVA_HOME/bin
    export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar

    export HADOOP_HOME=/usr/local/hadoop
    export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:
    export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
    export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib:$HADOOP_COMMON_LIB_NATIVE_DIR"
    export HADOOP_COMMON_HOME=${HADOOP_HOME}
    export HADOOP_HDFS_HOME=${HADOOP_HOME}
    export HADOOP_MAPRED_HOME=${HADOOP_HOME}
    export HADOOP_YARN_HOME=${HADOOP_HOME}
    export HADOOP_CONF_DIR=${HADOOP_HOME}/etc/hadoop
    export HDFS_CONF_DIR=${HADOOP_HOME}/etc/hadoop
    export YARN_CONF_DIR=${HADOOP_HOME}/etc/hadoop
    export LD_LIBRARY_PATH=$HADOOP_HOME/lib/native:/usr/lib64

    export HBASE_HOME=/usr/local/hbase
    export HIVE_HOME=/usr/local/hive
    export PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$HBASE_HOME/bin:$HIVE_HOME/bin:/usr/local/zookeeper/bin

    #Zookeeper
    export ZOOKEEPER_HOME=/usr/local/zookeeper
    export PATH=$PATH:$ZOOKEEPER_HOME/bin


    #set Hadoop_compile
    export MAVEN_HOME=/usr/local/apache-maven-3.3.9
    export FINDBUGS_HOME=/usr/local/findbugs-3.0.1
    export PROTOBUF_HOME=/usr/local/protobuf-2.5.0
    export ANT_HOME=/usr/local/apache-ant-1.9.14
    export PATH=$PATH:$ZOOKEEPER_HOME/bin:PATH=$JAVA_HOME/bin:$MAVEN_HOME/bin:$FINDBUGS_HOME/bin:$ANT_HOME/bin
    export MAVEN_OPTS="-Xmx2g -XX:MaxMetaspaceSize=512M -XX:ReservedCodeCacheSize=512m"





    <属性>
      <name> dfs.nameservices </ name>
      <value> mycluster </ value>
    </ property>

    <属性>
      <name> dfs.ha.namenodes.mycluster </ name>
      <value> nn1,nn2 </ value>
    </ property>

    <属性>
      <name> dfs.namenode.rpc-address.mycluster.nn1 </ name>
      <value> machine1.example.com:8020 </ value>
    </ property>
    <属性>
      <name> dfs.namenode.rpc-address.mycluster.nn2 </ name>
      <value> machine2.example.com:8020 </ value>
    </ property>

    <属性>
      <name> dfs.namenode.http-address.mycluster.nn1 </ name>
      <value> machine1.example.com:50070 </ value>
    </ property>
    <属性>
      <name> dfs.namenode.http-address.mycluster.nn2 </ name>
      <value> machine2.example.com:50070 </ value>
    </ property>


    <属性>
      <name> dfs.namenode.shared.edits.dir </ name>
      <value> qjournal://node1.example.com:8485; node2.example.com:8485; node3.example.com:8485 / mycluster </ value>
    </ property>


    <属性>
      <name> dfs.client.failover.proxy.provider.mycluster </ name>
      <value> org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider </ value>
    </ property>



    sudo mkdir -p /data/hadoopdata/
    sudo chown -R hadoop:hadoop /data

  • 相关阅读:
    百度云人脸识别API人脸库管理
    百度云人脸识别API人脸对比
    Linux之find命令
    Linux之grep命令
    Linux之sed命令
    LNMP Shell脚本发布
    auto_modify_ip Shell脚本安装
    JVM虚拟机详解+Tomcat性能优化
    Tomcat连接器详解
    Tomcat配置详解
  • 原文地址:https://www.cnblogs.com/walkersss/p/12781594.html
Copyright © 2011-2022 走看看