zoukankan      html  css  js  c++  java
  • hadoop 随笔1

    sed -i 's/#auth required pam_wheel.so/auth required pam_wheel.so/g' '/etc/pam.d/su'
    cp /etc/login.defs /etc/login.defs_bak
    echo "SU_WHEEL_ONLY yes" >> /etc/login.defs
    gpasswd -a hadoop wheel





    cat /etc/hosts
    192.168.80.166 nn1.hadoop
    192.168.80.167 nn2.hadoop
    192.168.80.168 s1.hadoop
    192.168.80.169 s2.hadoop
    192.168.80.170 s3.hadoop

    scp /etc/hosts root@192.168.80.167:/etc/
    scp /etc/hosts root@192.168.80.168:/etc/
    scp /etc/hosts root@192.168.80.169:/etc/
    scp /etc/hosts root@192.168.80.170:/etc/

    yum install -y ntp
    /usr/sbin/ntpdate ntp1.aliyun.com

    su - hadoop
    rm -rf .ssh
    mkdir .ssh
    chmod 700 ./.ssh
    ssh-keygen -t rsa


    mv id_rsa.pub id_rsa.pubs2
    mv id_rsa.pub id_rsa.pubs1
    mv id_rsa.pub id_rsa.pubnn1
    mv id_rsa.pub id_rsa.pubnn2

    nn1.hadoop
    touch authorized_keys
    chmod 600 authorized_keys

    scp id_rsa.pubnn1 hadoop@nn2.hadoop:$PWD
    scp id_rsa.pubnn1 hadoop@s1.hadoop:$PWD
    scp id_rsa.pubnn1 hadoop@s2.hadoop:$PWD

    nn2.hadoop
    cd ~/.ssh
    cat id_rsa.pubnn1 >> authorized_keys

    sudo chown -R hadoop:hadoop /home/hadoop

    zkServer.sh start
    zkServer.sh status
    jps


    yum -y install svn ncurses-devel gcc* lzo-devel zlib-devel autoconf automake libtool cmake openssl-devel bzip2



    export JAVA_HOME=/usr/local/jdk1.8.0_201
    export JRE_HOME=$JAVA_HOME/jre
    export PATH=$PATH:$JAVA_HOME/bin
    export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar


    #!/bin/bash
    #文件名:ssh_all.sh
    RUN_HOME=$(cd "$(dirname "$0")"; echo "${PWD}")

    NOW_LIST=(`cat ${RUN_HOME}/ips`)

    SSH_USER="hadoop"
    for i in ${NOW_LIST[@]}; do
        f_cmd="ssh $SSH_USER@$i "$*""
        echo $f_cmd
        if eval $f_cmd; then
            echo "OK"
        else
            echo "FAIL"
        fi
    done




    #!/bin/bash
    #文件名:ssh_root.sh
    RUN_HOME=$(cd "$(dirname "$0")"; echo "${PWD}")

    NOW_LIST=(`cat ${RUN_HOME}/ips`)

    SSH_USER="hadoop"
    for i in ${NOW_LIST[@]}; do
        f_cmd="ssh $SSH_USER@i ~/exe.sh "$*""
        echo $f_cmd
        if eval $f_cmd; then
            echo "OK"
        else
            echo "FAIL"
        fi
    done



    #文件名exe.sh
    cmd=$*

    su - <<EOF
    $cmd

    EOF



    #!/bin/bash
    RUN_HOME=$(cd "(dirname "$0")"; echo "${PWD}")

    NOW_LIST=(`cat ${UN_HOME}/ips`)

    SSH_USER="hadoop"
    for i in ${NOW_LIST[@]}; do
        f_cmd="scp $1 $SSH_USER@i:$2"
        echo $f_cmd
        if eval $f_cmd; then
            echo "ok"
        else
            echo "FAIL"
        fi
    done


    chown -R hadoop:hadoop /usr/local/hadoop-2.7.3
    chmod -R 770 /usr/local/hadoop-2.7.3
    ln -s /usr/local/hadoop-2.7.3 /usr/local/hadoop
    chown -R hadoop:hadoop /usr/local/hadoop



    export JAVA_HOME=/usr/local/jdk1.8.0_201
    export JRE_HOME=$JAVA_HOME/jre
    export PATH=$PATH:$JAVA_HOME/bin
    export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar

    export HADOOP_HOME=/usr/local/hadoop
    export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:
    export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
    export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib:$HADOOP_COMMON_LIB_NATIVE_DIR"
    #Zookeeper
    export ZOOKEEPER_HOME=/home/hadoop/apps/zookeeper-3.4.14
    #export PATH=$PATH:$ZOOKEEPER_HOME/bin


    #set Hadoop_compile
    export MAVEN_HOME=/usr/local/apache-maven-3.3.9
    export FINDBUGS_HOME=/usr/local/findbugs-3.0.1
    export PROTOBUF_HOME=/usr/local/protobuf-2.5.0
    export ANT_HOME=/usr/local/apache-ant-1.9.14
    export PATH=$PATH:$ZOOKEEPER_HOME/bin:PATH=$JAVA_HOME/bin:$MAVEN_HOME/bin:$FINDBUGS_HOME/bin:$ANT_HOME/bin
    export MAVEN_OPTS="-Xmx2g -XX:MaxMetaspaceSize=512M -XX:ReservedCodeCacheSize=512m"





    cp -a /etc/yum.repos.d/CentOS-Base.repo /etc/yum.repos.d/CentOS-Base.repo.bak
    wget -O /etc/yum.repos.d/CentOS-Base.repo https://repo.huaweicloud.com/repository/conf/CentOS-7-reg.repo
    yum clean all
    yum makecache


    yum install -y openssh-server vim gcc gcc-c++ glibc-headers bzip2-devel lzo-devel curl wget openssh-clients zlib-devel autoconf automake cmake libtool openssl-devel fuse-devel snappy-devel telnet unzip zip net-tools.x86_64 firewalld systemd



    dataDir=/data/zkdata
    logDir=/data/log/zklog

    sudo mkdir -p /data/zkdata

    sudo mkdir -p /data/log/zklog

    sudo chown -R hadoop:hadoop /data
    sudo chown -R hadoop:hadoop /usr/local/zookeeper


    文件在末尾添加
    #set Hadoop Path
    export JAVA_HOME=/usr/local/jdk1.8.0_201
    export JRE_HOME=$JAVA_HOME/jre
    export PATH=$PATH:$JAVA_HOME/bin
    export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar

    export HADOOP_HOME=/usr/local/hadoop
    export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:
    export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
    export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib:$HADOOP_COMMON_LIB_NATIVE_DIR"
    export HADOOP_COMMON_HOME=${HADOOP_HOME}
    export HADOOP_HDFS_HOME=${HADOOP_HOME}
    export HADOOP_MAPRED_HOME=${HADOOP_HOME}
    export HADOOP_YARN_HOME=${HADOOP_HOME}
    export HADOOP_CONF_DIR=${HADOOP_HOME}/etc/hadoop
    export HDFS_CONF_DIR=${HADOOP_HOME}/etc/hadoop
    export YARN_CONF_DIR=${HADOOP_HOME}/etc/hadoop
    export LD_LIBRARY_PATH=$HADOOP_HOME/lib/native:/usr/lib64

    export HBASE_HOME=/usr/local/hbase
    export HIVE_HOME=/usr/local/hive
    export PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$HBASE_HOME/bin:$HIVE_HOME/bin:/usr/local/zookeeper/bin

    #Zookeeper
    export ZOOKEEPER_HOME=/usr/local/zookeeper
    export PATH=$PATH:$ZOOKEEPER_HOME/bin


    #set Hadoop_compile
    export MAVEN_HOME=/usr/local/apache-maven-3.3.9
    export FINDBUGS_HOME=/usr/local/findbugs-3.0.1
    export PROTOBUF_HOME=/usr/local/protobuf-2.5.0
    export ANT_HOME=/usr/local/apache-ant-1.9.14
    export PATH=$PATH:$ZOOKEEPER_HOME/bin:PATH=$JAVA_HOME/bin:$MAVEN_HOME/bin:$FINDBUGS_HOME/bin:$ANT_HOME/bin
    export MAVEN_OPTS="-Xmx2g -XX:MaxMetaspaceSize=512M -XX:ReservedCodeCacheSize=512m"





    <属性>
      <name> dfs.nameservices </ name>
      <value> mycluster </ value>
    </ property>

    <属性>
      <name> dfs.ha.namenodes.mycluster </ name>
      <value> nn1,nn2 </ value>
    </ property>

    <属性>
      <name> dfs.namenode.rpc-address.mycluster.nn1 </ name>
      <value> machine1.example.com:8020 </ value>
    </ property>
    <属性>
      <name> dfs.namenode.rpc-address.mycluster.nn2 </ name>
      <value> machine2.example.com:8020 </ value>
    </ property>

    <属性>
      <name> dfs.namenode.http-address.mycluster.nn1 </ name>
      <value> machine1.example.com:50070 </ value>
    </ property>
    <属性>
      <name> dfs.namenode.http-address.mycluster.nn2 </ name>
      <value> machine2.example.com:50070 </ value>
    </ property>


    <属性>
      <name> dfs.namenode.shared.edits.dir </ name>
      <value> qjournal://node1.example.com:8485; node2.example.com:8485; node3.example.com:8485 / mycluster </ value>
    </ property>


    <属性>
      <name> dfs.client.failover.proxy.provider.mycluster </ name>
      <value> org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider </ value>
    </ property>



    sudo mkdir -p /data/hadoopdata/
    sudo chown -R hadoop:hadoop /data

  • 相关阅读:
    mysql数据库常用指令
    解决windows的mysql无法启动 服务没有报告任何错误的经验。
    “Can't open file for writing”或“operation not permitted”的解决办法
    启动Apache出现错误Port 80 in use by "Unable to open process" with PID 4!
    如何打开windows的服务services.msc
    常见的HTTP状态码 404 500 301 200
    linux系统常用的重启、关机指令
    (wifi)wifi移植之命令行调试driver和supplicant
    linux(debian)安装USB无线网卡(tp-link TL-WN725N rtl8188eu )
    alloc_chrdev_region申请一个动态主设备号,并申请一系列次设备号
  • 原文地址:https://www.cnblogs.com/walkersss/p/12781594.html
Copyright © 2011-2022 走看看