zoukankan      html  css  js  c++  java
  • [Script] Hadoop使用docker搭建完全分布模式环境的脚本

      之前调试程序都是在本机用的单机模式,体系结构老师要求实验要用完全分布模式的环境运行程序。  

      不想用虚拟机,完全是出于对界面的不习惯。于是使用了docker。

      脚本会与之前的hadoop环境搭建有重叠。

      参考了博客:http://tashan10.com/yong-dockerda-jian-hadoopwei-fen-bu-shi-ji-qun/,并不完全一致,因为我希望和本机的java hadoop版本一致。

      1 # Reference http://tashan10.com/yong-dockerda-jian-hadoopwei-fen-bu-shi-ji-qun/
      2 
      3 # Ubuntu 14.04 LTS Hadoop 2.7 Fully Distributed with docker.
      4 
      5 # Install docker.
      6 sudo apt-get install apt-transport https
      7 sudo apt-key adv --keyserver hkp://keyserver.ubuntu.com:80 --recv-keys 36A1D7869245C8950F966E92D8576A8BA88D21E9
      8 sudo bash -c "echo deb https://get.docker.io/ubuntu docker main > /etc/apt/sources.list.d/docker.list"
      9 sudo apt-get update
     10 sudo apt-get install lxc-docker
     11 
     12 # Remove "sudo". kirai is my ubuntu's username.
     13 sudo groupadd docker
     14 sudo gpasswd -a kirai docker
     15 sudo reboot
     16 
     17 # Install ubuntu mirror.
     18 docker pull ubuntu:14.04
     19 
     20 # Run the ubuntu mirror.
     21 docker run -ti ubuntu:14.04
     22 
     23 # Change the package mirror.
     24 # mv /etc/apt/sources.list /etc/apt/sources.list.bk
     25 # echo "deb https://mirrors.tuna.tsinghua.edu.cn/ubuntu/ xenial main restricted universe multiverse" > /etc/apt/sources.list
     26 # echo "deb https://mirrors.tuna.tsinghua.edu.cn/ubuntu/ xenial-updates main restricted universe multiverse"  >> /etc/apt/sources.list
     27 # echo "deb https://mirrors.tuna.tsinghua.edu.cn/ubuntu/ xenial-backports main restricted universe multiverse" >> /etc/apt/sources.list
     28 # echo "deb https://mirrors.tuna.tsinghua.edu.cn/ubuntu/ xenial-security main restricted universe multiverse" >> /etc/apt/sources.list
     29 
     30 
     31 # Install java hadoop in the mirror.
     32 
     33 apt-get install software-properties-common python-software-properties
     34 add-apt-repository ppa:webupd8team/java
     35 apt-get update
     36 apt-get install oracle-java7-installer
     37 exit
     38 
     39 # Persistence (4631f498dec7 is the container's id with java)
     40 docker commit -m "java installed" 4631f498dec7 ubuntu:java
     41 
     42 # Restart the container with java.
     43 # Reference ubuntu 14.04 LTS hadoop installation tutorial by me.
     44 ##get & install hadoop 2.7.1    (under user : hadoop)
     45 cd ~
     46 mkdir hadoop
     47 sudo wget http://mirror.bit.edu.cn/apache/hadoop/common/hadoop-2.7.1/hadoop-2.7.1.tar.gz    #get hadoop2.7.1
     48 sudo tar xzf hadoop-2.7.1.tar.gz    #unzip the hadoop-*.*.*.tar.gz
     49 sudo rm hadoop-2.7.1.tar.gz    #remove the zip
     50 sudo mv hadoop-2.7.1 /usr/local/    #install hadoop at local
     51 sudo chmod 774 /usr/local/hadoop     #granted permissions to users (r&w)
     52 
     53 ##configure ~/.bashrc
     54 update-alternatives --config java    #get java's path (stared one. such as '/usr/lib/jvm/java-7-openjdk-amd64/jre/bin/java'. but we just need '/usr/lib/jvm/java-7-openjdk-amd64' this shorter path to set up JAVA_HOME environment variables)
     55 sudo vi ~/.bashrc    #edit bashrc file
     56 
     57 ##add the content below to the end of bashrc
     58 #--------------------------------------------------------------#
     59 
     60 export JAVA_HOME=/usr/lib/jvm/java-7-oracle
     61 export HADOOP_INSTALL=/usr/local/hadoop-2.7.1
     62 export PATH=$PATH:$HADOOP_INSTALL/bin
     63 export PATH=$PATH:$HADOOP_INSTALL/sbin
     64 export HADOOP_MAPRED_HOME=$HADOOP_INSTALL
     65 export HADOOP_COMMON_HOME=$HADOOP_INSTALL
     66 export HADOOP_HDFS_HOME=$HADOOP_INSTALL
     67 export YARN_HOME=$HADOOP_INSTALL
     68 export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_INSTALL/lib/native
     69 export HADOOP_OPTS="-Djava.library.path=$HADOOP_INSTALL/lib"
     70 #HADOOP VARIABLES END
     71 
     72 #--------------------------------------------------------------#
     73 source ~/.bashrc    #make the environment variables come into effect
     74 
     75 ##configure hadoop
     76 sudo vi /usr/local/hadoop-2.7.1/etc/hadoop/hadoop-env.sh     #edit hadoop-env.sh
     77 ?JAVA_HOME    #(in vim) locate JAVA_HOME
     78 #change 'export JAVA_HOME=${JAVA_HOME}' into 'export JAVA_HOME=/usr/lib/jvm/java-7-oracle'
     79 source /usr/local/hadoop-2.7.1/etc/hadoop/hadoop-env.sh #update
     80 
     81 ##test
     82 cd /usr/local/hadoop-2.7.1/    #go there
     83 sudo mkdir input
     84 sudo cp README.txt input
     85 bin/hadoop jar share/hadoop/mapreduce/sources/hadoop-mapreduce-examples-2.7.1-sources.jar org.apache.hadoop.examples.WordCount input output
     86 
     87 
     88 
     89 
     90 #----------------- FINISHED INSTALLATION ---------------------------#
     91 # Persistence (8d9a50a0ee10 is the container's id with hadoop)
     92 docker commit -m "hadoop installed" 8d9a50a0ee10 ubuntu:hadoop
     93 
     94 # Configure hadoop.
     95 # mainly core-site.xml、hdfs-site.xml、mapred-site.xml
     96 cd /usr/local/hadoop-2.7.1
     97 mkdir tmp
     98 mkdir namenode
     99 mkdir datanode
    100 
    101 cd etc/hadoop
    102 cp mapred-site.xml.template mapred-site.xml
    103 
    104 vi core-site.xml
    105 #--add contents between <configuration></configuration> --#
    106     <property>
    107         <name>hadoop.tmp.dir</name>
    108         <value>/usr/local/hadoop-2.7.1/tmp</value>
    109         <description>A base for other temporary directories.</description>
    110     </property>
    111 
    112     <property>
    113         <name>fs.default.name</name>
    114         <value>hdfs://master:9000</value>
    115         <final>true</final>
    116         <description>The name of the default file system.  A URI whose
    117         scheme and authority determine the FileSystem implementation.  The
    118         uri's scheme determines the config property (fs.SCHEME.impl) naming
    119         the FileSystem implementation class.  The uri's authority is used to
    120         determine the host, port, etc. for a filesystem.</description>
    121     </property>
    122 #----#
    123 
    124 vi hdfs-site.xml
    125 #--add contents between <configuration></configuration> one master, two slave--#
    126     <property>
    127     <name>dfs.replication</name>
    128     <value>2</value>
    129     <final>true</final>
    130     <description>Default block replication.
    131     The actual number of replications can be specified when the file is created.
    132     The default is used if replication is not specified in create time.
    133     </description>
    134     </property>
    135 
    136     <property>
    137     <name>dfs.namenode.name.dir</name>
    138     <value>/usr/local/hadoop-2.7.1/namenode</value>
    139     <final>true</final>
    140     </property>
    141 
    142     <property>
    143     <name>dfs.datanode.data.dir</name>
    144     <value>/usr/local/hadoop-2.7.1/datanode</value>
    145     <final>true</final>
    146     </property>
    147 #----#
    148 
    149 vi mapred-site.xml
    150 #--add contents between <configuration></configuration>--#
    151     <property>
    152     <name>mapred.job.tracker</name>
    153     <value>master:9001</value>
    154     <description>
    155     The host and port that the MapReduce job tracker runs
    156     at.  If "local", then jobs are run in-process as a single map
    157     and reduce task.
    158     </description>
    159     </property>
    160 #----#
    161 
    162 # Format namenode.
    163 hadoop namenode -format
    164 
    165 # Install ssh.
    166 apt-get install ssh
    167 cd ~
    168 ssh-keygen -t rsa -P '' -f ~/.ssh/id_dsa
    169 cd .ssh
    170 cat id_dsa.pub >> authorized_keys
    171 
    172 vi ~/.bashrc
    173 #--append--#
    174 #autorun
    175 /usr/sbin/sshd
    176 #----#
    177 
    178 # ifconfig enable.
    179 apt-get install net-tools
    180 exit
    181 
    182 # Persistence (342b9f9e1893 is the container's id with configured-hadoop)
    183 docker commit -m "configured hadoop" 342b9f9e1893 ubuntu:chadoop
    184 
    185 
    186 #-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-#
    187 # Build distributed environment.
    188 
    189 # Open three terminals and run:
    190 docker run -ti -h slave1 ubuntu:chadoop
    191 docker run -ti -h slave2 ubuntu:chadoop
    192 docker run -ti -h master ubuntu:chadoop
    193 
    194 # Write down the master and slaves' IP:
    195 # slave1:172.17.0.2
    196 # slave2:172.17.0.3
    197 # master:172.17.0.4
    198 
    199 # For each container:
    200 vi /etc/hosts
    201 
    202 #--write down--#
    203 172.17.0.2    slave1
    204 172.17.0.3    slave2
    205 172.17.0.4    master
    206 #----#
    207 
    208 # For master:
    209 vi /usr/local/hadoop-2.7.1/etc/hadoop/slaves
    210 
    211 #--write down--#
    212 slave1
    213 slave2
    214 #----#
    215 
    216 #-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-_-#
  • 相关阅读:
    【后缀数组】
    【后缀数组之height数组】
    【后缀数组之SA数组】【真难懂啊】
    【转】为何浮点数可能丢失精度
    UVa_Live 3664(精度坑)
    【火车出栈】ZOJ
    并发查询
    java基础知识1
    感悟__你总是要建立自己的价值观,世界观,人生观,信念的,总不能一直靠鸡汤,被外界,环境,他人,见闻等所掌控你的情绪,积极或者消极.
    batch、随机、Mini-batch梯度下降
  • 原文地址:https://www.cnblogs.com/kirai/p/6994150.html
Copyright © 2011-2022 走看看