zoukankan      html  css  js  c++  java
  • pof补充

    [初始化系统]

    sudo systemctl stop firewalld
    sudo systemctl disable firewalld
    sudo setenforce 0
    sudo vi /etc/sysconfig/selinux
    SELINUX=disabled

    sudo tar -zxf docker-18.06.3-ce.tgz -C /usr/local/
    cp /usr/local/docker/* /usr/bin/
    mkdir -p /home/docker

    [修正docker自启动问题]

    添加docker自启动需要文件

    sudo vi /usr/lib/systemd/system/docker.service
    [Unit]
    Description=Docker Application Container Engine
    Documentation=https://docs.docker.com
    After=network-online.target firewalld.service
    Wants=network-online.target
    [Service]
    Type=notify
    ExecStart=/usr/bin/dockerd --graph /home/docker
    ExecReload=/bin/kill -s HUP $MAINPID
    LimitNOFILE=infinity
    LimitNPROC=infinity
    TimeoutStartSec=0
    Delegate=yes
    KillMode=process
    Restart=on-failure
    StartLimitBurst=3
    StartLimitInterval=60s
    [Install]
    WantedBy=multi-user.target

    sudo groupadd docker

    将您的用户添加到该docker组

    sudo usermod -aG docker $USER

    将登陆用户加入到docker用户组中

    sudo gpasswd -a $USER docker

    更新用户组

    newgrp docker

    配置docker以systemctl方式启动

    sudo systemctl start docker

    配置docker开机启动

    sudo systemctl enable docker


    数据库是否换成maridb
    [数据库需确认]

    copy sql脚本到数据库容器
    [完善docker cp sql脚本到容器]

    使用docker cp方式拷贝dataexa-insight-microservice-poc.sql到容器根目录下

    docker cp /home/dataexa/insight-deploy/resources/mirrors/mysql/dataexa-insight-microservice-poc.sql mysql:/
    docker exec -it mysql bash
    mysql -uroot -pDataExa5528280
    create database dataexa-insight-microservice-poc default character set utf8 default collate utf8_general_ci;
    show databases;
    use dataexa-insight-microservice-poc;

    source生效 执行完整路径下的sql脚本

    source /dataexa-insight-microservice-poc.sql
    show tables;

    使用docker cp方式拷贝xxl-job-poc.sql到容器根目录下

    docker cp /home/dataexa/insight-deploy/resources/xxl/xxl-job-poc.sql mysql:/
    source /xxl-job-poc.sql

    平滑方式退出docker容器

    键盘: curl+ p +q 三个键位一起按出现 read escape sequence 即可平滑退出容器


    [unzip /home/dataexa/insight-deploy/resources/html/platform.zip -d] /home/dataexa/insight-microservice/workspace/nginx_workspace/html/

    需要先安装unzip

    unzip [缺unzip离线包]
    sudo rpm -ivh /home/dataexa/insight-deploy/resources/tools_package/offline_installer/centos/unzip-6.0-20.el7.x86_64.rpm

    查看unzip是否安装成功
    which unzip
    /usr/bin/unzip

    [粘贴复制的时候-v 没有空格]
    [修改完nginx.conf]
    docker restart nginx

    重启前需要curl+p+q 平滑退出容器

    重启elasticsearch容器

    docker restart elasticsearch

    离线安装bzip2

    sudo rpm -ivh insight-deploy/resources/tools_package/offline_installer/centos/bzip2-1.0.6-13.el7.x86_64.rpm

    检查 bzip2 是否安装

    which bzip2
    /usr/bin/bzip2

    conda使用的python3.5.tar.gz包变成了zip包
    [修正命令]
    unzip /home/dataexa/insight-deploy/resources/conda/python/python3.5.zip -d /home/dataexa/anaconda3/envs


    增加创建路径[python监控路径]

    mkdir -p /home/dataexa/insight-microservice/workspace/python_service_workspace/container_monitor
    mkdir -p /home/dataexa/insight-microservice/workspace/python_service_workspace/dlv5


    vim 补充 [待修改] vim 不能用

    sudo cp /home/dataexa/insight-deploy/resources/tools_package/offline_installer/ubuntu/vim.zip
    sudo cd /usr/local
    sudo unzip vim.zip
    sudo ln -s /usr/local/vim/bin/vim /usr/local/bin/vim

    查看vim是否安装

    sudo which vim

    python [修正]
    dlv5 修改[增加nohup python]
    nohup python /home/dataexa/insight-microservice/workspace/python_service_workspace/dlv5/insight-v5service/tornado_service.py >> /home/dataexa/insight-microservice/workspace/python_service_workspace/dlv5/insight-v5service/nohup.out 2>&1 &

    insight-tagging [创建目录,cp之前没目录]
    mkdir -p /home/dataexa/insight-microservice/workspace/python_service_workspace/data_labeling
    cd /home/dataexa/insight-microservice/workspace/python_service_workspace/data_labeling/insight-tagging/

    启动之前,先把缓存清除

    find . -name "pycache" |xargs rm -r

    启动命令

    nohup python tornado_service.py >/dev/null 2>&1 &

    解压前创建spark⼯作⽬录

    mkdir -p /home/dataexa/insight-microservice/workspace/spark_workspace

    解压前创建hadoop⼯作⽬录

    mkdir -p /home/dataexa/insight-microservice/workspace/hadoop_workspace/

    解压前创建 yarn-conf⼯作⽬录

    mkdir -p /home/dataexa/insight-microservice/workspace/hadoopworkspace/yarn-conf/

    Hadoop部署[修正环境变量] 修改jdk和hadoop环境变量都设置到dataexa用户下
    vi ~/.bashrc
    export JAVA_HOME=/home/dataexa/insight-deploy/resources/jdk/jdk1.8.0_231
    export PATH=$JAVA_HOME/bin:$PATH
    export HADOOP_HOME=/home/dataexa/insight-deploy/resources/hadoop/hadoop-2.9.0
    export PATH=$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH

    vi etc/hadoop/hadoop-env.sh
    export JAVA_HOME=/home/dataexa/insight-deploy/resources/jdk/jdk1.8.0_231
    export PATH=$JAVA_HOME/bin:$PATH

    如果出现格式化文件系统 hdfs namenode -format 出错

    cd /tmp
    将标识有hadoop文件名称的hadoop相关文件 备份移动到另一个目录下
    重新格式化文件系统

    出现20/05/21 07:23:26 INFO ipc.Client: Retrying connect to server: 0.0.0.0/0.0.0.0:8032. Already tried 0 time(s); retry policy is RetryUpT

    sbin/start-yarn.sh #启动yarn即可

    出现org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory hdfs://localhost:9000/user/dataexa/output already exists

    hdfs dfs -rm -r output
    再重新执行即可

    完整hadoop部署
    [修改了jdk和hadoop的环境变量到dataexa用户下]

    解压

    tar xf /home/dataexa/insight-deploy/resources/jdk/jdk-8u231-linux-x64.tar.gz -C /home/dataexa/insight-deploy/resources/jdk

    添加变量环境

    vi ~/.bashrc
    export JAVA_HOME=/home/dataexa/insight-deploy/resources/jdk/jdk1.8.0_231
    export PATH=$JAVA_HOME/bin:$PATH
    export HADOOP_HOME=/home/dataexa/insight-deploy/resources/hadoop/hadoop-2.9.0
    export PATH=$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH

    生效dataexa下的用户环境

    source ~/.bashrc

    进入hadoop环境

    cd /home/dataexa/insight-deploy/resources/hadoop/hadoop-2.9.0
    vi etc/hadoop/hadoop-env.sh
    export JAVA_HOME=/home/dataexa/insight-deploy/resources/jdk/jdk1.8.0_231
    export PATH=$JAVA_HOME/bin:$PATH

    在Hadoop的安装目录下,创建input目录

    mkdir input

    拷贝input文件到input目录下

    cp etc/hadoop/*.xml input

    执行Hadoop job

    hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-2.9.0.jar grep input output 'dfs[a-z.]+'
    上面的job是使用hadoop自带的样例,在input中统计含有dfs的字符串。

    确认执行结果

    cat output/*

    添加主机地址和映射

    sudo vi /etc/hosts
    192.168.1.237 localhost

    修改设定文件

    vi etc/hadoop/core-site.xml


    fs.defaultFS
    hdfs://localhost:9000

    vi etc/hadoop/hdfs-site.xml


    dfs.replication
    1

    设置本机⽆密码ssh登陆

    ssh-keygen -t rsa
    cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
    chmod 0600 ~/.ssh/authorized_keys

    测试

    ssh dataexa@192.168.1.237

    格式化文件系统

    hdfs namenode -format

    启动名称节点和数据节点后台进程

    sbin/start-dfs.sh

    确认

    jps

    访问NameNode的web页面

    http://192.168.1.237:50070/

    创建HDFS

    hdfs dfs -mkdir /user/test

    拷贝input文件到HDFS目录下

    hdfs dfs -put etc/hadoop /user/test/input

    确认,查看

    hadoop fs -ls /user/test/input

    执行Hadoop job

    hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-2.9.0.jar grep /user/test/input output 'dfs[a-z.]+'

    确认执行结果

    hdfs dfs -cat output/*

    或者从HDFS拷贝到本地查看

    bin/hdfs dfs -get output output
    cat output/*

    停止daemon

    sbin/stop-dfs.sh

    执行YARN job

    修改设定文件
    cp etc/hadoop/mapred-site.xml.template etc/hadoop/mapred-site.xml
    vi etc/hadoop/mapred-site.xml


    mapreduce.framework.name
    yarn

    vi etc/hadoop/yarn-site.xml


    yarn.nodemanager.aux-services
    mapreduce_shuffle

    启动ResourceManger和NodeManager后台进程

    sbin/start-yarn.sh

    访问ResourceManger的web页面

    http://192.168.1.237:8088/

    #执行hadoop job
    hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-2.9.0.jar grep /user/test/input output 'dfs[a-z.]+'

    如果这里报错,需要sbin/stop-yarn.sh再sbin/start-yarn.sh

    确认执行结果

    hdfs dfs -cat output/*

    停止daemon

    sbin/stop-yarn.sh

    启动jobhistory daemon

    sbin/mr-jobhistory-daemon.sh start historyserver

    访问Job History Server的web页面 出现数据即可

    http://localhost:19888/

  • 相关阅读:
    php 学习笔记 数组2
    php 学习笔记 数组1
    jQuery prop 全选和全不全
    jquery获取多个checkbox的值异步提交给php
    php 计算多维数组中所有值的总和
    系统状态码原型
    hadoop win10开发环境
    kafka spring整合版本匹配
    mac 安装brew
    hadoop2.8.5 idea2019.1.3插件安装
  • 原文地址:https://www.cnblogs.com/sxgaofeng/p/13111997.html
Copyright © 2011-2022 走看看