zoukankan      html  css  js  c++  java
  • spark安装 centos7

    1.安装前准备

    安装scala  https://www.cnblogs.com/yoyo1216/p/13365935.html

    安装java  https://www.cnblogs.com/yoyo1216/p/12668926.html

    mkdir -p /usr/local/spark/data
    cd /usr/local/spark
    

    2.下载 http://spark.apache.org/downloads.html

    wget https://apache.website-solution.net/spark/spark-2.4.6/spark-2.4.6-bin-hadoop2.6.tgz
    

    3.解压

    tar -zvxf spark-1.6.0-cdh5.9.0-src.tar.gz    # 解压
    

    4.配置环境变量

    vim /etc/profile
    export SPARK_HOME=/usr/local/spark/spark-2.4.6-bin-hadoop2.6
    export PATH=SPARK_HOME/bin:SPARK_HOME/sbin:$PATH
    
    source /etc/profile
    

    5.修改配置文件

    cd /usr/local/spark/spark-1.6.0-cdh5.9.0/conf
    cp spark-env.sh.template spark-env.sh
    vim spark-env.sh
    
    export JAVA_HOME=/usr/local/java1.8/jdk1.8.0_241
    export SPARK_HOME=/u/usr/local/spark/spark-2.4.6-bin-hadoop2.6
    export SPARK_MASTER_IP=192.168.104.94
    export SPARK_LOCAL_IP=192.168.104.94
    export SPARK_EXECUTOR_MEMORY=1G
    export SCALA_HOME=/usr/local/scala/scala-2.13.3
    export HADOOP_HOME=/usr/local/hadoop2.6/hadoop-2.6.0-cdh5.9.0
    export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop
    
    export SPARK_MASTER_PORT=7077
    export SPARK_MASTER_WEBUI_PORT=7070
    export SPARK_WORKER_CORES=2
    export SPARK_WORKER_MEMORY=1024m
    export SPARK_WORKER_INSTANCES=2
    export SPARK_YARN_USER_ENV="CLASSPATH=/usr/local/hadoop2.6/hadoop-2.6.0-cdh5.9.0/etc/hadoop"
    export SPARK_CLASSPATH=$HBASE_HOME/lib/hbase-protocol-1.2.0-cdh5.9.0.jar:$HBASE_HOME/lib/hbase-common-1.2.0-cdh5.9.0.jar:$HBASE_HOME/lib/htrace-core-3.2.0-incubating.jar:$HBASE_HOME/lib/hbase-server-1.2.0-cdh5.9.0.jar:$HBASE_HOME/lib/hbase-client-1.2.0-cdh5.9.0.jar:$HBASE_HOME/lib/metrics-core-2.2.0.jar:$SPARK_CLASSPATH
    export SPARK_LOCAL_DIR="/mnt/spark/tmp"
    export YARN_CONF_DIR=$HADOOP_HOME/etc/hadoop
    export SPARK_JAVA_OPTS="-Dspark.storage.blockManagerHeartBeatMs=60000-Dspark.local.dir=$SPARK_LOCAL_DIR -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -Xloggc:$SPARK_HOME/logs/gc.log -XX:+UseConcMarkSweepGC -XX:+UseCMSCompactAtFullCollection -XX:CMSInitiatingOccupancyFraction=60"
    export SPARK_DAEMON_JAVA_OPTS="-Dspark.deploy.recoveryMode=ZOOKEEPER -Dspark.deploy.zookeeper.url=192.168.104.94:2181,192.168.104.96:2181,192.168.104.97:2181 -Dspark.deploy.zookeeper.dir=/spark"
    export SPARK_DIST_CLASSPATH=$(${HADOOP_HOME}/bin/hadoop classpath)
    
    
    
    
    cd /usr/local/spark/spark-1.6.0-cdh5.9.0/conf
    cp log4j.properties.template log4j.properties
    vim log4j.properties
    
    log4j.logger.org.apache.spark.repl.Main=INFO
    

    6.启动

    cd /usr/local/spark/spark-1.6.0-cdh5.9.0/sbin
    bash start-all.sh  #  启动
    bash stop-all.sh  # 停止
    bash spark-daemon.sh start/stop/restart regionserver   # 启动/停止/重启
    

    7.查看是否启动成功

    jps
    
    Master
    Worker   出现这个两个启动正常
    

    8.Spark监控页面

    http://192.168.104.94:7070/
    

    9.验证是否安装成功

    cd /usr/local/spark/spark-2.4.6-bin-hadoop2.6/bin
    ./spark-shell --master local   # 本地模式运行
    http://192.168.104.94:4040/jobs/   # 监控页面
    

    10.安装过程中的报错

    Spark 启动java.lang.NoClassDefFoundError: com/fasterxml/jackson/databind/Module
    使用Maven下载以下依赖 : jackson-annotations-2.9.9.jar、 jackson-core-2.9.9.jar、 jackson-databind-2.9.9.jar
    放到 $HADOOP_HOME/share/hadoop/common下
    

      

  • 相关阅读:
    Pascal Analyzer 4 代码分析使用简要说明
    界面动态配置:持久化反持久化
    (Sql Server)数据的拆分和合并
    javascript中的promise和deferred:实践(二)
    对面向接口编程、按分层建项目的反思和新的分层结构思路
    revel框架教程之权限控制
    等待与通知机制
    实现代码编辑器
    自定义html标签
    javascript生成自定义的arcgis simpletoolbar
  • 原文地址:https://www.cnblogs.com/yoyo1216/p/13533223.html
Copyright © 2011-2022 走看看