zoukankan      html  css  js  c++  java
  • IDEA创建Spark开发环境

    1. 下面安装的六步和通过Maven安装Java项目的方式相同
      1. 如果刚安装完IDEA,直接点击[Create New Project]
      2. 左侧选择Maven,然后先勾选上部的[Create from archetype],然后右侧选择[maven-archetype-quickstart]
      3. 填写GroupID、ArtifactId,点击Next
      4. 配置一下maven的位置以及配置文件和本地仓库位置
      5. 配置项目存储的路径,接着点击Next即可
      6. 进入项目后在右下角选择[Enable Auto-Import],这样在修改配置文件时自动导入了 
    2. 导入 scala SDK
    3. 选择导入的scala SDK的地址
    4. 构建自己需要的项目结构(便于管理)
    5. 将对应的目录结构改成对应的源码包
    6. 导入需要的Spark开发的依赖
      <repositories>
          <repository>
            <id>central</id>
            <name>aliyun maven</name>
            <url>http://maven.aliyun.com/nexus/content/groups/public/</url>
            <layout>default</layout>
            <!-- 是否开启发布版构件下载 -->
            <releases>
              <enabled>true</enabled>
            </releases>
            <!-- 是否开启快照版构件下载 -->
            <snapshots>
              <enabled>false</enabled>
            </snapshots>
          </repository>
      
          <repository>
            <id>cloudera</id>
            <url>https://repository.cloudera.com/artifactory/cloudera-repos/</url>
          </repository>
          <repository>
            <id>jboss</id>
            <url>http://repository.jboss.com/nexus/content/groups/public</url>
          </repository>
      </repositories>
      
      
      <properties>
              <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
              <maven.compiler.source>1.7</maven.compiler.source>
              <maven.compiler.target>1.7</maven.compiler.target>
              <hadoop.version>2.6.0-cdh5.7.6</hadoop.version>
              <spark.version>2.2.0</spark.version>
              <mysq.version>5.1.27</mysq.version>
              <hbase.version>1.2.0-cdh5.7.6</hbase.version>
              <uasparser.version>0.6.1</uasparser.version>
      </properties>
      
      
      <dependencies>
              <dependency>
                  <groupId>org.apache.spark</groupId>
                  <artifactId>spark-core_2.11</artifactId>
                  <version>${spark.version}</version>
              </dependency>
              <dependency>
                  <groupId>org.apache.spark</groupId>
                  <artifactId>spark-sql_2.11</artifactId>
                  <version>${spark.version}</version>
              </dependency>
              <dependency>
                  <groupId>org.apache.spark</groupId>
                  <artifactId>spark-streaming_2.11</artifactId>
                  <version>${spark.version}</version>
              </dependency>
              <dependency>
                  <groupId>org.apache.spark</groupId>
                  <artifactId>spark-streaming-kafka-0-8_2.11</artifactId>
                  <version>2.2.0</version>
              </dependency>
              <dependency>
                  <groupId>org.apache.spark</groupId>
                  <artifactId>spark-hive_2.11</artifactId>
                  <version>${spark.version}</version>
              </dependency>
      
              <dependency>
                  <groupId>org.apache.hadoop</groupId>
                  <artifactId>hadoop-client</artifactId>
                  <version>${hadoop.version}</version>
              </dependency>
      
              <dependency>
                  <groupId>mysql</groupId>
                  <artifactId>mysql-connector-java</artifactId>
                  <version>${mysq.version}</version>
              </dependency>
      
              <dependency>
                  <groupId>org.apache.hbase</groupId>
                  <artifactId>hbase-server</artifactId>
                  <version>${hbase.version}</version>
              </dependency>
              <dependency>
                  <groupId>org.apache.hbase</groupId>
                  <artifactId>hbase-hadoop2-compat</artifactId>
                  <version>1.2.0-cdh5.7.6</version>
              </dependency>
              <dependency>
                  <groupId>org.apache.hbase</groupId>
                  <artifactId>hbase-client</artifactId>
                  <version>1.2.0-cdh5.7.6</version>
              </dependency>
      
              <dependency>
                  <groupId>cz.mallat.uasparser</groupId>
                  <artifactId>uasparser</artifactId>
                  <version>${uasparser.version}</version>
              </dependency>
      
              <!-- Spark MLlib依赖包-->
              <dependency>
                  <groupId>org.apache.spark</groupId>
                  <artifactId>spark-mllib_2.11</artifactId>
                  <version>2.2.0</version>
              </dependency>
              <dependency>
                  <groupId>org.scalanlp</groupId>
                  <artifactId>breeze_2.11</artifactId>
                  <version>0.13.1</version>
              </dependency>
      
      <!--        <dependency>
                  <groupId>com.github.fommil.netlib</groupId>
                  <artifactId>all</artifactId>
                  <version>1.1.2</version>
              </dependency>-->
              <dependency>
                  <groupId>org.jblas</groupId>
                  <artifactId>jblas</artifactId>
                  <version>1.2.3</version>
              </dependency>
      
              <dependency>
                  <groupId>org.mongodb.spark</groupId>
                  <artifactId>mongo-spark-connector_2.11</artifactId>
                  <version>2.3.1</version>
              </dependency>
      
              <dependency>
                  <groupId>redis.clients</groupId>
                  <artifactId>jedis</artifactId>
                  <version>2.8.0</version>
              </dependency>
      
              <dependency>
                  <groupId>org.apache.kafka</groupId>
                  <artifactId>kafka-clients</artifactId>
                  <version>0.8.2.1</version>
              </dependency>
      
              <dependency>
                  <groupId>com.alibaba</groupId>
                  <artifactId>fastjson</artifactId>
                  <version>1.2.31</version>
              </dependency>
      </dependencies>
  • 相关阅读:
    Eclipse安装Pydev插件时所遇到的问题
    打开Eclipse弹出“No java virtual machine was found..."的解决方法
    使用adb报错;error: unknown host service
    itools安装程序无法创建临时文件夹
    多线程十二之ConcurrentHashMap1.8实现分析
    多线程十一之ConcurrentHashMap1.7源码分析
    多线程十之CopyOnWriteArrayList源码分析
    多线程学习笔记九之ThreadLocal
    多线程学习笔记八之线程池ThreadPoolExecutor实现分析
    多线程学习笔记七之信号量Semaphore
  • 原文地址:https://www.cnblogs.com/WeiKing/p/13246637.html
Copyright © 2011-2022 走看看