zoukankan      html  css  js  c++  java
  • HIVE- 新建UDF范例

    首先pom文件导入依赖,Hadoop和hive的依赖导入自己机器的版本,hive记得导jdbc

    <dependency>
                <groupId>org.apache.hadoop</groupId>
                <artifactId>hadoop-client</artifactId>
                <version>2.6.1</version>
            </dependency>
            <dependency>
                <groupId>org.apache.hive</groupId>
                <artifactId>hive-service</artifactId>
                <version>1.2.1</version>
            </dependency>
            <dependency>
                <groupId>org.apache.hive</groupId>
                <artifactId>hive-jdbc</artifactId>
                <version>1.2.1</version>
            </dependency>
            <dependency>
                <groupId>org.apache.hive</groupId>
                <artifactId>hive-exec</artifactId>
                <version>1.2.1</version>
            </dependency>

    新建UDF的Java类

    package com.hive.udf;
    
    import org.apache.hadoop.hive.ql.exec.UDF;
    import org.apache.hadoop.io.Text;
    
    import java.text.ParseException;
    import java.text.SimpleDateFormat;
    import java.util.Date;
    
    /**
     * Created by lenovo on 2017/10/26.
     */
    public class DateTransforUDF extends UDF{
    
        public static final SimpleDateFormat inputFormat = new SimpleDateFormat("yyyy-MM-dd HH:mm:ss");  //匹配讲要格式化的日期格式
        public static final SimpleDateFormat outputFormat=new SimpleDateFormat("yyyyMMddHHmmss");  //要格式化的最终格式
        /*
        2013-02-17 10:38:34
        20130217103834
         */
    
        public Text evaluate(final Text input){
            Text output=new Text();
    
            if (null==input){
                return null;
            }
            String inputDate = input.toString().trim();
            if(null==inputDate){
                return null;
            }
            try {
                Date parseDate = inputFormat.parse(inputDate);
                String outputDate = outputFormat.format(parseDate);
                output.set(outputDate);
            } catch (ParseException e) {
                e.printStackTrace();
                return output;
            }
            return output;
        }
    
        public static void main(String[] args) {
            System.out.println(new DateTransforUDF().evaluate(new Text("2013-02-17 10:38:34")));
        }
    }

    打jar包,上传到Linux,

    hive (register_info)> add jar /root/hive_udf/BaseTest-1.0-SNAPSHOT.jar
                        > ;
    Added [/root/hive_udf/BaseTest-1.0-SNAPSHOT.jar] to class path
    Added resources: [/root/hive_udf/BaseTest-1.0-SNAPSHOT.jar]
    hive (register_info)> create temporary function Tdate as 'com.test.OnlyDateUDF';
    OK
    Time taken: 0.006 seconds
  • 相关阅读:
    Struts2-1.配置&与第一个应用
    1.rs.first()、rs.last()、rs.next()、rs.getRow()
    网页跳转
    js---DOM元素节点
    4、BufferedIn(out)putStream--->字节输入/输出流的缓冲区类(高效类:高效率读写)
    3、FileInputStream--->类文件输入流(读取文件数据)
    2、FileOutputStream--->文件输出流(向文件写入数据)
    1、IO输入&输出流 简介
    OutOfMemoryError系列
    Spark调优,性能优化
  • 原文地址:https://www.cnblogs.com/RzCong/p/7844975.html
Copyright © 2011-2022 走看看