zoukankan      html  css  js  c++  java
  • 10.Execution failed with exit status: 3

    错误信息:
    insert overwrite table t_mobile_mid_use_p_tmp4_rcf
    select '201411' as month_id,
    a.prov_id, a.city, a.client_imsi, a.os_version,
    b.install_status, b.install_date, b.unstall_status, b.unstall_date,
    a.label_name, a.package_name, a.app_version, a.app_type_id, a.type_label_name,
    b.run_time, monthSpace(b.install_date) as install_days,
    a.flow, a.use_time, a.run_count, a.active_days, a.is_from_plugin,
    from_unixtime(unix_timestamp(),'yyyy-MM-dd HH:mm:ss') as load_date
    from t_mobile_mid_use_p_tmp3_1_rcf a
    join t_mobile_client_p_rcf b on (a.client_imsi = b.client_imsi and a.label_name = b.label_name);

    Query ID = ca_20141218152020_9e4ebfa2-f663-47b8-a0cf-5303b9c0e482
    Total jobs = 1
    14/12/18 15:21:02 WARN conf.Configuration: file:/tmp/ca/hive_2014-12-18_15-20-54_155_1926187970964040123-1/-local-10005/jobconf.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval;  Ignoring.
    14/12/18 15:21:02 WARN conf.Configuration: file:/tmp/ca/hive_2014-12-18_15-20-54_155_1926187970964040123-1/-local-10005/jobconf.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts;  Ignoring.
    Execution log at: /tmp/ca/ca_20141218152020_9e4ebfa2-f663-47b8-a0cf-5303b9c0e482.log
    2014-12-18 03:21:03 Starting to launch local task to process map join; maximum memory = 1065484288
    2014-12-18 03:21:08 Processing rows: 200000 Hashtable size: 199999 Memory usage: 112049704 percentage: 0.105
    2014-12-18 03:21:09 Processing rows: 300000 Hashtable size: 299999 Memory usage: 160367688 percentage: 0.151
    2014-12-18 03:21:10 Processing rows: 400000 Hashtable size: 399999 Memory usage: 209294088 percentage: 0.196
    2014-12-18 03:21:11 Processing rows: 500000 Hashtable size: 499999 Memory usage: 257089944 percentage: 0.241
    2014-12-18 03:21:12 Processing rows: 600000 Hashtable size: 599999 Memory usage: 305440536 percentage: 0.287
    2014-12-18 03:21:14 Processing rows: 700000 Hashtable size: 699999 Memory usage: 347305664 percentage: 0.326
    2014-12-18 03:21:14 Processing rows: 800000 Hashtable size: 799999 Memory usage: 403916624 percentage: 0.379
    2014-12-18 03:21:16 Processing rows: 900000 Hashtable size: 899999 Memory usage: 452238592 percentage: 0.424
    2014-12-18 03:21:16 Processing rows: 1000000 Hashtable size: 999999 Memory usage: 499593552 percentage: 0.469
    2014-12-18 03:21:18 Processing rows: 1100000 Hashtable size: 1099999 Memory usage: 547966320 percentage: 0.514
    2014-12-18 03:21:19 Processing rows: 1200000 Hashtable size: 1199999 Memory usage: 593792800 percentage: 0.557
    2014-12-18 03:21:21 Processing rows: 1300000 Hashtable size: 1299999 Memory usage: 641564688 percentage: 0.602
    2014-12-18 03:21:21 Processing rows: 1400000 Hashtable size: 1399999 Memory usage: 690130432 percentage: 0.648
    2014-12-18 03:21:21 Processing rows: 1500000 Hashtable size: 1499999 Memory usage: 737340976 percentage: 0.692
    2014-12-18 03:21:24 Processing rows: 1600000 Hashtable size: 1599999 Memory usage: 793258352 percentage: 0.745
    2014-12-18 03:21:25 Processing rows: 1700000 Hashtable size: 1699999 Memory usage: 841009952 percentage: 0.789
    2014-12-18 03:21:25 Processing rows: 1800000 Hashtable size: 1799999 Memory usage: 887464680 percentage: 0.833
    2014-12-18 03:21:28 Processing rows: 1900000 Hashtable size: 1899999 Memory usage: 934581288 percentage: 0.877
    2014-12-18 03:21:28 Processing rows: 2000000 Hashtable size: 1999999 Memory usage: 984062056 percentage: 0.924
    Execution failed with exit status: 3
    Obtaining error information

    Task failed!
    Task ID:
      Stage-5

    官方FAQ解释:

    Hive converted a join into a locally running and faster 'mapjoin', but ran out of memory while doing so. There are two bugs responsible for this.

    hives metric for converting joins miscalculated the required amount of memory. This is especially true for compressed files and ORC files, as hive uses the filesize as metric, but compressed tables require more memory in their uncompressed 'in memory representation'.

    The later option may lead to bug number two if you happen to have a affected Hadoop version.

    Hive/Hadoop ignores 'hive.mapred.local.mem' ! (more exactly: bug in Hadoop 2.2 where hadoop-env.cmd sets the -xmx parameter multiple times, effectively overriding the user set hive.mapred.local.mem setting. see: 

    • 2) & 3) can be set in Big-Bench/engines/hive/conf/hiveSettings.sql

      原因:

      t_mobile_client_p_rcft_mobile_mid_use_p_tmp3_1_rcf 因此,Hive优化器认为 是小表,所以,会将这张表数据加到DistributeCache中,造成内存溢出。

    ======select count(1) from t_mobile_mid_use_p_tmp3_1_rcf;

    /**
     *MapReduce Jobs Launched:
     *Job 0: Map: 14  Reduce: 1   Cumulative CPU: 102.42 sec   HDFS Read: 172923550 HDFS Write: 9 SUCCESS
     *Total MapReduce CPU Time Spent: 1 minutes 42 seconds 420 msec
     *OK
     *34304843
     *Time taken: 33.022 seconds, Fetched: 1 row(s)
     */

    ======select count(*) from t_mobile_client_p_rcf;
    /**
     *MapReduce Jobs Launched:
     *Job 0: Map: 5  Reduce: 1   Cumulative CPU: 62.47 sec   HDFS Read: 116257926 HDFS Write: 10 SUCCESS
     *Total MapReduce CPU Time Spent: 1 minutes 2 seconds 470 msec
     *OK
     *165830880
     *Time taken: 37.75 seconds, Fetched: 1 row(s)
    */

    解决方法:

    set hive.auto.convert.join=false;关闭自动转化MapJoin,默认为true;

    set hive.ignore.mapjoin.hint=false; 关闭忽略mapjoin的hints(不忽略,hints有效),默认为true(忽略hints)。

  • 相关阅读:
    1001.A+B Format(20)
    大一下学期的自我目标
    re模块3
    re模块2
    re模块
    configParser模块
    logging模块
    hashlib模块
    sys模块
    isinstance函数
  • 原文地址:https://www.cnblogs.com/yejibigdata/p/6473812.html
Copyright © 2011-2022 走看看