java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row {"cal_dt":"2013-12-31","year_beg_dt":"2013-01-01","qtr_beg_dt":"2013-10-01","month_beg_dt":"2013-12-01","week_beg_dt":"2013-12-29","age_for_year_id":0,"age_for_qtr_id":0,"age_for_month_id":1,"age_for_week_id":5,"age_for_dt_id":34,"age_for_rtl_year_id":1,"age_for_rtl_qtr_id":1,"age_for_rtl_month_id":1,"age_for_rtl_week_id":5,"age_for_cs_week_id":5,"day_of_cal_id":41638,"day_of_year_id":365,"day_of_qtr_id":92,"day_of_month_id":31,"day_of_week_id":3,"week_of_year_id":53,"week_of_cal_id":5948,"month_of_qtr_id":3,"month_of_year_id":12,"month_of_cal_id":1368,"qtr_of_year_id":4,"qtr_of_cal_id":456,"year_of_cal_id":114,"year_end_dt":"2013-12-31","qtr_end_dt":"2013-12-31","month_end_dt":"2013-12-31","week_end_dt":"2013-12-31","cal_dt_name":"31-Dec-2013","cal_dt_desc":"Dec 31st 2013","cal_dt_short_name":"Tue 12-31-13","ytd_yn_id":0,"qtd_yn_id":0,"mtd_yn_id":0,"wtd_yn_id":0,"season_beg_dt":"2013-12-21","day_in_year_count":365,"day_in_qtr_count":92,"day_in_month_count":31,"day_in_week_count":3,"rtl_year_beg_dt":"2013-12-29","rtl_qtr_beg_dt":"2013-12-29","rtl_month_beg_dt":"2013-12-29","rtl_week_beg_dt":"2013-12-29","cs_week_beg_dt":"2013-12-30","cal_date":"2013-12-31","day_of_week":"Tue ","month_id":"2013M12","prd_desc":"Dec-2013","prd_flag":"N","prd_id":"2013M12 ","prd_ind":"N","qtr_desc":"Year 2013 - Quarter 04","qtr_id":"2013Q04 ","qtr_ind":"N","retail_week":"1","retail_year":"2014","retail_start_date":"2013-12-29","retail_wk_end_date":"2014-01-04","week_ind":"N","week_num_desc":"Wk.53 - 13","week_beg_date":"2013-12-29 00:00:00","week_end_date":"2013-12-31 00:00:00","week_in_year_id":"2013W53 ","week_id":"2013W53 ","week_beg_end_desc_mdy":"12/29/13 - 12/31/13","week_beg_end_desc_md":"12/29 - 12/31","year_id":"2013","year_ind":"N","cal_dt_mns_1year_dt":"2012-12-31","cal_dt_mns_2year_dt":"2011-12-31","cal_dt_mns_1qtr_dt":"2013-09-30","cal_dt_mns_2qtr_dt":"2013-06-30","cal_dt_mns_1month_dt":"2013-11-30","cal_dt_mns_2month_dt":"2013-10-31","cal_dt_mns_1week_dt":"2013-12-24","cal_dt_mns_2week_dt":"2013-12-17","curr_cal_dt_mns_1year_yn_id":0,"curr_cal_dt_mns_2year_yn_id":0,"curr_cal_dt_mns_1qtr_yn_id":0,"curr_cal_dt_mns_2qtr_yn_id":0,"curr_cal_dt_mns_1month_yn_id":0,"curr_cal_dt_mns_2month_yn_id":0,"curr_cal_dt_mns_1week_yn_ind":0,"curr_cal_dt_mns_2week_yn_ind":0,"rtl_month_of_rtl_year_id":"1","rtl_qtr_of_rtl_year_id":1,"rtl_week_of_rtl_year_id":1,"season_of_year_id":1,"ytm_yn_id":0,"ytq_yn_id":1,"ytw_yn_id":0,"cre_date":"2005-09-07","cre_user":"USER_X ","upd_date":"2013-11-27 00:16:56","upd_user":"USER_X"}

kylin-dev | 和风 | 11 months ago
  1. 0

    org.apache.hadoop.hive.ql.metadata.HiveException

    kylin-dev | 11 months ago | 和风
    java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row {"cal_dt":"2013-12-31","year_beg_dt":"2013-01-01","qtr_beg_dt":"2013-10-01","month_beg_dt":"2013-12-01","week_beg_dt":"2013-12-29","age_for_year_id":0,"age_for_qtr_id":0,"age_for_month_id":1,"age_for_week_id":5,"age_for_dt_id":34,"age_for_rtl_year_id":1,"age_for_rtl_qtr_id":1,"age_for_rtl_month_id":1,"age_for_rtl_week_id":5,"age_for_cs_week_id":5,"day_of_cal_id":41638,"day_of_year_id":365,"day_of_qtr_id":92,"day_of_month_id":31,"day_of_week_id":3,"week_of_year_id":53,"week_of_cal_id":5948,"month_of_qtr_id":3,"month_of_year_id":12,"month_of_cal_id":1368,"qtr_of_year_id":4,"qtr_of_cal_id":456,"year_of_cal_id":114,"year_end_dt":"2013-12-31","qtr_end_dt":"2013-12-31","month_end_dt":"2013-12-31","week_end_dt":"2013-12-31","cal_dt_name":"31-Dec-2013","cal_dt_desc":"Dec 31st 2013","cal_dt_short_name":"Tue 12-31-13","ytd_yn_id":0,"qtd_yn_id":0,"mtd_yn_id":0,"wtd_yn_id":0,"season_beg_dt":"2013-12-21","day_in_year_count":365,"day_in_qtr_count":92,"day_in_month_count":31,"day_in_week_count":3,"rtl_year_beg_dt":"2013-12-29","rtl_qtr_beg_dt":"2013-12-29","rtl_month_beg_dt":"2013-12-29","rtl_week_beg_dt":"2013-12-29","cs_week_beg_dt":"2013-12-30","cal_date":"2013-12-31","day_of_week":"Tue ","month_id":"2013M12","prd_desc":"Dec-2013","prd_flag":"N","prd_id":"2013M12 ","prd_ind":"N","qtr_desc":"Year 2013 - Quarter 04","qtr_id":"2013Q04 ","qtr_ind":"N","retail_week":"1","retail_year":"2014","retail_start_date":"2013-12-29","retail_wk_end_date":"2014-01-04","week_ind":"N","week_num_desc":"Wk.53 - 13","week_beg_date":"2013-12-29 00:00:00","week_end_date":"2013-12-31 00:00:00","week_in_year_id":"2013W53 ","week_id":"2013W53 ","week_beg_end_desc_mdy":"12/29/13 - 12/31/13","week_beg_end_desc_md":"12/29 - 12/31","year_id":"2013","year_ind":"N","cal_dt_mns_1year_dt":"2012-12-31","cal_dt_mns_2year_dt":"2011-12-31","cal_dt_mns_1qtr_dt":"2013-09-30","cal_dt_mns_2qtr_dt":"2013-06-30","cal_dt_mns_1month_dt":"2013-11-30","cal_dt_mns_2month_dt":"2013-10-31","cal_dt_mns_1week_dt":"2013-12-24","cal_dt_mns_2week_dt":"2013-12-17","curr_cal_dt_mns_1year_yn_id":0,"curr_cal_dt_mns_2year_yn_id":0,"curr_cal_dt_mns_1qtr_yn_id":0,"curr_cal_dt_mns_2qtr_yn_id":0,"curr_cal_dt_mns_1month_yn_id":0,"curr_cal_dt_mns_2month_yn_id":0,"curr_cal_dt_mns_1week_yn_ind":0,"curr_cal_dt_mns_2week_yn_ind":0,"rtl_month_of_rtl_year_id":"1","rtl_qtr_of_rtl_year_id":1,"rtl_week_of_rtl_year_id":1,"season_of_year_id":1,"ytm_yn_id":0,"ytq_yn_id":1,"ytw_yn_id":0,"cre_date":"2005-09-07","cre_user":"USER_X ","upd_date":"2013-11-27 00:16:56","upd_user":"USER_X"}
  2. 0

    What's the native snappy library when running jar with Hadoop

    Stack Overflow | 12 months ago | dennis
    java.lang.RuntimeException: native snappy library not available: this version of libhadoop was built without snappy support.
  3. 0

    Apahe storm : Snappy compression with SequencyFileBolt does not work

    Stack Overflow | 2 months ago | AvneeshAtri
    java.lang.RuntimeException: Error preparing HdfsBolt: native snappy library not available: this version of libhadoop was built without snappy support.
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Having issues with snappy compression

    flume-user | 1 year ago | Dinesh Narayanan
    org.apache.flume.EventDeliveryException: java.lang.RuntimeException: native snappy library not available: this version of libhadoop was built without snappy support.
  6. 0

    Native library lz4 not available for Spark

    Stack Overflow | 6 months ago | javadba
    java.lang.RuntimeException: native lz4 library not available

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.lang.RuntimeException

      native snappy library not available: this version of libhadoop was built without snappy support.

      at org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded()
    2. Hadoop
      SequenceFile.createWriter
      1. org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:65)
      2. org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:134)
      3. org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:150)
      4. org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:165)
      5. org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1201)
      6. org.apache.hadoop.io.SequenceFile$Writer.<init>(SequenceFile.java:1094)
      7. org.apache.hadoop.io.SequenceFile$BlockCompressWriter.<init>(SequenceFile.java:1444)
      8. org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:277)
      9. org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:530)
      9 frames
    3. Hive Query Language
      ExecMapper.map
      1. org.apache.hadoop.hive.ql.exec.Utilities.createSequenceWriter(Utilities.java:1508)
      2. org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat.getHiveRecordWriter(HiveSequenceFileOutputFormat.java:64)
      3. org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getRecordWriter(HiveFileFormatUtils.java:261)
      4. org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:246)
      5. org.apache.hadoop.hive.ql.exec.FileSinkOperator.createBucketForFileIdx(FileSinkOperator.java:622)
      6. org.apache.hadoop.hive.ql.exec.FileSinkOperator.createBucketFiles(FileSinkOperator.java:566)
      7. org.apache.hadoop.hive.ql.exec.FileSinkOperator.process(FileSinkOperator.java:675)
      8. org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:837)
      9. org.apache.hadoop.hive.ql.exec.SelectOperator.process(SelectOperator.java:88)
      10. org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:837)
      11. org.apache.hadoop.hive.ql.exec.FilterOperator.process(FilterOperator.java:122)
      12. org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:837)
      13. org.apache.hadoop.hive.ql.exec.TableScanOperator.process(TableScanOperator.java:97)
      14. org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:162)
      15. org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:508)
      16. org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:163)
      16 frames
    4. Hadoop
      LocalContainerLauncher$EventHandler$1.run
      1. org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
      2. org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
      3. org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
      4. org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:380)
      5. org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:301)
      6. org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:187)
      7. org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:230)
      7 frames
    5. Java RT
      Thread.run
      1. java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
      2. java.util.concurrent.FutureTask.run(FutureTask.java:262)
      3. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
      4. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
      5. java.lang.Thread.run(Thread.java:745)
      5 frames