java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row {"name":"John","age":35.1}

JIRA | Siyuan Zhou | 3 years ago
  1. 0

    First reported here: https://groups.google.com/forum/#!topic/mongodb-user/lZaFgOdrnYk Steps to reproduce: Insert the below document into the persons collection {code} db.persons.insert({name: "John", age: 35.1}); {code} Create a table in hive from the persons collection in mongo {code} create table double (name STRING, age DOUBLE) STORED BY 'com.mongodb.hadoop.hive.MongoStorageHandler' TBLPROPERTIES('mongo.uri'='mongodb://localhost:27017/test.persons'); {code} Setup a table in hive to write to a mongo collection {code} create external table double_ext (name string, age double) stored by 'com.mongodb.hadoop.hive.MongoStorageHandler' TBLPROPERTIES('mongo.uri'='mongodb://localhost:27017/test.ext'); {code} Insert data into that table {code} insert into table double_ext select "John", 35.1 from double; {code} The operation fails and this stack trace is printed {noformat} Diagnostic Messages for this Task: java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row {"name":"John","age":35.1} at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:162) at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:430) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:366) at org.apache.hadoop.mapred.Child$4.run(Child.java:255) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:394) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1190) at org.apache.hadoop.mapred.Child.main(Child.java:249) Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row {"name":"John","age":35.1} at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:671) at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) ... 8 more Caused by: java.lang.ClassCastException: org.apache.hadoop.hive.serde2.io.DoubleWritable cannot be cast to java.lang.Double at com.mongodb.hadoop.hive.BSONSerDe.serializePrimitive(BSONSerDe.java:592) at com.mongodb.hadoop.hive.BSONSerDe.serializeObject(BSONSerDe.java:454) at com.mongodb.hadoop.hive.BSONSerDe.serializeStruct(BSONSerDe.java:525) at com.mongodb.hadoop.hive.BSONSerDe.serialize(BSONSerDe.java:442) at org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:617) at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832) at org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832) at org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:90) at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832) at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:652) ... 9 more {noformat}

    JIRA | 3 years ago | Siyuan Zhou
    java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row {"name":"John","age":35.1}
  2. 0

    First reported here: https://groups.google.com/forum/#!topic/mongodb-user/lZaFgOdrnYk Steps to reproduce: Insert the below document into the persons collection {code} db.persons.insert({name: "John", age: 35.1}); {code} Create a table in hive from the persons collection in mongo {code} create table double (name STRING, age DOUBLE) STORED BY 'com.mongodb.hadoop.hive.MongoStorageHandler' TBLPROPERTIES('mongo.uri'='mongodb://localhost:27017/test.persons'); {code} Setup a table in hive to write to a mongo collection {code} create external table double_ext (name string, age double) stored by 'com.mongodb.hadoop.hive.MongoStorageHandler' TBLPROPERTIES('mongo.uri'='mongodb://localhost:27017/test.ext'); {code} Insert data into that table {code} insert into table double_ext select "John", 35.1 from double; {code} The operation fails and this stack trace is printed {noformat} Diagnostic Messages for this Task: java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row {"name":"John","age":35.1} at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:162) at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:430) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:366) at org.apache.hadoop.mapred.Child$4.run(Child.java:255) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:394) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1190) at org.apache.hadoop.mapred.Child.main(Child.java:249) Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row {"name":"John","age":35.1} at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:671) at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) ... 8 more Caused by: java.lang.ClassCastException: org.apache.hadoop.hive.serde2.io.DoubleWritable cannot be cast to java.lang.Double at com.mongodb.hadoop.hive.BSONSerDe.serializePrimitive(BSONSerDe.java:592) at com.mongodb.hadoop.hive.BSONSerDe.serializeObject(BSONSerDe.java:454) at com.mongodb.hadoop.hive.BSONSerDe.serializeStruct(BSONSerDe.java:525) at com.mongodb.hadoop.hive.BSONSerDe.serialize(BSONSerDe.java:442) at org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:617) at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832) at org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832) at org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:90) at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832) at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:652) ... 9 more {noformat}

    JIRA | 3 years ago | Siyuan Zhou
    java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row {"name":"John","age":35.1}
  3. 0

    {noformat} SELECT SUM(L_QUANTITY), (SUM(L_QUANTITY) + -1.30000000000000000000E+000), (-2.20000000000000020000E+000 % (SUM(L_QUANTITY) + -1.30000000000000000000E+000)), MIN(L_EXTENDEDPRICE) FROM lineitem_orc WHERE ((L_EXTENDEDPRICE <= L_LINENUMBER) OR (L_TAX > L_EXTENDEDPRICE)); {noformat} executed over tpch line item with scale factor 1gb {noformat} 13/06/15 11:19:17 WARN conf.HiveConf: DEPRECATED: Configuration property hive.metastore.local no longer has any effect. Make sure to provide a valid value for hive.metastore.uris if you are connecting to a remote metastore. Logging initialized using configuration in file:/C:/Hadoop/hive-0.9.0/conf/hive-log4j.properties Hive history file=c:\hadoop\hive-0.9.0\logs\history/hive_job_log_jenkinsuser_5292@SLAVE23-WIN_201306151119_1652846565.txt Total MapReduce jobs = 1 Launching Job 1 out of 1 Number of reduce tasks determined at compile time: 1 In order to change the average load for a reducer (in bytes): set hive.exec.reducers.bytes.per.reducer=<number> In order to limit the maximum number of reducers: set hive.exec.reducers.max=<number> In order to set a constant number of reducers: set mapred.reduce.tasks=<number> Starting Job = job_201306142329_0098, Tracking URL = http://localhost:50030/jobdetails.jsp?jobid=job_201306142329_0098 Kill Command = c:\Hadoop\hadoop-1.1.0-SNAPSHOT\bin\hadoop.cmd job -kill job_201306142329_0098 Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 1 2013-06-15 11:19:47,490 Stage-1 map = 0%, reduce = 0% 2013-06-15 11:20:29,801 Stage-1 map = 76%, reduce = 0% 2013-06-15 11:20:32,849 Stage-1 map = 0%, reduce = 0% 2013-06-15 11:20:35,880 Stage-1 map = 100%, reduce = 100% Ended Job = job_201306142329_0098 with errors Error during job, obtaining debugging information... Job Tracking URL: http://localhost:50030/jobdetails.jsp?jobid=job_201306142329_0098 Examining task ID: task_201306142329_0098_m_000002 (and more) from job job_201306142329_0098 Task with the most failures(4): ----- Task ID: task_201306142329_0098_m_000000 URL: http://localhost:50030/taskdetails.jsp?jobid=job_201306142329_0098&tipid=task_201306142329_0098_m_000000 ----- Diagnostic Messages for this Task: java.lang.RuntimeException: Hive Runtime Error while closing operators at org.apache.hadoop.hive.ql.exec.vector.VectorExecMapper.close(VectorExecMapper.java:229) at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:57) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) at org.apache.hadoop.mapred.Child$4.run(Child.java:271) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135) at org.apache.hadoop.mapred.Child.main(Child.java:265) Caused by: java.lang.ClassCastException: org.apache.hadoop.io.NullWritable cannot be cast to org.apache.hadoop.hive.serde2.io.DoubleWritable at org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableDoubleObjectInspector.get(WritableDoubleObjectInspector.java:35) at org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe.serialize(LazyBinarySerDe.java:340) at org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe.serializeStruct(LazyBinarySerDe.java:257) at org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe.serialize(LazyBinarySerDe.java:204) at org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:245) at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832) at org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator.flush(VectorGroupByOperator.java:281) at org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator.closeOp(VectorGroupByOperator.java:423) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:588) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.vector.VectorExecMapper.close(VectorExecMapper.java:196) ... 8 more FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.MapRedTask MapReduce Jobs Launched: Job 0: Map: 1 Reduce: 1 HDFS Read: 0 HDFS Write: 0 FAIL Total MapReduce CPU Time Spent: 0 msec {noformat} Similar issues seen with other queries: {noformat} SELECT VAR_SAMP(L_SUPPKEY), (VAR_SAMP(L_SUPPKEY) - -2.20000000000000020000E+000), (-(VAR_SAMP(L_SUPPKEY))) FROM lineitem_orc WHERE ((L_SUPPKEY = -1)); {noformat} {noformat} 13/06/15 11:41:08 WARN conf.HiveConf: DEPRECATED: Configuration property hive.metastore.local no longer has any effect. Make sure to provide a valid value for hive.metastore.uris if you are connecting to a remote metastore. Logging initialized using configuration in file:/C:/Hadoop/hive-0.9.0/conf/hive-log4j.properties Hive history file=c:\hadoop\hive-0.9.0\logs\history/hive_job_log_jenkinsuser_6976@SLAVE23-WIN_201306151141_1255577417.txt Total MapReduce jobs = 1 Launching Job 1 out of 1 Number of reduce tasks determined at compile time: 1 In order to change the average load for a reducer (in bytes): set hive.exec.reducers.bytes.per.reducer=<number> In order to limit the maximum number of reducers: set hive.exec.reducers.max=<number> In order to set a constant number of reducers: set mapred.reduce.tasks=<number> Starting Job = job_201306142329_0109, Tracking URL = http://localhost:50030/jobdetails.jsp?jobid=job_201306142329_0109 Kill Command = c:\Hadoop\hadoop-1.1.0-SNAPSHOT\bin\hadoop.cmd job -kill job_201306142329_0109 Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 1 2013-06-15 11:41:38,753 Stage-1 map = 0%, reduce = 0% 2013-06-15 11:42:17,959 Stage-1 map = 100%, reduce = 100% Ended Job = job_201306142329_0109 with errors Error during job, obtaining debugging information... Job Tracking URL: http://localhost:50030/jobdetails.jsp?jobid=job_201306142329_0109 Examining task ID: task_201306142329_0109_m_000002 (and more) from job job_201306142329_0109 Task with the most failures(4): ----- Task ID: task_201306142329_0109_m_000000 URL: http://localhost:50030/taskdetails.jsp?jobid=job_201306142329_0109&tipid=task_201306142329_0109_m_000000 ----- Diagnostic Messages for this Task: java.lang.RuntimeException: Hive Runtime Error while closing operators at org.apache.hadoop.hive.ql.exec.vector.VectorExecMapper.close(VectorExecMapper.java:229) at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:57) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) at org.apache.hadoop.mapred.Child$4.run(Child.java:271) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135) at org.apache.hadoop.mapred.Child.main(Child.java:265) Caused by: java.lang.ClassCastException: org.apache.hadoop.io.NullWritable cannot be cast to [Ljava.lang.Object; at org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldData(StandardStructObjectInspector.java:166) at org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe.serializeStruct(LazyBinarySerDe.java:248) at org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe.serialize(LazyBinarySerDe.java:534) at org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe.serializeStruct(LazyBinarySerDe.java:257) at org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe.serialize(LazyBinarySerDe.java:204) at org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:245) at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832) at org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator.flush(VectorGroupByOperator.java:281) at org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator.closeOp(VectorGroupByOperator.java:423) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:588) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.vector.VectorExecMapper.close(VectorExecMapper.java:196) ... 8 more FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.MapRedTask MapReduce Jobs Launched: Job 0: Map: 1 Reduce: 1 HDFS Read: 0 HDFS Write: 0 FAIL Total MapReduce CPU Time Spent: 0 msec {noformat} {noformat} SELECT MAX(L_ORDERKEY), (MAX(L_ORDERKEY) + 2) FROM lineitem_orc WHERE ((L_ORDERKEY <= L_DISCOUNT)); {noformat} {noformat} 13/06/15 11:55:02 WARN conf.HiveConf: DEPRECATED: Configuration property hive.metastore.local no longer has any effect. Make sure to provide a valid value for hive.metastore.uris if you are connecting to a remote metastore. Logging initialized using configuration in file:/C:/Hadoop/hive-0.9.0/conf/hive-log4j.properties Hive history file=c:\hadoop\hive-0.9.0\logs\history/hive_job_log_jenkinsuser_3208@SLAVE23-WIN_201306151155_919702960.txt Total MapReduce jobs = 1 Launching Job 1 out of 1 Number of reduce tasks determined at compile time: 1 In order to change the average load for a reducer (in bytes): set hive.exec.reducers.bytes.per.reducer=<number> In order to limit the maximum number of reducers: set hive.exec.reducers.max=<number> In order to set a constant number of reducers: set mapred.reduce.tasks=<number> Starting Job = job_201306142329_0114, Tracking URL = http://localhost:50030/jobdetails.jsp?jobid=job_201306142329_0114 Kill Command = c:\Hadoop\hadoop-1.1.0-SNAPSHOT\bin\hadoop.cmd job -kill job_201306142329_0114 Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 1 2013-06-15 11:55:34,191 Stage-1 map = 0%, reduce = 0% 2013-06-15 11:56:14,464 Stage-1 map = 100%, reduce = 100% Ended Job = job_201306142329_0114 with errors Error during job, obtaining debugging information... Job Tracking URL: http://localhost:50030/jobdetails.jsp?jobid=job_201306142329_0114 Examining task ID: task_201306142329_0114_m_000002 (and more) from job job_201306142329_0114 Task with the most failures(4): ----- Task ID: task_201306142329_0114_m_000000 URL: http://localhost:50030/taskdetails.jsp?jobid=job_201306142329_0114&tipid=task_201306142329_0114_m_000000 ----- Diagnostic Messages for this Task: java.lang.RuntimeException: Hive Runtime Error while closing operators at org.apache.hadoop.hive.ql.exec.vector.VectorExecMapper.close(VectorExecMapper.java:229) at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:57) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) at org.apache.hadoop.mapred.Child$4.run(Child.java:271) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135) at org.apache.hadoop.mapred.Child.main(Child.java:265) Caused by: java.lang.ClassCastException: org.apache.hadoop.io.NullWritable cannot be cast to org.apache.hadoop.io.LongWritable at org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableLongObjectInspector.get(WritableLongObjectInspector.java:35) at org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe.serialize(LazyBinarySerDe.java:325) at org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe.serializeStruct(LazyBinarySerDe.java:257) at org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe.serialize(LazyBinarySerDe.java:204) at org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:245) at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502) at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832) at org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator.flush(VectorGroupByOperator.java:281) at org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator.closeOp(VectorGroupByOperator.java:423) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:588) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597) at org.apache.hadoop.hive.ql.exec.vector.VectorExecMapper.close(VectorExecMapper.java:196) ... 8 more FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.MapRedTask MapReduce Jobs Launched: Job 0: Map: 1 Reduce: 1 HDFS Read: 0 HDFS Write: 0 FAIL Total MapReduce CPU Time Spent: 0 msec {noformat}

    Apache's JIRA Issue Tracker | 3 years ago | Tony Murphy
    java.lang.RuntimeException: Hive Runtime Error while closing operators
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    org.apache.hadoop.io.DoubleWritable cannot be cast to org.apache.hadoop.hive.serde2.io.DoubleWritable

    GitHub | 3 years ago | xelllee
    org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.ClassCastException: org.apache.hadoop.io.DoubleWritable cannot be cast to org.apache.hadoop.hive.serde2.io.DoubleWritable

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.lang.ClassCastException

      org.apache.hadoop.hive.serde2.io.DoubleWritable cannot be cast to java.lang.Double

      at com.mongodb.hadoop.hive.BSONSerDe.serializePrimitive()
    2. com.mongodb.hadoop
      BSONSerDe.serialize
      1. com.mongodb.hadoop.hive.BSONSerDe.serializePrimitive(BSONSerDe.java:592)
      2. com.mongodb.hadoop.hive.BSONSerDe.serializeObject(BSONSerDe.java:454)
      3. com.mongodb.hadoop.hive.BSONSerDe.serializeStruct(BSONSerDe.java:525)
      4. com.mongodb.hadoop.hive.BSONSerDe.serialize(BSONSerDe.java:442)
      4 frames
    3. Hive Query Language
      ExecMapper.map
      1. org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:617)
      2. org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502)
      3. org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832)
      4. org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
      5. org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502)
      6. org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832)
      7. org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:90)
      8. org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502)
      9. org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832)
      10. org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:652)
      11. org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144)
      11 frames
    4. Hadoop
      Child$4.run
      1. org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
      2. org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:430)
      3. org.apache.hadoop.mapred.MapTask.run(MapTask.java:366)
      4. org.apache.hadoop.mapred.Child$4.run(Child.java:255)
      4 frames
    5. Java RT
      Subject.doAs
      1. java.security.AccessController.doPrivileged(Native Method)
      2. javax.security.auth.Subject.doAs(Subject.java:394)
      2 frames
    6. Hadoop
      UserGroupInformation.doAs
      1. org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1190)
      1 frame
    7. Hadoop
      Child.main
      1. org.apache.hadoop.mapred.Child.main(Child.java:249)
      1 frame