com.datastax.spark.connector.types.TypeConversionException: Failed to convert column X of vassil_test.general_6 to java.lang.Long: null

DataStax JIRA | Vassil Lunchev | 11 months ago
  1. 0

    Issue related to Connector 1.5.0-RC1. The same code was tested with Connecter 1.4.1 and it works fine. (see the 'Environment' field) When trying to read from a Cassandra table with CassandraJavaUtil.mapRowTo(): {code:java} JavaRDD<Iterable<General6>> shardSessions = CassandraJavaUtil .javaFunctions(context) .cassandraTable(LIVE_SESSIONS_KEYSPACE_NAME, LIVE_SESSIONS_TABLE_NAME, CassandraJavaUtil.mapRowTo(General6.class)) {code} It cannot map a nullable column with value null to java.lang.Long. {code:java} com.datastax.spark.connector.types.TypeConversionException: Failed to convert column X of vassil_test.general_6 to java.lang.Long: null at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.tryConvert(GettableDataToMappedTypeConverter.scala:135) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.convertedColumnValue(GettableDataToMappedTypeConverter.scala:161) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$setterParamValue(GettableDataToMappedTypeConverter.scala:210) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter$$anonfun$com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$invokeSetters$2.apply(GettableDataToMappedTypeConverter.scala:250) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter$$anonfun$com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$invokeSetters$2.apply(GettableDataToMappedTypeConverter.scala:249) at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108) at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$invokeSetters(GettableDataToMappedTypeConverter.scala:249) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter$$anonfun$convertPF$1.applyOrElse(GettableDataToMappedTypeConverter.scala:263) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.convert(GettableDataToMappedTypeConverter.scala:20) at com.datastax.spark.connector.rdd.reader.ClassBasedRowReader.read(ClassBasedRowReader.scala:37) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$12.apply(CassandraTableScanRDD.scala:208) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$12.apply(CassandraTableScanRDD.scala:208) at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) at scala.collection.Iterator$$anon$13.next(Iterator.scala:372) at com.datastax.spark.connector.util.CountingIterator.next(CountingIterator.scala:16) at com.datastax.spark.connector.util.BufferedIterator2.advance(BufferedIterator2.scala:24) at com.datastax.spark.connector.util.BufferedIterator2.<init>(BufferedIterator2.scala:13) at com.datastax.spark.connector.util.SpanningIterator.<init>(SpanningIterator.scala:21) at com.datastax.spark.connector.rdd.SpannedRDD.compute(SpannedRDD.scala:29) at com.datastax.spark.connector.rdd.SpannedRDD.compute(SpannedRDD.scala:22) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:88) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Caused by: com.datastax.spark.connector.types.TypeConversionException: Cannot convert object null to java.lang.Long. at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:47) at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$TimeTypeConverter$$anonfun$convertPF$21.applyOrElse(TypeConverter.scala:458) at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:166) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.types.ChainedTypeConverter.convert(TypeConverter.scala:63) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.tryConvert(GettableDataToMappedTypeConverter.scala:132) ... 229 more {code}

    DataStax JIRA | 11 months ago | Vassil Lunchev
    com.datastax.spark.connector.types.TypeConversionException: Failed to convert column X of vassil_test.general_6 to java.lang.Long: null
  2. 0

    Issue related to Connector 1.5.0-RC1. The same code was tested with Connecter 1.4.1 and it works fine. (see the 'Environment' field) When trying to read from a Cassandra table with CassandraJavaUtil.mapRowTo(): {code:java} JavaRDD<Iterable<General6>> shardSessions = CassandraJavaUtil .javaFunctions(context) .cassandraTable(LIVE_SESSIONS_KEYSPACE_NAME, LIVE_SESSIONS_TABLE_NAME, CassandraJavaUtil.mapRowTo(General6.class)) {code} It cannot map a nullable column with value null to java.lang.Long. {code:java} com.datastax.spark.connector.types.TypeConversionException: Failed to convert column X of vassil_test.general_6 to java.lang.Long: null at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.tryConvert(GettableDataToMappedTypeConverter.scala:135) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.convertedColumnValue(GettableDataToMappedTypeConverter.scala:161) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$setterParamValue(GettableDataToMappedTypeConverter.scala:210) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter$$anonfun$com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$invokeSetters$2.apply(GettableDataToMappedTypeConverter.scala:250) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter$$anonfun$com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$invokeSetters$2.apply(GettableDataToMappedTypeConverter.scala:249) at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108) at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$invokeSetters(GettableDataToMappedTypeConverter.scala:249) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter$$anonfun$convertPF$1.applyOrElse(GettableDataToMappedTypeConverter.scala:263) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.convert(GettableDataToMappedTypeConverter.scala:20) at com.datastax.spark.connector.rdd.reader.ClassBasedRowReader.read(ClassBasedRowReader.scala:37) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$12.apply(CassandraTableScanRDD.scala:208) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$12.apply(CassandraTableScanRDD.scala:208) at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) at scala.collection.Iterator$$anon$13.next(Iterator.scala:372) at com.datastax.spark.connector.util.CountingIterator.next(CountingIterator.scala:16) at com.datastax.spark.connector.util.BufferedIterator2.advance(BufferedIterator2.scala:24) at com.datastax.spark.connector.util.BufferedIterator2.<init>(BufferedIterator2.scala:13) at com.datastax.spark.connector.util.SpanningIterator.<init>(SpanningIterator.scala:21) at com.datastax.spark.connector.rdd.SpannedRDD.compute(SpannedRDD.scala:29) at com.datastax.spark.connector.rdd.SpannedRDD.compute(SpannedRDD.scala:22) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300) at org.apache.spark.rdd.RDD.iterator(RDD.scala:264) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:88) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Caused by: com.datastax.spark.connector.types.TypeConversionException: Cannot convert object null to java.lang.Long. at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:47) at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$TimeTypeConverter$$anonfun$convertPF$21.applyOrElse(TypeConverter.scala:458) at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:166) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.types.ChainedTypeConverter.convert(TypeConverter.scala:63) at com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.tryConvert(GettableDataToMappedTypeConverter.scala:132) ... 229 more {code}

    DataStax JIRA | 11 months ago | Vassil Lunchev
    com.datastax.spark.connector.types.TypeConversionException: Failed to convert column X of vassil_test.general_6 to java.lang.Long: null
  3. 0

    Cassandra is not working with UDT

    Stack Overflow | 1 year ago | kaushal
    com.datastax.spark.connector.types.TypeConversionException: Cannot convert object null to com.datastax.spark.connector.UDTValue.
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    SPARKC-260 reported the issue when push down VARINT column filters. {code} CREATE TABLE linkcurrent.time_series_counters_2015_09 ( id int, series varint, rollup_minutes varint, period_stamp timestamp, event_type varint, value counter, PRIMARY KEY ((id, series, rollup_minutes), period_stamp, event_type) ) >>> df = sqlContext.read.format("org.apache.spark.sql.cassandra").load(table="time_series_counters_2015_09", keyspace="linkcurrent") >>> test = df.filter("id = 1 AND series = 0 AND rollup_minutes = 60") >>> test.take(1) WARN 2015-10-06 21:19:23 org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 16.0 (TID 22, 127.0.0.1): java.io.IOException: Exception during preparation of SELECT "id", "series", "rollup_minutes", "period_stamp", "event_type", "value" FROM "linkcurrent"."time_series_counters_2015_09" WHERE "id" = ? AND "series" = ? AND "rollup_minutes" = ? ALLOW FILTERING: Cannot convert object 0.0 of type class org.apache.spark.sql.types.Decimal to java.math.BigInteger. at com.datastax.spark.connector.rdd.CassandraTableScanRDD.createStatement(CassandraTableScanRDD.scala:188) at com.datastax.spark.connector.rdd.CassandraTableScanRDD.com$datastax$spark$connector$rdd$CassandraTableScanRDD$$fetchTokenRange(CassandraTableScanRDD.scala:202) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$13.apply(CassandraTableScanRDD.scala:229) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$13.apply(CassandraTableScanRDD.scala:229) at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) at com.datastax.spark.connector.util.CountingIterator.hasNext(CountingIterator.scala:12) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) at scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:388) at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:308) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:207) at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:62) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:70) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at org.apache.spark.scheduler.Task.run(Task.scala:70) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Caused by: com.datastax.spark.connector.types.TypeConversionException: Cannot convert object 0.0 of type class org.apache.spark.sql.types.Decimal to java.math.BigInteger. at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:42) at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:40) at com.datastax.spark.connector.types.TypeConverter$JavaBigIntegerConverter$$anonfun$convertPF$15.applyOrElse(TypeConverter.scala:354) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:40) at com.datastax.spark.connector.types.TypeConverter$JavaBigIntegerConverter$.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:352) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:53) at com.datastax.spark.connector.types.TypeConverter$JavaBigIntegerConverter$.convert(TypeConverter.scala:352) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter$$anonfun$convertPF$26.applyOrElse(TypeConverter.scala:702) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:40) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:695) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:53) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.convert(TypeConverter.scala:695) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$9.apply(CassandraTableScanRDD.scala:181) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$9.apply(CassandraTableScanRDD.scala:180) at scala.collection.TraversableLike$WithFilter$$anonfun$map$2.apply(TraversableLike.scala:722) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at scala.collection.TraversableLike$WithFilter.map(TraversableLike.scala:721) at com.datastax.spark.connector.rdd.CassandraTableScanRDD.createStatement(CassandraTableScanRDD.scala:180) ... 18 more ERROR 2015-10-06 21:19:23 org.apache.spark.scheduler.TaskSetManager: Task 0 in stage 16.0 failed 4 times; aborting job Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/Users/sebastianestevez/Documents/dse-4.8.0/resources/spark/python/pyspark/sql/dataframe.py", line 307, in take return self.limit(num).collect() File "/Users/sebastianestevez/Documents/dse-4.8.0/resources/spark/python/pyspark/sql/dataframe.py", line 281, in collect port = self._sc._jvm.PythonRDD.collectAndServe(self._jdf.javaToPython().rdd()) File "/Users/sebastianestevez/Documents/dse-4.8.0/resources/spark/python/lib/py4j-0.8.2.1-src.zip/py4j/java_gateway.py", line 538, in __call__ File "/Users/sebastianestevez/Documents/dse-4.8.0/resources/spark/python/lib/py4j-0.8.2.1-src.zip/py4j/protocol.py", line 300, in get_return_value py4j.protocol.Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.collectAndServe. : org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 16.0 failed 4 times, most recent failure: Lost task 0.3 in stage 16.0 (TID 25, 127.0.0.1): java.io.IOException: Exception during preparation of SELECT "id", "series", "rollup_minutes", "period_stamp", "event_type", "value" FROM "linkcurrent"."time_series_counters_2015_09" WHERE "id" = ? AND "series" = ? AND "rollup_minutes" = ? ALLOW FILTERING: Cannot convert object 0.0 of type class org.apache.spark.sql.types.Decimal to java.math.BigInteger. at com.datastax.spark.connector.rdd.CassandraTableScanRDD.createStatement(CassandraTableScanRDD.scala:188) at com.datastax.spark.connector.rdd.CassandraTableScanRDD.com$datastax$spark$connector$rdd$CassandraTableScanRDD$$fetchTokenRange(CassandraTableScanRDD.scala:202) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$13.apply(CassandraTableScanRDD.scala:229) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$13.apply(CassandraTableScanRDD.scala:229) at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) at com.datastax.spark.connector.util.CountingIterator.hasNext(CountingIterator.scala:12) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) at scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:388) at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:308) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:207) at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:62) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:70) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at org.apache.spark.scheduler.Task.run(Task.scala:70) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Caused by: com.datastax.spark.connector.types.TypeConversionException: Cannot convert object 0.0 of type class org.apache.spark.sql.types.Decimal to java.math.BigInteger. at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:42) at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:40) at com.datastax.spark.connector.types.TypeConverter$JavaBigIntegerConverter$$anonfun$convertPF$15.applyOrElse(TypeConverter.scala:354) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:40) at com.datastax.spark.connector.types.TypeConverter$JavaBigIntegerConverter$.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:352) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:53) at com.datastax.spark.connector.types.TypeConverter$JavaBigIntegerConverter$.convert(TypeConverter.scala:352) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter$$anonfun$convertPF$26.applyOrElse(TypeConverter.scala:702) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:40) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:695) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:53) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.convert(TypeConverter.scala:695) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$9.apply(CassandraTableScanRDD.scala:181) at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$9.apply(CassandraTableScanRDD.scala:180) at scala.collection.TraversableLike$WithFilter$$anonfun$map$2.apply(TraversableLike.scala:722) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at scala.collection.TraversableLike$WithFilter.map(TraversableLike.scala:721) at com.datastax.spark.connector.rdd.CassandraTableScanRDD.createStatement(CassandraTableScanRDD.scala:180) ... 18 more Driver stacktrace: at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1273) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1264) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1263) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1263) at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:730) at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:730) at scala.Option.foreach(Option.scala:236) at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:730) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1457) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1418) at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) {code}

    DataStax JIRA | 1 year ago | Alex Liu
    org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 16.0 failed 4 times, most recent failure: Lost task 0.3 in stage 16.0 (TID 25, 127.0.0.1): java.io.IOException: Exception during preparation of SELECT "id", "series", "rollup_minutes", "period_stamp", "event_type", "value" FROM "linkcurrent"."time_series_counters_2015_09" WHERE "id" = ? AND "series" = ? AND "rollup_minutes" = ? ALLOW FILTERING: Cannot convert object 0.0 of type class org.apache.spark.sql.types.Decimal to java.math.BigInteger.
  6. 0

    My Cassandra Table is: {code:java} CREATE TABLE keyspace.wish_counter ( wish_date date, wish_published_time timeuuid, wish_counter_value counter, PRIMARY KEY (wish_date, wish_published_time) ) WITH CLUSTERING ORDER BY (wish_published_time ASC) {code} I'm loading data from Cassandra into a class 'WishCountTable' : {code:java} class WishCountTable extends Serializable { var wish_date: DateTime = new DateTime(0) var wish_published_time: UUID = new UUID(0L, 0L) var wish_counter_value: Long = 0L } {code} Everything is alright but whenever I try to save data into cassandra, I get an error. {code:java} saveRDD.saveToCassandra(keyspace, "wish_counter") {code} h4. ERROR: {code:java} 16/03/07 19:18:08 ERROR Executor: Exception in task 0.0 in stage 1.0 (TID 5) com.datastax.spark.connector.types.TypeConversionException: Cannot convert object 2016-02-21T06:00:00.000+06:00 of type class org.joda.time.DateTime to com.datastax.driver.core.LocalDate. at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:45) at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$$anonfun$convertPF$20.applyOrElse(TypeConverter.scala:447) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:437) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:56) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$.convert(TypeConverter.scala:437) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter$$anonfun$convertPF$28.applyOrElse(TypeConverter.scala:756) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:749) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:56) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.convert(TypeConverter.scala:749) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1$$anonfun$convertPF$1$$anonfun$applyOrElse$1.apply$mcVI$sp(MappedToGettableDataConverter.scala:170) at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1$$anonfun$convertPF$1.applyOrElse(MappedToGettableDataConverter.scala:169) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1.convert(MappedToGettableDataConverter.scala:18) at com.datastax.spark.connector.writer.DefaultRowWriter.readColumnValues(DefaultRowWriter.scala:21) at com.datastax.spark.connector.writer.BoundStatementBuilder.bind(BoundStatementBuilder.scala:35) at com.datastax.spark.connector.writer.GroupingBatchBuilder.next(GroupingBatchBuilder.scala:106) at com.datastax.spark.connector.writer.GroupingBatchBuilder.next(GroupingBatchBuilder.scala:31) at scala.collection.Iterator$class.foreach(Iterator.scala:727) at com.datastax.spark.connector.writer.GroupingBatchBuilder.foreach(GroupingBatchBuilder.scala:31) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:155) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:139) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:110) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:109) at com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:139) at com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:109) at com.datastax.spark.connector.writer.TableWriter.write(TableWriter.scala:139) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:37) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:37) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:88) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) 16/03/07 19:18:08 ERROR TaskSetManager: Task 0 in stage 1.0 failed 1 times; aborting job Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 1.0 failed 1 times, most recent failure: Lost task 0.0 in stage 1.0 (TID 5, localhost): com.datastax.spark.connector.types.TypeConversionException: Cannot convert object 2016-02-21T06:00:00.000+06:00 of type class org.joda.time.DateTime to com.datastax.driver.core.LocalDate. at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:45) at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$$anonfun$convertPF$20.applyOrElse(TypeConverter.scala:447) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:437) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:56) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$.convert(TypeConverter.scala:437) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter$$anonfun$convertPF$28.applyOrElse(TypeConverter.scala:756) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:749) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:56) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.convert(TypeConverter.scala:749) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1$$anonfun$convertPF$1$$anonfun$applyOrElse$1.apply$mcVI$sp(MappedToGettableDataConverter.scala:170) at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1$$anonfun$convertPF$1.applyOrElse(MappedToGettableDataConverter.scala:169) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1.convert(MappedToGettableDataConverter.scala:18) at com.datastax.spark.connector.writer.DefaultRowWriter.readColumnValues(DefaultRowWriter.scala:21) at com.datastax.spark.connector.writer.BoundStatementBuilder.bind(BoundStatementBuilder.scala:35) at com.datastax.spark.connector.writer.GroupingBatchBuilder.next(GroupingBatchBuilder.scala:106) at com.datastax.spark.connector.writer.GroupingBatchBuilder.next(GroupingBatchBuilder.scala:31) at scala.collection.Iterator$class.foreach(Iterator.scala:727) at com.datastax.spark.connector.writer.GroupingBatchBuilder.foreach(GroupingBatchBuilder.scala:31) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:155) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:139) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:110) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:109) at com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:139) at com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:109) at com.datastax.spark.connector.writer.TableWriter.write(TableWriter.scala:139) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:37) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:37) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:88) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Driver stacktrace: at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1283) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1271) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1270) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1270) at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697) at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697) at scala.Option.foreach(Option.scala:236) at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:697) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1496) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1458) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1447) at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:567) at org.apache.spark.SparkContext.runJob(SparkContext.scala:1822) at org.apache.spark.SparkContext.runJob(SparkContext.scala:1835) at org.apache.spark.SparkContext.runJob(SparkContext.scala:1912) at com.datastax.spark.connector.RDDFunctions.saveToCassandra(RDDFunctions.scala:37) at org.qm.UpdateWishTable$.main(UpdateWishTable.scala:93) at org.qm.UpdateWishTable.main(UpdateWishTable.scala) Caused by: com.datastax.spark.connector.types.TypeConversionException: Cannot convert object 2016-02-21T06:00:00.000+06:00 of type class org.joda.time.DateTime to com.datastax.driver.core.LocalDate. at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:45) at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$$anonfun$convertPF$20.applyOrElse(TypeConverter.scala:447) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:437) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:56) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$.convert(TypeConverter.scala:437) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter$$anonfun$convertPF$28.applyOrElse(TypeConverter.scala:756) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:749) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:56) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.convert(TypeConverter.scala:749) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1$$anonfun$convertPF$1$$anonfun$applyOrElse$1.apply$mcVI$sp(MappedToGettableDataConverter.scala:170) at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1$$anonfun$convertPF$1.applyOrElse(MappedToGettableDataConverter.scala:169) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1.convert(MappedToGettableDataConverter.scala:18) at com.datastax.spark.connector.writer.DefaultRowWriter.readColumnValues(DefaultRowWriter.scala:21) at com.datastax.spark.connector.writer.BoundStatementBuilder.bind(BoundStatementBuilder.scala:35) at com.datastax.spark.connector.writer.GroupingBatchBuilder.next(GroupingBatchBuilder.scala:106) at com.datastax.spark.connector.writer.GroupingBatchBuilder.next(GroupingBatchBuilder.scala:31) at scala.collection.Iterator$class.foreach(Iterator.scala:727) at com.datastax.spark.connector.writer.GroupingBatchBuilder.foreach(GroupingBatchBuilder.scala:31) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:155) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:139) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:110) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:109) at com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:139) at com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:109) at com.datastax.spark.connector.writer.TableWriter.write(TableWriter.scala:139) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:37) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:37) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:88) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) 16/03/07 19:18:08 ERROR Executor: Exception in task 1.0 in stage 1.0 (TID 6) com.datastax.spark.connector.types.TypeConversionException: Cannot convert object 2016-02-28T06:00:00.000+06:00 of type class org.joda.time.DateTime to com.datastax.driver.core.LocalDate. at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:45) at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$$anonfun$convertPF$20.applyOrElse(TypeConverter.scala:447) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:437) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:56) at com.datastax.spark.connector.types.TypeConverter$LocalDateConverter$.convert(TypeConverter.scala:437) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter$$anonfun$convertPF$28.applyOrElse(TypeConverter.scala:756) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.com$datastax$spark$connector$types$NullableTypeConverter$$super$convert(TypeConverter.scala:749) at com.datastax.spark.connector.types.NullableTypeConverter$class.convert(TypeConverter.scala:56) at com.datastax.spark.connector.types.TypeConverter$OptionToNullConverter.convert(TypeConverter.scala:749) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1$$anonfun$convertPF$1$$anonfun$applyOrElse$1.apply$mcVI$sp(MappedToGettableDataConverter.scala:170) at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1$$anonfun$convertPF$1.applyOrElse(MappedToGettableDataConverter.scala:169) at com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43) at com.datastax.spark.connector.writer.MappedToGettableDataConverter$$anon$1.convert(MappedToGettableDataConverter.scala:18) at com.datastax.spark.connector.writer.DefaultRowWriter.readColumnValues(DefaultRowWriter.scala:21) at com.datastax.spark.connector.writer.BoundStatementBuilder.bind(BoundStatementBuilder.scala:35) at com.datastax.spark.connector.writer.GroupingBatchBuilder.next(GroupingBatchBuilder.scala:106) at com.datastax.spark.connector.writer.GroupingBatchBuilder.next(GroupingBatchBuilder.scala:31) at scala.collection.Iterator$class.foreach(Iterator.scala:727) at com.datastax.spark.connector.writer.GroupingBatchBuilder.foreach(GroupingBatchBuilder.scala:31) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:155) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:139) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:110) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:109) at com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:139) at com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:109) at com.datastax.spark.connector.writer.TableWriter.write(TableWriter.scala:139) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:37) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:37) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:88) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) {code}

    DataStax JIRA | 9 months ago | Safat Siddiqui
    com.datastax.spark.connector.types.TypeConversionException: Cannot convert object 2016-02-21T06:00:00.000+06:00 of type class org.joda.time.DateTime to com.datastax.driver.core.LocalDate.

    1 unregistered visitors
    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. com.datastax.spark.connector.types.TypeConversionException

      Cannot convert object null to java.lang.Long.

      at com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply()
    2. spark-cassandra-connector
      TypeConverter$TimeTypeConverter$$anonfun$convertPF$21.applyOrElse
      1. com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:47)
      2. com.datastax.spark.connector.types.TypeConverter$$anonfun$convert$1.apply(TypeConverter.scala:43)
      3. com.datastax.spark.connector.types.TypeConverter$TimeTypeConverter$$anonfun$convertPF$21.applyOrElse(TypeConverter.scala:458)
      3 frames
    3. Scala
      PartialFunction$OrElse.applyOrElse
      1. scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:166)
      1 frame
    4. spark-cassandra-connector
      GettableDataToMappedTypeConverter$$anonfun$com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$invokeSetters$2.apply
      1. com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43)
      2. com.datastax.spark.connector.types.ChainedTypeConverter.convert(TypeConverter.scala:63)
      3. com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.tryConvert(GettableDataToMappedTypeConverter.scala:132)
      4. com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.convertedColumnValue(GettableDataToMappedTypeConverter.scala:161)
      5. com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$setterParamValue(GettableDataToMappedTypeConverter.scala:210)
      6. com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter$$anonfun$com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$invokeSetters$2.apply(GettableDataToMappedTypeConverter.scala:250)
      7. com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter$$anonfun$com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$invokeSetters$2.apply(GettableDataToMappedTypeConverter.scala:249)
      7 frames
    5. Scala
      TraversableLike$WithFilter.foreach
      1. scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772)
      2. scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
      3. scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
      4. scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771)
      4 frames
    6. spark-cassandra-connector
      CassandraTableScanRDD$$anonfun$12.apply
      1. com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.com$datastax$spark$connector$rdd$reader$GettableDataToMappedTypeConverter$$invokeSetters(GettableDataToMappedTypeConverter.scala:249)
      2. com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter$$anonfun$convertPF$1.applyOrElse(GettableDataToMappedTypeConverter.scala:263)
      3. com.datastax.spark.connector.types.TypeConverter$class.convert(TypeConverter.scala:43)
      4. com.datastax.spark.connector.rdd.reader.GettableDataToMappedTypeConverter.convert(GettableDataToMappedTypeConverter.scala:20)
      5. com.datastax.spark.connector.rdd.reader.ClassBasedRowReader.read(ClassBasedRowReader.scala:37)
      6. com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$12.apply(CassandraTableScanRDD.scala:208)
      7. com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$12.apply(CassandraTableScanRDD.scala:208)
      7 frames
    7. Scala
      Iterator$$anon$13.next
      1. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
      2. scala.collection.Iterator$$anon$13.next(Iterator.scala:372)
      2 frames
    8. spark-cassandra-connector
      SpannedRDD.compute
      1. com.datastax.spark.connector.util.CountingIterator.next(CountingIterator.scala:16)
      2. com.datastax.spark.connector.util.BufferedIterator2.advance(BufferedIterator2.scala:24)
      3. com.datastax.spark.connector.util.BufferedIterator2.<init>(BufferedIterator2.scala:13)
      4. com.datastax.spark.connector.util.SpanningIterator.<init>(SpanningIterator.scala:21)
      5. com.datastax.spark.connector.rdd.SpannedRDD.compute(SpannedRDD.scala:29)
      6. com.datastax.spark.connector.rdd.SpannedRDD.compute(SpannedRDD.scala:22)
      6 frames
    9. Spark
      Executor$TaskRunner.run
      1. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      2. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      3. org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      4. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      5. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      6. org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      7. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      8. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      9. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      10. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      11. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      12. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      13. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      14. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      15. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      16. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      17. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      18. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      19. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      20. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      21. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      22. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      23. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      24. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      25. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      26. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      27. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      28. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      29. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      30. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      31. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      32. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      33. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      34. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      35. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      36. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      37. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      38. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      39. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      40. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      41. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      42. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      43. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      44. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      45. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      46. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      47. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      48. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      49. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      50. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      51. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      52. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      53. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      54. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      55. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      56. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      57. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      58. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      59. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      60. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      61. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      62. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      63. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      64. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      65. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      66. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      67. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      68. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      69. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      70. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      71. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      72. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      73. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      74. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      75. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      76. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      77. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      78. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      79. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      80. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      81. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      82. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      83. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      84. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      85. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      86. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      87. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      88. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      89. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      90. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      91. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      92. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      93. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      94. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      95. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      96. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      97. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      98. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      99. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      100. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      101. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      102. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      103. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      104. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      105. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      106. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      107. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      108. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      109. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      110. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      111. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      112. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      113. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      114. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      115. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      116. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      117. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      118. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      119. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      120. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      121. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      122. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      123. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      124. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      125. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      126. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      127. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      128. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      129. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      130. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      131. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      132. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      133. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      134. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      135. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      136. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      137. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      138. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      139. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      140. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      141. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      142. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      143. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      144. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      145. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      146. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      147. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      148. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      149. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      150. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      151. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      152. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      153. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      154. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      155. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      156. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      157. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      158. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      159. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      160. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      161. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      162. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      163. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      164. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      165. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      166. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      167. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      168. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      169. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      170. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      171. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      172. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      173. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      174. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      175. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      176. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      177. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      178. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      179. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      180. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      181. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      182. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      183. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      184. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      185. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      186. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      187. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      188. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      189. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      190. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      191. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      192. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      193. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      194. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      195. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      196. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      197. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      198. org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
      199. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
      200. org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
      201. org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
      202. org.apache.spark.scheduler.Task.run(Task.scala:88)
      203. org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
      203 frames
    10. Java RT
      Thread.run
      1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
      2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
      3. java.lang.Thread.run(Thread.java:745)
      3 frames