org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 2.0 (TID 2, 172.26.28.101): scala.MatchError: UUIDType (of class org.apache.spark.sql.cassandra.types.UUIDType$)

DataStax JIRA | Alexander Sedov | 1 year ago
tip
Do you know that we can give you better hits? Get more relevant results from Samebug’s stack trace search.
  1. 0

    My C* has tables {code} CREATE TABLE csod.role ( object_id uuid, code text, description text, level int, name text, solr_query text, PRIMARY KEY (object_id) ) {code} and {code} CREATE TABLE csod.user_role ( role uuid, user uuid, role_name text, solr_query text, PRIMARY KEY (role, user) ) {code} When I try to use CassandraSQLContext in the Spark shell for joining this tables I get an exception: {code} scala> csc.sql("select * from role r join user_role ur on r.object_id = ur.role").collect WARN 2016-02-10 16:44:46 org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 2.0 (TID 2, 172.26.28.101): scala.MatchError: UUIDType (of class org.apache.spark.sql.cassandra.types.UUIDType$) at org.apache.spark.sql.execution.SparkSqlSerializer2$$anonfun$createSerializationFunction$1.apply(SparkSqlSerializer2.scala:232) at org.apache.spark.sql.execution.SparkSqlSerializer2$$anonfun$createSerializationFunction$1.apply(SparkSqlSerializer2.scala:227) at org.apache.spark.sql.execution.Serializer2SerializationStream.writeKey(SparkSqlSerializer2.scala:65) at org.apache.spark.storage.DiskBlockObjectWriter.write(BlockObjectWriter.scala:206) at org.apache.spark.util.collection.WritablePartitionedIterator$$anon$3.writeNext(WritablePartitionedPairCollection.scala:104) at org.apache.spark.util.collection.ExternalSorter.spillToPartitionFiles(ExternalSorter.scala:375) at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:208) at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:62) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:70) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at org.apache.spark.scheduler.Task.run(Task.scala:70) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) {code} If I understand right the join should work just like a string composition but it doesn't.

    DataStax JIRA | 1 year ago | Alexander Sedov
    org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 2.0 (TID 2, 172.26.28.101): scala.MatchError: UUIDType (of class org.apache.spark.sql.cassandra.types.UUIDType$)
  2. 0

    My C* has tables {code} CREATE TABLE csod.role ( object_id uuid, code text, description text, level int, name text, solr_query text, PRIMARY KEY (object_id) ) {code} and {code} CREATE TABLE csod.user_role ( role uuid, user uuid, role_name text, solr_query text, PRIMARY KEY (role, user) ) {code} When I try to use CassandraSQLContext in the Spark shell for joining this tables I get an exception: {code} scala> csc.sql("select * from role r join user_role ur on r.object_id = ur.role").collect WARN 2016-02-10 16:44:46 org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 2.0 (TID 2, 172.26.28.101): scala.MatchError: UUIDType (of class org.apache.spark.sql.cassandra.types.UUIDType$) at org.apache.spark.sql.execution.SparkSqlSerializer2$$anonfun$createSerializationFunction$1.apply(SparkSqlSerializer2.scala:232) at org.apache.spark.sql.execution.SparkSqlSerializer2$$anonfun$createSerializationFunction$1.apply(SparkSqlSerializer2.scala:227) at org.apache.spark.sql.execution.Serializer2SerializationStream.writeKey(SparkSqlSerializer2.scala:65) at org.apache.spark.storage.DiskBlockObjectWriter.write(BlockObjectWriter.scala:206) at org.apache.spark.util.collection.WritablePartitionedIterator$$anon$3.writeNext(WritablePartitionedPairCollection.scala:104) at org.apache.spark.util.collection.ExternalSorter.spillToPartitionFiles(ExternalSorter.scala:375) at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:208) at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:62) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:70) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at org.apache.spark.scheduler.Task.run(Task.scala:70) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) {code} If I understand right the join should work just like a string composition but it doesn't.

    DataStax JIRA | 1 year ago | Alexander Sedov
    org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 2.0 (TID 2, 172.26.28.101): scala.MatchError: UUIDType (of class org.apache.spark.sql.cassandra.types.UUIDType$)
  3. 0

    Spark SQL Join error on cassandra UUID types

    Stack Overflow | 1 year ago | jguerra
    org.apache.spark.scheduler.TaskSetManager: Lost task 3.0 in stage 0.0 (TID 6, 161.72.45.76): scala.MatchError: UUIDType (of class org.apache.spark.sql.cassandra.types.UUIDType$)
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. org.apache.spark.scheduler.TaskSetManager

      Lost task 0.0 in stage 2.0 (TID 2, 172.26.28.101): scala.MatchError: UUIDType (of class org.apache.spark.sql.cassandra.types.UUIDType$)

      at org.apache.spark.sql.execution.SparkSqlSerializer2$$anonfun$createSerializationFunction$1.apply()
    2. Spark Project SQL
      Serializer2SerializationStream.writeKey
      1. org.apache.spark.sql.execution.SparkSqlSerializer2$$anonfun$createSerializationFunction$1.apply(SparkSqlSerializer2.scala:232)
      2. org.apache.spark.sql.execution.SparkSqlSerializer2$$anonfun$createSerializationFunction$1.apply(SparkSqlSerializer2.scala:227)
      3. org.apache.spark.sql.execution.Serializer2SerializationStream.writeKey(SparkSqlSerializer2.scala:65)
      3 frames
    3. Spark
      Executor$TaskRunner.run
      1. org.apache.spark.storage.DiskBlockObjectWriter.write(BlockObjectWriter.scala:206)
      2. org.apache.spark.util.collection.WritablePartitionedIterator$$anon$3.writeNext(WritablePartitionedPairCollection.scala:104)
      3. org.apache.spark.util.collection.ExternalSorter.spillToPartitionFiles(ExternalSorter.scala:375)
      4. org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:208)
      5. org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:62)
      6. org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:70)
      7. org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
      8. org.apache.spark.scheduler.Task.run(Task.scala:70)
      9. org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
      9 frames
    4. Java RT
      Thread.run
      1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
      2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
      3. java.lang.Thread.run(Thread.java:745)
      3 frames