org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, test-cluster-w-1.c.test-project.internal): java.lang.AbstractMethodError: uk.co.test.CalculateScore$$Lambda$10/1666820030.call(Ljava/lang/Object;)Ljava/util/Iterator;

Stack Overflow | cahen | 6 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    Spark job became incompatible with Google Dataproc

    Stack Overflow | 6 months ago | cahen
    org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, test-cluster-w-1.c.test-project.internal): java.lang.AbstractMethodError: uk.co.test.CalculateScore$$Lambda$10/1666820030.call(Ljava/lang/Object;)Ljava/util/Iterator;

    Root Cause Analysis

    1. org.apache.spark.scheduler.TaskSetManager

      Lost task 0.0 in stage 0.0 (TID 0, test-cluster-w-1.c.test-project.internal): java.lang.AbstractMethodError: uk.co.test.CalculateScore$$Lambda$10/1666820030.call(Ljava/lang/Object;)Ljava/util/Iterator;

      at org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$3$1.apply()
    2. Spark
      JavaRDDLike$$anonfun$fn$3$1.apply
      1. org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$3$1.apply(JavaRDDLike.scala:142)
      2. org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$3$1.apply(JavaRDDLike.scala:142)
      2 frames
    3. Scala
      AbstractIterator.toArray
      1. scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
      2. scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
      3. scala.collection.Iterator$class.foreach(Iterator.scala:893)
      4. scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
      5. scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:59)
      6. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:104)
      7. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:48)
      8. scala.collection.TraversableOnce$class.to(TraversableOnce.scala:310)
      9. scala.collection.AbstractIterator.to(Iterator.scala:1336)
      10. scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:302)
      11. scala.collection.AbstractIterator.toBuffer(Iterator.scala:1336)
      12. scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:289)
      13. scala.collection.AbstractIterator.toArray(Iterator.scala:1336)
      13 frames
    4. Spark
      Executor$TaskRunner.run
      1. org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$13.apply(RDD.scala:893)
      2. org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$13.apply(RDD.scala:893)
      3. org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1897)
      4. org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1897)
      5. org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
      6. org.apache.spark.scheduler.Task.run(Task.scala:85)
      7. org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
      7 frames
    5. Java RT
      Thread.run
      1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
      2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
      3. java.lang.Thread.run(Thread.java:745)
      3 frames