scheduler.TaskSetManager: Lost task 18.0 in stage 0.0 (TID 1, ulnode373.pv05.siri.apple.com): java.io.IOException: com.esotericsoftware.kryo.KryoException: Error during Java deserialization.

GitHub | liangzhaozeng | 4 months ago
  1. 0

    GitHub comment 147#237326312

    GitHub | 4 months ago | liangzhaozeng
    scheduler.TaskSetManager: Lost task 18.0 in stage 0.0 (TID 1, ulnode373.pv05.siri.apple.com): java.io.IOException: com.esotericsoftware.kryo.KryoException: Error during Java deserialization.
  2. 0

    How to deserialize Pipeline model in spark.ml?

    Stack Overflow | 2 years ago | Tim Malt
    java.lang.ClassNotFoundException: com.myCompany.spark.classifier.LabelTransformer
  3. 0

    ClassNotFoundException during readObject()

    GitHub | 4 years ago | harrah
    java.lang.ClassNotFoundException: ObjType$
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    ClassNotFoundException during readObject()

    GitHub | 5 years ago | ihji
    java.lang.ClassNotFoundException: ObjType$
  6. 0

    The Label compare model for English is broken

    GitHub | 2 years ago | an-fbk
    java.lang.ClassNotFoundException: org.dmilne.weka.wrapper.TypedAttribute

    10 unregistered visitors
    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.lang.ClassNotFoundException

      com.databricks.spark.avro.DefaultSource$SerializableConfiguration

      at java.net.URLClassLoader$1.run()
    2. Java RT
      ObjectInputStream.readObject
      1. java.net.URLClassLoader$1.run(URLClassLoader.java:366)
      2. java.net.URLClassLoader$1.run(URLClassLoader.java:355)
      3. java.security.AccessController.doPrivileged(Native Method)
      4. java.net.URLClassLoader.findClass(URLClassLoader.java:354)
      5. java.lang.ClassLoader.loadClass(ClassLoader.java:425)
      6. sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
      7. java.lang.ClassLoader.loadClass(ClassLoader.java:358)
      8. java.lang.Class.forName0(Native Method)
      9. java.lang.Class.forName(Class.java:278)
      10. java.io.ObjectInputStream.resolveClass(ObjectInputStream.java:625)
      11. java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1612)
      12. java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
      13. java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
      14. java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
      15. java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
      15 frames
    3. Kryo
      Kryo.readClassAndObject
      1. com.esotericsoftware.kryo.serializers.JavaSerializer.read(JavaSerializer.java:63)
      2. com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:790)
      2 frames
    4. Spark
      Broadcast.value
      1. org.apache.spark.serializer.KryoDeserializationStream.readObject(KryoSerializer.scala:229)
      2. org.apache.spark.broadcast.TorrentBroadcast$$anonfun$10.apply(TorrentBroadcast.scala:254)
      3. org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1287)
      4. org.apache.spark.broadcast.TorrentBroadcast$.unBlockifyObject(TorrentBroadcast.scala:255)
      5. org.apache.spark.broadcast.TorrentBroadcast$$anonfun$readBroadcastBlock$1.apply(TorrentBroadcast.scala:189)
      6. org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1253)
      7. org.apache.spark.broadcast.TorrentBroadcast.readBroadcastBlock(TorrentBroadcast.scala:174)
      8. org.apache.spark.broadcast.TorrentBroadcast._value$lzycompute(TorrentBroadcast.scala:65)
      9. org.apache.spark.broadcast.TorrentBroadcast._value(TorrentBroadcast.scala:65)
      10. org.apache.spark.broadcast.TorrentBroadcast.getValue(TorrentBroadcast.scala:89)
      11. org.apache.spark.broadcast.Broadcast.value(Broadcast.scala:70)
      11 frames
    5. com.databricks.spark
      DefaultSource$$anonfun$buildReader$1.apply
      1. com.databricks.spark.avro.DefaultSource$$anonfun$buildReader$1.apply(DefaultSource.scala:146)
      2. com.databricks.spark.avro.DefaultSource$$anonfun$buildReader$1.apply(DefaultSource.scala:145)
      2 frames
    6. org.apache.spark
      FileScanRDD$$anon$1.hasNext
      1. org.apache.spark.sql.execution.datasources.FileFormat$$anon$1.apply(fileSourceInterfaces.scala:279)
      2. org.apache.spark.sql.execution.datasources.FileFormat$$anon$1.apply(fileSourceInterfaces.scala:263)
      3. org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.nextIterator(FileScanRDD.scala:116)
      4. org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:91)
      4 frames
    7. Spark Project Catalyst
      GeneratedClass$GeneratedIterator.processNext
      1. org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
      1 frame
    8. Spark Project SQL
      WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext
      1. org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
      2. org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:370)
      2 frames
    9. Scala
      Iterator$$anon$11.hasNext
      1. scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
      2. scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
      3. scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
      3 frames
    10. Spark
      Executor$TaskRunner.run
      1. org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1682)
      2. org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:1115)
      3. org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:1115)
      4. org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1897)
      5. org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1897)
      6. org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
      7. org.apache.spark.scheduler.Task.run(Task.scala:85)
      8. org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
      8 frames
    11. Java RT
      Thread.run
      1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
      2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
      3. java.lang.Thread.run(Thread.java:745)
      3 frames