Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via Stack Overflow by apolak
, 1 year ago
Lost task 1.0 in stage 0.0 (TID 1, localhost): java.lang.NoClassDefFoundError: com/esotericsoftware/kryo/KryoSerializable
java.lang.ClassNotFoundException: com.esotericsoftware.kryo.KryoSerializable	at java.net.URLClassLoader.findClass(URLClassLoader.java:381)	at java.lang.ClassLoader.loadClass(ClassLoader.java:424)	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)	at java.lang.ClassLoader.loadClass(ClassLoader.java:357)	at java.lang.ClassLoader.defineClass1(Native Method)	at java.lang.ClassLoader.defineClass(ClassLoader.java:760)	at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)	at java.net.URLClassLoader.defineClass(URLClassLoader.java:467)	at java.net.URLClassLoader.access$100(URLClassLoader.java:73)	at java.net.URLClassLoader$1.run(URLClassLoader.java:368)	at java.net.URLClassLoader$1.run(URLClassLoader.java:362)	at java.security.AccessController.doPrivileged(Native Method)	at java.net.URLClassLoader.findClass(URLClassLoader.java:361)	at java.lang.ClassLoader.loadClass(ClassLoader.java:424)	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)	at java.lang.ClassLoader.loadClass(ClassLoader.java:357)	at org.apache.spark.rdd.SqlNewHadoopRDDState$.setInputFileName(SqlNewHadoopRDDState.scala:37)	at org.apache.spark.rdd.HadoopRDD$$anon$1.(HadoopRDD.scala:218)	at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:208)	at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)	at org.apache.spark.scheduler.Task.run(Task.scala:89)	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)	at java.lang.Thread.run(Thread.java:745)