Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Samebug tips

  1. ,
    Expert tip

    This might be an issue with the file location in the Spark submit command. Try it with

    spark-submit --master spark://master:7077 \
         hello_world_from_pyspark.py {file location}
    
    
  2. ,
    Expert tip

    Check if you've set a name in Application -> Run. If you didn't, the generated XML is gonna have missing information and then this exception will be thrown.

Solutions on the web

via gmane.org by Unknown author, 1 year ago
java.lang.ClassNotFoundException: scala.Some	at java.net.URLClassLoader.findClass(URLClassLoader.java:381)	at java.lang.ClassLoader.loadClass(ClassLoader.java:424)	at java.lang.ClassLoader.loadClass(ClassLoader.java:357)	at java.lang.Class.forName0(Native Method)	at java.lang.Class.forName(Class.java:348)	at org.apache.spark.serializer.JavaDeserializationStream$$anon$1.resolveClass(JavaSerializer.scala:67)	at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1613)	at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1518)	at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1774)	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351)	at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1993)	at java.io.ObjectInputStream.defaultReadObject(ObjectInputStream.java:501)	at org.apache.spark.Accumulable$$anonfun$readObject$1.apply$mcV$sp(Accumulators.scala:152)	at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1160)