Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via GitHub by aashishrtyagi
, 1 month ago
cannot assign instance of scala.collection.immutable.List$SerializationProxy to field org.apache.spark.rdd.RDD.org$apache$spark$rdd$RDD$$dependencies_ of type scala.collection.Seq in instance of org.apache.spark.rdd.MapPartitionsRDD
via Stack Overflow by moun
, 3 weeks ago
cannot assign instance of scala.collection.immutable.List$SerializationProxy to field org.apache.spark.rdd.RDD.org$apache$spark$rdd$RDD$$dependencies_ of type scala.collection.Seq in instance of org.apache.spark.rdd.MapPartitionsRDD
via Google Groups by ** rafael **, 1 month ago
cannot assign instance of scala.collection.immutable.List$SerializationProxy to field org.apache.spark.rdd.RDD.org$apache$spark$rdd$RDD$$dependencies_ of type scala.collection.Seq in instance of org.apache.spark.rdd.MapPartitionsRDD
via Stack Overflow by renegademonkey
, 8 months ago
cannot assign instance of scala.collection.immutable.List$SerializationProxy to field org.apache.spark.rdd.RDD.org$apache$spark$rdd$RDD$$dependencies_ of type scala.collection.Seq in instance of org.apache.spark.rdd.MapPartitionsRDD
via Stack Overflow by user3709612
, 6 months ago
cannot assign instance of scala.collection.immutable.List$SerializationProxy to field org.apache.spark.rdd.RDD.org$apache$spark$rdd$RDD$$dependencies_ of type scala.collection.Seq in instance of org.apache.spark.rdd.MapPartitionsRDD
via Stack Overflow by user1870400
, 1 year ago
cannot assign instance of scala.collection.immutable.List$SerializationProxy to field org.apache.spark.rdd.RDD.org$apache$spark$rdd$RDD$$dependencies_ of type scala.collection.Seq in instance of org.apache.spark.rdd.MapPartitionsRDD
java.lang.ClassCastException: cannot assign instance of scala.collection.immutable.List$SerializationProxy to field org.apache.spark.rdd.RDD.org$apache$spark$rdd$RDD$$dependencies_ of type scala.collection.Seq in instance of org.apache.spark.rdd.MapPartitionsRDD	at java.io.ObjectStreamClass$FieldReflector.setObjFieldValues(ObjectStreamClass.java:2133)	at java.io.ObjectStreamClass.setObjFieldValues(ObjectStreamClass.java:1305)	at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2006)	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1924)	at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801)	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351)	at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2000)	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1924)	at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801)	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351)	at java.io.ObjectInputStream.readObject(ObjectInputStream.java:371)	at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:75)	at org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:114)	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)	at org.apache.spark.scheduler.Task.run(Task.scala:85)	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)	at java.lang.Thread.run(Thread.java:745)