Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via Stack Overflow by geoHeil
, 1 year ago
java.lang.ClassCastException: cannot assign instance of scala.collection.immutable.HashMap$SerializationProxy to field org.apache.spark.executor.TaskMetrics._accumulatorUpdates of type scala.collection.immutable.Map in instance of org.apache.spark.executor.TaskMetrics
via Google Groups by Naresh Dulam, 7 months ago
java.lang.ClassCastException: cannot assign instance of scala.collection.immutable.HashMap$SerializationProxy to field org.apache.spark.executor.TaskMetrics._accumulatorUpdates of type scala.collection.immutable.Map in instance of org.apache.spark.executor.TaskMetrics
via spark-issues by Matt Butler (JIRA), 1 year ago
java.lang.ClassCastException: cannot assign instance of scala.collection.immutable.HashMap$SerializationProxy to field org.apache.spark.executor.TaskMetrics._accumulatorUpdates of type scala.collection.immutable.Map in instance of org.apache.spark.executor.TaskMetrics
via GitHub by AliTajeldin
, 1 year ago
java.lang.ClassCastException: cannot assign instance of scala.collection.immutable.HashMap$SerializationProxy to field org.apache.spark.executor.TaskMetrics._accumulatorUpdates of type scala.collection.immutable.Map in instance of org.apache.spark.executor.TaskMetrics
via spark-issues by Matt Butler (JIRA), 1 year ago
java.lang.ClassCastException: cannot assign instance of scala.collection.immutable.HashMap$SerializationProxy to field org.apache.spark.executor.TaskMetrics._accumulatorUpdates of type scala.collection.immutable.Map in instance of org.apache.spark.executor.TaskMetrics
via Stack Overflow by user2038119
, 2 years ago
java.lang.ClassCastException: cannot assign instance of scala.collection.immutable.HashMap$SerializationProxy to field org.apache.spark.executor.TaskMetrics._accumulatorUpdates of type scala.collection.immutable.Map in instance of org.apache.spark.executor.TaskMetrics
java.lang.ClassCastException: cannot assign instance of scala.collection.immutable.HashMap$SerializationProxy to field org.apache.spark.executor.TaskMetrics._accumulatorUpdates of type scala.collection.immutable.Map in instance of org.apache.spark.executor.TaskMetrics	at java.io.ObjectStreamClass$FieldReflector.setObjFieldValues(ObjectStreamClass.java:2089)	at java.io.ObjectStreamClass.setObjFieldValues(ObjectStreamClass.java:1261)	at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1999)	at java.io.ObjectInputStream.defaultReadObject(ObjectInputStream.java:501)	at org.apache.spark.executor.TaskMetrics$$anonfun$readObject$1.apply$mcV$sp(TaskMetrics.scala:220)	at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1160)	at org.apache.spark.executor.TaskMetrics.readObject(TaskMetrics.scala:219)	at sun.reflect.GeneratedMethodAccessor53.invoke(Unknown Source)	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)	at java.lang.reflect.Method.invoke(Method.java:483)	at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017)	at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1896)	at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801)	at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351)	at java.io.ObjectInputStream.readObject(ObjectInputStream.java:371)	at org.apache.spark.util.Utils$.deserialize(Utils.scala:91)	at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$reportHeartBeat$1$$anonfun$apply$6.apply(Executor.scala:440)	at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$reportHeartBeat$1$$anonfun$apply$6.apply(Executor.scala:430)	at scala.Option.foreach(Option.scala:236)	at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$reportHeartBeat$1.apply(Executor.scala:430)	at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$reportHeartBeat$1.apply(Executor.scala:428)	at scala.collection.Iterator$class.foreach(Iterator.scala:727)	at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)	at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)	at scala.collection.AbstractIterable.foreach(Iterable.scala:54)	at org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$reportHeartBeat(Executor.scala:428)	at org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply$mcV$sp(Executor.scala:472)	at org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:472)	at org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:472)	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1699)	at org.apache.spark.executor.Executor$$anon$1.run(Executor.scala:472)	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)	at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)	at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)	at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)	at java.lang.Thread.run(Thread.java:745)