org.apache.spark.SparkException: Task not serializable

github.com | 4 months ago
  1. 0

    Scala Spark dataframe : Task not serilizable exception even with Broadcast variables

    Stack Overflow | 7 months ago | Himaprasoon
    org.apache.spark.SparkException: Task not serializable
  2. Speed up your debug routine!

    Automated exception search integrated into your IDE

  3. 0

    Test fails with SparkContext has been shutdown

    GitHub | 2 years ago | jamborta
    org.apache.spark.SparkException: SparkContext has been shutdown
  4. 0

    GitHub comment 54#71801487

    GitHub | 2 years ago | velvia
    org.apache.spark.SparkException: SparkContext has been shutdown

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. org.apache.spark.SparkException

      Task not serializable

      at org.apache.spark.util.ClosureCleaner$.ensureSerializable()
    2. Spark
      RDD.collect
      1. org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:298)
      2. org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:288)
      3. org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:108)
      4. org.apache.spark.SparkContext.clean(SparkContext.scala:2038)
      5. org.apache.spark.SparkContext.runJob(SparkContext.scala:1897)
      6. org.apache.spark.SparkContext.runJob(SparkContext.scala:1912)
      7. org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:884)
      8. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
      9. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
      10. org.apache.spark.rdd.RDD.withScope(RDD.scala:357)
      11. org.apache.spark.rdd.RDD.collect(RDD.scala:883)
      11 frames
    3. Spark Project SQL
      SparkPlan.executeCollect
      1. org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:290)
      1 frame