org.apache.spark.SparkException: Job aborted due to stage failure: Task 2 in stage 40431.0 failed 1 times, most recent failure: Lost task 2.0 in stage 40431.0 (TID 8029, localhost): java.lang.ArrayIndexOutOfBoundsException Driver stacktrace:

Stack Overflow | Zach Moshe | 4 months ago
  1. 0

    ArrayIndexOutOfBoundsException when accessing triplets of a Graph

    Stack Overflow | 4 months ago | Zach Moshe
    org.apache.spark.SparkException: Job aborted due to stage failure: Task 2 in stage 40431.0 failed 1 times, most recent failure: Lost task 2.0 in stage 40431.0 (TID 8029, localhost): java.lang.ArrayIndexOutOfBoundsException Driver stacktrace:
  2. 0

    GitHub comment 86#226820790

    GitHub | 6 months ago | anfeng
    org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3, hkcloud-node1): ExecutorLostFailure (executor 753a7274-173a-41bf-b800-0345adc29be0-S0 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages. Driver stacktrace:
  3. 0

    an error on Mesos

    GitHub | 6 months ago | jsongcse
    org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3, hkcloud-node1): ExecutorLostFailure (executor 753a7274-173a-41bf-b800-0345adc29be0-S0 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages. Driver stacktrace:
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Unable to find class: org.apache.spark.h2o.package$StringHolder

    Stack Overflow | 1 day ago | lserlohn
    org.apache.spark.SparkException: Job aborted due to stage failure: Exception while getting task result: com.esotericsoftware.kryo.KryoException: Unable to find class: org.apache.spark.h2o.package$StringHolder
  6. 0

    KryoException: Unable to find class: org.apache.spark.h2o.package$StringHolder

    Google Groups | 2 days ago | Unknown author
    org.apache.spark.SparkException: Job aborted due to stage failure: Exception while getting task result: com.esotericsoftware.kryo.KryoException: Unable to find class: org.apache.spark.h2o.package$StringHolder

  1. Nikolay Rybak 1 times, last 1 month ago
  2. tyson925 2 times, last 2 months ago
  3. tyson925 1 times, last 4 months ago
  4. meneal 1 times, last 4 months ago
20 unregistered visitors
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. org.apache.spark.SparkException

    Job aborted due to stage failure: Task 2 in stage 40431.0 failed 1 times, most recent failure: Lost task 2.0 in stage 40431.0 (TID 8029, localhost): java.lang.ArrayIndexOutOfBoundsException Driver stacktrace:

    at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages()
  2. Spark
    DAGScheduler$$anonfun$abortStage$1.apply
    1. org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
    2. org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
    3. org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
    3 frames
  3. Scala
    ArrayBuffer.foreach
    1. scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
    2. scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
    2 frames
  4. Spark
    DAGScheduler$$anonfun$handleTaskSetFailed$1.apply
    1. org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
    2. org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
    3. org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
    3 frames
  5. Scala
    Option.foreach
    1. scala.Option.foreach(Option.scala:236)
    1 frame
  6. Spark
    RDD.collect
    1. org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
    2. org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640)
    3. org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599)
    4. org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588)
    5. org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
    6. org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
    7. org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
    8. org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
    9. org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
    10. org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
    11. org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
    12. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
    13. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
    14. org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
    15. org.apache.spark.rdd.RDD.collect(RDD.scala:926)
    15 frames