org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): java.lang.ArrayIndexOutOfBoundsException: 0

GitHub | chrimiway | 7 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    GitHub comment 93#237154478

    GitHub | 7 months ago | chrimiway
    org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): java.lang.ArrayIndexOutOfBoundsException: 0

    Root Cause Analysis

    1. org.apache.spark.SparkException

      Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): java.lang.ArrayIndexOutOfBoundsException: 0

      at org.dia.core.SciTensor.<init>()
    2. org.dia.core
      SciSparkContext$$anonfun$1.apply
      1. org.dia.core.SciTensor.<init>(SciTensor.scala:39)
      2. org.dia.core.SciSparkContext$$anonfun$1.apply(SciSparkContext.scala:123)
      3. org.dia.core.SciSparkContext$$anonfun$1.apply(SciSparkContext.scala:93)
      3 frames
    3. Scala
      AbstractIterator.reduceLeft
      1. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
      2. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
      3. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
      4. scala.collection.Iterator$class.foreach(Iterator.scala:727)
      5. scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
      6. scala.collection.TraversableOnce$class.reduceLeft(TraversableOnce.scala:172)
      7. scala.collection.AbstractIterator.reduceLeft(Iterator.scala:1157)
      7 frames
    4. Spark
      Executor$TaskRunner.run
      1. org.apache.spark.rdd.RDD$$anonfun$reduce$1$$anonfun$14.apply(RDD.scala:1011)
      2. org.apache.spark.rdd.RDD$$anonfun$reduce$1$$anonfun$14.apply(RDD.scala:1009)
      3. org.apache.spark.SparkContext$$anonfun$36.apply(SparkContext.scala:1951)
      4. org.apache.spark.SparkContext$$anonfun$36.apply(SparkContext.scala:1951)
      5. org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
      6. org.apache.spark.scheduler.Task.run(Task.scala:89)
      7. org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
      7 frames
    5. Java RT
      Thread.run
      1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
      2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
      3. java.lang.Thread.run(Thread.java:745)
      3 frames