org.apache.spark.SparkException: Job aborted.

GitHub | saurabh-tyagi | 2 months ago
  1. Speed up your debug routine!

    Automated exception search integrated into your IDE

  2. 0

    GitHub comment 227#249985046

    GitHub | 2 months ago | wilvancleve
    org.apache.spark.SparkException: Job aborted.
  3. 0

    GitHub comment 825#252825829

    GitHub | 2 months ago | BioDCH
    org.apache.spark.SparkException: Job aborted. .........

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. org.apache.spark.SparkException

      Job aborted.

      at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply$mcV$sp()
    2. org.apache.spark
      InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply
      1. org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply$mcV$sp(InsertIntoHadoopFsRelationCommand.scala:149)
      2. org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRelationCommand.scala:115)
      3. org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRelationCommand.scala:115)
      3 frames
    3. Spark Project SQL
      SQLExecution$.withNewExecutionId
      1. org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:57)
      1 frame
    4. org.apache.spark
      ExecutedCommandExec.doExecute
      1. org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:115)
      2. org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:60)
      3. org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:58)
      4. org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
      4 frames
    5. Spark Project SQL
      SparkPlan$$anonfun$executeQuery$1.apply
      1. org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115)
      2. org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115)
      3. org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:136)
      3 frames
    6. Spark
      RDDOperationScope$.withScope
      1. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
      1 frame
    7. Spark Project SQL
      QueryExecution.toRdd
      1. org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:133)
      2. org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:114)
      3. org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:86)
      4. org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:86)
      4 frames
    8. org.apache.spark
      DataSource.write
      1. org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:487)
      1 frame
    9. Spark Project SQL
      DataFrameWriter.save
      1. org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:211)
      2. org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:194)
      2 frames
    10. com.databricks.spark
      DefaultSource.createRelation
      1. com.databricks.spark.redshift.RedshiftWriter.unloadData(RedshiftWriter.scala:278)
      2. com.databricks.spark.redshift.RedshiftWriter.saveToRedshift(RedshiftWriter.scala:346)
      3. com.databricks.spark.redshift.DefaultSource.createRelation(DefaultSource.scala:106)
      3 frames
    11. org.apache.spark
      DataSource.write
      1. org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:429)
      1 frame
    12. Spark Project SQL
      DataFrameWriter.save
      1. org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:211)
      1 frame