Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

org.apache.spark.SparkException: Job aborted.	at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply$mcV$sp(InsertIntoHadoopFsRelationCommand.scala:149)	at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRelationCommand.scala:115)	at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRelationCommand.scala:115)	at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:57)	at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:115)	at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:60)	at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:58)	at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115)	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115)	at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:136)	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)	at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:133)	at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:114)	at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:86)	at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:86)	at org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:487)	at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:211)	at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:194)	at com.databricks.spark.redshift.RedshiftWriter.unloadData(RedshiftWriter.scala:278)	at com.databricks.spark.redshift.RedshiftWriter.saveToRedshift(RedshiftWriter.scala:346)	at com.databricks.spark.redshift.DefaultSource.createRelation(DefaultSource.scala:106)	at org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:429)	at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:211)