org.apache.spark.SparkException: Yarn application has already ended! It might have been killed or unable to launch application master.

Stack Overflow | Kami Wan | 3 months ago
  1. 0

    Run Spark-shell with error :SparkContext: Error initializing SparkContext

    Stack Overflow | 3 months ago | Kami Wan
    org.apache.spark.SparkException: Yarn application has already ended! It might have been killed or unable to launch application master.
  2. 0

    Spark on Yarn client run with errors

    Stack Overflow | 4 weeks ago | cherah30
    org.apache.spark.SparkException: Yarn application has already ended! It might have been killed or unable to launch application master.
  3. 0

    Where can I set the spark.eventLog.enabled?

    Stack Overflow | 4 months ago | A.Dumas
    org.apache.spark.SparkException: A master URL must be set in your configuration
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Setting spark.driver.memory on SparkSession

    Stack Overflow | 2 months ago | user1710158
    org.apache.spark.SparkException: A master URL must be set in your configuration
  6. 0

    Spark 2.0 Thrift server not started in yarn mode

    Stack Overflow | 6 months ago | Churchill vins
    org.apache.spark.SparkException: Yarn application has already ended! It might have been killed or unable to launch application master.
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. org.apache.spark.SparkException

    Yarn application has already ended! It might have been killed or unable to launch application master.

    at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication()
  2. Spark
    SparkContext$.getOrCreate
    1. org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:85)
    2. org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:62)
    3. org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:149)
    4. org.apache.spark.SparkContext.<init>(SparkContext.scala:500)
    5. org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2256)
    5 frames
  3. Spark Project SQL
    SparkSession$Builder$$anonfun$8.apply
    1. org.apache.spark.sql.SparkSession$Builder$$anonfun$8.apply(SparkSession.scala:831)
    2. org.apache.spark.sql.SparkSession$Builder$$anonfun$8.apply(SparkSession.scala:823)
    2 frames
  4. Scala
    Option.getOrElse
    1. scala.Option.getOrElse(Option.scala:121)
    1 frame
  5. Spark Project SQL
    SparkSession$Builder.getOrCreate
    1. org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:823)
    1 frame
  6. Spark REPL
    Main$.createSparkSession
    1. org.apache.spark.repl.Main$.createSparkSession(Main.scala:101)
    1 frame
  7. $line3
    $eval.$print
    1. $line3.$read$$iw$$iw.<init>(<console>:15)
    2. $line3.$read$$iw.<init>(<console>:31)
    3. $line3.$read.<init>(<console>:33)
    4. $line3.$read$.<init>(<console>:37)
    5. $line3.$read$.<clinit>(<console>)
    6. $line3.$eval$.$print$lzycompute(<console>:7)
    7. $line3.$eval$.$print(<console>:6)
    8. $line3.$eval.$print(<console>)
    8 frames
  8. Java RT
    Method.invoke
    1. sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    2. sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    3. sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    4. java.lang.reflect.Method.invoke(Method.java:498)
    4 frames
  9. Scala Compiler
    IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply
    1. scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:786)
    2. scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1047)
    3. scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:638)
    4. scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:637)
    4 frames
  10. Scala
    AbstractFileClassLoader.asContext
    1. scala.reflect.internal.util.ScalaClassLoader$class.asContext(ScalaClassLoader.scala:31)
    2. scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:19)
    2 frames
  11. Scala Compiler
    ILoop.processLine
    1. scala.tools.nsc.interpreter.IMain$WrappedRequest.loadAndRunReq(IMain.scala:637)
    2. scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:569)
    3. scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:565)
    4. scala.tools.nsc.interpreter.ILoop.interpretStartingWith(ILoop.scala:807)
    5. scala.tools.nsc.interpreter.ILoop.command(ILoop.scala:681)
    6. scala.tools.nsc.interpreter.ILoop.processLine(ILoop.scala:395)
    6 frames
  12. Spark REPL
    SparkILoop$$anonfun$initializeSpark$1.apply
    1. org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1.apply$mcV$sp(SparkILoop.scala:38)
    2. org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1.apply(SparkILoop.scala:37)
    3. org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1.apply(SparkILoop.scala:37)
    3 frames
  13. Scala Compiler
    IMain.beQuietDuring
    1. scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:214)
    1 frame
  14. Spark REPL
    SparkILoop.loadFiles
    1. org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:37)
    2. org.apache.spark.repl.SparkILoop.loadFiles(SparkILoop.scala:94)
    2 frames
  15. Scala Compiler
    ILoop$$anonfun$process$1.apply
    1. scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply$mcZ$sp(ILoop.scala:920)
    2. scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply(ILoop.scala:909)
    3. scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply(ILoop.scala:909)
    3 frames
  16. Scala
    ScalaClassLoader$.savingContextLoader
    1. scala.reflect.internal.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:97)
    1 frame
  17. Scala Compiler
    ILoop.process
    1. scala.tools.nsc.interpreter.ILoop.process(ILoop.scala:909)
    1 frame
  18. Spark REPL
    Main.main
    1. org.apache.spark.repl.Main$.doMain(Main.scala:68)
    2. org.apache.spark.repl.Main$.main(Main.scala:51)
    3. org.apache.spark.repl.Main.main(Main.scala)
    3 frames
  19. Java RT
    Method.invoke
    1. sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    2. sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    3. sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    4. java.lang.reflect.Method.invoke(Method.java:498)
    4 frames
  20. Spark
    SparkSubmit.main
    1. org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:729)
    2. org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:185)
    3. org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:210)
    4. org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:124)
    5. org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
    5 frames