org.apache.spark.SparkException: Yarn application has already ended! It might have been killed or unable to launch application master.

chinahadoop.cn | 3 months ago
tip
Do you find the tips below useful? Click on the to mark them and say thanks to rafael . Or join the community to write better ones.
  1. 0

    why sparkcontext stopped?

    Stack Overflow | 1 year ago | 최승호
    rg.apache.spark.SparkException: Yarn application has already ended! It might have been killed or unable to launch application master.
  2. 0

    Spark Config Files

    Stack Overflow | 1 year ago | Ashesh Nair
    org.apache.spark.SparkException: Yarn application has already ended! It might have been killed or unable to launch application master.
  3. 0

    Hadoop Mania

    blogspot.com | 1 year ago
    org.apache.spark.SparkException: Yarn application has already ended! It might have been killed or unable to launch application master.
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0
    samebug tip
    Change the queue name to hadoop
  6. 0

    Why does spark-shell --master yarn-client fail (yet pyspark --master yarn seems to work)?

    Stack Overflow | 3 years ago | fxm
    org.apache.spark.SparkException: Yarn application already ended,might be killed or not able to launch application master .

  1. tyson925 1 times, last 3 months ago
  2. muffinmannen 6 times, last 10 months ago
  3. zokni 1 times, last 12 months ago
  4. harshg 4 times, last 1 year ago
13 unregistered visitors
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. org.apache.spark.SparkException

    Yarn application has already ended! It might have been killed or unable to launch application master.

    at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication()
  2. Spark
    SparkContext.<init>
    1. org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:124)
    2. org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:64)
    3. org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:144)
    4. org.apache.spark.SparkContext.<init>(SparkContext.scala:530)
    4 frames
  3. Spark REPL
    SparkILoop.createSparkContext
    1. org.apache.spark.repl.SparkILoop.createSparkContext(SparkILoop.scala:1017)
    1 frame
  4. $line3
    $eval.$print
    1. $line3.$read$$iwC$$iwC.<init>(<console>:15)
    2. $line3.$read$$iwC.<init>(<console>:24)
    3. $line3.$read.<init>(<console>:26)
    4. $line3.$read$.<init>(<console>:30)
    5. $line3.$read$.<clinit>(<console>)
    6. $line3.$eval$.<init>(<console>:7)
    7. $line3.$eval$.<clinit>(<console>)
    8. $line3.$eval.$print(<console>)
    8 frames
  5. Java RT
    Method.invoke
    1. sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    2. sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    3. sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    4. java.lang.reflect.Method.invoke(Method.java:498)
    4 frames
  6. Spark REPL
    SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply
    1. org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
    2. org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346)
    3. org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840)
    4. org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871)
    5. org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819)
    6. org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857)
    7. org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902)
    8. org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)
    9. org.apache.spark.repl.SparkILoopInit$$anonfun$initializeSpark$1.apply(SparkILoopInit.scala:125)
    10. org.apache.spark.repl.SparkILoopInit$$anonfun$initializeSpark$1.apply(SparkILoopInit.scala:124)
    11. org.apache.spark.repl.SparkIMain.beQuietDuring(SparkIMain.scala:324)
    12. org.apache.spark.repl.SparkILoopInit$class.initializeSpark(SparkILoopInit.scala:124)
    13. org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:64)
    14. org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1$$anonfun$apply$mcZ$sp$5.apply$mcV$sp(SparkILoop.scala:974)
    15. org.apache.spark.repl.SparkILoopInit$class.runThunks(SparkILoopInit.scala:159)
    16. org.apache.spark.repl.SparkILoop.runThunks(SparkILoop.scala:64)
    17. org.apache.spark.repl.SparkILoopInit$class.postInitialization(SparkILoopInit.scala:108)
    18. org.apache.spark.repl.SparkILoop.postInitialization(SparkILoop.scala:64)
    19. org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:991)
    20. org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
    21. org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
    21 frames
  7. Scala Compiler
    ScalaClassLoader$.savingContextLoader
    1. scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
    1 frame
  8. Spark REPL
    Main.main
    1. org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945)
    2. org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059)
    3. org.apache.spark.repl.Main$.main(Main.scala:31)
    4. org.apache.spark.repl.Main.main(Main.scala)
    4 frames
  9. Java RT
    Method.invoke
    1. sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    2. sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    3. sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    4. java.lang.reflect.Method.invoke(Method.java:498)
    4 frames
  10. Spark
    SparkSubmit.main
    1. org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
    2. org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
    3. org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
    4. org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
    5. org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
    5 frames