java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext. This stopped SparkContext was created at: org.apache.spark.SparkContext.<init>(SparkContext.scala:83) org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:874) org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:81) org.apache.spark.streaming.api.java.JavaStreamingContext.<init>(JavaStreamingContext.scala:140) bigdata.spark_0_1.spark.runJobInTOS(spark.java:889) bigdata.spark_0_1.spark.main(spark.java:773) The currently active SparkContext was created at: org.apache.spark.SparkContext.<init>(SparkContext.scala:83) org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:874) org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:81) org.apache.spark.streaming.api.java.JavaStreamingContext.<init>(JavaStreamingContext.scala:140) bigdata.spark_0_1.spark.runJobInTOS(spark.java:889) bigdata.spark_0_1.spark.main(spark.java:773)

Talend Open Integration Solution | lei ju | 6 months ago
  1. 0

    Talend Open Integration Solution | 6 months ago | lei ju
    java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext. This stopped SparkContext was created at: org.apache.spark.SparkContext.<init>(SparkContext.scala:83) org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:874) org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:81) org.apache.spark.streaming.api.java.JavaStreamingContext.<init>(JavaStreamingContext.scala:140) bigdata.spark_0_1.spark.runJobInTOS(spark.java:889) bigdata.spark_0_1.spark.main(spark.java:773) The currently active SparkContext was created at: org.apache.spark.SparkContext.<init>(SparkContext.scala:83) org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:874) org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:81) org.apache.spark.streaming.api.java.JavaStreamingContext.<init>(JavaStreamingContext.scala:140) bigdata.spark_0_1.spark.runJobInTOS(spark.java:889) bigdata.spark_0_1.spark.main(spark.java:773)
  2. 0

    GitHub comment 572#246278541

    GitHub | 3 months ago | car2008
    java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext. This stopped SparkContext was created at: org.apache.spark.SparkContext.<init>(SparkContext.scala:82) org.hammerlab.guacamole.commands.SparkCommand.createSparkContext(SparkCommand.scala:65) org.hammerlab.guacamole.commands.SparkCommand.run(SparkCommand.scala:10) org.hammerlab.guacamole.commands.Command.run(Command.scala:27) org.hammerlab.guacamole.Main$.main(Main.scala:49) org.hammerlab.guacamole.Main.main(Main.scala) sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.lang.reflect.Method.invoke(Method.java:606) org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) The currently active SparkContext was created at: (No active SparkContext.)
  3. 0

    GitHub comment 195#172252690

    GitHub | 11 months ago | jramos
    java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext. This stopped SparkContext was created at: org.apache.spark.SparkContext.<init>(SparkContext.scala:147) io.prediction.workflow.SharedSparkContext$class.beforeAll(BaseTest.scala:65) io.prediction.controller.EngineSuite.beforeAll(EngineTest.scala:18) org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187) io.prediction.controller.EngineSuite.beforeAll(EngineTest.scala:18) org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253) io.prediction.controller.EngineSuite.run(EngineTest.scala:18) org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:444) org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:651) sbt.TestRunner.runTest$1(TestFramework.scala:76) sbt.TestRunner.run(TestFramework.scala:85) sbt.TestFramework$$anon$2$$anonfun$$init$$1$$anonfun$apply$8.apply(TestFramework.scala:202) sbt.TestFramework$$anon$2$$anonfun$$init$$1$$anonfun$apply$8.apply(TestFramework.scala:202) sbt.TestFramework$.sbt$TestFramework$$withContextLoader(TestFramework.scala:185) sbt.TestFramework$$anon$2$$anonfun$$init$$1.apply(TestFramework.scala:202) sbt.TestFramework$$anon$2$$anonfun$$init$$1.apply(TestFramework.scala:202) sbt.TestFunction.apply(TestFramework.scala:207) sbt.Tests$.sbt$Tests$$processRunnable$1(Tests.scala:239) sbt.Tests$$anonfun$makeSerial$1.apply(Tests.scala:245) sbt.Tests$$anonfun$makeSerial$1.apply(Tests.scala:245) The currently active SparkContext was created at: (No active SparkContext.)
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    GitHub comment 572#246357384

    GitHub | 3 months ago | arahuja
    java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext. This stopped SparkContext was created at: org.apache.spark.SparkContext.(SparkContext.scala:82) org.hammerlab.guacamole.commands.SparkCommand.createSparkContext( SparkCommand.scala:65) org.hammerlab.guacamole.commands.SparkCommand.run(SparkCommand.scala:10) org.hammerlab.guacamole.commands.Command.run(Command.scala:27) org.hammerlab.guacamole.Main$.main(Main.scala:49) org.hammerlab.guacamole.Main.main(Main.scala) sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java: 57) sun.reflect.DelegatingMethodAccessorImpl.invoke( DelegatingMethodAccessorImpl.java:43) java.lang.reflect.Method.invoke(Method.java:606) org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) The currently active SparkContext was created at: (No active SparkContext.)
  6. 0

    GitHub comment 572#246357384

    GitHub | 3 months ago | arahuja
    java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext. This stopped SparkContext was created at: org.apache.spark.SparkContext.(SparkContext.scala:82) org.hammerlab.guacamole.commands.SparkCommand.createSparkContext( SparkCommand.scala:65) org.hammerlab.guacamole.commands.SparkCommand.run(SparkCommand.scala:10) org.hammerlab.guacamole.commands.Command.run(Command.scala:27) org.hammerlab.guacamole.Main$.main(Main.scala:49) org.hammerlab.guacamole.Main.main(Main.scala) sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java: 57) sun.reflect.DelegatingMethodAccessorImpl.invoke( DelegatingMethodAccessorImpl.java:43) java.lang.reflect.Method.invoke(Method.java:606) org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) The currently active SparkContext was created at: (No active SparkContext.)

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.lang.IllegalStateException

      Cannot call methods on a stopped SparkContext. This stopped SparkContext was created at: org.apache.spark.SparkContext.<init>(SparkContext.scala:83) org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:874) org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:81) org.apache.spark.streaming.api.java.JavaStreamingContext.<init>(JavaStreamingContext.scala:140) bigdata.spark_0_1.spark.runJobInTOS(spark.java:889) bigdata.spark_0_1.spark.main(spark.java:773) The currently active SparkContext was created at: org.apache.spark.SparkContext.<init>(SparkContext.scala:83) org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:874) org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:81) org.apache.spark.streaming.api.java.JavaStreamingContext.<init>(JavaStreamingContext.scala:140) bigdata.spark_0_1.spark.runJobInTOS(spark.java:889) bigdata.spark_0_1.spark.main(spark.java:773)

      at org.apache.spark.SparkContext.org$apache$spark$SparkContext$$assertNotStopped()
    2. Spark
      SparkContext.makeRDD
      1. org.apache.spark.SparkContext.org$apache$spark$SparkContext$$assertNotStopped(SparkContext.scala:107)
      2. org.apache.spark.SparkContext$$anonfun$parallelize$1.apply(SparkContext.scala:740)
      3. org.apache.spark.SparkContext$$anonfun$parallelize$1.apply(SparkContext.scala:739)
      4. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
      5. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
      6. org.apache.spark.SparkContext.withScope(SparkContext.scala:725)
      7. org.apache.spark.SparkContext.parallelize(SparkContext.scala:739)
      8. org.apache.spark.SparkContext$$anonfun$makeRDD$1.apply(SparkContext.scala:823)
      9. org.apache.spark.SparkContext$$anonfun$makeRDD$1.apply(SparkContext.scala:823)
      10. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
      11. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
      12. org.apache.spark.SparkContext.withScope(SparkContext.scala:725)
      13. org.apache.spark.SparkContext.makeRDD(SparkContext.scala:822)
      13 frames
    3. Spark Project Streaming
      ReceiverTracker$ReceiverTrackerEndpoint$$anonfun$receive$1.applyOrElse
      1. org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverTrackerEndpoint.org$apache$spark$streaming$scheduler$ReceiverTracker$ReceiverTrackerEndpoint$$startReceiver(ReceiverTracker.scala:585)
      2. org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverTrackerEndpoint$$anonfun$receive$1.applyOrElse(ReceiverTracker.scala:477)
      2 frames
    4. org.apache.spark
      Dispatcher$MessageLoop.run
      1. org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:116)
      2. org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:204)
      3. org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:100)
      4. org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:215)
      4 frames
    5. Java RT
      Thread.run
      1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
      2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
      3. java.lang.Thread.run(Thread.java:745)
      3 frames