org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout

Google Groups | hart jo | 4 months ago
  1. 0

    Sparkling water executor error

    Google Groups | 4 months ago | hart jo
    org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout
  2. 0

    How to investigate failing dataproc worker processes?

    Stack Overflow | 8 months ago | sthomps
    org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout
  3. 0

    Spark Application Not Recovering when Executor Lost

    Stack Overflow | 6 months ago | user481a
    org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Spark - Reading .gz file faster than flat files on s3 - Flat file take way too long and never completes with 200+GB memory

    Stack Overflow | 6 months ago | SpringStarter
    org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout
  6. 0

    SPARK Job returning ExitCodeException exitCode=1 : Scala

    Stack Overflow | 1 month ago | Newbie
    org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout

  1. Nikolay Rybak 4 times, last 1 month ago
  2. tyson925 55 times, last 1 week ago
  3. poroszd 12 times, last 2 months ago
  4. kid 1 times, last 3 months ago
  5. Handemelindo 1 times, last 4 months ago
4 more registered users
28 unregistered visitors
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. java.util.concurrent.TimeoutException

    Futures timed out after [120 seconds]

    at scala.concurrent.impl.Promise$DefaultPromise.ready()
  2. Scala
    Await$.result
    1. scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
    2. scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
    3. scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
    4. scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
    5. scala.concurrent.Await$.result(package.scala:107)
    5 frames
  3. org.apache.spark
    RpcEndpointRef.askWithRetry
    1. org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
    2. org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:101)
    3. org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:77)
    3 frames
  4. Spark
    SparkShutdownHookManager$$anonfun$runAll$1.apply
    1. org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stopExecutors(CoarseGrainedSchedulerBackend.scala:335)
    2. org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stop(CoarseGrainedSchedulerBackend.scala:344)
    3. org.apache.spark.scheduler.cluster.SparkDeploySchedulerBackend.org$apache$spark$scheduler$cluster$SparkDeploySchedulerBackend$$stop(SparkDeploySchedulerBackend.scala:197)
    4. org.apache.spark.scheduler.cluster.SparkDeploySchedulerBackend.stop(SparkDeploySchedulerBackend.scala:101)
    5. org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:446)
    6. org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:1582)
    7. org.apache.spark.SparkContext$$anonfun$stop$9.apply$mcV$sp(SparkContext.scala:1740)
    8. org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1219)
    9. org.apache.spark.SparkContext.stop(SparkContext.scala:1739)
    10. org.apache.spark.SparkContext$$anonfun$3.apply$mcV$sp(SparkContext.scala:596)
    11. org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:267)
    12. org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ShutdownHookManager.scala:239)
    13. org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239)
    14. org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239)
    15. org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1801)
    16. org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply$mcV$sp(ShutdownHookManager.scala:239)
    17. org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239)
    18. org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239)
    18 frames
  5. Scala
    Try$.apply
    1. scala.util.Try$.apply(Try.scala:161)
    1 frame
  6. Spark
    SparkShutdownHookManager$$anon$2.run
    1. org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:239)
    2. org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:218)
    2 frames
  7. Hadoop
    ShutdownHookManager$1.run
    1. org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)
    1 frame