Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via Google Groups by hart jo, 1 year ago
Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout
via Stack Overflow by SpringStarter
, 1 year ago
Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout
via Stack Overflow by user481a
, 1 year ago
Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout
via Stack Overflow by sthomps
, 2 years ago
Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout
via Stack Overflow by Newbie
, 1 year ago
Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout
via Stack Overflow by Knows Not Much
, 2 years ago
Futures timed out after [120 seconds]. This timeout is controlled by spark.rpc.askTimeout
java.util.concurrent.TimeoutException: Futures timed out after [120 seconds]	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)	at scala.concurrent.Await$.result(package.scala:107)	at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)	at org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:101)	at org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:77)	at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stopExecutors(CoarseGrainedSchedulerBackend.scala:335)	at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stop(CoarseGrainedSchedulerBackend.scala:344)	at org.apache.spark.scheduler.cluster.SparkDeploySchedulerBackend.org$apache$spark$scheduler$cluster$SparkDeploySchedulerBackend$$stop(SparkDeploySchedulerBackend.scala:197)	at org.apache.spark.scheduler.cluster.SparkDeploySchedulerBackend.stop(SparkDeploySchedulerBackend.scala:101)	at org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:446)	at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:1582)	at org.apache.spark.SparkContext$$anonfun$stop$9.apply$mcV$sp(SparkContext.scala:1740)	at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1219)	at org.apache.spark.SparkContext.stop(SparkContext.scala:1739)	at org.apache.spark.SparkContext$$anonfun$3.apply$mcV$sp(SparkContext.scala:596)	at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:267)	at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ShutdownHookManager.scala:239)	at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239)	at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239)	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1801)	at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply$mcV$sp(ShutdownHookManager.scala:239)	at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239)	at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239)	at scala.util.Try$.apply(Try.scala:161)	at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:239)	at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:218)	at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)