org.apache.spark.SparkException: Job aborted due to stage failure: Task 5 in stage 0.0 failed 1 times, most recent failure: Lost task 5.0 in stage 0.0 (TID 5, localhost): jcuda.CudaException: CUDA_ERROR_INVALID_VALUE


Solutions on the web

Solution icon of github
via GitHub by TPolzer
, 1 year ago
Job aborted due to stage failure: Task 5 in stage 0.0 failed 1 times, most recent failure: Lost task 5.0 in stage 0.0 (TID 5, localhost): jcuda.CudaException: CUDA_ERROR_INVALID_VALUE

Solution icon of github
The owner protected this stack trace.

Stack trace

org.apache.spark.SparkException: Job aborted due to stage failure: Task 5 in stage 0.0 failed 1 times, most recent failure: Lost task 5.0 in stage 0.0 (TID 5, localhost): jcuda.CudaException: CUDA_ERROR_INVALID_VALUE
	at jcuda.driver.JCudaDriver.checkResult(JCudaDriver.java:312)
	at jcuda.driver.JCudaDriver.cuMemcpyDtoHAsync(JCudaDriver.java:5732)
	at com.ibm.gpuenabler.HybridIterator$$anonfun$copyGpuToCpu$1.apply(HybridIterator.scala:198)
	at com.ibm.gpuenabler.HybridIterator$$anonfun$copyGpuToCpu$1.apply(HybridIterator.scala:162)
	at scala.runtime.Tuple2Zipped$$anonfun$map$extension$1.apply(Tuple2Zipped.scala:40)
	at scala.runtime.Tuple2Zipped$$anonfun$map$extension$1.apply(Tuple2Zipped.scala:38)
	at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
	at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
	at scala.runtime.Tuple2Zipped$.map$extension(Tuple2Zipped.scala:38)
	at com.ibm.gpuenabler.HybridIterator.copyGpuToCpu(HybridIterator.scala:162)
	at com.ibm.gpuenabler.HybridIterator.freeGPUMemory(HybridIterator.scala:121)
	at com.ibm.gpuenabler.CUDAFunction.compute(CUDAFunction.scala:421)
	at com.ibm.gpuenabler.MapGPUPartitionsRDD.compute(CUDARDDUtils.scala:108)
	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
	at com.ibm.gpuenabler.MapGPUPartitionsRDD.compute(CUDARDDUtils.scala:95)
	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
	at org.apache.spark.scheduler.Task.run(Task.scala:89)
	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)

Write tip

You have a different solution? A short tip here would help you and many other users who saw this issue last week.

Users with the same issue

You are the first who have seen this exception. Write a tip to help other users and build your expert profile.