Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via Stack Overflow by user1753235
, 1 year ago
) File "/data/3/tmp/hadoop-hadoop/nm-local-dir/usercache/user/appcache/application_1468851295159_0020/container_1468851295159_0020_01_000016/pyspark.zip/pyspark/serializers.py", line 263, in dump_stream vs = list(itertools.islice(iterator, batch)) File "/usr/local/spark/python/pyspark/rdd.py", line 1898, in <lambda> IndexError: list index out of range
via Stack Overflow by kiseliu
, 1 year ago
process serializer.dump_stream(func(split_index, iterator), outfile) File "/Users/lyj/Programs/Apache/spark/python/lib/pyspark.zip/pyspark/serializers.py", line 263, in dump_stream vs = list(itertools.islice(iterator, batch)) File "/mypath
via Stack Overflow by Nightfury
, 2 months ago
process serializer.dump_stream(func(split_index, iterator), outfile) File "/opt/spark-2.2.1-bin-hadoop2.7/python/lib/pyspark.zip/pyspark/serializers.py", line 268, in dump_stream vs = list(itertools.islice(iterator, batch)) File "<stdin>", line 1, in <lambda> IndexError: list index out of range
via Stack Overflow by Aru
, 5 days ago
(func(split_index, iterator), outfile) File "/usr/lib/spark/python/lib/pyspark.zip/pyspark/serializers.py", line 263, in dump_stream vs = list(itertools.islice(iterator, batch)) File "<stdin>", line 1, in <lambda> IndexError: list index out of range
via GitHub by lossyrob
, 10 months ago
Traceback (most recent call last): File "/usr/local/spark-2.1.0-bin-hadoop2.7/python/lib/pyspark.zip/pyspark/worker.py", line 174, in main process() File "/usr/local/spark-2.1.0-bin-hadoop2.7/python/lib/pyspark.zip/pyspark/worker.py", line
org.apache.spark.api.python.PythonException: Traceback (most recent call last):
File "/data/3/tmp/hadoop-hadoop/nm-local-dir/usercache/user/appcache/application_1468851295159_0020/container_1468851295159_0020_01_000016/pyspark.zip/pyspark/worker.py", line 111, in main
    process()
  File "/data/3/tmp/hadoop-hadoop/nm-local-dir/usercache/user/appcache/application_1468851295159_0020/container_1468851295159_0020_01_000016/pyspark.zip/pyspark/worker.py", line 106, in process
    serializer.dump_stream(func(split_index, iterator), outfile)
  File "/data/3/tmp/hadoop-hadoop/nm-local-dir/usercache/user/appcache/application_1468851295159_0020/container_1468851295159_0020_01_000016/pyspark.zip/pyspark/serializers.py", line 263, in dump_stream
    vs = list(itertools.islice(iterator, batch))
  File "/usr/local/spark/python/pyspark/rdd.py", line 1898, in <lambda>
IndexError: list index out of range	at org.apache.spark.api.python.PythonRunner$$anon$1.read(PythonRDD.scala:166)	at org.apache.spark.api.python.PythonRunner$$anon$1.next(PythonRDD.scala:129)	at org.apache.spark.api.python.PythonRunner$$anon$1.next(PythonRDD.scala:125)	at org.apache.spark.InterruptibleIterator.next(InterruptibleIterator.scala:43)	at scala.collection.Iterator$class.foreach(Iterator.scala:727)	at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)	at org.apache.spark.api.python.PythonRDD$.writeIteratorToStream(PythonRDD.scala:452)	at org.apache.spark.api.python.PythonRunner$WriterThread$$anonfun$run$3.apply(PythonRDD.scala:280)	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1765)	at org.apache.spark.api.python.PythonRunner$WriterThread.run(PythonRDD.scala:239)