org.apache.spark.SparkException

There are no available Samebug tips for this exception. Do you have an idea how to solve this issue? A short tip would help users who saw this issue last week.

  • Apache Spark Not able to view output
    via Stack Overflow by John Constantine
    ,
  • Re: pyspark not working
    via by prateek arora,
  • Spark: Length of List Tuple
    via Stack Overflow by Stephan Kristyn
    ,
    • org.apache.spark.SparkException: Job aborted due to stage failure: Task 4 in stage 24.0 failed 1 times, most recent failure: Lost task 4.0 in stage 24.0 (TID 76, localhost): org.apache.spark.api.python.PythonException: Traceback (most recent call last): File "/databricks/spark/python/pyspark/worker.py", line 111, in main process() File "/databricks/spark/python/pyspark/worker.py", line 106, in process serializer.dump_stream(func(split_index, iterator), outfile) File "/databricks/spark/python/pyspark/rdd.py", line 2346, in pipeline_func return func(split, prev_func(split, iterator)) File "/databricks/spark/python/pyspark/rdd.py", line 2346, in pipeline_func return func(split, prev_func(split, iterator)) File "/databricks/spark/python/pyspark/rdd.py", line 317, in func return f(iterator) File "/databricks/spark/python/pyspark/rdd.py", line 1776, in combineLocally merger.mergeValues(iterator) File "/databricks/spark/python/pyspark/shuffle.py", line 236, in mergeValues for k, v in iterator: File "<ipython-input-31-4b09041aa30b>", line 1, in <lambda> File "<ipython-input-28-f43debc22073>", line 3, in parseNeighbors File "/databricks/python/lib/python2.7/re.py", line 171, in split return _compile(pattern, flags).split(string, maxsplit) TypeError: expected string or buffer at org.apache.spark.api.python.PythonRunner$$anon$1.read(PythonRDD.scala:166) at org.apache.spark.api.python.PythonRunner$$anon$1.<init>(PythonRDD.scala:207) at org.apache.spark.api.python.PythonRunner.compute(PythonRDD.scala:125) at org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:70) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.api.python.PairwiseRDD.compute(PythonRDD.scala:342) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:79) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:46) at org.apache.spark.scheduler.Task.run(Task.scala:96) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:222) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)

    Users with the same issue

    Unknown visitor1 times, last one,
    Unknown visitor1 times, last one,
    tyson925
    2 times, last one,
    Unknown visitor1 times, last one,
    Unknown visitor1 times, last one,
    2 more bugmates