org.apache.spark.api.python.PythonException

Traceback (most recent call last): File "c:\spark\python\lib\pyspark.zip\pyspark\worker.py", line 172, in main File "c:\spark\python\lib\pyspark.zip\pyspark\worker.py", line 167, in process File "c:\spark\python\lib\pyspark.zip\pyspark\serializers.py", line 263, in dump_stream vs = list(itertools.islice(iterator, batch)) File "c:\spark\python\lib\pyspark.zip\pyspark\rdd.py", line 1306, in takeUpToNumLeft File "c:/sparkcourse/test-recommendation.py", line 8, in get_counts_and_averages return ID_and_ratings_tuple[0], (nratings, float(sum(x for x in ID_and_ratings_tuple[1]))/nratings) TypeError: unsupported operand type(s) for +: 'int' and 'str'

Samebug tips0

We couldn't find tips for this exception.

Don't give up yet. Paste your full stack trace to get a solution.

Solutions on the web70

  • via Stack Overflow by Ruofan Kong
    ,
  • via Unknown by Unknown author,
  • Stack trace

    • org.apache.spark.api.python.PythonException: Traceback (most recent call last): File "c:\spark\python\lib\pyspark.zip\pyspark\worker.py", line 172, in main File "c:\spark\python\lib\pyspark.zip\pyspark\worker.py", line 167, in process File "c:\spark\python\lib\pyspark.zip\pyspark\serializers.py", line 263, in dump_stream vs = list(itertools.islice(iterator, batch)) File "c:\spark\python\lib\pyspark.zip\pyspark\rdd.py", line 1306, in takeUpToNumLeft File "c:/sparkcourse/test-recommendation.py", line 8, in get_counts_and_averages return ID_and_ratings_tuple[0], (nratings, float(sum(x for x in ID_and_ratings_tuple[1]))
    • ratings) TypeError: unsupported operand type(s) for +: 'int' and 'str' at org.apache.spark.api.python.PythonRunner$$anon$1.read(PythonRDD.scala:193) at org.apache.spark.api.python.PythonRunner$$anon$1.<init>(PythonRDD.scala:234) at org.apache.spark.api.python.PythonRunner.compute(PythonRDD.scala:152) at org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:63) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319) at org.apache.spark.rdd.RDD.iterator(RDD.scala:283) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70) at org.apache.spark.scheduler.Task.run(Task.scala:85) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274) at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.lang.Thread.run(Unknown Source)

    Write tip

    You have a different solution? A short tip here would help you and many other users who saw this issue last week.

    Users with the same issue

    We couldn't find other users who have seen this exception.