ache.spark.api.python.PythonException: Traceback (most recent call last): File "c:\spark\python\lib\pyspark.zip\pyspark\worker.py", line 172, in main File "c:\spark\python\lib\pyspark.zip\pyspark\worker.py", line 167, in process File "c:\spark\python\lib\pyspark.zip\pyspark\serializers.py", line 263, in dump_stream vs = list(itertools.islice(iterator, batch)) File "c:\spark\python\lib\pyspark.zip\pyspark\rdd.py", line 1306, in takeUpToNumLeft File "c:/sparkcourse/test-recommendation.py", line 8, in get_counts_and_averages return ID_and_ratings_tuple[0], (nratings, float(sum(x for x in ID_and_ratings_tuple[1]))/nratings) TypeError: unsupported operand type(s) for +: 'int' and 'str'

There are no available Samebug tips for this exception. Do you have an idea how to solve this issue? A short tip would help users who saw this issue last week.

  • ache.spark.api.python.PythonException: Traceback (most recent call last): File "c:\spark\python\lib\pyspark.zip\pyspark\worker.py", line 172, in main File "c:\spark\python\lib\pyspark.zip\pyspark\worker.py", line 167, in process File "c:\spark\python\lib\pyspark.zip\pyspark\serializers.py", line 263, in dump_stream vs = list(itertools.islice(iterator, batch)) File "c:\spark\python\lib\pyspark.zip\pyspark\rdd.py", line 1306, in takeUpToNumLeft File "c:/sparkcourse/test-recommendation.py", line 8, in get_counts_and_averages return ID_and_ratings_tuple[0], (nratings, float(sum(x for x in ID_and_ratings_tuple[1]))
  • ratings) TypeError: unsupported operand type(s) for +: 'int' and 'str' at org.apache.spark.api.python.PythonRunner$$anon$1.read(PythonRDD.scala:193) at org.apache.spark.api.python.PythonRunner$$anon$1.<init>(PythonRDD.scala:234) at org.apache.spark.api.python.PythonRunner.compute(PythonRDD.scala:152) at org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:63) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319) at org.apache.spark.rdd.RDD.iterator(RDD.scala:283) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70) at org.apache.spark.scheduler.Task.run(Task.scala:85) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274) at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.lang.Thread.run(Unknown Source)
No Bugmate found.