Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via Stack Overflow by Thiago Marzagão
, 1 year ago
/pyspark/worker.py", line 167, in process # serializer.dump_stream(func(split_index, iterator), outfile) # File "/home/cgu.local/thiagovm/spark-2.0.0-bin-hadoop2.7/python/lib/pyspark.zip/pyspark/serializers.py", line 263, in dump_stream # vs = list
via GitHub by lossyrob
, 9 months ago
Traceback (most recent call last): File "/usr/local/spark-2.1.0-bin-hadoop2.7/python/lib/pyspark.zip/pyspark/worker.py", line 174, in main process() File "/usr/local/spark-2.1.0-bin-hadoop2.7/python/lib/pyspark.zip/pyspark/worker.py", line
via GitHub by jpdna
, 1 year ago
Traceback (most recent call last): File "/home/paschallj/Spark/1.6.3/spark-1.6.3-bin-hadoop2.6/python/lib/pyspark.zip/pyspark/worker.py", line 98, in main command = pickleSer._read_with_length(infile) File "/home/paschallj/Spark/1.6.3/spark
via Stack Overflow by user1836155
, 2 years ago
Traceback (most recent call last): File "spark-1.5.1-bin-hadoop2.6/python/lib/pyspark.zip/pyspark/worker.py", line 111, in main process() File "spark-1.5.1-bin-hadoop2.6/python/lib/pyspark.zip/pyspark/worker.py", line 106, in
via Stack Overflow by Algina
, 1 year ago
Traceback (most recent call last): File "/home/alg/programs/spark-2.0.2-bin-hadoop2.7/python/lib/pyspark.zip/pyspark/worker.py", line 175, in main process() File "/home/alg/programs/spark-2.0.2-bin-hadoop2.7/python/lib/pyspark.zip/pyspark
via Stack Overflow by 谢一男
, 11 months ago
Traceback (most recent call last): File "/usr/local/spark/ python/lib/pyspark.zip/pyspark/worker.py", line 111, in main process() File "/usr/local/spark/python/lib/pyspark.zip/pyspark/worker.py", line 106, in process
org.apache.spark.api.python.PythonException: Traceback (most recent call last):
# File "/home/cgu.local/thiagovm/spark-2.0.0-bin-hadoop2.7/python/lib/pyspark.zip/pyspark/worker.py", line 172, in main
# process()
# File "/home/cgu.local/thiagovm/spark-2.0.0-bin-hadoop2.7/python/lib/pyspark.zip/pyspark/worker.py", line 167, in process
# serializer.dump_stream(func(split_index, iterator), outfile)
# File "/home/cgu.local/thiagovm/spark-2.0.0-bin-hadoop2.7/python/lib/pyspark.zip/pyspark/serializers.py", line 263, in dump_stream
# vs = list(itertools.islice(iterator, batch))
# File "/usr/local/lib/python3.5/site-packages/splearn/feature_extraction/text.py", line 289, in <lambda>
# A = Z.transform(lambda X: list(map(analyze, X)), column='X').persist()
# File "/usr/local/lib/python3.5/site-packages/sklearn/feature_extraction/text.py", line 238, in <lambda>
# tokenize(preprocess(self.decode(doc))), stop_words)
# File "/usr/local/lib/python3.5/site-packages/sklearn/feature_extraction/text.py", line 204, in <lambda>
# return lambda x: strip_accents(x.lower())
# AttributeError: 'numpy.ndarray' object has no attribute 'lower'	at org.apache.spark.api.python.PythonRunner$$anon$1.read(PythonRDD.scala:193)	at org.apache.spark.api.python.PythonRunner$$anon$1.(PythonRDD.scala:234)	at org.apache.spark.api.python.PythonRunner.compute(PythonRDD.scala:152)	at org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:63)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319)	at org.apache.spark.rdd.RDD$$anonfun$8.apply(RDD.scala:332)	at org.apache.spark.rdd.RDD$$anonfun$8.apply(RDD.scala:330)	at org.apache.spark.storage.BlockManager$$anonfun$doPutIterator$1.apply(BlockManager.scala:935)	at org.apache.spark.storage.BlockManager$$anonfun$doPutIterator$1.apply(BlockManager.scala:910)	at org.apache.spark.storage.BlockManager.doPut(BlockManager.scala:866)	at org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:910)	at org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:668)	at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:330)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:281)	at org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:63)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:283)	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)	at org.apache.spark.scheduler.Task.run(Task.scala:85)	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)	at java.lang.Thread.run(Thread.java:745)