Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via GitHub by orencp
, 1 year ago
pyspark_cassandra.DeferringRowReader.read(Lcom/datastax/driver/core/Row;Lcom/datastax/spark/connector/CassandraRowMetadata;)Ljava/lang/Object;
via Stack Overflow by orenco
, 1 year ago
pyspark_cassandra.DeferringRowReader.read(Lcom/datastax/driver/core/Row;Lcom/datastax/spark/connector/CassandraRowMetadata;)Ljava/lang/Object;
java.lang.AbstractMethodError: pyspark_cassandra.DeferringRowReader.read(Lcom/datastax/driver/core/Row;Lcom/datastax/spark/connector/CassandraRowMetadata;)Ljava/lang/Object;	at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$17.apply(CassandraTableScanRDD.scala:315)	at com.datastax.spark.connector.rdd.CassandraTableScanRDD$$anonfun$17.apply(CassandraTableScanRDD.scala:315)	at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)	at scala.collection.Iterator$$anon$13.next(Iterator.scala:372)	at com.datastax.spark.connector.util.CountingIterator.next(CountingIterator.scala:16)	at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)	at scala.collection.Iterator$GroupedIterator.takeDestructively(Iterator.scala:914)	at scala.collection.Iterator$GroupedIterator.go(Iterator.scala:929)	at scala.collection.Iterator$GroupedIterator.fill(Iterator.scala:968)	at scala.collection.Iterator$GroupedIterator.hasNext(Iterator.scala:972)	at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)	at scala.collection.Iterator$class.foreach(Iterator.scala:727)	at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)	at org.apache.spark.api.python.PythonRDD$.writeIteratorToStream(PythonRDD.scala:452)	at org.apache.spark.api.python.PythonRunner$WriterThread$$anonfun$run$3.apply(PythonRDD.scala:280)	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1766)	at org.apache.spark.api.python.PythonRunner$WriterThread.run(PythonRDD.scala:239)