java.io.NotSerializableException

org.neo4j.driver.internal.InternalNode Serialization stack: - object not serializable (class: org.neo4j.driver.internal.InternalNode, value: node<10516047>) - element of array (index: 0) - array (class [Ljava.lang.Object;, size 1) - field (class: org.apache.spark.sql.catalyst.expressions.GenericRow, name: values, type: class [Ljava.lang.Object;) - object (class org.apache.spark.sql.catalyst.expressions.GenericRowWithSchema, [node<10516047>]) - element of array (index: 0) - array (class [Lorg.apache.spark.sql.Row;, size 1)

Solutions on the web251

  • ) - field (class: org.apache.spark.sql.catalyst.expressions.GenericRow, name: values, type: class [Ljava.lang.Object;) - object (class org.apache.spark.sql.catalyst.expressions.GenericRowWithSchema, [node<10516047>]) - element of array (index: 0) - array (class [Lorg.apache.spark.sql.Row;, size 1)
  • org.neo4j.driver.internal.InternalNode Serialization stack: - object not serializable (class: org.neo4j.driver.internal.InternalNode, value: node<10516047>) - element of array (index: 0) - array (class [Ljava.lang.Object;, size 1
  • via Stack Overflow by Jas
    , 1 year ago
    org.apache.hadoop.io.LongWritable Serialization stack: - object not serializable (class: org.apache.hadoop.io.LongWritable, value: 0) - field (class: scala.Tuple2, name: _1, type: class java.lang.Object) - object (class scala.Tuple2, (0
  • Stack trace

    • java.io.NotSerializableException: org.neo4j.driver.internal.InternalNode Serialization stack: - object not serializable (class: org.neo4j.driver.internal.InternalNode, value: node<10516047>) - element of array (index: 0) - array (class [Ljava.lang.Object;, size 1) - field (class: org.apache.spark.sql.catalyst.expressions.GenericRow, name: values, type: class [Ljava.lang.Object;) - object (class org.apache.spark.sql.catalyst.expressions.GenericRowWithSchema, [node<10516047>]) - element of array (index: 0) - array (class [Lorg.apache.spark.sql.Row;, size 1) at org.apache.spark.serializer.SerializationDebugger$.improveException(SerializationDebugger.scala:40) at org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:46) at org.apache.spark.serializer.JavaSerializerInstance.serialize(JavaSerializer.scala:100) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:313) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)

    Write tip

    You have a different solution? A short tip here would help you and many other users who saw this issue last week.

    Users with the same issue

    30 times, 6 months ago
    14 times, 4 months ago
    Unknown user
    Once, 11 months ago
    Unknown user
    Once, 1 year ago
    Once, 1 month ago
    7 more bugmates