java.lang.RuntimeException: Unsupported datatype StructType(List())

github.com | 3 months ago
  1. 0

    Can I create a function in Spark SQL?

    Stack Overflow | 2 years ago | user3826955
    java.lang.RuntimeException: [1.1] failure:</code>INSERT'' expected but identifier CREATE found</p> <p>CREATE OR REPLACE FUNCTION apply_rules (pcode VARCHAR2) RETURN BOOLEAN AS LANGUAGE JAVA NAME 'main.scala.GroovyIntegrator.applyRules (java.lang.String) return java.lang.Boolean'; ^
  2. 0

    Append Mode is not Enabled for AvroSaver

    GitHub | 1 year ago | mkanchwala
    java.lang.RuntimeException: Append mode is not supported by com.databricks.spark.avro.DefaultSource
  3. Speed up your debug routine!

    Automated exception search integrated into your IDE

  4. 0

    Saving / exporting transformed DataFrame back to JDBC / MySQL

    Stack Overflow | 1 year ago | Matt Zukowski
    java.lang.RuntimeException: org.apache.spark.sql.execution.datasources.jdbc.DefaultSource does not allow create table as select.
  5. 0

    GC segfaults on macOS 10.12

    GitHub | 2 months ago | LukasKellenberger
    java.lang.RuntimeException: Nonzero exit code: 139

  1. Handemelindo 1 times, last 2 weeks ago
  2. rp 4 times, last 1 month ago
  3. rp 1 times, last 2 months ago
  4. balintn 3 times, last 2 months ago
  5. max_samebug 2 times, last 2 months ago
13 more registered users
70 unregistered visitors
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. java.lang.RuntimeException

    Unsupported datatype StructType(List())

    at scala.sys.package$.error()
  2. Scala
    package$.error
    1. scala.sys.package$.error(package.scala:27)
    1 frame
  3. Spark Project SQL
    ParquetTypesConverter$$anonfun$1.apply
    1. org.apache.spark.sql.parquet.ParquetTypesConverter$.fromDataType(ParquetRelation.scala:201)
    2. org.apache.spark.sql.parquet.ParquetTypesConverter$$anonfun$1.apply(ParquetRelation.scala:235)
    3. org.apache.spark.sql.parquet.ParquetTypesConverter$$anonfun$1.apply(ParquetRelation.scala:235)
    3 frames
  4. Scala
    AbstractTraversable.map
    1. scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
    2. scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
    3. scala.collection.immutable.List.foreach(List.scala:318)
    4. scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
    5. scala.collection.AbstractTraversable.map(Traversable.scala:105)
    5 frames
  5. Spark Project SQL
    SparkStrategies$ParquetOperations$.apply
    1. org.apache.spark.sql.parquet.ParquetTypesConverter$.convertFromAttributes(ParquetRelation.scala:234)
    2. org.apache.spark.sql.parquet.ParquetTypesConverter$.writeMetaData(ParquetRelation.scala:267)
    3. org.apache.spark.sql.parquet.ParquetRelation$.createEmpty(ParquetRelation.scala:143)
    4. org.apache.spark.sql.parquet.ParquetRelation$.create(ParquetRelation.scala:122)
    5. org.apache.spark.sql.execution.SparkStrategies$ParquetOperations$.apply(SparkStrategies.scala:139)
    5 frames
  6. Spark Project Catalyst
    QueryPlanner$$anonfun$1.apply
    1. org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
    2. org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
    2 frames
  7. Scala
    Iterator$$anon$13.hasNext
    1. scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
    1 frame
  8. Spark Project Catalyst
    QueryPlanner.apply
    1. org.apache.spark.sql.catalyst.planning.QueryPlanner.apply(QueryPlanner.scala:59)
    1 frame
  9. Spark Project SQL
    SchemaRDD.saveAsParquetFile
    1. org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:264)
    2. org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:264)
    3. org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:265)
    4. org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:265)
    5. org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:268)
    6. org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:268)
    7. org.apache.spark.sql.SchemaRDDLike$class.saveAsParquetFile(SchemaRDDLike.scala:66)
    8. org.apache.spark.sql.SchemaRDD.saveAsParquetFile(SchemaRDD.scala:98)
    8 frames