org.apache.avro.SchemaParseException: Illegal character in: ordercontainer.paymentinfo

GitHub | samelamin | 8 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    GitHub comment 234#232210707

    GitHub | 8 months ago | samelamin
    org.apache.avro.SchemaParseException: Illegal character in: ordercontainer.paymentinfo

    Root Cause Analysis

    1. org.apache.avro.SchemaParseException

      Illegal character in: ordercontainer.paymentinfo

      at org.apache.avro.Schema.validateName()
    2. Apache Avro
      SchemaBuilder$FieldDefault.noDefault
      1. org.apache.avro.Schema.validateName(Schema.java:1083)
      2. org.apache.avro.Schema.access$200(Schema.java:79)
      3. org.apache.avro.Schema$Field.<init>(Schema.java:372)
      4. org.apache.avro.SchemaBuilder$FieldBuilder.completeField(SchemaBuilder.java:2124)
      5. org.apache.avro.SchemaBuilder$FieldBuilder.completeField(SchemaBuilder.java:2120)
      6. org.apache.avro.SchemaBuilder$FieldBuilder.access$5200(SchemaBuilder.java:2034)
      7. org.apache.avro.SchemaBuilder$FieldDefault.noDefault(SchemaBuilder.java:2146)
      7 frames
    3. com.databricks.spark
      SchemaConverters$$anonfun$convertStructToAvro$1.apply
      1. com.databricks.spark.avro.SchemaConverters$$anonfun$convertStructToAvro$1.apply(SchemaConverters.scala:110)
      2. com.databricks.spark.avro.SchemaConverters$$anonfun$convertStructToAvro$1.apply(SchemaConverters.scala:105)
      2 frames
    4. Scala
      ArrayOps$ofRef.foreach
      1. scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
      2. scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
      2 frames
    5. com.databricks.spark
      AvroRelation.prepareJobForWrite
      1. com.databricks.spark.avro.SchemaConverters$.convertStructToAvro(SchemaConverters.scala:105)
      2. com.databricks.spark.avro.AvroRelation.prepareJobForWrite(AvroRelation.scala:83)
      2 frames
    6. org.apache.spark
      InsertIntoHadoopFsRelation$$anonfun$run$1.apply
      1. org.apache.spark.sql.execution.datasources.BaseWriterContainer.driverSideSetup(WriterContainer.scala:103)
      2. org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelation$$anonfun$run$1.apply$mcV$sp(InsertIntoHadoopFsRelation.scala:147)
      3. org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelation$$anonfun$run$1.apply(InsertIntoHadoopFsRelation.scala:108)
      4. org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelation$$anonfun$run$1.apply(InsertIntoHadoopFsRelation.scala:108)
      4 frames
    7. Spark Project SQL
      SQLExecution$.withNewExecutionId
      1. org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:56)
      1 frame
    8. org.apache.spark
      InsertIntoHadoopFsRelation.run
      1. org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelation.run(InsertIntoHadoopFsRelation.scala:108)
      1 frame
    9. Spark Project SQL
      SparkPlan$$anonfun$execute$5.apply
      1. org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
      2. org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
      3. org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
      4. org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
      5. org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
      5 frames
    10. Spark
      RDDOperationScope$.withScope
      1. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
      1 frame
    11. Spark Project SQL
      QueryExecution.toRdd
      1. org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
      2. org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
      3. org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
      3 frames
    12. org.apache.spark
      ResolvedDataSource$.apply
      1. org.apache.spark.sql.execution.datasources.ResolvedDataSource$.apply(ResolvedDataSource.scala:256)
      1 frame
    13. Spark Project SQL
      DataFrameWriter.save
      1. org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:148)
      2. org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:139)
      2 frames
    14. com.databricks.spark
      DefaultSource.createRelation
      1. com.databricks.spark.redshift.RedshiftWriter.unloadData(RedshiftWriter.scala:278)
      2. com.databricks.spark.redshift.RedshiftWriter.saveToRedshift(RedshiftWriter.scala:346)
      3. com.databricks.spark.redshift.DefaultSource.createRelation(DefaultSource.scala:106)
      3 frames
    15. org.apache.spark
      ResolvedDataSource$.apply
      1. org.apache.spark.sql.execution.datasources.ResolvedDataSource$.apply(ResolvedDataSource.scala:222)
      1 frame
    16. Spark Project SQL
      DataFrameWriter.save
      1. org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:148)
      1 frame
    17. Unknown
      Importer.main
      1. Importer$.main(Importer.scala:46)
      2. Importer.main(Importer.scala)
      2 frames
    18. Java RT
      Method.invoke
      1. sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
      2. sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
      3. sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
      4. java.lang.reflect.Method.invoke(Method.java:498)
      4 frames
    19. IDEA
      AppMain.main
      1. com.intellij.rt.execution.application.AppMain.main(AppMain.java:144)
      1 frame