scala.MatchError: ArrayType(StringType,false) (of class org.apache.spark.sql.catalyst.types.ArrayType)

nabble.com | 8 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    UDF not working in Spark SQL

    Stack Overflow | 3 years ago | visakh
    scala.MatchError: ArrayType(StringType,false) (of class org.apache.spark.sql.catalyst.types.ArrayType)
  2. 0

    shark-sql

    solutionscore.com | 2 years ago
    scala.MatchError: ArrayType(StringType,false) (of class org.apache.spark.sql.catalyst.types.ArrayType)
  3. 0

    Apache Spark User List - Running Hive UDF from spark-shell fails due to datatype issue

    nabble.com | 8 months ago
    scala.MatchError: ArrayType(StringType,false) (of class org.apache.spark.sql.catalyst.types.ArrayType)
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. scala.MatchError

      ArrayType(StringType,false) (of class org.apache.spark.sql.catalyst.types.ArrayType)

      at org.apache.spark.sql.hive.HiveInspectors$typeInfoConversions.toTypeInfo()
    2. Spark Project Hive
      HiveFunctionRegistry$$anonfun$2.apply
      1. org.apache.spark.sql.hive.HiveInspectors$typeInfoConversions.toTypeInfo(HiveInspectors.scala:216)
      2. org.apache.spark.sql.hive.HiveFunctionRegistry$$anonfun$2.apply(hiveUdfs.scala:52)
      3. org.apache.spark.sql.hive.HiveFunctionRegistry$$anonfun$2.apply(hiveUdfs.scala:52)
      3 frames
    3. Scala
      AbstractTraversable.map
      1. scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
      2. scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
      3. scala.collection.immutable.List.foreach(List.scala:318)
      4. scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
      5. scala.collection.AbstractTraversable.map(Traversable.scala:105)
      5 frames
    4. Spark Project Hive
      HiveContext$$anon$3.org$apache$spark$sql$catalyst$analysis$OverrideFunctionRegistry$$super$lookupFunction
      1. org.apache.spark.sql.hive.HiveFunctionRegistry.lookupFunction(hiveUdfs.scala:52)
      2. org.apache.spark.sql.hive.HiveContext$$anon$3.org$apache$spark$sql$catalyst$analysis$OverrideFunctionRegistry$$super$lookupFunction(HiveContext.scala:253)
      2 frames
    5. Spark Project Catalyst
      OverrideFunctionRegistry$$anonfun$lookupFunction$2.apply
      1. org.apache.spark.sql.catalyst.analysis.OverrideFunctionRegistry$$anonfun$lookupFunction$2.apply(FunctionRegistry.scala:41)
      2. org.apache.spark.sql.catalyst.analysis.OverrideFunctionRegistry$$anonfun$lookupFunction$2.apply(FunctionRegistry.scala:41)
      2 frames
    6. Scala
      Option.getOrElse
      1. scala.Option.getOrElse(Option.scala:120)
      1 frame
    7. Spark Project Catalyst
      OverrideFunctionRegistry$class.lookupFunction
      1. org.apache.spark.sql.catalyst.analysis.OverrideFunctionRegistry$class.lookupFunction(FunctionRegistry.scala:41)
      1 frame
    8. Spark Project Hive
      HiveContext$$anon$3.lookupFunction
      1. org.apache.spark.sql.hive.HiveContext$$anon$3.lookupFunction(HiveContext.scala:253)
      1 frame
    9. Spark Project Catalyst
      TreeNode$$anonfun$4.apply
      1. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveFunctions$$anonfun$apply$5$$anonfun$applyOrElse$3.applyOrElse(Analyzer.scala:131)
      2. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveFunctions$$anonfun$apply$5$$anonfun$applyOrElse$3.applyOrElse(Analyzer.scala:129)
      3. org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:165)
      4. org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:183)
      4 frames
    10. Scala
      AbstractIterator.toArray
      1. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
      2. scala.collection.Iterator$class.foreach(Iterator.scala:727)
      3. scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
      4. scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
      5. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
      6. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
      7. scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
      8. scala.collection.AbstractIterator.to(Iterator.scala:1157)
      9. scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
      10. scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
      11. scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
      12. scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
      12 frames
    11. Spark Project Catalyst
      QueryPlan$$anonfun$1$$anonfun$apply$1.apply
      1. org.apache.spark.sql.catalyst.trees.TreeNode.transformChildrenDown(TreeNode.scala:212)
      2. org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:168)
      3. org.apache.spark.sql.catalyst.plans.QueryPlan.org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1(QueryPlan.scala:52)
      4. org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$1$$anonfun$apply$1.apply(QueryPlan.scala:66)
      4 frames
    12. Scala
      AbstractTraversable.map
      1. scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
      2. scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
      3. scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
      4. scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
      5. scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
      6. scala.collection.AbstractTraversable.map(Traversable.scala:105)
      6 frames
    13. Spark Project Catalyst
      QueryPlan$$anonfun$1.apply
      1. org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$1.apply(QueryPlan.scala:65)
      1 frame
    14. Scala
      AbstractIterator.toArray
      1. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
      2. scala.collection.Iterator$class.foreach(Iterator.scala:727)
      3. scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
      4. scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
      5. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
      6. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
      7. scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
      8. scala.collection.AbstractIterator.to(Iterator.scala:1157)
      9. scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
      10. scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
      11. scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
      12. scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
      12 frames
    15. Spark Project Catalyst
      RuleExecutor$$anonfun$apply$1$$anonfun$apply$2.apply
      1. org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpressionsDown(QueryPlan.scala:70)
      2. org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpressions(QueryPlan.scala:41)
      3. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveFunctions$$anonfun$apply$5.applyOrElse(Analyzer.scala:129)
      4. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveFunctions$$anonfun$apply$5.applyOrElse(Analyzer.scala:127)
      5. org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:165)
      6. org.apache.spark.sql.catalyst.trees.TreeNode.transform(TreeNode.scala:156)
      7. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveFunctions$.apply(Analyzer.scala:127)
      8. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveFunctions$.apply(Analyzer.scala:126)
      9. org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$apply$1$$anonfun$apply$2.apply(RuleExecutor.scala:61)
      10. org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$apply$1$$anonfun$apply$2.apply(RuleExecutor.scala:59)
      10 frames
    16. Scala
      List.foldLeft
      1. scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:111)
      2. scala.collection.immutable.List.foldLeft(List.scala:84)
      2 frames
    17. Spark Project Catalyst
      RuleExecutor$$anonfun$apply$1.apply
      1. org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$apply$1.apply(RuleExecutor.scala:59)
      2. org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$apply$1.apply(RuleExecutor.scala:51)
      2 frames
    18. Scala
      List.foreach
      1. scala.collection.immutable.List.foreach(List.scala:318)
      1 frame
    19. Spark Project Catalyst
      RuleExecutor.apply
      1. org.apache.spark.sql.catalyst.rules.RuleExecutor.apply(RuleExecutor.scala:51)
      1 frame
    20. Spark Project SQL
      SQLContext$QueryExecution.analyzed
      1. org.apache.spark.sql.SQLContext$QueryExecution.analyzed$lzycompute(SQLContext.scala:394)
      2. org.apache.spark.sql.SQLContext$QueryExecution.analyzed(SQLContext.scala:394)
      2 frames
    21. Spark Project Hive
      HiveContext$QueryExecution.optimizedPlan
      1. org.apache.spark.sql.hive.HiveContext$QueryExecution.optimizedPlan$lzycompute(HiveContext.scala:350)
      2. org.apache.spark.sql.hive.HiveContext$QueryExecution.optimizedPlan(HiveContext.scala:349)
      2 frames
    22. Spark Project SQL
      SQLContext$QueryExecution.executedPlan
      1. org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:399)
      2. org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:397)
      3. org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:403)
      4. org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:403)
      4 frames
    23. Spark Project Hive
      HiveContext$QueryExecution.simpleString
      1. org.apache.spark.sql.hive.HiveContext$QueryExecution.simpleString(HiveContext.scala:419)
      1 frame
    24. Spark Project SQL
      SchemaRDD.toString
      1. org.apache.spark.sql.SchemaRDDLike$class.toString(SchemaRDDLike.scala:67)
      2. org.apache.spark.sql.SchemaRDD.toString(SchemaRDD.scala:103)
      2 frames
    25. Scala
      ScalaRunTime$.replStringOf
      1. scala.runtime.ScalaRunTime$.scala$runtime$ScalaRunTime$$inner$1(ScalaRunTime.scala:324)
      2. scala.runtime.ScalaRunTime$.stringOf(ScalaRunTime.scala:329)
      3. scala.runtime.ScalaRunTime$.replStringOf(ScalaRunTime.scala:337)
      3 frames