org.apache.spark.sql.AnalysisException: no such table actresses_and_actors_perm; line 1 pos 61

github.com | 7 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    Merge pull request #14 from fluxcapacitor/nlp · fluxcapacitor/pipeline@b6c4fd6 · GitHub

    github.com | 7 months ago
    org.apache.spark.sql.AnalysisException: no such table actresses_and_actors_perm; line 1 pos 61
  2. 0

    Spark ML indexer cannot resolve DataFrame column name with dots?

    Stack Overflow | 1 year ago | Joshua Taylor
    org.apache.spark.sql.AnalysisException: cannot resolve 'a.b' given input columns a.b;
  3. 0

    ERROR server.TThreadPoolServer: Error occurred during processing of message

    Stack Overflow | 1 year ago | Techie
    org.apache.spark.sql.AnalysisException: no such table service; line 1 pos 14
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    ERROR server.TThreadPoolServer: Error occurred during processing of message

    spark-user | 1 year ago | Dasun Hegoda
    org.apache.spark.sql.AnalysisException: no such table service; line 1 pos 14
  6. 0

    ERROR server.TThreadPoolServer: Error occurred during processing of message

    spark-user | 1 year ago | Dasun Hegoda
    org.apache.spark.sql.AnalysisException: no such table service; line 1 pos 14

  1. harshg 1 times, last 11 months ago
4 unregistered visitors
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. org.apache.spark.sql.AnalysisException

    no such table actresses_and_actors_perm; line 1 pos 61

    at org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis()
  2. Spark Project Catalyst
    TreeNode$$anonfun$4.apply
    1. org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42)
    2. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.getTable(Analyzer.scala:260)
    3. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$7.applyOrElse(Analyzer.scala:268)
    4. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$7.applyOrElse(Analyzer.scala:264)
    5. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:57)
    6. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:57)
    7. org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:51)
    8. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:56)
    9. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54)
    10. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54)
    11. org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:249)
    11 frames
  3. Scala
    AbstractIterator.toArray
    1. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
    2. scala.collection.Iterator$class.foreach(Iterator.scala:727)
    3. scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
    4. scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
    5. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
    6. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
    7. scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
    8. scala.collection.AbstractIterator.to(Iterator.scala:1157)
    9. scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
    10. scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
    11. scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
    12. scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
    12 frames
  4. Spark Project Catalyst
    TreeNode$$anonfun$4.apply
    1. org.apache.spark.sql.catalyst.trees.TreeNode.transformChildren(TreeNode.scala:279)
    2. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:54)
    3. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54)
    4. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54)
    5. org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:249)
    5 frames
  5. Scala
    AbstractIterator.toArray
    1. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
    2. scala.collection.Iterator$class.foreach(Iterator.scala:727)
    3. scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
    4. scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
    5. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
    6. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
    7. scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
    8. scala.collection.AbstractIterator.to(Iterator.scala:1157)
    9. scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
    10. scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
    11. scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
    12. scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
    12 frames
  6. Spark Project Catalyst
    TreeNode$$anonfun$4.apply
    1. org.apache.spark.sql.catalyst.trees.TreeNode.transformChildren(TreeNode.scala:279)
    2. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:54)
    3. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54)
    4. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54)
    5. org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:249)
    5 frames
  7. Scala
    AbstractIterator.toArray
    1. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
    2. scala.collection.Iterator$class.foreach(Iterator.scala:727)
    3. scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
    4. scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
    5. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
    6. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
    7. scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
    8. scala.collection.AbstractIterator.to(Iterator.scala:1157)
    9. scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
    10. scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
    11. scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
    12. scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
    12 frames
  8. Spark Project Catalyst
    TreeNode$$anonfun$4.apply
    1. org.apache.spark.sql.catalyst.trees.TreeNode.transformChildren(TreeNode.scala:279)
    2. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:54)
    3. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54)
    4. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54)
    5. org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:249)
    5 frames
  9. Scala
    AbstractIterator.toArray
    1. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
    2. scala.collection.Iterator$class.foreach(Iterator.scala:727)
    3. scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
    4. scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
    5. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
    6. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
    7. scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
    8. scala.collection.AbstractIterator.to(Iterator.scala:1157)
    9. scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
    10. scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
    11. scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
    12. scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
    12 frames
  10. Spark Project Catalyst
    RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply
    1. org.apache.spark.sql.catalyst.trees.TreeNode.transformChildren(TreeNode.scala:279)
    2. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:54)
    3. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:264)
    4. org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:254)
    5. org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:83)
    6. org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:80)
    6 frames
  11. Scala
    List.foldLeft
    1. scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:111)
    2. scala.collection.immutable.List.foldLeft(List.scala:84)
    2 frames
  12. Spark Project Catalyst
    RuleExecutor$$anonfun$execute$1.apply
    1. org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:80)
    2. org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:72)
    2 frames
  13. Scala
    List.foreach
    1. scala.collection.immutable.List.foreach(List.scala:318)
    1 frame
  14. Spark Project Catalyst
    RuleExecutor.execute
    1. org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:72)
    1 frame
  15. Spark Project SQL
    SQLContext$QueryExecution.assertAnalyzed
    1. org.apache.spark.sql.SQLContext$QueryExecution.analyzed$lzycompute(SQLContext.scala:916)
    2. org.apache.spark.sql.SQLContext$QueryExecution.analyzed(SQLContext.scala:916)
    3. org.apache.spark.sql.SQLContext$QueryExecution.assertAnalyzed(SQLContext.scala:914)
    3 frames