org.apache.spark.sql.AnalysisException: Reference 'var' is ambiguous, could be: var#166, var#167.;

GitHub | hiltonmbr | 5 months ago
  1. 0

    GitHub comment 111#233146888

    GitHub | 5 months ago | hiltonmbr
    org.apache.spark.sql.AnalysisException: Reference 'var' is ambiguous, could be: var#166, var#167.;
  2. 0

    GitHub comment 111#233147755

    GitHub | 5 months ago | kevinushey
    org.apache.spark.sql.AnalysisException: Reference 'var' is ambiguous, could be: var#6, var#7.;
  3. 0

    Duplicate columns in Spark Dataframe

    Stack Overflow | 1 year ago | Bamqf
    org.apache.spark.sql.AnalysisException: Reference 'Email' is ambiguous, could be: Email#350, Email#361.;
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    GitHub comment 141#156318742

    GitHub | 1 year ago | jerryivanhoe
    org.apache.spark.sql.AnalysisException: Non-local session path expected to be non-null;
  6. 0

    Table in Cassandra: create table myTable ( ts timestamp, value text ) I'm trying to use CassandraSQLContext.sql("select * from myTable where ts IN (1466708400000,1466643600000)") to query the table, and I got the exception indicating the filter value type of Long is not the same as column type of timestamp. This was not an issue with connector 1.5.1-M2. The exception: ] star_ws - Couldn't perform the aggregation cache - event_agg_h_acme_pull_acme_pull_sduh: org.apache.spark.sql.AnalysisException: cannot resolve 'ts IN (1466708400000,1466643600000)' due to data type mismatch: Arguments must be same type; org.apache.spark.sql.AnalysisException: cannot resolve 'ts IN (1466708400000,1466643600000)' due to data type mismatch: Arguments must be same type; at org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1$$anonfun$apply$2.applyOrElse(CheckAnalysis.scala:65) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1$$anonfun$apply$2.applyOrElse(CheckAnalysis.scala:57) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformUp$1.apply(TreeNode.scala:335) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformUp$1.apply(TreeNode.scala:335) at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:69) at org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:334) at org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpressionUp$1(QueryPlan.scala:108) at org.apache.spark.sql.catalyst.plans.QueryPlan.org$apache$spark$sql$catalyst$plans$QueryPlan$$recursiveTransform$2(QueryPlan.scala:118) at org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$2.apply(QueryPlan.scala:127) at scala.collection.Iterator$$anon$11.next(Iterator.scala:370) at scala.collection.Iterator$class.foreach(Iterator.scala:742) at scala.collection.AbstractIterator.foreach(Iterator.scala:1194) at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:59) at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:104) at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:48) at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:308) at scala.collection.AbstractIterator.to(Iterator.scala:1194) at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:300) at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1194) at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:287) at scala.collection.AbstractIterator.toArray(Iterator.scala:1194) at org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpressionsUp(QueryPlan.scala:127) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1.apply(CheckAnalysis.scala:57) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1.apply(CheckAnalysis.scala:50) at org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:121) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$foreachUp$1.apply(TreeNode.scala:120) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$foreachUp$1.apply(TreeNode.scala:120) at scala.collection.immutable.List.foreach(List.scala:381) at org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:120) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.checkAnalysis(CheckAnalysis.scala:50) at org.apache.spark.sql.catalyst.analysis.Analyzer.checkAnalysis(Analyzer.scala:44) at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:34) at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:133) at org.apache.spark.sql.cassandra.CassandraSQLContext.cassandraSql(CassandraSQLContext.scala:70)

    DataStax JIRA | 5 months ago | Stephen Qi
    org.apache.spark.sql.AnalysisException: cannot resolve 'ts IN (1466708400000,1466643600000)' due to data type mismatch: Arguments must be same type;

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. org.apache.spark.sql.AnalysisException

      Reference 'var' is ambiguous, could be: var#166, var#167.;

      at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolve()
    2. Spark Project Catalyst
      LogicalPlan$$anonfun$resolve$1.apply
      1. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolve(LogicalPlan.scala:264)
      2. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveQuoted(LogicalPlan.scala:168)
      3. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolve$1.apply(LogicalPlan.scala:130)
      4. org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolve$1.apply(LogicalPlan.scala:129)
      4 frames
    3. Scala
      IterableLike$class.foreach
      1. scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
      2. scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
      3. scala.collection.Iterator$class.foreach(Iterator.scala:893)
      4. scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
      5. scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
      5 frames
    4. org.apache.spark
      StructType.foreach
      1. org.apache.spark.sql.types.StructType.foreach(StructType.scala:94)
      1 frame