java.lang.UnsupportedOperationException: Cannot evaluate expression: parse_df_to_string(input[1, int, true], input[2, int, true], input[3, int, true], input[4, int, true], input[5, int, true])

Stack Overflow | aks | 2 months ago
  1. 0

    UnsupportedOperationException: Cannot evalute expression: .. when adding new column withColumn() and udf()

    Stack Overflow | 2 months ago | aks
    java.lang.UnsupportedOperationException: Cannot evaluate expression: parse_df_to_string(input[1, int, true], input[2, int, true], input[3, int, true], input[4, int, true], input[5, int, true])
  2. 0

    python+pyspark: error on inner join with multiple column comparison in pyspark

    Stack Overflow | 3 months ago | Satya
    java.lang.UnsupportedOperationException: Cannot evaluate expression: count(1)
  3. 0

    Spark-Submit python file on cluster

    Stack Overflow | 5 months ago | user5147250
    java.lang.UnsupportedOperationException: Cannot evaluate expression: PythonUDF#<lambda>(input[2, StringType])
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Spark-Submit python file on cluster

    Stack Overflow | 5 months ago | user5147250
    java.lang.UnsupportedOperationException: Cannot evaluate expression: PythonUDF#<lambda>(input[2, StringType])
  6. 0

    Spark-Submit python file on cluster

    Stack Overflow | 5 months ago | user5147250
    org.apache.spark.api.python.PythonException: Traceback (most recent call last): File "/ephemeral/usr/hdp/2.3.4.33-1/spark/python/lib/pyspark.zip/pyspark/worker.py", line 98, in main command = pickleSer._read_with_length(infile) File "/ephemeral/usr/hdp/2.3.4.33-1/spark/python/lib/pyspark.zip/pyspark/serializers.py", line 156, in _read_with_length length = read_int(stream) File "/ephemeral/usr/hdp/2.3.4.33-1/spark/python/lib/pyspark.zip/pyspark/serializers.py", line 545, in read_int raise EOFError EOFError

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.lang.UnsupportedOperationException

      Cannot evaluate expression: parse_df_to_string(input[1, int, true], input[2, int, true], input[3, int, true], input[4, int, true], input[5, int, true])

      at org.apache.spark.sql.catalyst.expressions.Unevaluable$class.doGenCode()
    2. Spark Project Catalyst
      Unevaluable$class.doGenCode
      1. org.apache.spark.sql.catalyst.expressions.Unevaluable$class.doGenCode(Expression.scala:224)
      1 frame
    3. org.apache.spark
      PythonUDF.doGenCode
      1. org.apache.spark.sql.execution.python.PythonUDF.doGenCode(PythonUDF.scala:27)
      1 frame
    4. Spark Project Catalyst
      Expression$$anonfun$genCode$2.apply
      1. org.apache.spark.sql.catalyst.expressions.Expression$$anonfun$genCode$2.apply(Expression.scala:104)
      2. org.apache.spark.sql.catalyst.expressions.Expression$$anonfun$genCode$2.apply(Expression.scala:101)
      2 frames
    5. Scala
      Option.getOrElse
      1. scala.Option.getOrElse(Option.scala:121)
      1 frame
    6. Spark Project Catalyst
      CodegenContext$$anonfun$generateExpressions$1.apply
      1. org.apache.spark.sql.catalyst.expressions.Expression.genCode(Expression.scala:101)
      2. org.apache.spark.sql.catalyst.expressions.codegen.CodegenContext$$anonfun$generateExpressions$1.apply(CodeGenerator.scala:740)
      3. org.apache.spark.sql.catalyst.expressions.codegen.CodegenContext$$anonfun$generateExpressions$1.apply(CodeGenerator.scala:740)
      3 frames