java.lang.NoSuchMethodError: com.google.common.base.Splitter.splitToList(Ljava/lang/CharSequence;)Ljava/util/List;

GitHub | samelamin | 4 months ago
  1. 0

    GitHub comment 10#238589343

    GitHub | 4 months ago | samelamin
    java.lang.NoSuchMethodError: com.google.common.base.Splitter.splitToList(Ljava/lang/CharSequence;)Ljava/util/List;
  2. 0

    GitHub comment 12#258612097

    GitHub | 1 month ago | samelamin
    java.lang.NoSuchMethodError: com.google.common.base.Splitter.splitToList(Ljava/lang/CharSequence;)Ljava/util/List; and the stack trace ```
  3. 0

    Cannot create a grails project by calling 'grails create-app PROJECTNAME'

    Stack Overflow | 2 years ago
    java.lang.NoSuchMethodError: com.google.common.base.Splitter.splitToList(Ljava/lang/CharSequence;)Ljava/util/List;
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    When upgrading the previous tika language detector to use the the Optimaize language-detector (https://github.com/optimaize/language-detector) it turns out the 0.5 version depends on the guava-18.jar library. However, all pig versions up to 0.16.0 is bundled with guava-11.jar library. which has precedence to guava-18.jar REGISTER'ed in the script or .pigbootup Thus we get the error: {code} java.lang.NoSuchMethodError: com.google.common.base.Splitter.splitToList(Ljava/lang/CharSequence;)Ljava/util/List; at com.optimaize.langdetect.i18n.LdLocale.fromString(LdLocale.java:77) at com.optimaize.langdetect.profiles.BuiltInLanguages.<clinit>(BuiltInLanguages.java:21) at com.optimaize.langdetect.profiles.LanguageProfileReader.readAllBuiltIn(LanguageProfileReader.java:118) at org.apache.tika.langdetect.OptimaizeLangDetector.loadModels(OptimaizeLangDetector.java:63) at dk.kb.webdanica.criteria.C4.computeNewC4(C4.java:56) at dk.kb.webdanica.criteria.CombinedCombo.exec(CombinedCombo.java:118) at dk.kb.webdanica.criteria.CombinedCombo.exec(CombinedCombo.java:85) at org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.getNext(POUserFunc.java:326) at org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.getNextString(POUserFunc.java:426) at org.apache.pig.backend.hadoop.executionengine.physicalLayer.PhysicalOperator.getNext(PhysicalOperator.java:341) at org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach.processPlan(POForEach.java:404) at org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach.getNextTuple(POForEach.java:321) at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapBase.runPipeline(PigGenericMapBase.java:280) at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapBase.map(PigGenericMapBase.java:275) at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapBase.map(PigGenericMapBase.java:65) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370) at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:212) {code}

    JIRA | 4 months ago | Søren Vejrup Carlsen
    java.lang.NoSuchMethodError: com.google.common.base.Splitter.splitToList(Ljava/lang/CharSequence;)Ljava/util/List;
  6. 0

    Bug

    GitHub | 2 years ago | Bouki
    java.lang.NoSuchMethodError: com.google.common.base.Splitter.splitToList(Ljava/lang/CharSequence;)Ljava/util/List;

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.lang.NoSuchMethodError

      com.google.common.base.Splitter.splitToList(Ljava/lang/CharSequence;)Ljava/util/List;

      at com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystemBase$ParentTimestampUpdateIncludePredicate.create()
    2. com.google.cloud
      GoogleHadoopFileSystemBase.initialize
      1. com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystemBase$ParentTimestampUpdateIncludePredicate.create(GoogleHadoopFileSystemBase.java:572)
      2. com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystemBase.createOptionsBuilderFromConfig(GoogleHadoopFileSystemBase.java:1890)
      3. com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystemBase.configure(GoogleHadoopFileSystemBase.java:1587)
      4. com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystemBase.initialize(GoogleHadoopFileSystemBase.java:793)
      5. com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystemBase.initialize(GoogleHadoopFileSystemBase.java:756)
      5 frames
    3. Hadoop
      Path.getFileSystem
      1. org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2433)
      2. org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
      3. org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
      4. org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
      5. org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
      6. org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
      6 frames
    4. com.google.cloud
      AbstractBigQueryInputFormat.getSplits
      1. com.google.cloud.hadoop.io.bigquery.AbstractBigQueryInputFormat.extractExportPathRoot(AbstractBigQueryInputFormat.java:247)
      2. com.google.cloud.hadoop.io.bigquery.AbstractBigQueryInputFormat.getSplits(AbstractBigQueryInputFormat.java:107)
      2 frames
    5. Spark
      RDD$$anonfun$partitions$2.apply
      1. org.apache.spark.rdd.NewHadoopRDD.getPartitions(NewHadoopRDD.scala:113)
      2. org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)
      3. org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)
      3 frames
    6. Scala
      Option.getOrElse
      1. scala.Option.getOrElse(Option.scala:120)
      1 frame
    7. Spark
      RDD$$anonfun$partitions$2.apply
      1. org.apache.spark.rdd.RDD.partitions(RDD.scala:237)
      2. org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
      3. org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)
      4. org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)
      4 frames
    8. Scala
      Option.getOrElse
      1. scala.Option.getOrElse(Option.scala:120)
      1 frame
    9. Spark
      RDD$$anonfun$partitions$2.apply
      1. org.apache.spark.rdd.RDD.partitions(RDD.scala:237)
      2. org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
      3. org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)
      4. org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)
      4 frames
    10. Scala
      Option.getOrElse
      1. scala.Option.getOrElse(Option.scala:120)
      1 frame
    11. Spark
      RDD.first
      1. org.apache.spark.rdd.RDD.partitions(RDD.scala:237)
      2. org.apache.spark.rdd.RDD$$anonfun$take$1.apply(RDD.scala:1293)
      3. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
      4. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
      5. org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
      6. org.apache.spark.rdd.RDD.take(RDD.scala:1288)
      7. org.apache.spark.rdd.RDD$$anonfun$first$1.apply(RDD.scala:1328)
      8. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
      9. org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
      10. org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
      11. org.apache.spark.rdd.RDD.first(RDD.scala:1327)
      11 frames
    12. com.spotify.spark
      package$BigQuerySQLContext.bigQuerySelect
      1. com.spotify.spark.bigquery.package$BigQuerySQLContext.bigQueryTable(package.scala:112)
      2. com.spotify.spark.bigquery.package$BigQuerySQLContext.bigQuerySelect(package.scala:93)
      2 frames
    13. Unknown
      $iwC.<init>
      1. $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:28)
      2. $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:33)
      3. $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:35)
      4. $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:37)
      5. $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:39)
      6. $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:41)
      7. $iwC$$iwC$$iwC$$iwC.<init>(<console>:43)
      8. $iwC$$iwC$$iwC.<init>(<console>:45)
      9. $iwC$$iwC.<init>(<console>:47)
      10. $iwC.<init>(<console>:49)
      10 frames