java.io.FileNotFoundException: File file:/home/file/new/ALL.adam/part-r-00027.gz.parquet does not exist

GitHub | car2008 | 3 months ago
  1. 0

    spark-submit throw exception in spark-standalone using .adam which transformed from .vcf

    GitHub | 3 months ago | car2008
    java.io.FileNotFoundException: File file:/home/file/new/ALL.adam/part-r-00027.gz.parquet does not exist
  2. 0

    Spark query of parquet files - error traceback says file is missing... but the file is there. Why?

    Stack Overflow | 1 year ago | GrahamM
    java.io.FileNotFoundException: File file:/home/s26e-5a5fbda111ac17-5edfd8a0d95d/notebook/notebooks/correctedDoppler.parquet/part-r-00015.parquet does not exist
  3. 0

    GitHub comment 1193#264024786

    GitHub | 3 days ago | nowell-jana
    java.io.FileNotFoundException: File /var/gobblin-data/work/working/gobblin-counter/job_gobblin-counter_1480546570415/job.state does not exist
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    FileNotFound Exception when trying to store file in hadoop distributed cache

    Stack Overflow | 3 years ago | Ophir Yoktan
    java.io.FileNotFoundException: File /Workflow/data does not exist
  6. 0

    Getting an error while running mapreduce program in Eclipse

    Stack Overflow | 2 years ago | Prathap Kb
    java.io.FileNotFoundException: File C:/home/hadoop/tmp/mapred/local/taskTracker/jobcache/job_201412090852_0015/attempt_201412090852_0015_m_000003_2/work/tmp does not exist.

    2 unregistered visitors
    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.io.FileNotFoundException

      File file:/home/file/new/ALL.adam/part-r-00027.gz.parquet does not exist

      at org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus()
    2. Hadoop
      FilterFileSystem.getFileStatus
      1. org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:534)
      2. org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:747)
      3. org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:524)
      4. org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:409)
      4 frames
    3. Parquet
      ParquetRecordReader.initialize
      1. parquet.hadoop.ParquetFileReader.readFooter(ParquetFileReader.java:385)
      2. parquet.hadoop.ParquetRecordReader.initializeInternalReader(ParquetRecordReader.java:157)
      3. parquet.hadoop.ParquetRecordReader.initialize(ParquetRecordReader.java:140)
      3 frames
    4. Spark
      Executor$TaskRunner.run
      1. org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:158)
      2. org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:129)
      3. org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:64)
      4. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      5. org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      6. org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      7. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      8. org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      9. org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      10. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      11. org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      12. org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      13. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      14. org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      15. org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      16. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      17. org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      18. org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
      19. org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
      20. org.apache.spark.scheduler.Task.run(Task.scala:89)
      21. org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
      21 frames
    5. Java RT
      Thread.run
      1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
      2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
      3. java.lang.Thread.run(Thread.java:745)
      3 frames