java.io.IOException

Block 5922357248 is not available in Alluxio

Samebug tips0

There are no available Samebug tips for this exception. If you know how to solve this issue, help other users by writing a short tip.

Don't give up yet. Paste your full stack trace to get a solution.

Solutions on the web12936

  • via Google Groups by Chanh Le, 11 months ago
    Block 5922357248 is not available in Alluxio
  • via Google Groups by Amran Chen, 7 months ago
    No available Alluxio worker found
  • Block blk_2440422069461309270_3925117 is not valid.
  • Stack trace

    • java.io.IOException: Block 5922357248 is not available in Alluxio at alluxio.client.block.AlluxioBlockStore.getInStream(AlluxioBlockStore.java:115) at alluxio.client.file.FileInStream.updateBlockInStream(FileInStream.java:508) at alluxio.client.file.FileInStream.updateStreams(FileInStream.java:415) at alluxio.client.file.FileInStream.close(FileInStream.java:147) at alluxio.hadoop.HdfsFileInputStream.close(HdfsFileInputStream.java:115) at java.io.FilterInputStream.close(FilterInputStream.java:181) at org.apache.parquet.hadoop.ParquetFileReader.readFooter(ParquetFileReader.java:432) at org.apache.parquet.hadoop.ParquetFileReader.readFooter(ParquetFileReader.java:385) at org.apache.parquet.hadoop.ParquetRecordReader.initializeInternalReader(ParquetRecordReader.java:157) at org.apache.parquet.hadoop.ParquetRecordReader.initialize(ParquetRecordReader.java:140) at org.apache.spark.rdd.SqlNewHadoopRDD$$anon$1.<init>(SqlNewHadoopRDD.scala:180) at org.apache.spark.rdd.SqlNewHadoopRDD.compute(SqlNewHadoopRDD.scala:126) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at org.apache.spark.scheduler.Task.run(Task.scala:89) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)

    Write tip

    You have a different solution? A short tip here would help you and many other users who saw this issue last week.

    Users with the same issue

    You’re the first here who have seen this exception.