java.io.IOException: File from recovered queue is nowhere to be found

Google Groups | ankit beohar | 8 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    Hbase Region Server Down

    Google Groups | 8 months ago | ankit beohar
    java.io.IOException: File from recovered queue is nowhere to be found
  2. 0

    Trouble in RecommenderJob on hadoop

    Stack Overflow | 2 years ago | qianda66
    java.io.FileNotFoundException: File does not exist: /user/hduser/temp/preparePreferenceMatrix/numUsers.bin
  3. 0

    Spark cluster computing framework

    gmane.org | 2 years ago
    java.io.FileNotFoundException: File does not exist: /user/marcel/outputs/output_spark/log0
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Crunch, mail # user - Re: LeaseExpiredExceptions and temp side effect files - 2015-08-21, 20:03

    search-hadoop.com | 1 year ago
    org.apache.crunch.CrunchRuntimeException: Could not read runtime node information
  6. 0

    Failed to run Kafka MapReduce example

    GitHub | 1 year ago | marianomirabelli
    java.io.FileNotFoundException: File does not exist: /home/hduser/gobblin/gobblin-dist/conf/gobblin-mapreduce.properties

  1. jf-ast 3 times, last 1 month ago
2 unregistered visitors
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. java.io.FileNotFoundException

    File does not exist: /hbase/oldWALs/xxxxxxxxxx

    at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf()
  2. Apache Hadoop HDFS
    ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod
    1. org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66)
    2. org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56)
    3. org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1932)
    4. org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1873)
    5. org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1853)
    6. org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1825)
    7. org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:559)
    8. org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.getBlockLocations(AuthorizationProviderProxyClientProtocol.java:87)
    9. org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:363)
    10. org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
    10 frames
  3. Hadoop
    Server$Handler$1.run
    1. org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
    2. org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060)
    3. org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044)
    4. org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2040)
    4 frames
  4. Java RT
    Subject.doAs
    1. java.security.AccessController.doPrivileged(Native Method)
    2. javax.security.auth.Subject.doAs(Subject.java:415)
    2 frames
  5. Hadoop
    Server$Handler.run
    1. org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1707)
    2. org.apache.hadoop.ipc.Server$Handler.run(Server.java:2038)
    2 frames
  6. Java RT
    Constructor.newInstance
    1. sun.reflect.NativeConstructorAccessorImpl.newInstance0(NativeMethod)
    2. sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
    3. sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    4. java.lang.reflect.Constructor.newInstance(Constructor.java:526)
    4 frames
  7. Hadoop
    RemoteException.unwrapRemoteException
    1. org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
    2. org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
    2 frames
  8. Apache Hadoop HDFS
    DistributedFileSystem$3.doCall
    1. org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:1215)
    2. org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:1203)
    3. org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:1193)
    4. org.apache.hadoop.hdfs.DFSInputStream.fetchLocatedBlocksAndGetLastBlockLength(DFSInputStream.java:299)
    5. org.apache.hadoop.hdfs.DFSInputStream.openInfo(DFSInputStream.java:265)
    6. org.apache.hadoop.hdfs.DFSInputStream.<init>(DFSInputStream.java:257)
    7. org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:1492)
    8. org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:302)
    9. org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:298)
    9 frames
  9. Hadoop
    FileSystemLinkResolver.resolve
    1. org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
    1 frame
  10. Apache Hadoop HDFS
    DistributedFileSystem.open
    1. org.apache.hadoop.hdfs.DistributedFileSystem.open(DistributedFileSystem.java:298)
    1 frame
  11. Hadoop
    FileSystem.open
    1. org.apache.hadoop.fs.FilterFileSystem.open(FilterFileSystem.java:161)
    2. org.apache.hadoop.fs.FileSystem.open(FileSystem.java:766)
    2 frames
  12. org.apache.hadoop
    WALFactory.createReader
    1. org.apache.hadoop.hbase.wal.WALFactory.createReader(WALFactory.java:291)
    2. org.apache.hadoop.hbase.wal.WALFactory.createReader(WALFactory.java:267)
    3. org.apache.hadoop.hbase.wal.WALFactory.createReader(WALFactory.java:255)
    4. org.apache.hadoop.hbase.wal.WALFactory.createReader(WALFactory.java:397)
    4 frames
  13. HBase
    ReplicationSource$ReplicationSourceWorkerThread.run
    1. org.apache.hadoop.hbase.replication.regionserver.ReplicationWALReaderManager.openReader(ReplicationWALReaderManager.java:69)
    2. org.apache.hadoop.hbase.replication.regionserver.ReplicationSource$ReplicationSourceWorkerThread.openReader(ReplicationSource.java:746)
    3. org.apache.hadoop.hbase.replication.regionserver.ReplicationSource$ReplicationSourceWorkerThread.run(ReplicationSource.java:542)
    3 frames