org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1244 ms. row count = 63113 2016-08-19 12:38:01,832 [Thread-0] ERROR org.apache.hadoop.hdfs.DFSClient - Failed to close inode 457368033 org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException): No lease on /tmp/temp-1982281463/tmp1114763885/_temporary/0/_temporary/attempt__0001_m_000001_1/part-m-00001 (inode 457368033): File does not exist. Holder DFSClient_NONMAPREDUCE_-797544746_1 does not have any open files.

Stack Overflow | Matt Pollock | 9 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    How to avoid UnsatisfiedLinkError when loading Parquet data into PIG

    Stack Overflow | 9 months ago | Matt Pollock
    org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 1244 ms. row count = 63113 2016-08-19 12:38:01,832 [Thread-0] ERROR org.apache.hadoop.hdfs.DFSClient - Failed to close inode 457368033 org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException): No lease on /tmp/temp-1982281463/tmp1114763885/_temporary/0/_temporary/attempt__0001_m_000001_1/part-m-00001 (inode 457368033): File does not exist. Holder DFSClient_NONMAPREDUCE_-797544746_1 does not have any open files.

    Root Cause Analysis

    1. org.apache.parquet.hadoop.InternalParquetRecordReader

      block read in memory in 1244 ms. row count = 63113 2016-08-19 12:38:01,832 [Thread-0] ERROR org.apache.hadoop.hdfs.DFSClient - Failed to close inode 457368033 org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException): No lease on /tmp/temp-1982281463/tmp1114763885/_temporary/0/_temporary/attempt__0001_m_000001_1/part-m-00001 (inode 457368033): File does not exist. Holder DFSClient_NONMAPREDUCE_-797544746_1 does not have any open files.

      at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease()
    2. Apache Hadoop HDFS
      ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod
      1. org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:3481)
      2. org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFileInternal(FSNamesystem.java:3571)
      3. org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFile(FSNamesystem.java:3538)
      4. org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.complete(NameNodeRpcServer.java:884)
      5. org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.complete(ClientNamenodeProtocolServerSideTranslatorPB.java:544)
      6. org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
      6 frames
    3. Hadoop
      Server$Handler$1.run
      1. org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
      2. org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
      3. org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2206)
      4. org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2202)
      4 frames
    4. Java RT
      Subject.doAs
      1. java.security.AccessController.doPrivileged(Native Method)
      2. javax.security.auth.Subject.doAs(Subject.java:422)
      2 frames
    5. Hadoop
      ProtobufRpcEngine$Invoker.invoke
      1. org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1709)
      2. org.apache.hadoop.ipc.Server$Handler.run(Server.java:2200)
      3. org.apache.hadoop.ipc.Client.call(Client.java:1426)
      4. org.apache.hadoop.ipc.Client.call(Client.java:1363)
      5. org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
      5 frames
    6. com.sun.proxy
      $Proxy12.complete
      1. com.sun.proxy.$Proxy12.complete(Unknown Source)
      1 frame
    7. Apache Hadoop HDFS
      ClientNamenodeProtocolTranslatorPB.complete
      1. org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.complete(ClientNamenodeProtocolTranslatorPB.java:464)
      1 frame
    8. Java RT
      Method.invoke
      1. sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
      2. sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
      3. sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
      4. java.lang.reflect.Method.invoke(Method.java:497)
      4 frames
    9. Hadoop
      RetryInvocationHandler.invoke
      1. org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256)
      2. org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104)
      2 frames
    10. com.sun.proxy
      $Proxy13.complete
      1. com.sun.proxy.$Proxy13.complete(Unknown Source)
      1 frame
    11. Apache Hadoop HDFS
      DistributedFileSystem.close
      1. org.apache.hadoop.hdfs.DFSOutputStream.completeFile(DFSOutputStream.java:2354)
      2. org.apache.hadoop.hdfs.DFSOutputStream.closeImpl(DFSOutputStream.java:2336)
      3. org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream.java:2300)
      4. org.apache.hadoop.hdfs.DFSClient.closeAllFilesBeingWritten(DFSClient.java:951)
      5. org.apache.hadoop.hdfs.DFSClient.closeOutputStreams(DFSClient.java:983)
      6. org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSystem.java:1134)
      6 frames
    12. Hadoop
      ShutdownHookManager$1.run
      1. org.apache.hadoop.fs.FileSystem$Cache.closeAll(FileSystem.java:2744)
      2. org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer.run(FileSystem.java:2761)
      3. org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)
      3 frames