java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[10.200.5.162:50010], original=[10.200.5.162:50010]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration.

Google Groups | Shannon Ma | 2 months ago
  1. 0

    hdfs connect error

    Google Groups | 2 months ago | Shannon Ma
    java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[10.200.5.162:50010], original=[10.200.5.162:50010]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration.
  2. 0

    java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good - 蔡先生的专栏 - 博客频道 - CSDN.NET

    csdn.net | 6 months ago
    java.lang.Exception: java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[127.0.0.1:50010], original=[127.0.0.1:50010]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration.
  3. 0

    Error while copying a file from local to hdfs in cloudlab | Simplilearn - Discussions on Certifications

    simplilearn.com | 4 months ago
    java.io.IOException: Got error, status message , ack with firstBadLink as 139.162.22.151:50010
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Hadoop bad connect ack exception

    Stack Overflow | 2 years ago | Istvan
    java.io.IOException: Bad connect ack with firstBadLink as 10.90.80.44:50010
  6. 0

    HDFS some datanodes of cluster are suddenly disconnected while reducers are running

    Stack Overflow | 5 years ago | user1429825
    java.io.IOException: Bad connect ack with firstBadLink as ***.***.***.148:20010

    6 unregistered visitors
    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.io.IOException

      Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[10.200.5.162:50010], original=[10.200.5.162:50010]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration.

      at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode()
    2. Apache Hadoop HDFS
      DFSOutputStream$DataStreamer.run
      1. org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode(DFSOutputStream.java:1040)
      2. org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1106)
      3. org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1253)
      4. org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:594)
      4 frames