GitHub | mattshma | 8 months ago
Do you know that we can give you better hits? Get more relevant results from Samebug’s stack trace search.
  1. 0

    Error while copying a file from local to hdfs in cloudlab | Simplilearn - Discussions on Certifications | 8 months ago Got error, status message , ack with firstBadLink as
  2. 0

    Hadoop bad connect ack exception

    Stack Overflow | 2 years ago | Istvan Bad connect ack with firstBadLink as
  3. Speed up your debug routine!

    Automated exception search integrated into your IDE

  4. 0

    HDFS some datanodes of cluster are suddenly disconnected while reducers are running

    Stack Overflow | 5 years ago | user1429825 Bad connect ack with firstBadLink as ***.***.***.148:20010
  5. 0

    Exception in createBlockOutputStream when copying data into HDFS

    Stack Overflow | 3 years ago | Naveen R Bad connect ack with firstBadLink as

    6 unregistered visitors
    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis


      Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[DatanodeInfoWithStorage[,DS-a67e1753-7160-4c87-8a87-330732f6ac30,DISK], DatanodeInfoWithStorage[,DS-f42cdbb3-7981-4630-8b98-0ac04bdf92a2,DISK]], original=[DatanodeInfoWithStorage[,DS-a67e1753-7160-4c87-8a87-330732f6ac30,DISK], DatanodeInfoWithStorage[,DS-f42cdbb3-7981-4630-8b98-0ac04bdf92a2,DISK]]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration.

      at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode()
    2. Apache Hadoop HDFS
      1. org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode(
      2. org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(
      3. org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(
      4. org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(
      5. org.apache.hadoop.hdfs.DFSOutputStream$
      5 frames