java.io.IOException: Could not obtain block: blk_3380512596555557728_1002 file=/hbase/hbase.version

hadoop-hdfs-dev | Bassam Tabbara (JIRA) | 7 years ago
  1. 0

    [jira] Created: (HDFS-872) DFSClient 0.20.1 is incompatible with HDFS 0.20.2

    hadoop-hdfs-dev | 7 years ago | Bassam Tabbara (JIRA)
    java.io.IOException: Could not obtain block: blk_3380512596555557728_1002 file=/hbase/hbase.version
  2. 0

    Reading file from HDFS using HDFS Java API

    Stack Overflow | 3 years ago | dnivra
    java.io.IOException: Could not obtain block: blk_-747325769320762541_16269493 file=/user/s3t.txt
  3. Speed up your debug routine!

    Automated exception search integrated into your IDE

  4. 0

    Re: [Aurelius] Faunus: Incremental loading for nodes and edges?

    Google Groups | 3 years ago | David
    java.io.IOException: Blocklist for /user/graphie/output/job-1/part-r-00000 has changed!
  5. 0

    DataXceiver java.io.InterruptedIOException error on scannning Hbase table

    Google Groups | 3 years ago | AnushaGuntaka
    java.io.IOException: Could not seek StoreFileScanner[HFileScanner for reader reader=hdfs://172.20.193.234:9000/assortmentLinking/performance_weekly_sku/fa0fb91bd58f2117443db90278c3a3fe/cf/1597dcfc99e540 25bc7b848cfb998b1f, compression=none, cacheConf=CacheConfig:enabled [cacheDataOnRead=true] [cacheDataOnWrite=false] [cacheIndexesOnWrite=false] [cacheBloomsOnWrite=false] [cacheEvictOnClose=false] [cacheCompressed=false], firstKey=SKU128331STORE3942WEEK37/cf:facings/1397826519184/Put, lastKey=SKU129999STORE3966WEEK9/cf:week_id/1397827347036/Put, avgKeyLen=53, avgValueLen=3, entries=120178401, length=7838467097, cur=null] to key SKU128331STORE3942WEEK37/cf:/LATEST_TIMESTAMP/DeleteFamily/vlen=0/ts=0

  1. rp 1 times, last 3 weeks ago
1 unregistered visitors
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. java.io.IOException

    Could not obtain block: blk_3380512596555557728_1002 file=/hbase/hbase.version

    at org.apache.hadoop.hdfs.DFSClient$DFSInputStream.chooseDataNode()
  2. Apache Hadoop HDFS
    DFSClient$DFSInputStream.read
    1. org.apache.hadoop.hdfs.DFSClient$DFSInputStream.chooseDataNode(DFSClient.java:1788)
    2. org.apache.hadoop.hdfs.DFSClient$DFSInputStream.blockSeekTo(DFSClient.java:1616)
    3. org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1743)
    4. org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1673)
    4 frames
  3. Java RT
    DataInputStream.readUTF
    1. java.io.DataInputStream.readUnsignedShort(DataInputStream.java:320)
    2. java.io.DataInputStream.readUTF(DataInputStream.java:572)
    2 frames
  4. HBase
    FSUtils.checkVersion
    1. org.apache.hadoop.hbase.util.FSUtils.getVersion(FSUtils.java:189)
    2. org.apache.hadoop.hbase.util.FSUtils.checkVersion(FSUtils.java:208)
    2 frames
  5. HBase - Client
    HMaster.<init>
    1. org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:208)
    1 frame
  6. Java RT
    Constructor.newInstance
    1. sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    2. sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
    3. sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
    4. java.lang.reflect.Constructor.newInstance(Constructor.java:513)
    4 frames
  7. HBase - Client
    HMaster.main
    1. org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:1241)
    2. org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:1282)
    2 frames