java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOException: Cannot run program "/bin/ls": java.io.IOException: error=24, Too many open files

hadoop-hdfs-dev | Apache Hudson Server | 6 years ago
  1. 0

    Hadoop-Hdfs-trunk - Build # 554 - Still Failing

    hadoop-hdfs-dev | 6 years ago | Apache Hudson Server
    java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOException: Cannot run program "/bin/ls": java.io.IOException: error=24, Too many open files
  2. 0

    Hadoop-Hdfs-trunk - Build # 555 - Still Failing

    hadoop-hdfs-dev | 6 years ago | Apache Hudson Server
    java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOException: Cannot run program "/bin/ls": java.io.IOException: error=24, Too many open files
  3. 0

    [JENKINS-1921] Too many open files - Jenkins JIRA

    jenkins-ci.org | 7 months ago
    java.io.IOException: error=24, Too many open files
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Too many open files with svn

    GitHub | 11 months ago | dhireng
    java.io.IOException: error=24, Too many open files
  6. 0

    Too many open files

    areca | 8 years ago | nimdae
    java.io.IOException: Cannot run program "ls": java.io.IOException: error=24, Too many open files

  1. eti22 1 times, last 1 month ago
  2. asdasd 24 times, last 3 months ago
  3. rexgreenza 8 times, last 3 months ago
  4. Benj-AD 1 times, last 4 months ago
  5. Akshay 4 times, last 6 months ago
95 unregistered visitors
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. java.io.IOException

    java.io.IOException: error=24, Too many open files

    at java.lang.UNIXProcess.<init>()
  2. Java RT
    ProcessBuilder.start
    1. java.lang.UNIXProcess.<init>(UNIXProcess.java:148)
    2. java.lang.ProcessImpl.start(ProcessImpl.java:65)
    3. java.lang.ProcessBuilder.start(ProcessBuilder.java:452)
    3 frames
  3. Hadoop
    DiskChecker.checkDir
    1. org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:516)
    2. org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSystem.java:466)
    3. org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:131)
    4. org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:148)
    4 frames
  4. Apache Hadoop HDFS
    TestFileConcurrentReader.setUp
    1. org.apache.hadoop.hdfs.server.datanode.DataNode.getDataDirsFromURIs(DataNode.java:1592)
    2. org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1572)
    3. org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1518)
    4. org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1485)
    5. org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:630)
    6. org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:464)
    7. org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:186)
    8. org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:71)
    9. org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:178)
    10. org.apache.hadoop.hdfs.TestFileConcurrentReader.init(TestFileConcurrentReader.java:88)
    11. org.apache.hadoop.hdfs.TestFileConcurrentReader.setUp(TestFileConcurrentReader.java:73)
    11 frames