java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.OutOfMemoryError: Java heap space at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.getDiskReport(DirectoryScanner.java:549) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.scan(DirectoryScanner.java:422) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.reconcile(DirectoryScanner.java:403) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.run(DirectoryScanner.java:359) at java.util.concurrent.Executors$RunnableAdapter.cal l(Executors.java:471) at java.util.concurrent.FutureTask.runAndReset(Future Task.java:304) at java.util.concurrent.ScheduledThreadPoolExecutor$S cheduledFutureTask.access$301(ScheduledThreadPoolE xecutor.java:178) at java.util.concurrent.ScheduledThreadPoolExecutor$S cheduledFutureTask.run(ScheduledThreadPoolExecutor .java:293)

cloudera.com | 3 months ago
  1. 0

    Datanode shut down when running Hive - Cloudera Community

    cloudera.com | 1 year ago
    java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.OutOfMemoryError: Java heap space at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.getDiskReport(DirectoryScanner.java:549) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.scan(DirectoryScanner.java:422) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.reconcile(DirectoryScanner.java:403) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.run(DirectoryScanner.java:359) at java.util.concurrent.Executors$RunnableAdapter.cal l(Executors.java:471) at java.util.concurrent.FutureTask.runAndReset(Future Task.java:304) at java.util.concurrent.ScheduledThreadPoolExecutor$S cheduledFutureTask.access$301(ScheduledThreadPoolE xecutor.java:178) at java.util.concurrent.ScheduledThreadPoolExecutor$S cheduledFutureTask.run(ScheduledThreadPoolExecutor .java:293)
  2. 0

    Solved: Datanode shut down when running Hive - Cloudera Community

    cloudera.com | 3 months ago
    java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.OutOfMemoryError: Java heap space at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.getDiskReport(DirectoryScanner.java:549) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.scan(DirectoryScanner.java:422) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.reconcile(DirectoryScanner.java:403) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.run(DirectoryScanner.java:359) at java.util.concurrent.Executors$RunnableAdapter.cal l(Executors.java:471) at java.util.concurrent.FutureTask.runAndReset(Future Task.java:304) at java.util.concurrent.ScheduledThreadPoolExecutor$S cheduledFutureTask.access$301(ScheduledThreadPoolE xecutor.java:178) at java.util.concurrent.ScheduledThreadPoolExecutor$S cheduledFutureTask.run(ScheduledThreadPoolExecutor .java:293)
  3. 0

    UIB does not reconnect connection to LDAP when the connection breaks

    GitHub | 2 years ago | StigLau
    java.lang.RuntimeException: usernameExist failed for username=bjartebjart
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Grails 2.0.0-console errors

    GitHub | 5 years ago | bronoman
    java.lang.RuntimeException: It looks like you are missing some calls to the r:layoutResources tag. After rendering your page the following have not been rendered: [defer]
  6. 0

    Cannot create blade or bladeset while the app server is running

    GitHub | 3 years ago | thecapdan
    java.lang.RuntimeException: java.nio.file.NoSuchFileException: C:\Users\danielo\Desktop\TESTING\BladeRunnerJS-v0 .8-131-gc0618d0-DEV\BladeRunnerJS\apps\ted\bap1-bladeset\src\@appns\@bladeset

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.lang.RuntimeException

      java.util.concurrent.ExecutionException: java.lang.OutOfMemoryError: Java heap space at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.getDiskReport(DirectoryScanner.java:549) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.scan(DirectoryScanner.java:422) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.reconcile(DirectoryScanner.java:403) at org.apache.hadoop.hdfs.server.datanode.DirectorySc anner.run(DirectoryScanner.java:359) at java.util.concurrent.Executors$RunnableAdapter.cal l(Executors.java:471) at java.util.concurrent.FutureTask.runAndReset(Future Task.java:304) at java.util.concurrent.ScheduledThreadPoolExecutor$S cheduledFutureTask.access$301(ScheduledThreadPoolE xecutor.java:178) at java.util.concurrent.ScheduledThreadPoolExecutor$S cheduledFutureTask.run(ScheduledThreadPoolExecutor .java:293)

      at java.util.concurrent.ThreadPoolExecutor.runWorker()
    2. Java RT
      Thread.run
      1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
      2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
      3. java.lang.Thread.run(Thread.java:745)
      3 frames