java.lang.OutOfMemoryError: Unable to acquire 262144 bytes of memory, got 220032

Apache's JIRA Issue Tracker | Josh Rosen | 1 year ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    NullPointerException produced by com.mchange.v2.async.ThreadPoolAsynchronousRunner$PoolThread-#5

    Stack Overflow | 3 years ago | Vishal Singh
    java.lang.OutOfMemoryError: Java heap space </code></pre> <p>And </p> <pre><code> Exception in thread "http-8080-149" at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489)
  2. 0

    java.lang.OutOfMemoryError: Java heap space | Oracle Community

    oracle.com | 12 months ago
    java.lang.OutOfMemoryError: Java heap space - - - -
  3. Speed up your debug routine!

    Automated exception search integrated into your IDE

  4. 0

    [elasticsearch] server failed excpetion - Grokbase

    grokbase.com | 9 months ago
    java.lang.OutOfMemoryError: Java heap space at org.elasticsearch.common.compress.lzf.BufferRecycler.allocDecodeBuffer(BufferRecycler.java:150) at org.elasticsearch.common.io.stream.LZFStreamInput.<init>(LZFStreamInput.java:91) at org.elasticsearch.common.io.stream.CachedStreamInput.instance(CachedStreamInput.java:48) at org.elasticsearch.common.io.stream.CachedStreamInput.getCharArray(CachedStreamInput.java:79) at org.elasticsearch.common.io.stream.StreamInput.readUTF(StreamInput.java:160) at org.elasticsearch.common.io.stream.HandlesStreamInput.readUTF(HandlesStreamInput.java:49) at org.elasticsearch.action.index.IndexResponse.readFrom(IndexResponse.java:142) at org.elasticsearch.transport.netty.MessageChannelHandler.handleResponse(MessageChannelHandler.java:254) at org.elasticsearch.transport.netty.MessageChannelHandler.process(MessageChannelHandler.java:233) at org.elasticsearch.transport.netty.MessageChannelHandler.callDecode(MessageChannelHandler.java:141) at org.elasticsearch.transport.netty.MessageChannelHandler.messageReceived(MessageChannelHandler.java:95) at org.elasticsearch.common.netty.channel.SimpleChannelUpstreamHandler.handleUpstream(SimpleChannelUpstreamHandler.java:75) at org.elasticsearch.common.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:564) at org.elasticsearch.common.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:559) at org.elasticsearch.common.netty.channel.Channels.fireMessageReceived(Channels.java:268) at org.elasticsearch.common.netty.channel.Channels.fireMessageReceived(Channels.java:255) at org.elasticsearch.common.netty.channel.socket.nio.NioWorker.read(NioWorker.java:94) at org.elasticsearch.common.netty.channel.socket.nio.AbstractNioWorker.processSelectedKeys(AbstractNioWorker.java:364) at org.elasticsearch.common.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:238) at org.elasticsearch.common.netty.channel.socket.nio.NioWorker.run(NioWorker.java:38) at org.elasticsearch.common.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:885) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:907)
  5. 0

    RE: unknown issue in submitting a spark job

    apache.org | 1 year ago
    java.lang.OutOfMemoryError: GC overhead limit exceeded at > akka.actor.LightArrayRevolverScheduler$$anon$12.nextTick(Scheduler.scala:397) at > akka.actor.LightArrayRevolverScheduler$$anon$12.run(Scheduler.scala:363)

    1 unregistered visitors
    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.lang.OutOfMemoryError

      Unable to acquire 262144 bytes of memory, got 220032

      at org.apache.spark.memory.MemoryConsumer.allocateArray()
    2. org.apache.spark
      BytesToBytesMap.<init>
      1. org.apache.spark.memory.MemoryConsumer.allocateArray(MemoryConsumer.java:91)
      2. org.apache.spark.unsafe.map.BytesToBytesMap.allocate(BytesToBytesMap.java:735)
      3. org.apache.spark.unsafe.map.BytesToBytesMap.<init>(BytesToBytesMap.java:197)
      4. org.apache.spark.unsafe.map.BytesToBytesMap.<init>(BytesToBytesMap.java:212)
      4 frames
    3. Spark Project SQL
      UnsafeFixedWidthAggregationMap.<init>
      1. org.apache.spark.sql.execution.UnsafeFixedWidthAggregationMap.<init>(UnsafeFixedWidthAggregationMap.java:103)
      1 frame
    4. org.apache.spark
      TungstenAggregate$$anonfun$doExecute$1$$anonfun$2.apply
      1. org.apache.spark.sql.execution.aggregate.TungstenAggregationIterator.<init>(TungstenAggregationIterator.scala:483)
      2. org.apache.spark.sql.execution.aggregate.TungstenAggregate$$anonfun$doExecute$1$$anonfun$2.apply(TungstenAggregate.scala:95)
      3. org.apache.spark.sql.execution.aggregate.TungstenAggregate$$anonfun$doExecute$1$$anonfun$2.apply(TungstenAggregate.scala:86)
      3 frames
    5. Spark
      Executor$TaskRunner.run
      1. org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$20.apply(RDD.scala:710)
      2. org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$20.apply(RDD.scala:710)
      3. org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      4. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      5. org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      6. org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      7. org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      8. org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      9. org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
      10. org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
      11. org.apache.spark.scheduler.Task.run(Task.scala:89)
      12. org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
      12 frames
    6. Java RT
      Thread.run
      1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
      2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
      3. java.lang.Thread.run(Thread.java:745)
      3 frames