java.lang.RuntimeException: java.lang.IllegalArgumentException: Size exceeds Integer.MAX_VALUE

Stack Overflow | clay | 6 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    Spark IllegalArgumentException: Size exceeds Integer.MAX_VALUE

    Stack Overflow | 6 months ago | clay
    java.lang.RuntimeException: java.lang.IllegalArgumentException: Size exceeds Integer.MAX_VALUE
  2. 0

    hadoop for spark: Increase number of partitions

    Stack Overflow | 1 year ago | Adetiloye Philip Kehinde
    java.lang.RuntimeException: java.lang.IllegalArgumentException: Size exceeds Integer.MAX_VALUE
  3. 0

    Size exceeds Integer.MAX_VALUE

    GitHub | 4 months ago | giaosudau
    java.lang.RuntimeException: java.lang.reflect.InvocationTargetException
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.lang.RuntimeException

      java.lang.IllegalArgumentException: Size exceeds Integer.MAX_VALUE

      at sun.nio.ch.FileChannelImpl.map()
    2. Java RT
      FileChannelImpl.map
      1. sun.nio.ch.FileChannelImpl.map(FileChannelImpl.java:869)
      1 frame
    3. Spark
      BlockManager$$anonfun$getLocalBytes$2.apply
      1. org.apache.spark.storage.DiskStore$$anonfun$getBytes$2.apply(DiskStore.scala:103)
      2. org.apache.spark.storage.DiskStore$$anonfun$getBytes$2.apply(DiskStore.scala:91)
      3. org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1287)
      4. org.apache.spark.storage.DiskStore.getBytes(DiskStore.scala:105)
      5. org.apache.spark.storage.BlockManager.org$apache$spark$storage$BlockManager$$doGetLocalBytes(BlockManager.scala:497)
      6. org.apache.spark.storage.BlockManager$$anonfun$getLocalBytes$2.apply(BlockManager.scala:475)
      7. org.apache.spark.storage.BlockManager$$anonfun$getLocalBytes$2.apply(BlockManager.scala:475)
      7 frames
    4. Scala
      Option.map
      1. scala.Option.map(Option.scala:146)
      1 frame
    5. Spark
      NettyBlockRpcServer$$anonfun$2.apply
      1. org.apache.spark.storage.BlockManager.getLocalBytes(BlockManager.scala:475)
      2. org.apache.spark.storage.BlockManager.getBlockData(BlockManager.scala:280)
      3. org.apache.spark.network.netty.NettyBlockRpcServer$$anonfun$2.apply(NettyBlockRpcServer.scala:60)
      4. org.apache.spark.network.netty.NettyBlockRpcServer$$anonfun$2.apply(NettyBlockRpcServer.scala:60)
      4 frames
    6. Scala
      ArrayOps$ofRef.map
      1. scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
      2. scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
      3. scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
      4. scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
      5. scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
      6. scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:186)
      6 frames
    7. Spark
      NettyBlockRpcServer.receive
      1. org.apache.spark.network.netty.NettyBlockRpcServer.receive(NettyBlockRpcServer.scala:60)
      1 frame
    8. Spark
      TransportChannelHandler.channelRead0
      1. org.apache.spark.network.server.TransportRequestHandler.processRpcRequest(TransportRequestHandler.java:158)
      2. org.apache.spark.network.server.TransportRequestHandler.handle(TransportRequestHandler.java:106)
      3. org.apache.spark.network.server.TransportChannelHandler.channelRead0(TransportChannelHandler.java:119)
      4. org.apache.spark.network.server.TransportChannelHandler.channelRead0(TransportChannelHandler.java:51)
      4 frames
    9. Netty
      AbstractChannelHandlerContext.fireChannelRead
      1. io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
      2. io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
      3. io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
      4. io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:266)
      5. io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
      6. io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
      7. io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
      8. io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
      9. io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
      9 frames
    10. Spark
      TransportFrameDecoder.channelRead
      1. org.apache.spark.network.util.TransportFrameDecoder.channelRead(TransportFrameDecoder.java:85)
      1 frame
    11. Netty
      SingleThreadEventExecutor$2.run
      1. io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
      2. io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
      3. io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846)
      4. io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131)
      5. io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511)
      6. io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468)
      7. io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382)
      8. io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354)
      9. io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111)
      9 frames
    12. Java RT
      Thread.run
      1. java.lang.Thread.run(Thread.java:745)
      1 frame