storage.MemoryStore: ensureFreeSpace(14540) called with curMem=7140, maxMem=27370192816/06/07 23:15:53 INFO storage.MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 14.2 KB, free 261.0 MB)16/06/07 23:15:53 INFO broadcast.TorrentBroadcast: Reading broadcast variable 0 took 54 ms16/06/07 23:15:54 INFO storage.MemoryStore: ensureFreeSpace(204280) called with curMem=21680, maxMem=27370192816/06/07 23:15:54 INFO storage.MemoryStore: Block broadcast_0 stored as values in memory (estimated size 199.5 KB, free 260.8 MB)16/06/07 23:15:58 INFO Configuration.deprecation: mapred.tip.id is deprecated. Instead, use mapreduce.task.id16/06/07 23:15:58 INFO Configuration.deprecation: mapred.task.id is deprecated. Instead, use mapreduce.task.attempt.id16/06/07 23:15:58 INFO Configuration.deprecation: mapred.task.is.map is deprecated. Instead, use mapreduce.task.ismap16/06/07 23:15:58 INFO Configuration.deprecation: mapred.task.partition is deprecated. Instead, use mapreduce.task.partition16/06/07 23:15:58 INFO Configuration.deprecation: mapred.job.id is deprecated. Instead, use mapreduce.job.id16/06/07 23:16:01 ERROR executor.Executor: Exception in task 0.0 in stage 0.0 (TID 0) at com.ghost.scala.DataTranslate$.AnalyData( )

aboutyun.com | 4 months ago
  1. 0

    spark运行中问题。-Hadoop2|YARN-about云开发

    aboutyun.com | 4 months ago
    storage.MemoryStore: ensureFreeSpace(14540) called with curMem=7140, maxMem=27370192816/06/07 23:15:53 INFO storage.MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 14.2 KB, free 261.0 MB)16/06/07 23:15:53 INFO broadcast.TorrentBroadcast: Reading broadcast variable 0 took 54 ms16/06/07 23:15:54 INFO storage.MemoryStore: ensureFreeSpace(204280) called with curMem=21680, maxMem=27370192816/06/07 23:15:54 INFO storage.MemoryStore: Block broadcast_0 stored as values in memory (estimated size 199.5 KB, free 260.8 MB)16/06/07 23:15:58 INFO Configuration.deprecation: mapred.tip.id is deprecated. Instead, use mapreduce.task.id16/06/07 23:15:58 INFO Configuration.deprecation: mapred.task.id is deprecated. Instead, use mapreduce.task.attempt.id16/06/07 23:15:58 INFO Configuration.deprecation: mapred.task.is.map is deprecated. Instead, use mapreduce.task.ismap16/06/07 23:15:58 INFO Configuration.deprecation: mapred.task.partition is deprecated. Instead, use mapreduce.task.partition16/06/07 23:15:58 INFO Configuration.deprecation: mapred.job.id is deprecated. Instead, use mapreduce.job.id16/06/07 23:16:01 ERROR executor.Executor: Exception in task 0.0 in stage 0.0 (TID 0) at com.ghost.scala.DataTranslate$.AnalyData( )
  2. 0

    restarting streaming context

    Google Groups | 3 years ago | Arindam Paul
    storage.MemoryStore: ensureFreeSpace(3717) called with curMem=16907806, maxMem=339585269 13/08/22 15:16:24 INFO storage.MemoryStore: Block input-0-1377164784400 stored as bytes to memory (size 3.6 KB, free 307.7 MB) 13/08/22 15:16:24 WARN storage.BlockManager: Putting block input-0-1377164784400 failed spark.SparkException: Error sending message to BlockManager as driverActor is null [message = spark.storage.UpdateBlockInfo@38a038e6]
  3. 0

    Reading from Hbase

    Google Groups | 4 years ago | Håvard Wahl Kongsgård
    storage.MemoryStore: ensureFreeSpace(164832) called with curMem=0, maxMem=339585269 13/05/03 22:46:55 INFO storage.MemoryStore: Block broadcast_0 stored as values to memory (estimated size 161.0 KB, free 323.7 MB) hBaseRDD: spark.rdd.NewHadoopRDD[org.apache.hadoop.hbase.io.ImmutableBytesWritable,org.apache.hadoop.hbase.client.Result] = NewHadoopRDD[0] at NewHadoopRDD at <console>:31  hBaseRDD.count() java.lang.InstantiationException: MyTableInputFormat
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Spark driver memory keeps growing - Hortonworks

    hortonworks.com | 4 weeks ago
    storage.MemoryStore: Not enough space to cache broadcast_90 in memory! (computed 2.3 GB so far) 16/08/04 03:18:37 WARN storage.MemoryStore: Not enough space to cache broadcast_90 in memory! (computed 2.3 GB so far) 16/08/04 03:19:22 WARN storage.MemoryStore: Not enough space to cache broadcast_90 in memory! (computed 2.4 GB so far) 16/08/04 03:20:07 WARN storage.MemoryStore: Not enough space to cache broadcast_90 in memory! (computed 2.5 GB so far) 16/08/04 03:20:53 WARN storage.MemoryStore: Not enough space to cache broadcast_90 in memory! (computed 2.6 GB so far) 16/08/04 03:21:11 WARN storage.MemoryStore: Not enough space to cache broadcast_90 in memory! (computed 2.7 GB so far) 16/08/04 03:21:15 WARN storage.MemoryStore: Not enough space to cache broadcast_90 in memory! (computed 2.5 GB so far) 16/08/04 03:44:22 WARN storage.MemoryStore: Not enough space to cache broadcast_90 in memory! (computed 3.4 GB so far) 16/08/04 03:53:03 WARN storage.MemoryStore: Not enough space to cache broadcast_90 in memory! (computed 3.8 GB so far) 16/08/04 04:02:00 WARN storage.MemoryStore: Not enough space to cache broadcast_90 in memory! (computed 4.1 GB so far) 16/08/04 04:20:52 ERROR executor.CoarseGrainedExecutorBackend: RECEIVED SIGNAL 15: SIGTERM 16/08/04 04:20:52 ERROR executor.Executor: Exception in task 1.0 in stage 62.1 (TID 1109) java.lang.OutOfMemoryError: Java heap space
  6. 0

    CDH5.0: Spark shell cannot work when enable LZO in core-site.xml - Grokbase

    grokbase.com | 1 year ago
    storage.MemoryStore: Block broadcast_0 stored as values to memory (estimated size 200.9 KB, free 294.7 MB) file: org.apache.spark.rdd.RDD[String] = MappedRDD[1] at textFile at <console>:12scala> val counts = file.flatMap(line => line.split(" ")).map(word => (word, 1)).reduceByKey(_ + _)java.lang.RuntimeException: Error in configuring object at org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:109) at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:75) at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133) at org.apache.spark.rdd.HadoopRDD.getInputFormat(HadoopRDD.scala:123)

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. storage.MemoryStore

      ensureFreeSpace(14540) called with curMem=7140, maxMem=27370192816/06/07 23:15:53 INFO storage.MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 14.2 KB, free 261.0 MB)16/06/07 23:15:53 INFO broadcast.TorrentBroadcast: Reading broadcast variable 0 took 54 ms16/06/07 23:15:54 INFO storage.MemoryStore: ensureFreeSpace(204280) called with curMem=21680, maxMem=27370192816/06/07 23:15:54 INFO storage.MemoryStore: Block broadcast_0 stored as values in memory (estimated size 199.5 KB, free 260.8 MB)16/06/07 23:15:58 INFO Configuration.deprecation: mapred.tip.id is deprecated. Instead, use mapreduce.task.id16/06/07 23:15:58 INFO Configuration.deprecation: mapred.task.id is deprecated. Instead, use mapreduce.task.attempt.id16/06/07 23:15:58 INFO Configuration.deprecation: mapred.task.is.map is deprecated. Instead, use mapreduce.task.ismap16/06/07 23:15:58 INFO Configuration.deprecation: mapred.task.partition is deprecated. Instead, use mapreduce.task.partition16/06/07 23:15:58 INFO Configuration.deprecation: mapred.job.id is deprecated. Instead, use mapreduce.job.id16/06/07 23:16:01 ERROR executor.Executor: Exception in task 0.0 in stage 0.0 (TID 0) at com.ghost.scala.DataTranslate$.AnalyData( )

      at com.ghost.scala.DataTranslate$$anonfun$1.apply()
    2. com.ghost.scala
      DataTranslate$$anonfun$1.apply
      1. com.ghost.scala.DataTranslate$$anonfun$1.apply(DataTranslate.scala:29)
      2. com.ghost.scala.DataTranslate$$anonfun$1.apply(DataTranslate.scala:29)
      2 frames
    3. Scala
      Iterator$$anon$11.hasNext
      1. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
      2. scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
      3. scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:388)
      4. scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:388)
      5. scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
      6. scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
      6 frames
    4. Spark
      Executor$TaskRunner.run
      1. org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:203)
      2. org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:73)
      3. org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
      4. org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
      5. org.apache.spark.scheduler.Task.run(Task.scala:88)
      6. org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
      6 frames
    5. Java RT
      ThreadPoolExecutor$Worker.run
      1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
      2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
      2 frames