Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Samebug tips

  1. ,

    Java does not handle IPv6 nameservers before 1.8.0_60. Upgrade to the latest JVM.

  2. ,
    Expert tip

    This is a bug in Wise, updating to v4.10 seems to solve the problem.

java.lang.NumberFormatException: For input string: "id"	at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)	at java.lang.Long.parseLong(Long.java:441)	at java.lang.Long.parseLong(Long.java:483)	at scala.collection.immutable.StringLike$class.toLong(StringLike.scala:230)	at scala.collection.immutable.StringOps.toLong(StringOps.scala:31)	at $line65.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.parseTransactions(:38)	at $line67.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1.apply(:42)	at $line67.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1.apply(:42)	at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)	at org.apache.spark.storage.MemoryStore.unrollSafely(MemoryStore.scala:285)	at org.apache.spark.CacheManager.putInBlockManager(CacheManager.scala:171)	at org.apache.spark.CacheManager.getOrCompute(CacheManager.scala:78)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:268)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)	at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)	at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)	at org.apache.spark.scheduler.Task.run(Task.scala:89)	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)	at java.lang.Thread.run(Thread.java:745)