Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via Stack Overflow by Charan Adabala
, 2 years ago
ArrayBuffer(org.apache.spark.SparkException: Couldn't find leaders for Set([normalized-tenant4,0]))
via Google Groups by Charan Adabala, 1 year ago
ArrayBuffer(org.apache.spark.SparkException: Couldn't find leaders for Set([normalized-tenant4,0]))
via Stack Overflow by AKC
, 1 year ago
ArrayBuffer(kafka.common.NotLeaderForPartitionException, org.apache.spark.SparkException: Couldn't find leader offsets for Set([MyTopic,11]))
via Stack Overflow by Cheater
, 2 months ago
ArrayBuffer(kafka.common.NotLeaderForPartitionException, org.apache.spark.SparkException: Couldn't find leader offsets for Set([test,0]))
via Stack Overflow by Utkarsh Kajaria
, 10 months ago
ArrayBuffer(org.apache.spark.SparkException: Couldn't find leaders for Set([topic5,0], [topic5,1]))
via Stack Overflow by Kaushal
, 1 year ago
ArrayBuffer(kafka.common.NotLeaderForPartitionException, kafka.common.NotLeaderForPartitionException, org.apache.spark.SparkException: Couldn't find leader offsets for Set([Test1,4], [Test2,1],[Test2,3], [Test3,4], [Test4,1], [Test2,3], [Test5,2], [Test2,0], [Test1,5], [Test2,5], [Test2,2]))
org.apache.spark.SparkException: ArrayBuffer(org.apache.spark.SparkException: Couldn't find leaders for 
Set([normalized-tenant4,0]))	at org.apache.spark.streaming.kafka.DirectKafkaInputDStream.latestLeaderOffsets(DirectKafkaInputDStream.scala:123)	at org.apache.spark.streaming.kafka.DirectKafkaInputDStream.compute(DirectKafkaInputDStream.scala:145)	at org.apache.spark.streaming.dstream.DStream$$anonfun$getOrCompute$1$$anonfun$1$$anonfun$apply$7.apply(DStream.scala:350)	at org.apache.spark.streaming.dstream.DStream$$anonfun$getOrCompute$1$$anonfun$1$$anonfun$apply$7.apply(DStream.scala:350)	at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)	at org.apache.spark.streaming.dstream.DStream$$anonfun$getOrCompute$1$$anonfun$1.apply(DStream.scala:349)	at org.apache.spark.streaming.dstream.DStream$$anonfun$getOrCompute$1$$anonfun$1.apply(DStream.scala:349)	at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:399)	at org.apache.spark.streaming.dstream.DStream$$anonfun$getOrCompute$1.apply(DStream.scala:344)	at org.apache.spark.streaming.dstream.DStream$$anonfun$getOrCompute$1.apply(DStream.scala:342)	at scala.Option.orElse(Option.scala:257)	at org.apache.spark.streaming.dstream.DStream.getOrCompute(DStream.scala:339)	at org.apache.spark.streaming.dstream.ForEachDStream.generateJob(ForEachDStream.scala:38)	at org.apache.spark.streaming.DStreamGraph$$anonfun$1.apply(DStreamGraph.scala:120)	at org.apache.spark.streaming.DStreamGraph$$anonfun$1.apply(DStreamGraph.scala:120)	at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:251)	at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:251)	at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)	at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)	at scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:251)	at scala.collection.AbstractTraversable.flatMap(Traversable.scala:105)	at org.apache.spark.streaming.DStreamGraph.generateJobs(DStreamGraph.scala:120)	at org.apache.spark.streaming.scheduler.JobGenerator$$anonfun$2.apply(JobGenerator.scala:247)	at org.apache.spark.streaming.scheduler.JobGenerator$$anonfun$2.apply(JobGenerator.scala:245)	at scala.util.Try$.apply(Try.scala:161)	at org.apache.spark.streaming.scheduler.JobGenerator.generateJobs(JobGenerator.scala:245)	at org.apache.spark.streaming.scheduler.JobGenerator.org$apache$spark$streaming$scheduler$JobGenerator$$processEvent(JobGenerator.scala:181)	at org.apache.spark.streaming.scheduler.JobGenerator$$anon$1.onReceive(JobGenerator.scala:87)	at org.apache.spark.streaming.scheduler.JobGenerator$$anon$1.onReceive(JobGenerator.scala:86)	at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)