org.apache.kafka.common.config.ConfigException: No bootstrap urls given in bootstrap.servers

Stack Overflow | pritid | 6 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    Apache kafka cluster using MapR Spark streaming not working

    Stack Overflow | 6 months ago | pritid
    org.apache.kafka.common.config.ConfigException: No bootstrap urls given in bootstrap.servers

    Root Cause Analysis

    1. org.apache.kafka.common.config.ConfigException

      No bootstrap urls given in bootstrap.servers

      at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses()
    2. Apache Kafka
      KafkaConsumer.partitionsFor
      1. org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:57)
      2. org.apache.kafka.clients.consumer.KafkaConsumer.initializeConsumer(KafkaConsumer.java:606)
      3. org.apache.kafka.clients.consumer.KafkaConsumer.partitionsFor(KafkaConsumer.java:1563)
      3 frames
    3. org.apache.spark
      KafkaCluster$$anonfun$getPartitions$1$$anonfun$1.apply
      1. org.apache.spark.streaming.kafka.v09.KafkaCluster$$anonfun$getPartitions$1$$anonfun$1.apply(KafkaCluster.scala:54)
      2. org.apache.spark.streaming.kafka.v09.KafkaCluster$$anonfun$getPartitions$1$$anonfun$1.apply(KafkaCluster.scala:54)
      2 frames
    4. Scala
      AbstractTraversable.flatMap
      1. scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:251)
      2. scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:251)
      3. scala.collection.immutable.Set$Set1.foreach(Set.scala:74)
      4. scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:251)
      5. scala.collection.AbstractTraversable.flatMap(Traversable.scala:105)
      5 frames
    5. org.apache.spark
      KafkaUtils.createDirectStream
      1. org.apache.spark.streaming.kafka.v09.KafkaCluster$$anonfun$getPartitions$1.apply(KafkaCluster.scala:53)
      2. org.apache.spark.streaming.kafka.v09.KafkaCluster$$anonfun$getPartitions$1.apply(KafkaCluster.scala:52)
      3. org.apache.spark.streaming.kafka.v09.KafkaCluster.withConsumer(KafkaCluster.scala:164)
      4. org.apache.spark.streaming.kafka.v09.KafkaCluster.getPartitions(KafkaCluster.scala:52)
      5. org.apache.spark.streaming.kafka.v09.KafkaUtils$.getFromOffsets(KafkaUtils.scala:421)
      6. org.apache.spark.streaming.kafka.v09.KafkaUtils$.createDirectStream(KafkaUtils.scala:292)
      7. org.apache.spark.streaming.kafka.v09.KafkaUtils$.createDirectStream(KafkaUtils.scala:397)
      8. org.apache.spark.streaming.kafka.v09.KafkaUtils.createDirectStream(KafkaUtils.scala)
      8 frames
    6. com.cisco.it
      KafkaDirectStreamin2.main
      1. com.cisco.it.log.KafkaDirectStreamin2.main(KafkaDirectStreamin2.java:111)
      1 frame
    7. Java RT
      Method.invoke
      1. sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
      2. sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
      3. sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
      4. java.lang.reflect.Method.invoke(Method.java:606)
      4 frames
    8. Spark
      SparkSubmit.main
      1. org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:742)
      2. org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
      3. org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
      4. org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
      5. org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
      5 frames