org.apache.kafka.common.config.ConfigException: No bootstrap urls given in bootstrap.servers

Stack Overflow | pritid | 8 months ago
tip
Click on the to mark the solution that helps you, Samebug will learn from it.
As a community member, you’ll be rewarded for you help.
  1. 0

    Apache kafka cluster using MapR Spark streaming not working

    Stack Overflow | 8 months ago | pritid
    org.apache.kafka.common.config.ConfigException: No bootstrap urls given in bootstrap.servers

    Root Cause Analysis

    1. org.apache.kafka.common.config.ConfigException

      No bootstrap urls given in bootstrap.servers

      at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses()
    2. Apache Kafka
      KafkaConsumer.partitionsFor
      1. org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:57)
      2. org.apache.kafka.clients.consumer.KafkaConsumer.initializeConsumer(KafkaConsumer.java:606)
      3. org.apache.kafka.clients.consumer.KafkaConsumer.partitionsFor(KafkaConsumer.java:1563)
      3 frames
    3. org.apache.spark
      KafkaCluster$$anonfun$getPartitions$1$$anonfun$1.apply
      1. org.apache.spark.streaming.kafka.v09.KafkaCluster$$anonfun$getPartitions$1$$anonfun$1.apply(KafkaCluster.scala:54)
      2. org.apache.spark.streaming.kafka.v09.KafkaCluster$$anonfun$getPartitions$1$$anonfun$1.apply(KafkaCluster.scala:54)
      2 frames
    4. Scala
      AbstractTraversable.flatMap
      1. scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:251)
      2. scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:251)
      3. scala.collection.immutable.Set$Set1.foreach(Set.scala:74)
      4. scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:251)
      5. scala.collection.AbstractTraversable.flatMap(Traversable.scala:105)
      5 frames
    5. org.apache.spark
      KafkaUtils.createDirectStream
      1. org.apache.spark.streaming.kafka.v09.KafkaCluster$$anonfun$getPartitions$1.apply(KafkaCluster.scala:53)
      2. org.apache.spark.streaming.kafka.v09.KafkaCluster$$anonfun$getPartitions$1.apply(KafkaCluster.scala:52)
      3. org.apache.spark.streaming.kafka.v09.KafkaCluster.withConsumer(KafkaCluster.scala:164)
      4. org.apache.spark.streaming.kafka.v09.KafkaCluster.getPartitions(KafkaCluster.scala:52)
      5. org.apache.spark.streaming.kafka.v09.KafkaUtils$.getFromOffsets(KafkaUtils.scala:421)
      6. org.apache.spark.streaming.kafka.v09.KafkaUtils$.createDirectStream(KafkaUtils.scala:292)
      7. org.apache.spark.streaming.kafka.v09.KafkaUtils$.createDirectStream(KafkaUtils.scala:397)
      8. org.apache.spark.streaming.kafka.v09.KafkaUtils.createDirectStream(KafkaUtils.scala)
      8 frames
    6. com.cisco.it
      KafkaDirectStreamin2.main
      1. com.cisco.it.log.KafkaDirectStreamin2.main(KafkaDirectStreamin2.java:111)
      1 frame
    7. Java RT
      Method.invoke
      1. sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
      2. sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
      3. sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
      4. java.lang.reflect.Method.invoke(Method.java:606)
      4 frames
    8. Spark
      SparkSubmit.main
      1. org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:742)
      2. org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
      3. org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
      4. org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
      5. org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
      5 frames