Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via Stack Overflow by D.Asare
, 1 year ago
Couldn't connect and authenticate to get collection
via Stack Overflow by vedat
, 2 years ago
via Google Groups by Alessandro Gelormini, 2 years ago
Couldn't connect and authenticate to get collection
via Stack Overflow by Kevin
, 2 years ago
Couldn't connect and authenticate to get collection
via Stack Overflow by Shisoft
, 2 years ago
Couldn't connect and authenticate to get collection
via Google Groups by Adnan Fiaz, 2 years ago
Couldn't connect and authenticate to get collection
java.lang.NullPointerException: 	at com.mongodb.Mongo.createCluster(Mongo.java:613)	at com.mongodb.Mongo.(Mongo.java:283)	at com.mongodb.MongoClient.(MongoClient.java:265)	at com.mongodb.hadoop.util.MongoConfigUtil.getMongoClient(MongoConfigUtil.java:999)	at com.mongodb.hadoop.util.MongoConfigUtil.getCollection(MongoConfigUtil.java:439)	at com.mongodb.hadoop.splitter.MongoSplitterFactory.getSplitterByStats(MongoSplitterFactory.java:72)	at com.mongodb.hadoop.splitter.MongoSplitterFactory.getSplitter(MongoSplitterFactory.java:113)	at com.mongodb.hadoop.MongoInputFormat.getSplits(MongoInputFormat.java:56)	at org.apache.spark.rdd.NewHadoopRDD.getPartitions(NewHadoopRDD.scala:95)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.rdd.MapPartitionsWithPreparationRDD.getPartitions(MapPartitionsWithPreparationRDD.scala:40)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.ShuffleDependency.(Dependency.scala:82)	at org.apache.spark.sql.execution.ShuffledRowRDD.getDependencies(ShuffledRowRDD.scala:59)	at org.apache.spark.rdd.RDD$$anonfun$dependencies$2.apply(RDD.scala:226)	at org.apache.spark.rdd.RDD$$anonfun$dependencies$2.apply(RDD.scala:224)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.dependencies(RDD.scala:224)	at org.apache.spark.scheduler.DAGScheduler.visit$1(DAGScheduler.scala:351)	at org.apache.spark.scheduler.DAGScheduler.getParentStages(DAGScheduler.scala:363)	at org.apache.spark.scheduler.DAGScheduler.getParentStagesAndId(DAGScheduler.scala:266)	at org.apache.spark.scheduler.DAGScheduler.newResultStage(DAGScheduler.scala:300)	at org.apache.spark.scheduler.DAGScheduler.handleJobSubmitted(DAGScheduler.scala:734)	at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1463)	at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1455)	at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1444)	at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)