Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via Stack Overflow by Idan Fischman
, 1 year ago
This exception has no message.
via Stack Overflow by Salil Kulkarni
, 11 months ago
This exception has no message.
via Stack Overflow by mgaido
, 1 year ago
java.lang.NullPointerException: 	at org.apache.spark.sql.hive.client.ClientWrapper.conf(ClientWrapper.scala:205)	at org.apache.spark.sql.hive.HiveContext.hiveconf$lzycompute(HiveContext.scala:554)	at org.apache.spark.sql.hive.HiveContext.hiveconf(HiveContext.scala:553)	at org.apache.spark.sql.hive.HiveContext$$anonfun$configure$1.apply(HiveContext.scala:540)	at org.apache.spark.sql.hive.HiveContext$$anonfun$configure$1.apply(HiveContext.scala:539)	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)	at scala.collection.immutable.List.foreach(List.scala:318)	at scala.collection.TraversableLike$class.map(TraversableLike.scala:244)	at scala.collection.AbstractTraversable.map(Traversable.scala:105)	at org.apache.spark.sql.hive.HiveContext.configure(HiveContext.scala:539)	at org.apache.spark.sql.hive.HiveContext.metadataHive$lzycompute(HiveContext.scala:252)	at org.apache.spark.sql.hive.HiveContext.metadataHive(HiveContext.scala:239)	at org.apache.spark.sql.hive.HiveContext$$anon$2.(HiveContext.scala:459)	at org.apache.spark.sql.hive.HiveContext.catalog$lzycompute(HiveContext.scala:459)	at org.apache.spark.sql.hive.HiveContext.catalog(HiveContext.scala:458)	at org.apache.spark.sql.hive.HiveContext$$anon$3.(HiveContext.scala:475)	at org.apache.spark.sql.hive.HiveContext.analyzer$lzycompute(HiveContext.scala:475)	at org.apache.spark.sql.hive.HiveContext.analyzer(HiveContext.scala:474)	at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:34)	at org.apache.spark.sql.DataFrame.(DataFrame.scala:133)	at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:52)	at org.apache.spark.sql.SQLContext.createDataFrame(SQLContext.scala:417)	at org.apache.spark.sql.SQLImplicits.rddToDataFrameHolder(SQLImplicits.scala:155)	at $line46.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1.apply(:58)	at $line46.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1.apply(:48)	at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661)	at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661)	at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:50)	at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50)	at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50)	at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:426)	at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:49)	at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49)	at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49)	at scala.util.Try$.apply(Try.scala:161)	at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39)	at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:224)	at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224)	at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224)	at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)	at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:223)	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)	at java.lang.Thread.run(Thread.java:745)