Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via GitHub by martinstuder
, 1 year ago
java.net.UnknownHostException: namenode1.hdfs.mesos
via Stack Overflow by ktaube
, 1 year ago
via GitHub by yogeshnath
, 1 year ago
java.net.UnknownHostException: namenode1.hdfs.mesos
via Stack Overflow by Anup Ash
, 1 year ago
via GitHub by anupash147
, 1 year ago
java.net.UnknownHostException: nameservice1
via GitHub by radek1st
, 1 year ago
java.net.UnknownHostException: namenode1.hdfs.mesos
java.lang.IllegalArgumentException: java.net.UnknownHostException: namenode1.hdfs.mesos	at org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:377)	at org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:240)	at org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.getProxy(ConfiguredFailoverProxyProvider.java:124)	at org.apache.hadoop.io.retry.RetryInvocationHandler.(RetryInvocationHandler.java:74)	at org.apache.hadoop.io.retry.RetryInvocationHandler.(RetryInvocationHandler.java:65)	at org.apache.hadoop.io.retry.RetryProxy.create(RetryProxy.java:58)	at org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:152)	at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:579)	at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:524)	at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:146)	at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2397)	at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:89)	at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2431)	at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2413)	at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:368)	at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:167)	at org.apache.hadoop.mapred.JobConf.getWorkingDirectory(JobConf.java:653)	at org.apache.hadoop.mapred.FileInputFormat.setInputPaths(FileInputFormat.java:427)	at org.apache.hadoop.mapred.FileInputFormat.setInputPaths(FileInputFormat.java:400)	at org.apache.spark.SparkContext$$anonfun$hadoopFile$1$$anonfun$33.apply(SparkContext.scala:1015)	at org.apache.spark.SparkContext$$anonfun$hadoopFile$1$$anonfun$33.apply(SparkContext.scala:1015)	at org.apache.spark.rdd.HadoopRDD$$anonfun$getJobConf$6.apply(HadoopRDD.scala:176)	at org.apache.spark.rdd.HadoopRDD$$anonfun$getJobConf$6.apply(HadoopRDD.scala:176)	at scala.Option.map(Option.scala:145)	at org.apache.spark.rdd.HadoopRDD.getJobConf(HadoopRDD.scala:176)	at org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:195)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239)	at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237)	at scala.Option.getOrElse(Option.scala:120)	at org.apache.spark.rdd.RDD.partitions(RDD.scala:237)	at org.apache.spark.rdd.RDD$$anonfun$take$1.apply(RDD.scala:1307)	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)	at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)	at org.apache.spark.rdd.RDD.take(RDD.scala:1302)	at org.apache.spark.rdd.RDD$$anonfun$first$1.apply(RDD.scala:1342)	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)	at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)	at org.apache.spark.rdd.RDD.first(RDD.scala:1341)	at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:36)	at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:41)	at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:43)	at $iwC$$iwC$$iwC$$iwC$$iwC.(:45)	at $iwC$$iwC$$iwC$$iwC.(:47)	at $iwC$$iwC$$iwC.(:49)	at $iwC$$iwC.(:51)	at $iwC.(:53)