Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via qiita.com by Unknown author, 1 year ago
Lost task 0.0 in stage 0.0 (TID 0, spark003.example.com): java.lang.IllegalArgumentException: java.net.UnknownHostException: nameservice1
java.net.UnknownHostException: nameservice1	at org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:374)	at org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:312)	at org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:178)	at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:665)	at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:601)	at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:148)	at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2596)	at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:91)	at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2630)	at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2612)	at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:370)	at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:169)	at org.apache.hadoop.mapred.JobConf.getWorkingDirectory(JobConf.java:656)	at org.apache.hadoop.mapred.FileInputFormat.setInputPaths(FileInputFormat.java:436)	at org.apache.hadoop.mapred.FileInputFormat.setInputPaths(FileInputFormat.java:409)	at org.apache.spark.SparkContext$$anonfun$hadoopFile$1$$anonfun$32.apply(SparkContext.scala:1016)	at org.apache.spark.SparkContext$$anonfun$hadoopFile$1$$anonfun$32.apply(SparkContext.scala:1016)	at org.apache.spark.rdd.HadoopRDD$$anonfun$getJobConf$6.apply(HadoopRDD.scala:176)	at org.apache.spark.rdd.HadoopRDD$$anonfun$getJobConf$6.apply(HadoopRDD.scala:176)	at scala.Option.map(Option.scala:145)	at org.apache.spark.rdd.HadoopRDD.getJobConf(HadoopRDD.scala:176)	at org.apache.spark.rdd.HadoopRDD$$anon$1.(HadoopRDD.scala:220)	at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:216)	at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:297)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:297)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:297)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:297)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)	at org.apache.spark.scheduler.Task.run(Task.scala:88)	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)	at java.lang.Thread.run(Thread.java:745)