Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via GitHub by aitzkowi
, 1 year ago
Unable to construct journal, qjournal://journalnode1.service.consul:8485;journalnode2.service.consul:8485;journalnode3.service.consul:8485/hdfs
via iyunv.com by Unknown author, 2 years ago
Unable to construct journal, qjournal://spark-1421-0000:8485;spark-1421-0003:8485;spark-1421-0004:8485;spark-1421-0005:8485;spark-1421-0006:8485/hadoop-journal
via grokbase.com by Unknown author, 2 years ago
Unable to construct journal, bookkeeper://10.18.40.155:2183;10.18.40.155:2182;10.18.40.155:2181/NN
via Apache's JIRA Issue Tracker by Colin Patrick McCabe, 2 years ago
Unable to construct journal, qjournal://hadoop-mm:8485;hadoop-nn-0:8485;hadoop-nn-1:8485/hadoop
via grokbase.com by Unknown author, 2 years ago
Unable to construct journal, qjournal://cdh4master01:8485;cdh4master02:8485;cdh4worker03:8485/hdfscluster
java.lang.NullPointerException: 	at org.apache.hadoop.hdfs.qjournal.client.IPCLoggerChannelMetrics.getName(IPCLoggerChannelMetrics.java:107)	at org.apache.hadoop.hdfs.qjournal.client.IPCLoggerChannelMetrics.create(IPCLoggerChannelMetrics.java:91)	at org.apache.hadoop.hdfs.qjournal.client.IPCLoggerChannel.(IPCLoggerChannel.java:178)	at org.apache.hadoop.hdfs.qjournal.client.IPCLoggerChannel$1.createLogger(IPCLoggerChannel.java:156)	at org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager.createLoggers(QuorumJournalManager.java:367)	at org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager.createLoggers(QuorumJournalManager.java:149)	at org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager.(QuorumJournalManager.java:116)	at org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager.(QuorumJournalManager.java:105)	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)	at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)	at java.lang.reflect.Constructor.newInstance(Unknown Source)	at org.apache.hadoop.hdfs.server.namenode.FSEditLog.createJournal(FSEditLog.java:1580)	at org.apache.hadoop.hdfs.server.namenode.FSEditLog.initJournals(FSEditLog.java:275)	at org.apache.hadoop.hdfs.server.namenode.FSEditLog.initSharedJournalsForRead(FSEditLog.java:253)	at org.apache.hadoop.hdfs.server.namenode.FSImage.initEditLog(FSImage.java:772)	at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:620)	at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:281)	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1005)	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:735)	at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:531)	at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:587)	at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:754)	at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:738)	at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1427)	at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1493)