org.apache.kafka.connect.errors.ConnectException: Error creating writer for log file hdfs://hadoop-master03/tmp/kafka-connect/logs/actions_order/89/log

GitHub | krisskross | 5 months ago
  1. 0

    GitHub comment 88#232030415

    GitHub | 5 months ago | krisskross
    org.apache.kafka.connect.errors.ConnectException: Error creating writer for log file hdfs://hadoop-master03/tmp/kafka-connect/logs/actions_order/89/log
  2. 0

    Where can I find hadoop example jar files

    Stack Overflow | 1 year ago | Gavin Niu
    java.lang.Exception: java.io.FileNotFoundException: Path is not a file: /user/hduser/Text/Text
  3. 0

    Trouble in RecommenderJob on hadoop

    Stack Overflow | 1 year ago | qianda66
    java.io.FileNotFoundException: File does not exist: /user/hduser/temp/preparePreferenceMatrix/numUsers.bin
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    RegionServer failed in logsplitting, wal.HLogSplitter: Got while writing log entry to log

    Google Groups | 2 years ago | sreenivasulu y
    java.io.IOException: cannot get log writer
  6. 0

    Spark cluster computing framework

    gmane.org | 1 year ago
    java.io.FileNotFoundException: File does not exist: /user/marcel/outputs/output_spark/log0

    2 unregistered visitors
    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. java.io.FileNotFoundException

      failed to append to non-existent file /tmp/kafka-connect/logs/actions_order/89/log for client 10.100.1.52

      at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFileInternal()
    2. Apache Hadoop HDFS
      ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod
      1. org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFileInternal(FSNamesystem.java:2672)
      2. org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFileInt(FSNamesystem.java:2991)
      3. org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFile(FSNamesystem.java:2960)
      4. org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.append(NameNodeRpcServer.java:719)
      5. org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.append(NameNodeRpcServer.java:719)
      6. org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.append(ClientNamenodeProtocolServerSideTranslatorPB.java:421)
      7. org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
      7 frames
    3. Hadoop
      Server$Handler$1.run
      1. org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
      2. org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
      3. org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2151)
      4. org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2147)
      4 frames
    4. Java RT
      Subject.doAs
      1. java.security.AccessController.doPrivileged(Native Method)
      2. javax.security.auth.Subject.doAs(Subject.java:422)
      2 frames
    5. Hadoop
      Server$Handler.run
      1. org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
      2. org.apache.hadoop.ipc.Server$Handler.run(Server.java:2145)
      2 frames
    6. Java RT
      Constructor.newInstance
      1. sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
      2. sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
      3. sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
      4. java.lang.reflect.Constructor.newInstance(Constructor.java:423)
      4 frames
    7. Hadoop
      RemoteException.unwrapRemoteException
      1. org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
      2. org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
      2 frames
    8. Apache Hadoop HDFS
      DistributedFileSystem$4.doCall
      1. org.apache.hadoop.hdfs.DFSClient.callAppend(DFSClient.java:1769)
      2. org.apache.hadoop.hdfs.DFSClient.append(DFSClient.java:1803)
      3. org.apache.hadoop.hdfs.DFSClient.append(DFSClient.java:1796)
      4. org.apache.hadoop.hdfs.DistributedFileSystem$4.doCall(DistributedFileSystem.java:323)
      5. org.apache.hadoop.hdfs.DistributedFileSystem$4.doCall(DistributedFileSystem.java:319)
      5 frames
    9. Hadoop
      FileSystemLinkResolver.resolve
      1. org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
      1 frame
    10. Apache Hadoop HDFS
      DistributedFileSystem.append
      1. org.apache.hadoop.hdfs.DistributedFileSystem.append(DistributedFileSystem.java:319)
      1 frame
    11. Hadoop
      FileSystem.append
      1. org.apache.hadoop.fs.FileSystem.append(FileSystem.java:1173)
      1 frame
    12. io.confluent.connect
      HdfsSinkTask.open
      1. io.confluent.connect.hdfs.wal.WALFile$Writer.<init>(WALFile.java:221)
      2. io.confluent.connect.hdfs.wal.WALFile.createWriter(WALFile.java:67)
      3. io.confluent.connect.hdfs.wal.FSWAL.acquireLease(FSWAL.java:73)
      4. io.confluent.connect.hdfs.wal.FSWAL.apply(FSWAL.java:105)
      5. io.confluent.connect.hdfs.TopicPartitionWriter.applyWAL(TopicPartitionWriter.java:441)
      6. io.confluent.connect.hdfs.TopicPartitionWriter.recover(TopicPartitionWriter.java:197)
      7. io.confluent.connect.hdfs.DataWriter.recover(DataWriter.java:239)
      8. io.confluent.connect.hdfs.DataWriter.open(DataWriter.java:281)
      9. io.confluent.connect.hdfs.HdfsSinkTask.open(HdfsSinkTask.java:104)
      9 frames
    13. org.apache.kafka
      ConsumerNetworkClient$RequestFutureCompletionHandler.onComplete
      1. org.apache.kafka.connect.runtime.WorkerSinkTask.openPartitions(WorkerSinkTask.java:417)
      2. org.apache.kafka.connect.runtime.WorkerSinkTask.access$1000(WorkerSinkTask.java:54)
      3. org.apache.kafka.connect.runtime.WorkerSinkTask$HandleRebalance.onPartitionsAssigned(WorkerSinkTask.java:453)
      4. org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.onJoinComplete(ConsumerCoordinator.java:222)
      5. org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess(AbstractCoordinator.java:232)
      6. org.apache.kafka.clients.consumer.internals.AbstractCoordinator$1.onSuccess(AbstractCoordinator.java:227)
      7. org.apache.kafka.clients.consumer.internals.RequestFuture.fireSuccess(RequestFuture.java:133)
      8. org.apache.kafka.clients.consumer.internals.RequestFuture.complete(RequestFuture.java:107)
      9. org.apache.kafka.clients.consumer.internals.RequestFuture$2.onSuccess(RequestFuture.java:182)
      10. org.apache.kafka.clients.consumer.internals.RequestFuture.fireSuccess(RequestFuture.java:133)
      11. org.apache.kafka.clients.consumer.internals.RequestFuture.complete(RequestFuture.java:107)
      12. org.apache.kafka.clients.consumer.internals.AbstractCoordinator$SyncGroupResponseHandler.handle(AbstractCoordinator.java:436)
      13. org.apache.kafka.clients.consumer.internals.AbstractCoordinator$SyncGroupResponseHandler.handle(AbstractCoordinator.java:422)
      14. org.apache.kafka.clients.consumer.internals.AbstractCoordinator$CoordinatorResponseHandler.onSuccess(AbstractCoordinator.java:679)
      15. org.apache.kafka.clients.consumer.internals.AbstractCoordinator$CoordinatorResponseHandler.onSuccess(AbstractCoordinator.java:658)
      16. org.apache.kafka.clients.consumer.internals.RequestFuture$1.onSuccess(RequestFuture.java:167)
      17. org.apache.kafka.clients.consumer.internals.RequestFuture.fireSuccess(RequestFuture.java:133)
      18. org.apache.kafka.clients.consumer.internals.RequestFuture.complete(RequestFuture.java:107)
      19. org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient$RequestFutureCompletionHandler.onComplete(ConsumerNetworkClient.java:426)
      19 frames
    14. Apache Kafka
      NetworkClient.poll
      1. org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:278)
      1 frame
    15. org.apache.kafka
      ConsumerCoordinator.ensurePartitionAssignment
      1. org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.clientPoll(ConsumerNetworkClient.java:360)
      2. org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:224)
      3. org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:192)
      4. org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:163)
      5. org.apache.kafka.clients.consumer.internals.AbstractCoordinator.ensureActiveGroup(AbstractCoordinator.java:243)
      6. org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.ensurePartitionAssignment(ConsumerCoordinator.java:345)
      6 frames
    16. Apache Kafka
      KafkaConsumer.poll
      1. org.apache.kafka.clients.consumer.KafkaConsumer.pollOnce(KafkaConsumer.java:977)
      2. org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:937)
      2 frames
    17. org.apache.kafka
      WorkerTask.run
      1. org.apache.kafka.connect.runtime.WorkerSinkTask.pollConsumer(WorkerSinkTask.java:305)
      2. org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:222)
      3. org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:170)
      4. org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:142)
      5. org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:140)
      6. org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:175)
      6 frames
    18. Java RT
      Thread.run
      1. java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
      2. java.util.concurrent.FutureTask.run(FutureTask.java:266)
      3. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
      4. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
      5. java.lang.Thread.run(Thread.java:745)
      5 frames