javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)]

Stack Overflow | Ian | 10 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    Spark/Phoenix with Kerberos on YARN

    Stack Overflow | 10 months ago | Ian
    javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)]
  2. 0

    Spark Streaming and Phoenix Kerberos issue

    Stack Overflow | 5 months ago | nilesh1212
    javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)]
  3. 0

    Can not connect to secure hbase cluster via apache phoenix sqlline on OSX

    Stack Overflow | 2 years ago | Roman Rogozhnikov
    javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)]
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Apache NiFi Developer List - Using NiFI with a Secured HBASE setup

    nabble.com | 12 months ago
    javax.security.sasl.SaslException: GSS initiate failed
  6. 0

    Accessing HBase through Spark with Security enabled

    Stack Overflow | 9 months ago | aneela saleem
    javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)]

    3 unregistered visitors
    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. javax.security.sasl.SaslException

      GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)]

      at com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge()
    2. Java RT
      GssKrb5Client.evaluateChallenge
      1. com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:211)
      1 frame
    3. HBase
      HBaseSaslRpcClient.saslConnect
      1. org.apache.hadoop.hbase.security.HBaseSaslRpcClient.saslConnect(HBaseSaslRpcClient.java:179)
      1 frame
    4. HBase - Client
      RpcClientImpl$Connection$2.run
      1. org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupSaslConnection(RpcClientImpl.java:611)
      2. org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.access$600(RpcClientImpl.java:156)
      3. org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection$2.run(RpcClientImpl.java:737)
      4. org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection$2.run(RpcClientImpl.java:734)
      4 frames
    5. Java RT
      Subject.doAs
      1. java.security.AccessController.doPrivileged(Native Method)
      2. javax.security.auth.Subject.doAs(Subject.java:422)
      2 frames
    6. Hadoop
      UserGroupInformation.doAs
      1. org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
      1 frame
    7. HBase - Client
      AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod
      1. org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:734)
      2. org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
      3. org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
      4. org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
      5. org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
      6. org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
      6 frames
    8. HBase
      MasterProtos$MasterService$BlockingStub.isMasterRunning
      1. org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:50918)
      1 frame
    9. HBase - Client
      HBaseAdmin.getTableDescriptor
      1. org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.isMasterRunning(ConnectionManager.java:1564)
      2. org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1502)
      3. org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1524)
      4. org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1553)
      5. org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1704)
      6. org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
      7. org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
      8. org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:3917)
      9. org.apache.hadoop.hbase.client.HBaseAdmin.getTableDescriptor(HBaseAdmin.java:441)
      10. org.apache.hadoop.hbase.client.HBaseAdmin.getTableDescriptor(HBaseAdmin.java:463)
      10 frames
    10. Phoenix Core
      PhoenixDriver.connect
      1. org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:815)
      2. org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1215)
      3. org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112)
      4. org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1902)
      5. org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:744)
      6. org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186)
      7. org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:304)
      8. org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:296)
      9. org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
      10. org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:294)
      11. org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1243)
      12. org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1893)
      13. org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1862)
      14. org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77)
      15. org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1862)
      16. org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180)
      17. org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132)
      18. org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151)
      18 frames
    11. Java RT
      DriverManager.getConnection
      1. java.sql.DriverManager.getConnection(DriverManager.java:664)
      2. java.sql.DriverManager.getConnection(DriverManager.java:208)
      2 frames
    12. org.apache.phoenix
      SparkSqlContextFunctions.phoenixTableAsDataFrame
      1. org.apache.phoenix.mapreduce.util.ConnectionUtil.getConnection(ConnectionUtil.java:99)
      2. org.apache.phoenix.mapreduce.util.ConnectionUtil.getInputConnection(ConnectionUtil.java:57)
      3. org.apache.phoenix.mapreduce.util.ConnectionUtil.getInputConnection(ConnectionUtil.java:45)
      4. org.apache.phoenix.mapreduce.util.PhoenixConfigurationUtil.getSelectColumnMetadataList(PhoenixConfigurationUtil.java:263)
      5. org.apache.phoenix.spark.PhoenixRDD.toDataFrame(PhoenixRDD.scala:109)
      6. org.apache.phoenix.spark.SparkSqlContextFunctions.phoenixTableAsDataFrame(SparkSqlContextFunctions.scala:37)
      6 frames
    13. com.bosch.asc
      SMTProcessMonitor$delayedInit$body.apply
      1. com.bosch.asc.utils.HBaseUtils$.scanPhoenix(HBaseUtils.scala:123)
      2. com.bosch.asc.SMTProcess.addLookup(SMTProcess.scala:1125)
      3. com.bosch.asc.SMTProcess.saveMountTraceLogToPhoenix(SMTProcess.scala:1039)
      4. com.bosch.asc.SMTProcess.runETL(SMTProcess.scala:87)
      5. com.bosch.asc.SMTProcessMonitor$delayedInit$body.apply(SMTProcessMonitor.scala:20)
      5 frames
    14. Scala
      App$class.main
      1. scala.Function0$class.apply$mcV$sp(Function0.scala:40)
      2. scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12)
      3. scala.App$$anonfun$main$1.apply(App.scala:71)
      4. scala.App$$anonfun$main$1.apply(App.scala:71)
      5. scala.collection.immutable.List.foreach(List.scala:318)
      6. scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:32)
      7. scala.App$class.main(App.scala:71)
      7 frames
    15. com.bosch.asc
      SMTProcessMonitor.main
      1. com.bosch.asc.SMTProcessMonitor$.main(SMTProcessMonitor.scala:5)
      2. com.bosch.asc.SMTProcessMonitor.main(SMTProcessMonitor.scala)
      2 frames
    16. Java RT
      Method.invoke
      1. sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
      2. sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
      3. sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
      4. java.lang.reflect.Method.invoke(Method.java:498)
      4 frames
    17. Spark Project YARN Stable API
      ApplicationMaster$$anon$2.run
      1. org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:486)
      1 frame