org.apache.phoenix.exception.PhoenixIOException: Failed after attempts=35, exceptions: Wed Aug 03 18:56:17 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.net.ConnectTimeoutException: 10000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000] Wed Aug 03 18:56:17 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000 Wed Aug 03 18:56:18 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000 Wed Aug 03 18:56:29 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.net.ConnectTimeoutException: 10000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000] Wed Aug 03 18:56:31 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000 Wed Aug 03 19:10:37 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.net.ConnectTimeoutException: 10000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000] at org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:108)

Stack Overflow | ravinder | 4 months ago
  1. 0

    Remote JDBC Connection to Phoenix fails

    Stack Overflow | 4 months ago | ravinder
    org.apache.phoenix.exception.PhoenixIOException: Failed after attempts=35, exceptions: Wed Aug 03 18:56:17 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.net.ConnectTimeoutException: 10000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000] Wed Aug 03 18:56:17 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000 Wed Aug 03 18:56:18 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000 Wed Aug 03 18:56:29 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.net.ConnectTimeoutException: 10000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000] Wed Aug 03 18:56:31 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000 Wed Aug 03 19:10:37 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.net.ConnectTimeoutException: 10000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000] at org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:108)
  2. 0

    [jira] [Updated] (PHOENIX-2302) Can't connect to secure hbase cluster on MacOS

    phoenix-dev | 1 year ago | Roman Rogozhnikov (JIRA)
    util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable Debug is true storeKey false useTicketCache false useKeyTab true doNotPrompt false ticketCache is null isInitiator true KeyTab is service.keytab refreshKrb5Config is false principal is user/hmaster@EXAMPLE.COM tryFirstPass is false useFirstPass is false storePass is false clearPass is false principal is user/hmaster@EXAMPLE.COM Will use keytab Commit Succeeded ....
  3. 0

    HBase standalone connection refused error

    Google Groups | 2 years ago | Erdinç Bereketoğlu
    org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=1, exceptions: Mon Apr 27 21:04:14 UTC 2015, RpcRetryingCaller{globalStartTime=1430168636637, pause=100, retries=1}, org.apache.hadoop.hbase.MasterNotRunningException: ZooKeeper: KeeperErrorCode = ConnectionLoss for /hbase
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Bulk loading HFiles via LoadIncrementalHFiles fails at a region that is being compacted, a bug?

    Google Groups | 3 years ago | Stanislav Barton
    org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=10, exceptions: Wed Jun 26 18:18:35 CEST 2013, org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles$3@650e938c, org.apache.hadoop.hbase.RegionTooBusyException: in 60000ms
  6. 0

    Can not connect local java client to a remote Hbase

    Google Groups | 8 months ago | SOUFIANI Mustapha | السفياني مصطفى
    org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after > > > attempts=36, exceptions: > > > Wed Apr 20 10:32:43 WEST 2016, null, java.net.SocketTimeoutException: > > > callTimeout=60000, callDuration=75181: row 'pentaho_mappings,,' on > table > > > 'hbase:meta' at region=hbase:meta,,1.1588230740, > > > hostname=localhost,16020,1461071963695, seqNum=0 > > > at > > > com.pentaho.big.data.bundles.impl.shim.hbase.table.HBaseTableImpl.exists(HBaseTableImpl.java:71) at > > > org.pentaho.big.data.kettle.plugins.hbase.mapping.MappingAdmin.getMappedTables(MappingAdmin.java:502) at > > > org.pentaho.big.data.kettle.plugins.hbase.output.HBaseOutputDialog.setupMappedTableNames(HBaseOutputDialog.java:818) at > > > org.pentaho.big.data.kettle.plugins.hbase.output.HBaseOutputDialog.access$900(HBaseOutputDialog.java:88) at > > > org.pentaho.big.data.kettle.plugins.hbase.output.HBaseOutputDialog$7.widgetSelected(HBaseOutputDialog.java:398) at org.eclipse.swt.widgets.TypedListener.handleEvent(Unknown > Source)

    Not finding the right solution?
    Take a tour to get the most out of Samebug.

    Tired of useless tips?

    Automated exception search integrated into your IDE

    Root Cause Analysis

    1. org.apache.hadoop.hbase.client.RetriesExhaustedException

      Failed after attempts=35, exceptions: Wed Aug 03 18:56:17 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.net.ConnectTimeoutException: 10000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000] Wed Aug 03 18:56:17 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000 Wed Aug 03 18:56:18 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000 Wed Aug 03 18:56:29 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.net.ConnectTimeoutException: 10000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000] Wed Aug 03 18:56:31 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000 Wed Aug 03 18:56:43 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.net.ConnectTimeoutException: 10000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000] Wed Aug 03 19:10:37 IST 2016, RpcRetryingCaller{globalStartTime=1470230766277, pause=100, retries=35}, org.apache.hadoop.hbase.MasterNotRunningException: com.google.protobuf.ServiceException: org.apache.hadoop.net.ConnectTimeoutException: 10000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=ip-172-31-15-225.us-west-2.compute.internal/52.32.173.25:16000] at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)

      at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable()
    2. HBase - Client
      HBaseAdmin.getTableDescriptor
      1. org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:3917)
      2. org.apache.hadoop.hbase.client.HBaseAdmin.getTableDescriptor(HBaseAdmin.java:441)
      3. org.apache.hadoop.hbase.client.HBaseAdmin.getTableDescriptor(HBaseAdmin.java:463)
      3 frames
    3. Phoenix Core
      PhoenixDriver.connect
      1. org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:813)
      2. org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1213)
      3. org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112)
      4. org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1902)
      5. org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:744)
      6. org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186)
      7. org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:303)
      8. org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:295)
      9. org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
      10. org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:293)
      11. org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1236)
      12. org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1891)
      13. org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1860)
      14. org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77)
      15. org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1860)
      16. org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:162)
      17. org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:131)
      18. org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:133)
      18 frames
    4. Java RT
      DriverManager.getConnection
      1. java.sql.DriverManager.getConnection(DriverManager.java:571)
      2. java.sql.DriverManager.getConnection(DriverManager.java:233)
      2 frames
    5. com.lendingpoint.hadoop
      PhoenixTest.main
      1. com.lendingpoint.hadoop.phoenixconnect.PhoenixTest.main(PhoenixTest.java:28)
      1 frame