java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.Sessio nHiveMetaStoreClient

Stack Overflow | Michael | 6 months ago
tip
Your exception is missing from the Samebug knowledge base.
Here are the best solutions we found on the Internet.
Click on the to mark the helpful solution and get rewards for you help.
  1. 0

    Python pyspark error

    Stack Overflow | 6 months ago | Michael
    java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.Sessio nHiveMetaStoreClient
  2. 0

    Cannot create a Spark DataFrame in Jupyter notebook on a local Windows 7 machine

    Stack Overflow | 5 months ago | avinax
    java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
  3. 0

    Why does Spark report "java.net.URISyntaxException: Relative path in absolute URI" when working with DataFrames?

    Stack Overflow | 6 months ago | Dataminer
    org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Issues trying out example in Spark-shell

    Stack Overflow | 6 months ago | Ayu
    org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.Sess ionHiveMetaStoreClient
  6. 0

    hive installation error : URl

    Stack Overflow | 9 months ago | Punith R Kashi
    java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative path in absolute URI: ${system:java.io.tmpdir%7D/$%7Bsystem:user.name%7D
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. java.net.URISyntaxException

    Relative path in absolute URI: file:c:/Spark/bin/spark-warehouse

    at java.net.URI.checkPath()
  2. Java RT
    URI.<init>
    1. java.net.URI.checkPath(URI.java:1823)
    2. java.net.URI.<init>(URI.java:745)
    2 frames
  3. Hadoop
    Path.<init>
    1. org.apache.hadoop.fs.Path.initialize(Path.java:202)
    2. org.apache.hadoop.fs.Path.<init>(Path.java:171)
    2 frames
  4. Hive Metastore
    HiveMetaStoreClient.<init>
    1. org.apache.hadoop.hive.metastore.Warehouse.getWhRoot(Warehouse.java:159)
    2. org.apache.hadoop.hive.metastore.Warehouse.getDefaultDatabasePath(Warehouse.java:177)
    3. org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB_core(HiveMetaStore.java:600)
    4. org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:620)
    5. org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461)
    6. org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66)
    7. org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72)
    8. org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762)
    9. org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199)
    9 frames
  5. Hive Query Language
    SessionHiveMetaStoreClient.<init>
    1. org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
    1 frame
  6. Java RT
    Constructor.newInstance
    1. sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    2. sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    3. sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    4. java.lang.reflect.Constructor.newInstance(Constructor.java:423)
    4 frames
  7. Hive Metastore
    RetryingMetaStoreClient.getProxy
    1. org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521)
    2. org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
    3. org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132)
    4. org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
    4 frames
  8. Hive Query Language
    SessionState.start
    1. org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005)
    2. org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024)
    3. org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503)
    3 frames
  9. org.apache.spark
    HiveClientImpl.<init>
    1. org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:171)
    1 frame
  10. Java RT
    Constructor.newInstance
    1. sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    2. sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    3. sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    4. java.lang.reflect.Constructor.newInstance(Constructor.java:423)
    4 frames
  11. org.apache.spark
    IsolatedClientLoader.createClient
    1. org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:258)
    1 frame
  12. Spark Project Hive
    HiveSessionState.analyzer
    1. org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:359)
    2. org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:263)
    3. org.apache.spark.sql.hive.HiveSharedState.metadataHive$lzycompute(HiveSharedState.scala:39)
    4. org.apache.spark.sql.hive.HiveSharedState.metadataHive(HiveSharedState.scala:38)
    5. org.apache.spark.sql.hive.HiveSharedState.externalCatalog$lzycompute(HiveSharedState.scala:46)
    6. org.apache.spark.sql.hive.HiveSharedState.externalCatalog(HiveSharedState.scala:45)
    7. org.apache.spark.sql.hive.HiveSessionState.catalog$lzycompute(HiveSessionState.scala:50)
    8. org.apache.spark.sql.hive.HiveSessionState.catalog(HiveSessionState.scala:48)
    9. org.apache.spark.sql.hive.HiveSessionState$$anon$1.<init>(HiveSessionState.scala:63)
    10. org.apache.spark.sql.hive.HiveSessionState.analyzer$lzycompute(HiveSessionState.scala:63)
    11. org.apache.spark.sql.hive.HiveSessionState.analyzer(HiveSessionState.scala:62)
    11 frames
  13. Spark Project SQL
    DataFrameReader.load
    1. org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:49)
    2. org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
    3. org.apache.spark.sql.SparkSession.baseRelationToDataFrame(SparkSession.scala:382)
    4. org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:143)
    5. org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:132)
    5 frames
  14. Java RT
    Method.invoke
    1. sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    2. sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    3. sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    4. java.lang.reflect.Method.invoke(Method.java:498)
    4 frames
  15. Py4J
    GatewayConnection.run
    1. py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:237)
    2. py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
    3. py4j.Gateway.invoke(Gateway.java:280)
    4. py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:128)
    5. py4j.commands.CallCommand.execute(CallCommand.java:79)
    6. py4j.GatewayConnection.run(GatewayConnection.java:211)
    6 frames
  16. Java RT
    Thread.run
    1. java.lang.Thread.run(Thread.java:745)
    1 frame