java.lang.RuntimeException: Failure loading MapRClient.


Solutions on the web

Solution icon of web
via mapr.com by Unknown author, 1 year ago
Failure loading MapRClient.

Solution icon of stackoverflow
Failure loading MapRClient.

Solution icon of cask
via Cask Community Issue Tracker by Derek Wood, 1 year ago
Failure loading MapRClient.

Stack trace

java.lang.RuntimeException: Failure loading MapRClient.
	at com.mapr.fs.ShimLoader.injectNativeLoader(ShimLoader.java:305)
	at com.mapr.fs.ShimLoader.load(ShimLoader.java:223)
	at com.mapr.fs.MapRFileSystem.<clinit>(MapRFileSystem.java:107)
	at java.lang.Class.forName0(Native Method)
	at java.lang.Class.forName(Class.java:348)
	at org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.java:2147)
	at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:2112)
	at org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2206)
	at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2674)
	at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2687)
	at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94)
	at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2723)
	at org.apache.hadoop.fs.FileSystem$Cache.getUnique(FileSystem.java:2711)
	at org.apache.hadoop.fs.FileSystem.newInstance(FileSystem.java:454)
	at org.apache.hadoop.fs.FileSystem.newInstance(FileSystem.java:462)
	at org.apache.hadoop.fs.FileSystem.newInstance(FileSystem.java:444)
	at org.apache.hadoop.hive.shims.Hadoop23Shims.getNonCachedFileSystem(Hadoop23Shims.java:944)
	at org.apache.hadoop.hive.ql.exec.Utilities.createDirsWithPermission(Utilities.java:3687)
	at org.apache.hadoop.hive.ql.session.SessionState.createRootHDFSDir(SessionState.java:600)
	at org.apache.hadoop.hive.ql.session.SessionState.createSessionDirs(SessionState.java:554)
	at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:508)
	at org.apache.spark.sql.hive.client.HiveClientImpl.<init>(HiveClientImpl.scala:189)
	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
	at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
	at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:258)
	at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:359)
	at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:263)
	at org.apache.spark.sql.hive.HiveSharedState.metadataHive$lzycompute(HiveSharedState.scala:39)
	at org.apache.spark.sql.hive.HiveSharedState.metadataHive(HiveSharedState.scala:38)
	at org.apache.spark.sql.hive.HiveSharedState.externalCatalog$lzycompute(HiveSharedState.scala:46)
	at org.apache.spark.sql.hive.HiveSharedState.externalCatalog(HiveSharedState.scala:45)
	at org.apache.spark.sql.hive.HiveSessionState.catalog$lzycompute(HiveSessionState.scala:50)
	at org.apache.spark.sql.hive.HiveSessionState.catalog(HiveSessionState.scala:48)
	at org.apache.spark.sql.hive.HiveSessionState$$anon$1.<init>(HiveSessionState.scala:63)
	at org.apache.spark.sql.hive.HiveSessionState.analyzer$lzycompute(HiveSessionState.scala:63)
	at org.apache.spark.sql.hive.HiveSessionState.analyzer(HiveSessionState.scala:62)
	at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:49)
	at org.apache.spark.sql.Dataset.<init>(Dataset.scala:161)
	at org.apache.spark.sql.Dataset.<init>(Dataset.scala:167)
	at org.apache.spark.sql.Dataset$.apply(Dataset.scala:59)
	at org.apache.spark.sql.SparkSession.createDataset(SparkSession.scala:441)
	at org.apache.spark.sql.SQLContext.createDataset(SQLContext.scala:395)
	at org.apache.spark.sql.SQLImplicits.rddToDatasetHolder(SQLImplicits.scala:163)

Write tip

You have a different solution? A short tip here would help you and many other users who saw this issue last week.

Users with the same issue

You are the first who have seen this exception. Write a tip to help other users and build your expert profile.