-
Notifications
You must be signed in to change notification settings - Fork 80
Description
spark.sql("show databases").show
20/11/11 17:24:52 WARN HiveConf: HiveConf of name hive.metastore.event.db.notification.api.auth does not exist
20/11/11 17:24:53 WARN metastore: Failed to connect to the MetaStore Server...
20/11/11 17:24:54 ERROR PolicyRefresher: PolicyRefresher(serviceName=hiverepo): failed to refresh policies. Will continue to use last known version of policies (-1)
com.sun.jersey.api.client.ClientHandlerException: java.lang.RuntimeException: java.lang.NullPointerException
at com.sun.jersey.client.urlconnection.URLConnectionClientHandler.handle(URLConnectionClientHandler.java:155)
at com.sun.jersey.api.client.Client.handle(Client.java:652)
at com.sun.jersey.api.client.WebResource.handle(WebResource.java:682)
at com.sun.jersey.api.client.WebResource.access$200(WebResource.java:74)
at com.sun.jersey.api.client.WebResource$Builder.get(WebResource.java:509)
at org.apache.ranger.admin.client.RangerAdminRESTClient$3.run(RangerAdminRESTClient.java:134)
at org.apache.ranger.admin.client.RangerAdminRESTClient$3.run(RangerAdminRESTClient.java:126)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:360)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1709)
at org.apache.ranger.admin.client.RangerAdminRESTClient.getServicePoliciesIfUpdated(RangerAdminRESTClient.java:137)
at org.apache.ranger.plugin.util.PolicyRefresher.loadPolicyfromPolicyAdmin(PolicyRefresher.java:251)
at org.apache.ranger.plugin.util.PolicyRefresher.loadPolicy(PolicyRefresher.java:191)
at org.apache.ranger.plugin.util.PolicyRefresher.startRefresher(PolicyRefresher.java:137)
at org.apache.ranger.plugin.service.RangerBasePlugin.init(RangerBasePlugin.java:229)
at org.apache.submarine.spark.security.RangerSparkPlugin$.init(RangerSparkPlugin.scala:42)
at org.apache.submarine.spark.security.RangerSparkPlugin$.(RangerSparkPlugin.scala:57)
at org.apache.submarine.spark.security.RangerSparkPlugin$.(RangerSparkPlugin.scala)
at org.apache.submarine.spark.security.RangerSparkAuthorizer$.isAllowed(RangerSparkAuthorizer.scala:143)
at org.apache.spark.sql.execution.SubmarineShowDatabasesCommand$$anonfun$run$1.apply(SubmarineShowDatabasesCommand.scala:32)
at org.apache.spark.sql.execution.SubmarineShowDatabasesCommand$$anonfun$run$1.apply(SubmarineShowDatabasesCommand.scala:32)
at scala.collection.TraversableLike$$anonfun$filterImpl$1.apply(TraversableLike.scala:248)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
at scala.collection.TraversableLike$class.filterImpl(TraversableLike.scala:247)
at scala.collection.TraversableLike$class.filter(TraversableLike.scala:259)
at scala.collection.AbstractTraversable.filter(Traversable.scala:104)
at org.apache.spark.sql.execution.SubmarineShowDatabasesCommand.run(SubmarineShowDatabasesCommand.scala:32)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:194)