spark 执行 Hbase的 hive 报错

具体代码段如下:执行了.show()命令
---------------------------------------------------------------------------
Py4JJavaError                             Traceback (most recent call last)
<ipython-input-10-9f366d069c5f> in <module>
----> 1 user_profile_hbase1.show()

/home/hadoop/spark/python/pyspark/sql/dataframe.py in show(self, n, truncate, vertical)
    438         """
    439         if isinstance(truncate, bool) and truncate:
--> 440             print(self._jdf.showString(n, 20, vertical))
    441         else:
    442             print(self._jdf.showString(n, int(truncate), vertical))

/home/hadoop/spark/python/lib/py4j-0.10.9-src.zip/py4j/java_gateway.py in __call__(self, *args)
   1303         answer = self.gateway_client.send_command(command)
   1304         return_value = get_return_value(
-> 1305             answer, self.gateway_client, self.target_id, self.name)
   1306 
   1307         for temp_arg in temp_args:

/home/hadoop/spark/python/pyspark/sql/utils.py in deco(*a, **kw)
    126     def deco(*a, **kw):
    127         try:
--> 128             return f(*a, **kw)
    129         except py4j.protocol.Py4JJavaError as e:
    130             converted = convert_exception(e.java_exception)

/home/hadoop/spark/python/lib/py4j-0.10.9-src.zip/py4j/protocol.py in get_return_value(answer, gateway_client, target_id, name)
    326                 raise Py4JJavaError(
    327                     "An error occurred while calling {0}{1}{2}.\n".
--> 328                     format(target_id, ".", name), value)
    329             else:
    330                 raise Py4JError(

Py4JJavaError: An error occurred while calling o69.showString.
: java.io.IOException: Cannot create a record reader because of a previous error. Please look at the previous logs lines from the task's full log for more details.
	at org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getSplits(TableInputFormatBase.java:252)
	at org.apache.spark.rdd.NewHadoopRDD.getPartitions(NewHadoopRDD.scala:131)
	at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
	at scala.Option.getOrElse(Option.scala:189)
	at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
	at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
	at scala.Option.getOrElse(Option.scala:189)
	at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
	at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
	at scala.Option.getOrElse(Option.scala:189)
	at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
	at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
	at scala.Option.getOrElse(Option.scala:189)
	at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
	at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
	at scala.Option.getOrElse(Option.scala:189)
	at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
	at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
	at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
	at scala.Option.getOrElse(Option.scala:189)
	at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
	at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:442)
	at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:425)
	at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:47)
	at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:3627)
	at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:2697)
	at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3618)
	at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100)
	at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
	at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87)
	at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:764)
	at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
	at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3616)
	at org.apache.spark.sql.Dataset.head(Dataset.scala:2697)
	at org.apache.spark.sql.Dataset.take(Dataset.scala:2904)
	at org.apache.spark.sql.Dataset.getRows(Dataset.scala:300)
	at org.apache.spark.sql.Dataset.showString(Dataset.scala:337)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
	at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
	at py4j.Gateway.invoke(Gateway.java:282)
	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
	at py4j.commands.CallCommand.execute(CallCommand.java:79)
	at py4j.GatewayConnection.run(GatewayConnection.java:238)
	at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalStateException: The input format instance has not been properly initialized. Ensure you call initializeTable either in your constructor or initialize method
	at org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getTable(TableInputFormatBase.java:570)
	at org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getSplits(TableInputFormatBase.java:247)
	... 51 more

相应的包hbase包,hive包都移动到 $SPARK_HOME/jars,具体如下:

-rw-r--r-- 1 root root   183472 Mar 20 09:33 hive-beeline-2.3.7.jar
-rw-r--r-- 1 root root    43387 Mar 20 09:33 hive-cli-2.3.7.jar
-rw-r--r-- 1 root root   436980 Mar 20 09:33 hive-common-2.3.7.jar
-rw-r--r-- 1 root root 10839104 Mar 20 09:33 hive-exec-2.3.7-core.jar
-rw-r--r-- 1 root root   118060 Mar 20 09:33 hive-hbase-handler-3.1.2.jar
-rw-r--r-- 1 root root   268933 Mar 20 09:33 hive-hcatalog-core-3.1.2.jar
-rw-r--r-- 1 root root    55459 Mar 20 09:33 hive-hcatalog-pig-adapter-3.1.2.jar
-rw-r--r-- 1 root root    76615 Mar 20 09:33 hive-hcatalog-server-extensions-3.1.2.jar
-rw-r--r-- 1 root root   133382 Mar 20 09:33 hive-hcatalog-streaming-3.1.2.jar
-rw-r--r-- 1 root root   116311 Mar 20 09:33 hive-jdbc-2.3.7.jar
-rw-r--r-- 1 root root   326419 Mar 20 09:33 hive-llap-common-2.3.7.jar
-rw-r--r-- 1 root root  8194428 Mar 20 09:33 hive-metastore-2.3.7.jar
-rw-r--r-- 1 root root   916206 Mar 20 09:33 hive-serde-2.3.7.jar
-rw-r--r-- 1 root root    54116 Mar 20 09:33 hive-shims-0.23-2.3.7.jar
-rw-r--r-- 1 root root     8785 Mar 20 09:33 hive-shims-2.3.7.jar
-rw-r--r-- 1 root root   120305 Mar 20 09:33 hive-shims-common-2.3.7.jar
-rw-r--r-- 1 root root    12984 Mar 20 09:33 hive-shims-scheduler-2.3.7.jar
-rw-r--r-- 1 root root   234942 Mar 20 09:33 hive-storage-api-2.7.1.jar
-rw-r--r-- 1 root root    38352 Mar 20 09:33 hive-vector-code-gen-2.3.7.jar

 

-rw-r--r-- 1 root root    6666 Mar 20 10:17 hbase-annotations-2.3.4.jar
-rw-r--r-- 1 root root   14128 Mar 20 10:17 hbase-annotations-2.3.4-tests.jar
-rw-r--r-- 1 root root   90016 Mar 20 10:17 hbase-asyncfs-2.3.4.jar
-rw-r--r-- 1 root root   35593 Mar 20 10:17 hbase-asyncfs-2.3.4-tests.jar
-rw-r--r-- 1 root root 1932103 Mar 20 10:17 hbase-client-2.3.4.jar
-rw-r--r-- 1 root root  808377 Mar 20 10:17 hbase-common-2.3.4.jar
-rw-r--r-- 1 root root  353119 Mar 20 10:17 hbase-common-2.3.4-tests.jar
-rw-r--r-- 1 root root  471046 Mar 20 10:17 hbase-endpoint-2.3.4.jar
-rw-r--r-- 1 root root  187054 Mar 20 10:17 hbase-examples-2.3.4.jar
-rw-r--r-- 1 root root   20206 Mar 20 10:17 hbase-external-blockcache-2.3.4.jar
-rw-r--r-- 1 root root  165895 Mar 20 10:17 hbase-hadoop2-compat-2.3.4.jar
-rw-r--r-- 1 root root   40671 Mar 20 10:17 hbase-hadoop2-compat-2.3.4-tests.jar
-rw-r--r-- 1 root root   68103 Mar 20 10:17 hbase-hadoop-compat-2.3.4.jar
-rw-r--r-- 1 root root   23528 Mar 20 10:17 hbase-hadoop-compat-2.3.4-tests.jar
-rw-r--r-- 1 root root  145382 Mar 20 10:17 hbase-hbtop-2.3.4.jar
-rw-r--r-- 1 root root  111341 Mar 20 10:17 hbase-http-2.3.4.jar
-rw-r--r-- 1 root root   36071 Mar 20 10:17 hbase-it-2.3.4.jar
-rw-r--r-- 1 root root 2955484 Mar 20 10:17 hbase-it-2.3.4-tests.jar
-rw-r--r-- 1 root root   11956 Mar 20 10:17 hbase-logging-2.3.4.jar
-rw-r--r-- 1 root root  424934 Mar 20 10:17 hbase-mapreduce-2.3.4.jar
-rw-r--r-- 1 root root  539126 Mar 20 10:17 hbase-mapreduce-2.3.4-tests.jar
-rw-r--r-- 1 root root   28096 Mar 20 10:17 hbase-metrics-2.3.4.jar
-rw-r--r-- 1 root root   21423 Mar 20 10:17 hbase-metrics-api-2.3.4.jar
-rw-r--r-- 1 root root  204210 Mar 20 10:17 hbase-procedure-2.3.4.jar
-rw-r--r-- 1 root root  227016 Mar 20 10:17 hbase-procedure-2.3.4-tests.jar
-rw-r--r-- 1 root root 3278175 Mar 20 10:17 hbase-protocol-2.3.4.jar
-rw-r--r-- 1 root root 6855307 Mar 20 10:17 hbase-protocol-shaded-2.3.4.jar
-rw-r--r-- 1 root root   45509 Mar 20 10:17 hbase-replication-2.3.4.jar
-rw-r--r-- 1 root root   59910 Mar 20 10:17 hbase-resource-bundle-2.3.4.jar
-rw-r--r-- 1 root root  443987 Mar 20 10:17 hbase-rest-2.3.4.jar
-rw-r--r-- 1 root root  355923 Mar 20 10:17 hbase-rsgroup-2.3.4.jar
-rw-r--r-- 1 root root  138802 Mar 20 10:17 hbase-rsgroup-2.3.4-tests.jar
-rw-r--r-- 1 root root 5242321 Mar 20 10:17 hbase-server-2.3.4.jar
-rw-r--r-- 1 root root 9533375 Mar 20 10:17 hbase-server-2.3.4-tests.jar
-rw-r--r-- 1 root root  263391 Mar 20 10:17 hbase-shaded-gson-3.3.0.jar
-rw-r--r-- 1 root root 3895799 Mar 20 10:17 hbase-shaded-miscellaneous-3.3.0.jar
-rw-r--r-- 1 root root 4442390 Mar 20 10:17 hbase-shaded-netty-3.3.0.jar
-rw-r--r-- 1 root root 1754089 Mar 20 10:17 hbase-shaded-protobuf-3.3.0.jar
-rw-r--r-- 1 root root   14118 Mar 20 10:17 hbase-shell-2.3.4.jar
-rw-r--r-- 1 root root   12964 Mar 20 10:17 hbase-testing-util-2.3.4.jar
-rw-r--r-- 1 root root 3960818 Mar 20 10:17 hbase-thrift-2.3.4.jar
-rw-r--r-- 1 root root  109734 Mar 20 10:17 hbase-zookeeper-2.3.4.jar
-rw-r--r-- 1 root root   65412 Mar 20 10:17 hbase-zookeeper-2.3.4-tests.jar
 

(base) [root@node2 jars]# ll ht*
-rw-r--r-- 1 root root 1502280 Mar 20 09:33 htrace-core4-4.1.0-incubating.jar
-rw-r--r-- 1 root root  767140 Mar 20 09:33 httpclient-4.5.6.jar
-rw-r--r-- 1 root root  328347 Mar 20 09:33 httpcore-4.4.12.jar
 

(base) [root@node2 jars]# ll m*
-rw-r--r-- 1 root root   33786 Mar 20 09:33 machinist_2.12-0.6.8.jar
-rw-r--r-- 1 root root    3180 Mar 20 09:33 macro-compat_2.12-1.1.1.jar
-rw-r--r-- 1 root root 7343426 Mar 20 09:33 mesos-1.4.0-shaded-protobuf.jar
-rw-r--r-- 1 root root  105365 Mar 20 09:33 metrics-core-4.1.1.jar
-rw-r--r-- 1 root root   22042 Mar 20 09:33 metrics-graphite-4.1.1.jar
-rw-r--r-- 1 root root   20889 Mar 20 09:33 metrics-jmx-4.1.1.jar
-rw-r--r-- 1 root root   16642 Mar 20 09:33 metrics-json-4.1.1.jar
-rw-r--r-- 1 root root   23909 Mar 20 09:33 metrics-jvm-4.1.1.jar
-rw-r--r-- 1 root root    5711 Mar 20 09:33 minlog-1.3.0.jar
-rw-r--r-- 1 root root 2036609 Mar 20 09:33 mysql-connector-java-8.0.11.jar
-rw-r--r-- 1 root root   10476 Mar 20 09:33 mysql-metadata-storage-0.12.0.jar
 

但是在 jupyter notebook  执行 单独对 hive 表和 hbase表操作都没有问题,但是对 hive on hbase 表就不能,网络上一直也没有答案,寻求各位帮助,十分感谢~

解决了吗,同样的问题。。。