具体代码段如下:执行了.show()命令
---------------------------------------------------------------------------
Py4JJavaError Traceback (most recent call last)
<ipython-input-10-9f366d069c5f> in <module>
----> 1 user_profile_hbase1.show()
/home/hadoop/spark/python/pyspark/sql/dataframe.py in show(self, n, truncate, vertical)
438 """
439 if isinstance(truncate, bool) and truncate:
--> 440 print(self._jdf.showString(n, 20, vertical))
441 else:
442 print(self._jdf.showString(n, int(truncate), vertical))
/home/hadoop/spark/python/lib/py4j-0.10.9-src.zip/py4j/java_gateway.py in __call__(self, *args)
1303 answer = self.gateway_client.send_command(command)
1304 return_value = get_return_value(
-> 1305 answer, self.gateway_client, self.target_id, self.name)
1306
1307 for temp_arg in temp_args:
/home/hadoop/spark/python/pyspark/sql/utils.py in deco(*a, **kw)
126 def deco(*a, **kw):
127 try:
--> 128 return f(*a, **kw)
129 except py4j.protocol.Py4JJavaError as e:
130 converted = convert_exception(e.java_exception)
/home/hadoop/spark/python/lib/py4j-0.10.9-src.zip/py4j/protocol.py in get_return_value(answer, gateway_client, target_id, name)
326 raise Py4JJavaError(
327 "An error occurred while calling {0}{1}{2}.\n".
--> 328 format(target_id, ".", name), value)
329 else:
330 raise Py4JError(
Py4JJavaError: An error occurred while calling o69.showString.
: java.io.IOException: Cannot create a record reader because of a previous error. Please look at the previous logs lines from the task's full log for more details.
at org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getSplits(TableInputFormatBase.java:252)
at org.apache.spark.rdd.NewHadoopRDD.getPartitions(NewHadoopRDD.scala:131)
at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
at scala.Option.getOrElse(Option.scala:189)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
at scala.Option.getOrElse(Option.scala:189)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
at scala.Option.getOrElse(Option.scala:189)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
at scala.Option.getOrElse(Option.scala:189)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
at scala.Option.getOrElse(Option.scala:189)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:276)
at scala.Option.getOrElse(Option.scala:189)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:272)
at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:442)
at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:425)
at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:47)
at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:3627)
at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:2697)
at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3618)
at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100)
at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87)
at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:764)
at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3616)
at org.apache.spark.sql.Dataset.head(Dataset.scala:2697)
at org.apache.spark.sql.Dataset.take(Dataset.scala:2904)
at org.apache.spark.sql.Dataset.getRows(Dataset.scala:300)
at org.apache.spark.sql.Dataset.showString(Dataset.scala:337)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
at py4j.Gateway.invoke(Gateway.java:282)
at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
at py4j.commands.CallCommand.execute(CallCommand.java:79)
at py4j.GatewayConnection.run(GatewayConnection.java:238)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalStateException: The input format instance has not been properly initialized. Ensure you call initializeTable either in your constructor or initialize method
at org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getTable(TableInputFormatBase.java:570)
at org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getSplits(TableInputFormatBase.java:247)
... 51 more
相应的包hbase包,hive包都移动到 $SPARK_HOME/jars,具体如下:
-rw-r--r-- 1 root root 183472 Mar 20 09:33 hive-beeline-2.3.7.jar
-rw-r--r-- 1 root root 43387 Mar 20 09:33 hive-cli-2.3.7.jar
-rw-r--r-- 1 root root 436980 Mar 20 09:33 hive-common-2.3.7.jar
-rw-r--r-- 1 root root 10839104 Mar 20 09:33 hive-exec-2.3.7-core.jar
-rw-r--r-- 1 root root 118060 Mar 20 09:33 hive-hbase-handler-3.1.2.jar
-rw-r--r-- 1 root root 268933 Mar 20 09:33 hive-hcatalog-core-3.1.2.jar
-rw-r--r-- 1 root root 55459 Mar 20 09:33 hive-hcatalog-pig-adapter-3.1.2.jar
-rw-r--r-- 1 root root 76615 Mar 20 09:33 hive-hcatalog-server-extensions-3.1.2.jar
-rw-r--r-- 1 root root 133382 Mar 20 09:33 hive-hcatalog-streaming-3.1.2.jar
-rw-r--r-- 1 root root 116311 Mar 20 09:33 hive-jdbc-2.3.7.jar
-rw-r--r-- 1 root root 326419 Mar 20 09:33 hive-llap-common-2.3.7.jar
-rw-r--r-- 1 root root 8194428 Mar 20 09:33 hive-metastore-2.3.7.jar
-rw-r--r-- 1 root root 916206 Mar 20 09:33 hive-serde-2.3.7.jar
-rw-r--r-- 1 root root 54116 Mar 20 09:33 hive-shims-0.23-2.3.7.jar
-rw-r--r-- 1 root root 8785 Mar 20 09:33 hive-shims-2.3.7.jar
-rw-r--r-- 1 root root 120305 Mar 20 09:33 hive-shims-common-2.3.7.jar
-rw-r--r-- 1 root root 12984 Mar 20 09:33 hive-shims-scheduler-2.3.7.jar
-rw-r--r-- 1 root root 234942 Mar 20 09:33 hive-storage-api-2.7.1.jar
-rw-r--r-- 1 root root 38352 Mar 20 09:33 hive-vector-code-gen-2.3.7.jar
-rw-r--r-- 1 root root 6666 Mar 20 10:17 hbase-annotations-2.3.4.jar
-rw-r--r-- 1 root root 14128 Mar 20 10:17 hbase-annotations-2.3.4-tests.jar
-rw-r--r-- 1 root root 90016 Mar 20 10:17 hbase-asyncfs-2.3.4.jar
-rw-r--r-- 1 root root 35593 Mar 20 10:17 hbase-asyncfs-2.3.4-tests.jar
-rw-r--r-- 1 root root 1932103 Mar 20 10:17 hbase-client-2.3.4.jar
-rw-r--r-- 1 root root 808377 Mar 20 10:17 hbase-common-2.3.4.jar
-rw-r--r-- 1 root root 353119 Mar 20 10:17 hbase-common-2.3.4-tests.jar
-rw-r--r-- 1 root root 471046 Mar 20 10:17 hbase-endpoint-2.3.4.jar
-rw-r--r-- 1 root root 187054 Mar 20 10:17 hbase-examples-2.3.4.jar
-rw-r--r-- 1 root root 20206 Mar 20 10:17 hbase-external-blockcache-2.3.4.jar
-rw-r--r-- 1 root root 165895 Mar 20 10:17 hbase-hadoop2-compat-2.3.4.jar
-rw-r--r-- 1 root root 40671 Mar 20 10:17 hbase-hadoop2-compat-2.3.4-tests.jar
-rw-r--r-- 1 root root 68103 Mar 20 10:17 hbase-hadoop-compat-2.3.4.jar
-rw-r--r-- 1 root root 23528 Mar 20 10:17 hbase-hadoop-compat-2.3.4-tests.jar
-rw-r--r-- 1 root root 145382 Mar 20 10:17 hbase-hbtop-2.3.4.jar
-rw-r--r-- 1 root root 111341 Mar 20 10:17 hbase-http-2.3.4.jar
-rw-r--r-- 1 root root 36071 Mar 20 10:17 hbase-it-2.3.4.jar
-rw-r--r-- 1 root root 2955484 Mar 20 10:17 hbase-it-2.3.4-tests.jar
-rw-r--r-- 1 root root 11956 Mar 20 10:17 hbase-logging-2.3.4.jar
-rw-r--r-- 1 root root 424934 Mar 20 10:17 hbase-mapreduce-2.3.4.jar
-rw-r--r-- 1 root root 539126 Mar 20 10:17 hbase-mapreduce-2.3.4-tests.jar
-rw-r--r-- 1 root root 28096 Mar 20 10:17 hbase-metrics-2.3.4.jar
-rw-r--r-- 1 root root 21423 Mar 20 10:17 hbase-metrics-api-2.3.4.jar
-rw-r--r-- 1 root root 204210 Mar 20 10:17 hbase-procedure-2.3.4.jar
-rw-r--r-- 1 root root 227016 Mar 20 10:17 hbase-procedure-2.3.4-tests.jar
-rw-r--r-- 1 root root 3278175 Mar 20 10:17 hbase-protocol-2.3.4.jar
-rw-r--r-- 1 root root 6855307 Mar 20 10:17 hbase-protocol-shaded-2.3.4.jar
-rw-r--r-- 1 root root 45509 Mar 20 10:17 hbase-replication-2.3.4.jar
-rw-r--r-- 1 root root 59910 Mar 20 10:17 hbase-resource-bundle-2.3.4.jar
-rw-r--r-- 1 root root 443987 Mar 20 10:17 hbase-rest-2.3.4.jar
-rw-r--r-- 1 root root 355923 Mar 20 10:17 hbase-rsgroup-2.3.4.jar
-rw-r--r-- 1 root root 138802 Mar 20 10:17 hbase-rsgroup-2.3.4-tests.jar
-rw-r--r-- 1 root root 5242321 Mar 20 10:17 hbase-server-2.3.4.jar
-rw-r--r-- 1 root root 9533375 Mar 20 10:17 hbase-server-2.3.4-tests.jar
-rw-r--r-- 1 root root 263391 Mar 20 10:17 hbase-shaded-gson-3.3.0.jar
-rw-r--r-- 1 root root 3895799 Mar 20 10:17 hbase-shaded-miscellaneous-3.3.0.jar
-rw-r--r-- 1 root root 4442390 Mar 20 10:17 hbase-shaded-netty-3.3.0.jar
-rw-r--r-- 1 root root 1754089 Mar 20 10:17 hbase-shaded-protobuf-3.3.0.jar
-rw-r--r-- 1 root root 14118 Mar 20 10:17 hbase-shell-2.3.4.jar
-rw-r--r-- 1 root root 12964 Mar 20 10:17 hbase-testing-util-2.3.4.jar
-rw-r--r-- 1 root root 3960818 Mar 20 10:17 hbase-thrift-2.3.4.jar
-rw-r--r-- 1 root root 109734 Mar 20 10:17 hbase-zookeeper-2.3.4.jar
-rw-r--r-- 1 root root 65412 Mar 20 10:17 hbase-zookeeper-2.3.4-tests.jar
(base) [root@node2 jars]# ll ht*
-rw-r--r-- 1 root root 1502280 Mar 20 09:33 htrace-core4-4.1.0-incubating.jar
-rw-r--r-- 1 root root 767140 Mar 20 09:33 httpclient-4.5.6.jar
-rw-r--r-- 1 root root 328347 Mar 20 09:33 httpcore-4.4.12.jar
(base) [root@node2 jars]# ll m*
-rw-r--r-- 1 root root 33786 Mar 20 09:33 machinist_2.12-0.6.8.jar
-rw-r--r-- 1 root root 3180 Mar 20 09:33 macro-compat_2.12-1.1.1.jar
-rw-r--r-- 1 root root 7343426 Mar 20 09:33 mesos-1.4.0-shaded-protobuf.jar
-rw-r--r-- 1 root root 105365 Mar 20 09:33 metrics-core-4.1.1.jar
-rw-r--r-- 1 root root 22042 Mar 20 09:33 metrics-graphite-4.1.1.jar
-rw-r--r-- 1 root root 20889 Mar 20 09:33 metrics-jmx-4.1.1.jar
-rw-r--r-- 1 root root 16642 Mar 20 09:33 metrics-json-4.1.1.jar
-rw-r--r-- 1 root root 23909 Mar 20 09:33 metrics-jvm-4.1.1.jar
-rw-r--r-- 1 root root 5711 Mar 20 09:33 minlog-1.3.0.jar
-rw-r--r-- 1 root root 2036609 Mar 20 09:33 mysql-connector-java-8.0.11.jar
-rw-r--r-- 1 root root 10476 Mar 20 09:33 mysql-metadata-storage-0.12.0.jar
但是在 jupyter notebook 执行 单独对 hive 表和 hbase表操作都没有问题,但是对 hive on hbase 表就不能,网络上一直也没有答案,寻求各位帮助,十分感谢~
解决了吗,同样的问题。。。