`
ssydxa219
  • 浏览: 623639 次
  • 性别: Icon_minigender_1
  • 来自: 杭州
文章分类
社区版块
存档分类
最新评论

spark hive

 
阅读更多
org.apache.hive.service.cli.HiveSQLException: Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.spark.SparkTask. Failed to create spark client. at org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:380) at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:257) at org.apache.hive.service.cli.operation.SQLOperation.access$800(SQLOperation.java:91) at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork$1.run(SQLOperation.java:348) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork.run(SQLOperation.java:362) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create spark client. at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:64) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.java:115) at org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUtilities.java:126) at org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:103) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1232) at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:255) ... 11 more Caused by: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.RuntimeException: Cancel client '5e600ca4-e724-434c-a48b-c29097c39c7d'. Error: Child process exited before connecting back with error log Warning: Ignoring non-spark config property: hive.spark.client.server.connect.timeout=90000 Warning: Ignoring non-spark config property: hive.spark.client.rpc.threads=8 Warning: Ignoring non-spark config property: hive.spark.client.connect.timeout=1000 Warning: Ignoring non-spark config property: hive.spark.client.secret.bits=256 Warning: Ignoring non-spark config property: hive.spark.client.rpc.max.size=52428800 18/12/21 10:36:08 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 18/12/21 10:36:08 INFO RemoteDriver: Connecting to: dda:11184 Exception in thread "main" java.lang.NoSuchFieldError: SPARK_RPC_SERVER_ADDRESS at org.apache.hive.spark.client.rpc.RpcConfiguration.<clinit>(RpcConfiguration.java:47) at org.apache.hive.spark.client.RemoteDriver.<init>(RemoteDriver.java:134) at org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:516) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:894) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:198) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:228) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 18/12/21 10:36:08 INFO ShutdownHookManager: Shutdown hook called 18/12/21 10:36:08 INFO ShutdownHookManager: Deleting directory /tmp/spark-6a9d937a-5e5a-4314-846e-a2feb5d7fde9 at com.google.common.base.Throwables.propagate(Throwables.java:160) at org.apache.hive.spark.client.SparkClientImpl.<init>(SparkClientImpl.java:125) at org.apache.hive.spark.client.SparkClientFactory.createClient(SparkClientFactory.java:80) at org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.createRemoteClient(RemoteHiveSparkClient.java:101) at org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:97) at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:73) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:62) ... 22 more Caused by: java.util.concurrent.ExecutionException: java.lang.RuntimeException: Cancel client '5e600ca4-e724-434c-a48b-c29097c39c7d'. Error: Child process exited before connecting back with error log Warning: Ignoring non-spark config property: hive.spark.client.server.connect.timeout=90000 Warning: Ignoring non-spark config property: hive.spark.client.rpc.threads=8 Warning: Ignoring non-spark config property: hive.spark.client.connect.timeout=1000 Warning: Ignoring non-spark config property: hive.spark.client.secret.bits=256 Warning: Ignoring non-spark config property: hive.spark.client.rpc.max.size=52428800 18/12/21 10:36:08 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 18/12/21 10:36:08 INFO RemoteDriver: Connecting to: dda:11184 Exception in thread "main" java.lang.NoSuchFieldError: SPARK_RPC_SERVER_ADDRESS at org.apache.hive.spark.client.rpc.RpcConfiguration.<clinit>(RpcConfiguration.java:47) at org.apache.hive.spark.client.RemoteDriver.<init>(RemoteDriver.java:134) at org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:516) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:894) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:198) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:228) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 18/12/21 10:36:08 INFO ShutdownHookManager: Shutdown hook called 18/12/21 10:36:08 INFO ShutdownHookManager: Deleting directory /tmp/spark-6a9d937a-5e5a-4314-846e-a2feb5d7fde9 at io.netty.util.concurrent.AbstractFuture.get(AbstractFuture.java:41) at org.apache.hive.spark.client.SparkClientImpl.<init>(SparkClientImpl.java:109) ... 27 more Caused by: java.lang.RuntimeException: Cancel client '5e600ca4-e724-434c-a48b-c29097c39c7d'. Error: Child process exited before connecting back with error log Warning: Ignoring non-spark config property: hive.spark.client.server.connect.timeout=90000 Warning: Ignoring non-spark config property: hive.spark.client.rpc.threads=8 Warning: Ignoring non-spark config property: hive.spark.client.connect.timeout=1000 Warning: Ignoring non-spark config property: hive.spark.client.secret.bits=256 Warning: Ignoring non-spark config property: hive.spark.client.rpc.max.size=52428800 18/12/21 10:36:08 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 18/12/21 10:36:08 INFO RemoteDriver: Connecting to: dda:11184 Exception in thread "main" java.lang.NoSuchFieldError: SPARK_RPC_SERVER_ADDRESS at org.apache.hive.spark.client.rpc.RpcConfiguration.<clinit>(RpcConfiguration.java:47) at org.apache.hive.spark.client.RemoteDriver.<init>(RemoteDriver.java:134) at org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:516) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:894) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:198) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:228) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 18/12/21 10:36:08 INFO ShutdownHookManager: Shutdown hook called 18/12/21 10:36:08 INFO ShutdownHookManager: Deleting directory /tmp/spark-6a9d937a-5e5a-4314-846e-a2feb5d7fde9 at org.apache.hive.spark.client.rpc.RpcServer.cancelClient(RpcServer.java:212) at org.apache.hive.spark.client.SparkClientImpl$3.run(SparkClientImpl.java:503) ... 1 more
分享到:
评论

相关推荐

    简单的spark 读写hive以及mysql

    首先,`SparkToHive.java` 和 `SparkHive.java` 文件涉及的是 Spark 读取和写入 Hive 数据。Spark 支持通过 HiveContext(现在称为 HiveSession)连接到 Hive,这样可以使用 SQL 查询 Hive 表并将其结果转换为 Spark...

    项目实战——Spark将Hive表的数据写入ElasticSearch(Java版本)

    在本项目实战中,我们将探讨如何使用Java编程语言,结合Spark和Hive,将Hive中的数据高效地导入到ElasticSearch(ES)中,并利用ES的别名机制实现数据更新的平滑过渡。以下是对这个流程的详细解析: 1. **Hive数据...

    spark连接HIveDemo

    在大数据处理领域,Spark和Hive是两个非常重要的组件。Spark以其高效的计算性能和丰富的生态系统在实时处理和分析任务中占据重要地位,而Hive则作为数据仓库工具,提供了SQL接口进行大规模数据处理。本教程将详细...

    Spark不能使用hive自定义函数.doc

    ### Spark与Hive自定义函数兼容性问题解析 在大数据处理领域,Apache Spark 和 Apache Hive 都是非常重要的工具。Spark 是一种快速通用的大规模数据处理系统,而Hive 则是一种数据仓库工具,主要用于对存储在 ...

    建立Hive和Hbase的映射关系,通过Spark将Hive表中数据导入ClickHouse

    本话题关注的是如何建立Hive与HBase之间的映射关系,并利用Spark将Hive中的数据高效地导入到ClickHouse数据库。以下将详细介绍这一过程的关键步骤和涉及的技术点。 首先,Hive是基于Hadoop的数据仓库工具,用于存储...

    hive 面试宝典 spark hive优化

    hive 面试宝典,hive常见问题,hive优化非常详细

    spark-hive-udf:Spark Hive UDF示例

    Spark Hive UDF示例 建立项目 mvn clean package 将spark-hive-udf-1.0.0-SNAPSHOT.jar复制到边缘节点临时目录 spark-hive-udf]# cp target/spark-hive-udf-1.0.0-SNAPSHOT.jar /tmp 通过提供罐子来启动火花壳 spark...

    编译的spark-hive_2.11-2.3.0和 spark-hive-thriftserver_2.11-2.3.0.jar

    spark-hive_2.11-2.3.0 spark-hive-thriftserver_2.11-2.3.0.jar log4j-2.15.0.jar slf4j-api-1.7.7.jar slf4j-log4j12-1.7.25.jar curator-client-2.4.0.jar curator-framework-2.4.0.jar curator-recipes-2.4.0....

    eclipse集成hadoop+spark+hive开发源码实例

    这个压缩包"SparkHive"可能包含了示例代码或教程资源,用于演示如何在Eclipse中实现Hadoop、Spark和Hive的集成开发。通过这些例子,你可以学习如何创建数据处理项目,如何在Spark中运行SQL查询,以及如何在Hive中...

    spark-hive-2.11和spark-sql-以及spark-hadoop包另付下载地址

    在实际场景中,这个文件夹可能包含了如`spark-2.11.x-y.z-bin-hadoop2.x.y.tgz`这样的Spark发行版,`hive-2.3.x.y-z-bin.tar.gz`的Hive发行版,以及可能的`spark-hive_2.11-2.4.x.y.jar`和`spark-sql_2.11-2.4.x.y....

    spark-hive-thriftserver_2.11-2.1.3-SNAPSHOT-123456.jar

    spark-hive-thriftserver_2.11-2.1.spark-hive-thrift

    centOS 6.10 hadoop+spark+hive+HBase集群搭建

    hadoop+spark+hive Linux centos大数据集群搭建,简单易懂,从0到1搭建大数据集群

    spark操作hive表源码

    在大数据处理领域,Spark和Hive是两个非常重要的工具。Spark以其高效的计算性能和丰富的数据处理API,成为了大数据处理的首选框架,而Hive则以其SQL接口和对大规模数据仓库的支持,深受数据仓库和ETL工作的青睐。本...

    Spark和Hive的结合(让hive基于spark计算)

    spark和hive结合 1.安装mysql 2.在spark/conf中创建一个hive-site.xml    javax.jdo.option.ConnectionURL  jdbc:mysql://192.168.224.132:3306/hive?createDatabaseIfNotExist=true  JDBC connect string for ...

    spark2.3.1-with-hive

    Spark 2.3.1 是一个重要的大数据处理框架,它提供了高效的分布式计算能力,而与 Hive 的集成使得 Spark 可以充分利用 Hive 的元数据、SQL 查询功能和存储系统,为大数据分析提供更丰富的选择。Hive 是一种基于 ...

    spark-hive_2.11-2.1.4-SNAPSHOT.rar

    《Spark与Hive的融合:深入理解Spark-Hive 2.11-2.1.4-SNAPSHOT》 在大数据处理领域,Spark和Hive是两个极为重要的工具。Spark以其高效的内存计算和强大的分布式处理能力,成为了实时计算的首选;而Hive则凭借其SQL...

    Hadoop+Spark+Hive+HBase+Oozie+Kafka+Flume+Flink+ES+Redash等详细安装部署

    在大数据领域,构建一个完整的生态系统是至关重要的,其中包括多个组件,如Hadoop、Spark、Hive、HBase、Oozie、Kafka、Flume、Flink、Elasticsearch和Redash。这些组件协同工作,提供了数据存储、处理、调度、流...

    spark替代Hive实现ETL作业

    ### Spark替代Hive实现ETL作业的关键知识点 #### 使用Hive存在的问题 - **性能瓶颈**:Hive依赖于Hadoop MapReduce引擎进行计算逻辑的执行。尽管它具有较低的硬件需求和较大的吞吐量,但相较于现代DAG(有向无环图...

    SpringBoot整合hive-jdbc示例

    **SpringBoot整合Hive-JDBC详解** 在大数据处理领域,Hadoop生态中的Hive作为一个数据仓库工具,常常用于处理大规模的数据分析任务。而SpringBoot作为Java开发中的微服务框架,以其简洁的配置和快速的开发能力深受...

    hive3.x编译spark3.x包

    在IT行业中,Hive和Spark是两个非常重要的大数据处理框架。Hive主要提供了一种基于Hadoop的数据仓库工具,使得用户可以通过SQL语句对大规模数据进行查询和分析。而Spark则是一个快速、通用且可扩展的大数据处理引擎...

Global site tag (gtag.js) - Google Analytics