失败:执行错误,从org.apache.hadoop.hive.ql.exec.spark.sparktask返回代码2

mitkmikd  于 2021-05-29  发布在  Hadoop
关注(0)|答案(2)|浏览(336)

我在跑步
apache-hive-1.2.1-bin
hadoop-2.7.1版本
spark-1.5.1-bin-hadoop2.6版本
我能够在spark上配置配置单元,但是当我尝试执行查询时,它会给出下面的错误消息。

  1. hive> SELECT COUNT(*) AS rcount, yom From service GROUP BY yom;
  2. Query ID = hduser_20160110105649_4c90528a-76ba-4127-8849-54f2152be817
  3. Total jobs = 1
  4. Launching Job 1 out of 1
  5. In order to change the average load for a reducer (in bytes):
  6. set hive.exec.reducers.bytes.per.reducer=<number>
  7. In order to limit the maximum number of reducers:
  8. set hive.exec.reducers.max=<number>
  9. In order to set a constant number of reducers:
  10. set mapreduce.job.reduces=<number>
  11. Starting Spark Job = b9cbbd47-f41f-48b5-98c3-efcaa145390e
  12. Status: SENT
  13. FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.spark.SparkTask

我怎样才能解决这个问题?

rta7y2nd

rta7y2nd1#

我有同样的问题,但我没有配置Yarn,因为一些作业正在运行。我不确定那是问题的解决办法。

  1. yarn.resourcemanager.scheduler.class=org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler

你有没有像文件上说的那样配置Yarn?

lf3rwulv

lf3rwulv2#

yarn-site.xml:

  1. <property>
  2. <name>yarn.resourcemanager.scheduler.class</name>
  3. <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>
  4. </property>

相关问题