hadoop应用程序失败mapreduce成功

9rbhqvlz  于 2021-06-02  发布在  Hadoop
关注(0)|答案(2)|浏览(565)

我对hadoop 2(hadoop 2.2.0)比较陌生,不明白为什么资源管理器上的m/r job~应用程序被标记为失败:

  1. application_1399458460502_0015 pig Max temperature MAPREDUCE default Wed, 04 Jun 2014 17:16:52 GMT Wed, 04 Jun 2014 17:17:30 GMT FAILED FAILED History

当我知道m/r作业已成功完成,甚至作业历史服务器也声称已成功时:

  1. 2014.06.04 13:16:52 EDT 2014.06.04 13:17:19 EDT job_1399458460502_0015 Max temperature pig default SUCCEEDED 2 2 1 1

我不明白为什么申请被标记为失败。我在jobhistory服务器-日志上看到的唯一错误如下:

  1. 2014-06-04 13:17:19,628 INFO [Thread-62] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify JHEH isAMLastRetry: true
  2. 2014-06-04 13:17:19,628 INFO [Thread-62] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: JobHistoryEventHandler notified that forceJobCompletion is true
  3. 2014-06-04 13:17:19,628 INFO [Thread-62] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Calling stop for all the services
  4. 2014-06-04 13:17:19,629 INFO [Thread-62] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Stopping JobHistoryEventHandler. Size of the outstanding queue size is 0
  5. 2014-06-04 13:17:19,736 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copying hdfs://namenodeha/user/pig/.staging/job_1399458460502_0015/job_1399458460502_0015_1.jhist to hdfs://namenodeha/mr-history/tmp/pig/job_1399458460502_0015-1401902212831-pig-Max+temperature-1401902239623-2-1-SUCCEEDED-default.jhist_tmp
  6. 2014-06-04 13:17:19,812 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copied to done location: hdfs://namenodeha/mr-history/tmp/pig/job_1399458460502_0015-1401902212831-pig-Max+temperature-1401902239623-2-1-SUCCEEDED-default.jhist_tmp
  7. 2014-06-04 13:17:19,824 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copying hdfs://namenodeha/user/pig/.staging/job_1399458460502_0015/job_1399458460502_0015_1_conf.xml to hdfs://namenodeha/mr-history/tmp/pig/job_1399458460502_0015_conf.xml_tmp
  8. 2014-06-04 13:17:19,835 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before Scheduling: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:1 CompletedMaps:2 CompletedReds:1 ContAlloc:3 ContRel:0 HostLocal:2 RackLocal:0
  9. 2014-06-04 13:17:19,880 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Copied to done location: hdfs://namenodeha/mr-history/tmp/pig/job_1399458460502_0015_conf.xml_tmp
  10. 2014-06-04 13:17:19,914 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp to done: hdfs://namenodeha/mr-history/tmp/pig/job_1399458460502_0015.summary_tmp to hdfs://namenodeha/mr-history/tmp/pig/job_1399458460502_0015.summary
  11. 2014-06-04 13:17:19,925 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp to done: hdfs://namenodeha/mr-history/tmp/pig/job_1399458460502_0015_conf.xml_tmp to hdfs://namenodeha/mr-history/tmp/pig/job_1399458460502_0015_conf.xml
  12. 2014-06-04 13:17:19,937 INFO [eventHandlingThread] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Moved tmp to done: hdfs://namenodeha/mr-history/tmp/pig/job_1399458460502_0015-1401902212831-pig-Max+temperature-1401902239623-2-1-SUCCEEDED-default.jhist_tmp to hdfs://namenodeha/mr-history/tmp/pig/job_1399458460502_0015-1401902212831-pig-Max+temperature-1401902239623-2-1-SUCCEEDED-default.jhist
  13. 2014-06-04 13:17:19,938 INFO [Thread-62] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Stopped JobHistoryEventHandler. super.stop()
  14. 2014-06-04 13:17:19,940 INFO [Thread-62] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Setting job diagnostics to
  15. 2014-06-04 13:17:20,060 ERROR [Thread-62] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Exception while unregistering
  16. java.lang.NullPointerException
  17. at org.apache.hadoop.mapreduce.v2.util.MRWebAppUtil.getApplicationWebURLOnJHSWithoutScheme(MRWebAppUtil.java:133)
  18. at org.apache.hadoop.mapreduce.v2.util.MRWebAppUtil.getApplicationWebURLOnJHSWithScheme(MRWebAppUtil.java:148)
  19. at org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator.doUnregistration(RMCommunicator.java:207)
  20. at org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator.unregister(RMCommunicator.java:177)
  21. at org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator.serviceStop(RMCommunicator.java:250)
  22. at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.serviceStop(RMContainerAllocator.java:255)
  23. at org.apache.hadoop.service.AbstractService.stop(AbstractService.java:221)
  24. at org.apache.hadoop.service.ServiceOperations.stop(ServiceOperations.java:52)
  25. at org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter.serviceStop(MRAppMaster.java:817)
  26. at org.apache.hadoop.service.AbstractService.stop(AbstractService.java:221)
  27. at org.apache.hadoop.service.ServiceOperations.stop(ServiceOperations.java:52)
  28. at org.apache.hadoop.service.ServiceOperations.stopQuietly(ServiceOperations.java:80)
  29. at org.apache.hadoop.service.CompositeService.stop(CompositeService.java:159)
  30. at org.apache.hadoop.service.CompositeService.serviceStop(CompositeService.java:132)
  31. at org.apache.hadoop.service.AbstractService.stop(AbstractService.java:221)
  32. at org.apache.hadoop.mapreduce.v2.app.MRAppMaster.shutDownJob(MRAppMaster.java:548)
  33. at org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobFinishEventHandler$1.run(MRAppMaster.java:599)
  34. 2014-06-04 13:17:20,061 INFO [Thread-62] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Final Stats: PendingReds:0 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:0 AssignedReds:1 CompletedMaps:2 CompletedReds:1 ContAlloc:3 ContRel:0 HostLocal:2 RackLocal:0
  35. 2014-06-04 13:17:20,062 INFO [Thread-62] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Skipping cleaning up the staging dir. assuming AM will be retried.
  36. 2014-06-04 13:17:20,062 INFO [Thread-62] org.apache.hadoop.ipc.Server: Stopping server on 43851
  37. 2014-06-04 13:17:20,064 INFO [IPC Server listener on 43851] org.apache.hadoop.ipc.Server: Stopping IPC Server listener on 43851
  38. 2014-06-04 13:17:20,065 INFO [IPC Server Responder] org.apache.hadoop.ipc.Server: Stopping IPC Server Responder
  39. 2014-06-04 13:17:20,065 INFO [TaskHeartbeatHandler PingChecker] org.apache.hadoop.mapreduce.v2.app.TaskHeartbeatHandler: TaskHeartbeatHandler thread interrupted
  40. 2014-06-04 13:17:25,066 INFO [Thread-62] org.apache.hadoop.ipc.Server: Stopping server on 44771
  41. 2014-06-04 13:17:25,066 INFO [IPC Server listener on 44771] org.apache.hadoop.ipc.Server: Stopping IPC Server listener on 44771
  42. 2014-06-04 13:17:25,067 INFO [IPC Server Responder] org.apache.hadoop.ipc.Server: Stopping IPC Server Responder
  43. 2014-06-04 13:17:25,072 INFO [Thread-62] org.mortbay.log: Stopped SelectChannelConnector@0.0.0.0:0
  44. 2014-06-04 13:17:25,172 INFO [Thread-62] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Exiting MR AppMaster..GoodBye!
  45. 2014-06-04 13:17:25,173 INFO [Thread-1] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: MRAppMaster received a signal. Signaling RMCommunicator and JobHistoryEventHandler.
  46. 2014-06-04 13:17:25,173 INFO [Thread-1] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: RMCommunicator notified that iSignalled is: true
  47. 2014-06-04 13:17:25,173 INFO [Thread-1] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify RMCommunicator isAMLastRetry: false
  48. 2014-06-04 13:17:25,173 INFO [Thread-1] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: RMCommunicator notified that shouldUnregistered is: false
  49. 2014-06-04 13:17:25,173 INFO [Thread-1] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Notify JHEH isAMLastRetry: false
  50. 2014-06-04 13:17:25,174 INFO [Thread-1] org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: JobHistoryEventHandler notified that forceJobCompletion is false

或从资源管理器:

  1. Application application_1399458460502_0015 failed 2 times due to AM Container for appattempt_1399458460502_0015_000002 exited with exitCode: 1 due to: Exception from container-launch:
  2. org.apache.hadoop.util.Shell$ExitCodeException:
  3. at org.apache.hadoop.util.Shell.runCommand(Shell.java:464)
  4. at org.apache.hadoop.util.Shell.run(Shell.java:379)
  5. at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589)
  6. at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
  7. at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:283)
  8. at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:79)
  9. at java.util.concurrent.FutureTask.run(FutureTask.java:262)
  10. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  11. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
  12. at java.lang.Thread.run(Thread.java:744)
  13. .Failing this attempt.. Failing the application

这些错误都没有给我一点线索。我的配置如下:

  1. <?xml version="1.0"?>
  2. <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
  3. <configuration>
  4. <!--internal property-->
  5. <property>
  6. <name>bigdata.conf.identification</name>
  7. <value>cluster-DEV1</value>
  8. <final>true</final>
  9. </property>
  10. <!--hadoop properties-->
  11. <!-- Put site-specific property overrides in this file. -->
  12. <!--hbase-site-->
  13. <property>
  14. <name>hbase.zookeeper.quorum</name>
  15. <value>bd-prg-dev1-nn1,bd-prg-dev1-nn2,bd-prg-dev1-rm1</value>
  16. </property>
  17. <property>
  18. <name>zookeeper.session.timeout</name>
  19. <value>60000</value>
  20. </property>
  21. <property>
  22. <name>hbase.zookeeper.property.clientPort</name>
  23. <value>2181</value>
  24. </property>
  25. <property>
  26. <name>zookeeper.znode.parent</name>
  27. <value>/hbase-unsecure</value>
  28. </property>
  29. <!--core-site-->
  30. <property>
  31. <name>hadoop.security.authentication</name>
  32. <value>simple</value>
  33. </property>
  34. <property>
  35. <name>ipc.client.connect.max.retries</name>
  36. <value>50</value>
  37. </property>
  38. <property>
  39. <name>ipc.client.connection.maxidletime</name>
  40. <value>30000</value>
  41. </property>
  42. <property>
  43. <name>ipc.client.idlethreshold</name>
  44. <value>8000</value>
  45. </property>
  46. <property>
  47. <name>io.compression.codecs</name>
  48. <value>org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec</value>
  49. </property>
  50. <property>
  51. <name>io.serializations</name>
  52. <value>org.apache.hadoop.io.serializer.WritableSerialization</value>
  53. </property>
  54. <property>
  55. <name>hadoop.security.authorization</name>
  56. <value>false</value>
  57. </property>
  58. <property>
  59. <name>fs.defaultFS</name>
  60. <value>hdfs://namenodeha</value>
  61. </property>
  62. <property>
  63. <name>ha.zookeeper.quorum</name>
  64. <value>bd-prg-dev1-nn1:2181,bd-prg-dev1-nn2:2181,bd-prg-dev1-rm1:2181</value>
  65. </property>
  66. <!-- hdfs-site-->
  67. <property>
  68. <name>dfs.namenode.http-address</name>
  69. <value>bd-prg-dev1-nn1:50070</value>
  70. </property>
  71. <property>
  72. <name>dfs.datanode.ipc.address</name>
  73. <value>0.0.0.0:8010</value>
  74. </property>
  75. <property>
  76. <name>dfs.journalnode.http-address</name>
  77. <value>0.0.0.0:8480</value>
  78. </property>
  79. <property>
  80. <name>dfs.namenode.accesstime.precision</name>
  81. <value>0</value>
  82. </property>
  83. <property>
  84. <name>dfs.namenode.stale.datanode.interval</name>
  85. <value>30000</value>
  86. </property>
  87. <property>
  88. <name>dfs.datanode.address</name>
  89. <value>0.0.0.0:50010</value>
  90. </property>
  91. <property>
  92. <name>dfs.datanode.http.address</name>
  93. <value>0.0.0.0:50075</value>
  94. </property>
  95. <property>
  96. <name>dfs.namenode.secondary.http-address</name>
  97. <value>bd-prg-dev1-nn2:50090</value>
  98. </property>
  99. <property>
  100. <name>dfs.nameservices</name>
  101. <value>namenodeha</value>
  102. <description>Logical name for this new nameservice</description>
  103. </property>
  104. <property>
  105. <name>dfs.ha.namenodes.namenodeha</name>
  106. <value>nn1,nn2</value>
  107. <description>Unique identifiers for each NameNode in the nameservice</description>
  108. </property>
  109. <property>
  110. <name>dfs.namenode.rpc-address.namenodeha.nn1</name>
  111. <value>bd-prg-dev1-nn1:8020</value>
  112. </property>
  113. <property>
  114. <name>dfs.namenode.rpc-address.namenodeha.nn2</name>
  115. <value>bd-prg-dev1-nn2:8020</value>
  116. </property>
  117. <property>
  118. <name>dfs.namenode.http-address.namenodeha.nn1</name>
  119. <value>bd-prg-dev1-nn1:50070</value>
  120. </property>
  121. <property>
  122. <name>dfs.namenode.http-address.namenodeha.nn2</name>
  123. <value>bd-prg-dev1-nn2:50070</value>
  124. </property>
  125. <property>
  126. <name>dfs.client.failover.proxy.provider.namenodeha</name>
  127. <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
  128. </property>
  129. <!--yarn-site-->
  130. <property>
  131. <name>yarn.nodemanager.address</name>
  132. <value>0.0.0.0:45454</value>
  133. </property>
  134. <property>
  135. <name>yarn.nodemanager.container-monitor.interval-ms</name>
  136. <value>3000</value>
  137. </property>
  138. <property>
  139. <name>yarn.nodemanager.remote-app-log-dir</name>
  140. <value>/app-logs</value>
  141. </property>
  142. <property>
  143. <name>yarn.log.server.url</name>
  144. <value>bd-prg-dev1-rm1:19888/jobhistory/logs</value>
  145. </property>
  146. <property>
  147. <name>yarn.resourcemanager.admin.address</name>
  148. <value>bd-prg-dev1-rm1:8141</value>
  149. </property>
  150. <property>
  151. <name>yarn.resourcemanager.resource-tracker.address</name>
  152. <value>bd-prg-dev1-rm1:8025</value>
  153. </property>
  154. <property>
  155. <name>yarn.nodemanager.aux-services</name>
  156. <value>mapreduce_shuffle</value>
  157. </property>
  158. <property>
  159. <name>yarn.nodemanager.log-aggregation.compression-type</name>
  160. <value>gz</value>
  161. </property>
  162. <property>
  163. <name>yarn.nodemanager.health-checker.script.path</name>
  164. <value>/etc/hadoop/conf/health_check</value>
  165. </property>
  166. <property>
  167. <name>yarn.nodemanager.container-executor.class</name>
  168. <value>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor</value>
  169. </property>
  170. <property>
  171. <name>yarn.nodemanager.remote-app-log-dir-suffix</name>
  172. <value>logs</value>
  173. </property>
  174. <property>
  175. <name>yarn.resourcemanager.webapp.address</name>
  176. <value>bd-prg-dev1-rm1:8088</value>
  177. </property>
  178. <property>
  179. <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
  180. <value>org.apache.hadoop.mapred.ShuffleHandler</value>
  181. </property>
  182. <property>
  183. <name>yarn.resourcemanager.address</name>
  184. <value>bd-prg-dev1-rm1:8050</value>
  185. </property>
  186. <property>
  187. <name>yarn.log-aggregation-enable</name>
  188. <value>true</value>
  189. </property>
  190. <property>
  191. <name>yarn.resourcemanager.scheduler.address</name>
  192. <value>bd-prg-dev1-rm1:8030</value>
  193. </property>
  194. <!--mapred-site-->
  195. <property>
  196. <name>mapreduce.jobhistory.intermediate-done-dir</name>
  197. <value>/mr-history/tmp</value>
  198. </property>
  199. <property>
  200. <name>mapreduce.jobhistory.done-dir</name>
  201. <value>/mr-history/done</value>
  202. </property>
  203. <property>
  204. <name>mapreduce.jobhistory.webapp.address</name>
  205. <value>bd-prg-dev1-rm1:19888</value>
  206. </property>
  207. <property>
  208. <name>mapreduce.jobtracker.system.dir</name>
  209. <value>/mapred/system</value>
  210. </property>
  211. <property>
  212. <name>mapreduce.framework.name</name>
  213. <value>yarn</value>
  214. </property>
  215. </configuration>

客户端控制台的输出:

  1. [pig@hadooptools Hadoop]$ hadoop jar Jobs/hadoop-mr-poc-1.0-SNAPSHOT.jar com.ncr.bigdata.mr.MaxTemperatureDriver -conf clientConf/cluster/hadoop-client.xmlMRtest/input /MRtest/output/1
  2. 14/06/04 12:17:37 INFO mr.MaxTemperatureDriver: Job Max temperature submitted
  3. 14/06/04 12:17:37 INFO client.RMProxy: Connecting to ResourceManager at bd-prg-dev1-rm1/153.86.208.210:8050
  4. 14/06/04 12:17:37 INFO client.RMProxy: Connecting to ResourceManager at bd-prg-dev1-rm1/153.86.208.210:8050
  5. 14/06/04 12:17:38 INFO mapred.FileInputFormat: Total input paths to process : 1
  6. 14/06/04 12:17:38 INFO net.NetworkTopology: Adding a new node: /default-rack/153.86.208.208:50010
  7. 14/06/04 12:17:38 INFO net.NetworkTopology: Adding a new node: /default-rack/153.86.208.249:50010
  8. 14/06/04 12:17:39 INFO mapreduce.JobSubmitter: number of splits:2
  9. 14/06/04 12:17:39 INFO Configuration.deprecation: user.name is deprecated. Instead, use mapreduce.job.user.name
  10. 14/06/04 12:17:39 INFO Configuration.deprecation: mapred.jar is deprecated. Instead, use mapreduce.job.jar
  11. 14/06/04 12:17:39 INFO Configuration.deprecation: mapred.output.value.class is deprecated. Instead, use mapreduce.job.output.value.class
  12. 14/06/04 12:17:39 INFO Configuration.deprecation: mapred.job.name is deprecated. Instead, use mapreduce.job.name
  13. 14/06/04 12:17:39 INFO Configuration.deprecation: mapred.input.dir is deprecated. Instead, use mapreduce.input.fileinputformat.inputdir
  14. 14/06/04 12:17:39 INFO Configuration.deprecation: mapred.output.dir is deprecated. Instead, use mapreduce.output.fileoutputformat.outputdir
  15. 14/06/04 12:17:39 INFO Configuration.deprecation: mapred.map.tasks is deprecated. Instead, use mapreduce.job.maps
  16. 14/06/04 12:17:39 INFO Configuration.deprecation: mapred.output.key.class is deprecated. Instead, use mapreduce.job.output.key.class
  17. 14/06/04 12:17:39 INFO Configuration.deprecation: mapred.working.dir is deprecated. Instead, use mapreduce.job.working.dir
  18. 14/06/04 12:17:39 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1399458460502_0015
  19. 14/06/04 12:17:39 INFO impl.YarnClientImpl: Submitted application application_1399458460502_0015 to ResourceManager at bd-prg-dev1-rm1/153.86.208.210:8050
  20. 14/06/04 12:17:39 INFO mapreduce.Job: The url to track the job: http://bd-prg-dev1-rm1:8088/proxy/application_1399458460502_0015/
  21. 14/06/04 12:17:39 INFO mapreduce.Job: Running job: job_1399458460502_0015
  22. 14/06/04 12:17:46 INFO mapreduce.Job: Job job_1399458460502_0015 running in uber mode : false
  23. 14/06/04 12:17:46 INFO mapreduce.Job: map 0% reduce 0%
  24. 14/06/04 12:17:59 INFO mapreduce.Job: map 100% reduce 0%
  25. 14/06/04 12:18:07 INFO mapreduce.Job: map 100% reduce 100%
  26. 14/06/04 12:18:31 INFO ipc.Client: Retrying connect to server: bd-prg-dev1-dn1/153.86.208.208:44771. Already tried 0 time(s); maxRetries=45
  27. 14/06/04 12:18:51 INFO ipc.Client: Retrying connect to server: bd-prg-dev1-dn1/153.86.208.208:44771. Already tried 1 time(s); maxRetries=45
  28. 14/06/04 12:19:11 INFO ipc.Client: Retrying connect to server: bd-prg-dev1-dn1/153.86.208.208:44771. Already tried 2 time(s); maxRetries=45
  29. 14/06/04 12:33:14 INFO ipc.Client: Retrying connect to server: bd-prg-dev1-dn1/153.86.208.208:44771. Already tried 44 time(s); maxRetries=45
  30. 14/06/04 12:33:35 INFO mapreduce.Job: map 0% reduce 0%
  31. 14/06/04 12:33:35 INFO mapreduce.Job: Job job_1399458460502_0015 failed with state FAILED due to: Application application_1399458460502_0015 failed 2 times due to AM Container for appattempt_1399458460502_0015_000002 exited with exitCode: 1 due to: Exception from container-launch:
  32. org.apache.hadoop.util.Shell$ExitCodeException:
  33. at org.apache.hadoop.util.Shell.runCommand(Shell.java:464)
  34. at org.apache.hadoop.util.Shell.run(Shell.java:379)
  35. at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589)
  36. at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
  37. at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:283)
  38. at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:79)
  39. at java.util.concurrent.FutureTask.run(FutureTask.java:262)
  40. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  41. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
  42. at java.lang.Thread.run(Thread.java:744)
  43. .Failing this attempt.. Failing the application.
  44. 14/06/04 12:33:35 INFO mapreduce.Job: Counters: 0
  45. Exception in thread "main" java.io.IOException: Job failed!
  46. at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:836)
  47. at com.ncr.bigdata.mr.MaxTemperatureDriver.run(MaxTemperatureDriver.java:46)
  48. at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
  49. at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
  50. at com.ncr.bigdata.mr.MaxTemperatureDriver.main(MaxTemperatureDriver.java:54)
  51. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  52. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
  53. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  54. at java.lang.reflect.Method.invoke(Method.java:606)
  55. at org.apache.hadoop.util.RunJar.main(RunJar.java:212)

有人猜出什么问题了吗?

omqzjyyz

omqzjyyz1#

当无法从配置中找到某些内容时,似乎会出现这种问题(正如原始答案所暗示的那样)。
在我的情况下,我发现当我做一个Pig的行动,我需要参考一个 site.xml ,工作流的相关部分现在如下所示:

  1. <action name="read-into-table">
  2. <pig>
  3. <job-tracker>${jobTracker}</job-tracker>
  4. <name-node>${nameNode}</name-node>
  5. <job-xml>site.xml</job-xml>
  6. <configuration>
  7. <property>
  8. <name>mapred.job.queue.name</name>
  9. <value>${queueName}</value>
  10. </property>
  11. </configuration>
  12. <script>myFile.pig</script>
  13. </pig>
  14. <ok to="end" />
  15. <error to="fail" />
  16. </action>

缺少以下行:

  1. <job-xml>site.xml</job-xml>
展开查看全部
j5fpnvbx

j5fpnvbx2#

我不太清楚它是哪个配置属性,但是当我从集群中获取配置并用它创建配置对象时,它工作得很好。

相关问题