spark dataframe:spark dataframe saveastable leaseexpiredexception

xkftehaa  于 2021-06-02  发布在  Hadoop
关注(0)|答案(0)|浏览(374)

在spark应用程序中,我将Dataframe保存为Parquet文件,如下所示,

  1. comp_df.write.mode("overwrite").saveAsTable("cdr_step1", format="parquet", path="/data/intermediate_data/cdr_step1/")

如果我的Dataframe大小是小的,这工作正常。但是随着数据集大小的增加,我得到了以下错误。我在互联网上查过这个问题,在大多数地方,人们通过改变代码设计来解决这个问题。在我的情况下,我只有一行写操作,我不明白我需要改变什么。

  1. 17/02/02 13:22:56 ERROR datasources.DefaultWriterContainer: Job job_201702021228_0000 aborted.
  2. 17/02/02 13:22:56 INFO yarn.YarnAllocator: Driver requested a total number of 0 executor(s).
  3. 17/02/02 13:22:56 WARN scheduler.TaskSetManager: Lost task 1979.0 in stage 3.0 (TID 1984, slv3.cdh-prod.xxxx.com, executor 86): org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException): No lease on /data/intermediate_data/cdr_step1/_temporary/0/_temporary/attempt_201702021322_0003_m_001979_0/part-r-01979-9fe33b7c-0b14-4e63-8e96-6e83aabbe807.gz.parquet (inode 2144221): File does not exist. Holder DFSClient_NONMAPREDUCE_-1523564925_148 does not have any open files.
  4. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:3635)
  5. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.analyzeFileState(FSNamesystem.java:3438)
  6. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3294)
  7. at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:679)
  8. at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.addBlock(AuthorizationProviderProxyClientProtocol.java:214)
  9. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:489)
  10. at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  11. at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617)
  12. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073)
  13. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086)
  14. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082)
  15. at java.security.AccessController.doPrivileged(Native Method)
  16. at javax.security.auth.Subject.doAs(Subject.java:415)
  17. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
  18. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080)

暂无答案!

目前还没有任何答案,快来回答吧!

相关问题