来自dataframe的自定义分隔文本文件

eit6fx6z  于 2021-05-27  发布在  Hadoop
关注(0)|答案(0)|浏览(174)

我正在使用spark 1.6并尝试从Dataframe创建一个分隔文件。
字段分隔符是“| ^”,因此我在从temp表中选择时连接Dataframe中的列
现在,下面的代码每次都会因为这个错误而失败

ERROR scheduler.TaskSetManager: Task 172 in stage 9.0 failed 4 times; aborting job
19/03/01 09:10:15 ERROR datasources.InsertIntoHadoopFsRelation: Aborting job.
org.apache.spark.SparkException: Job aborted due to stage failure: Task 172 in stage 9.0 failed 4 times, most recent failure: Lost task 172.3 in stage 9.0 (TID 1397, tplhc01d104.iuser.iroot.adidom.com, executor 7): org.apache.spark.SparkException: Task failed while writing rows.
        at org.apache.spark.sql.execution.datasources.DefaultWriterContainer.writeRows(WriterContainer.scala:272

我用的代码是。。

tempDF.registerTempTable("BNUC_TEMP")

context.sql("select concat('VALID','|^', RECORD_ID,'|^', DATA_COL1,'|^', DATA_COL2,'|^','P','|^', DATA_COL4,'|^', DATA_COL5,'|^', DATA_COL6,'GBP','|^',from_unixtime(unix_timestamp( ACTION_DATE)),'|^',from_unixtime(unix_timestamp( UPDATED_DATE))) from BNUC_TEMP")
.write.mode("overwrite")
.text("/user/USERNAME/landing/staging/BNU/temp/")

暂无答案!

目前还没有任何答案,快来回答吧!

相关问题