当我们尝试插入覆盖配置单元表时,配置单元创建子文件夹-ext-10000。这些表格中的数据对于spark来说是不可见的。只有低行的表才有这些问题。
spark版本:版本3.1.1配置单元版本:配置单元3.1.0.3.1.4.0-315
我们试着
"hive.input.dir.recursive" = "TRUE"
"hive.mapred.supports.subdirectories" = "TRUE"
"hive.supports.subdirectories" = "TRUE"
"mapred.input.dir.recursive" = "TRUE"
它不会影响你
查询示例:
insert overwrite table categories
select
n2.id as category1_ccode
,n2.name as category1_name
,n3.id as category2_ccode
,n3.name as category2_name
from nomenclature as n1
left join nomenclature as n2
on n1.id = n2.parent_id
left nomenclature as n3
on n2.id = n3.parent_id
where
n1.name = 'Goods'
and n1.delete_mark = '00'
and n2.delete_mark = '00'
and n3.delete_mark = '00'
and n1.is_group = '00'
and n2.is_group = '00';
文件以orc格式存储
暂无答案!
目前还没有任何答案,快来回答吧!