DROP TABLE IF EXISTS tmp_dm_test_a.t_aa_orc;
USE tmp_dm_test_a;
CREATE EXTERNAL TABLE IF NOT EXISTS tmp_dm_test_a.t_aa_orc(
user_id string COMMENT '用户id'
,all_addr string COMMENT '常用地址'
)
PARTITIONED BY (
inc_day string COMMENT 'inc_day used by partition'
)
STORED AS orc
TBLPROPERTIES('orc.compress'='SNAPPY');
set hive.exec.dynamic.partition=true;
set hive.exec.dynamic.partition.mode=nonstrict;
set hive.fetch.task.conversion=more;
set hive.exec.parallel=true;
set mapreduce.output.fileoutputformat.compress.codec=org.apache.hadoop.io.compress.SnappyCodec;
set mapreduce.output.fileoutputformat.compress.type=BLOCK;
WITH tmp AS
(
SELECT 'sf1111' as user_id, '湖南省' as all_addr, '20180101'
union all
SELECT 'sf2222' as user_id, '江西省' as all_addr, '20180101'
union all
SELECT 'sf3333' as user_id, '上东省' as all_addr, '20180101'
union all
SELECT 'sf1111' as user_id, '湖南省' as all_addr, '20180102'
union all
SELECT 'sf2222' as user_id, '江西省' as all_addr, '20180102'
union all
SELECT 'sf3333' as user_id, '上东省' as all_addr, '20180102'
)
INSERT OVERWRITE TABLE tmp_dm_test_a.t_aa_orc PARTITION (inc_day)
SELECT * from tmp;
SELECT * from tmp_dm_test_a.t_aa_orc;
SELECT user_id,all_addr,original_union_id from tmp_dm_test_a.t_aa_orc;
SELECT user_id,all_addr,original_union_id from tmp_dm_test_a.t_aa_orc where inc_day='20180101';
SELECT user_id,all_addr,original_union_id from tmp_dm_test_a.t_aa_orc where inc_day='20180103';
----
SELECT phone_number,all_addr,original_union_id from tmp_dm_test_a.t_aa_orc;
--DDL语句最后添加CASCADE,否则新增的列在旧分区中不可见
alter table tmp_dm_test_a.t_aa_orc add columns(original_union_id string) cascade;
--新增多个字段
alter table `ods_wst`.`awd_pckt_in_sm`
add columns(
`rule_code` string COMMENT '规则编码'
, `bus_type` string COMMENT '扩展字段业务类型,用于关联扩展字段业务值1-4'
, `bus_attr1` string COMMENT '扩展字段业务值1'
, `bus_attr2` string COMMENT '扩展字段业务值2'
, `bus_attr3` string COMMENT '扩展字段业务值3'
, `bus_attr4` string COMMENT '扩展字段业务值4'
) cascade;
alter table tmp_dm_test_a.t_aa_orc partition(inc_day='20180101') add columns(original_union_id string);
DROP TABLE IF EXISTS tmp_dm_test_a.t_aa_orc;
USE tmp_dm_test_a;
CREATE EXTERNAL TABLE IF NOT EXISTS tmp_dm_test_a.t_aa_orc(
user_id string COMMENT '用户id'
,all_addr string COMMENT '常用地址'
,original_union_id string
)
PARTITIONED BY (
inc_day string COMMENT 'inc_day used by partition'
)
STORED AS orc
TBLPROPERTIES('orc.compress'='SNAPPY');
MSCK REPAIR TABLE tmp_dm_test_a.t_aa_orc;
WITH tmp AS
(
SELECT 'sf1111' as user_id, '湖南省' as all_addr,'sf0x1111' as original_union_id, '20180103'
union all
SELECT 'sf2222' as user_id, '江西省' as all_addr,'sf0x2211' as original_union_id, '20180103'
union all
SELECT 'sf3333' as user_id, '上东省' as all_addr,'sf0x3311' as original_union_id, '20180103'
union all
SELECT 'sf1111' as user_id, '湖南省' as all_addr,'sf0x4411' as original_union_id, '20180104'
union all
SELECT 'sf2222' as user_id, '江西省' as all_addr,'sf0x5511' as original_union_id, '20180104'
union all
SELECT 'sf3333' as user_id, '上东省' as all_addr,'sf0x6611' as original_union_id, '20180104'
)
INSERT OVERWRITE TABLE tmp_dm_test_a.t_aa_orc PARTITION (inc_day)
SELECT * from tmp;
******************** 之前分区的数据找不到phone_number,需要重跑历史数据
alter table tmp_dm_test_a.t_aa_orc change column user_id phone_number string;
alter table tmp_dm_test_a.t_aa_orc change column user_id phone_number string cascade;
------------- show create table tmp_dm_test_a.t_aa_orc;
ALTER TABLE table_name
[PARTITION partition_spec] -- (Note: Hive 0.14.0 and later)
ADD|REPLACE COLUMNS (col_name data_type [COMMENT col_comment], ...)
[CASCADE|RESTRICT] -- (Note: Hive 1.1.0 and later)
REPLACE列删除所有现有列并添加新的列集。这只能用于具有本机SerDe的表(DynamicSerDe、元数据类型pedcolumnsetserde、LazySimpleSerDe和ColumnarSerDe)。REPLACE列还可以用于删除列。
删除列示例:
原有Hive表test_change中有a,b,c,d,e这几个字段
将从test_change中删除“d”列:
ALTER TABLE test_change REPLACE COLUMNS (a int, b int,c string,e string) cascade;
将d和e两列一起删除:
ALTER TABLE test_change REPLACE COLUMNS (a int, b int,c string) cascade;
-- parque格式的数据,保留d,e字段的数据,但是replace删除后无法查询d,e
-- REPLACE也可以调整字段的顺序,原始数据不用变也可正常查询
parquet存储格式
DROP TABLE IF EXISTS tmp_dm_test_a.t_aa;
USE tmp_dm_test_a;
CREATE EXTERNAL TABLE IF NOT EXISTS tmp_dm_test_a.t_aa(
user_id string COMMENT '用户id'
,all_addr string COMMENT '常用地址'
)
PARTITIONED BY (
inc_day string COMMENT 'inc_day used by partition'
)
STORED AS parquet
TBLPROPERTIES('parquet.compression'='SNAPPY');
set hive.exec.dynamic.partition=true;
set hive.exec.dynamic.partition.mode=nonstrict;
set hive.fetch.task.conversion=more;
set hive.exec.parallel=true;
set mapreduce.output.fileoutputformat.compress.codec=org.apache.hadoop.io.compress.SnappyCodec;
set mapreduce.output.fileoutputformat.compress.type=BLOCK;
WITH tmp AS
(
SELECT 'sf1111' as user_id, '湖南省' as all_addr, '20180101'
union all
SELECT 'sf2222' as user_id, '江西省' as all_addr, '20180101'
union all
SELECT 'sf3333' as user_id, '上东省' as all_addr, '20180101'
union all
SELECT 'sf1111' as user_id, '湖南省' as all_addr, '20180102'
union all
SELECT 'sf2222' as user_id, '江西省' as all_addr, '20180102'
union all
SELECT 'sf3333' as user_id, '上东省' as all_addr, '20180102'
)
INSERT OVERWRITE TABLE tmp_dm_test_a.t_aa PARTITION (inc_day)
SELECT * from tmp;
SELECT * from tmp_dm_test_a.t_aa;
SELECT user_id,all_addr,original_union_id from tmp_dm_test_a.t_aa;
SELECT user_id,all_addr,original_union_id from tmp_dm_test_a.t_aa where inc_day='20180101';
----
SELECT phone_number,all_addr,original_union_id from tmp_dm_test_a.t_aa;
alter table tmp_dm_test_a.t_aa add columns(original_union_id string);
alter table tmp_dm_test_a.t_aa partition(inc_day='20180101') add columns(original_union_id string);
DROP TABLE IF EXISTS tmp_dm_test_a.t_aa;
USE tmp_dm_test_a;
CREATE EXTERNAL TABLE IF NOT EXISTS tmp_dm_test_a.t_aa(
user_id string COMMENT '用户id'
,all_addr string COMMENT '常用地址'
,original_union_id string
)
PARTITIONED BY (
inc_day string COMMENT 'inc_day used by partition'
)
STORED AS parquet
TBLPROPERTIES('parquet.compression'='SNAPPY');
MSCK REPAIR TABLE tmp_dm_test_a.t_aa;
WITH tmp AS
(
SELECT 'sf1111' as user_id, '湖南省' as all_addr,'sf0x1111' as original_union_id, '20180103'
union all
SELECT 'sf2222' as user_id, '江西省' as all_addr,'sf0x2211' as original_union_id, '20180103'
union all
SELECT 'sf3333' as user_id, '上东省' as all_addr,'sf0x3311' as original_union_id, '20180103'
union all
SELECT 'sf1111' as user_id, '湖南省' as all_addr,'sf0x4411' as original_union_id, '20180104'
union all
SELECT 'sf2222' as user_id, '江西省' as all_addr,'sf0x5511' as original_union_id, '20180104'
union all
SELECT 'sf3333' as user_id, '上东省' as all_addr,'sf0x6611' as original_union_id, '20180104'
)
INSERT OVERWRITE TABLE tmp_dm_test_a.t_aa PARTITION (inc_day)
SELECT * from tmp;
********************
alter table tmp_dm_test_a.t_aa change column user_id phone_number string;
alter table tmp_dm_test_a.t_aa change column user_id phone_number string cascade;
*********************************************
结论:
1、parquet和orc格式,旧分区中数据文件内容不可变。
2、parquet和orc格式:字段增加后,旧数据文件中无新字段内容;新产生的分区中数据文件才会有新字段内容。
3、parquet和orc格式:通过add语句末尾追加新增字段后,旧分区和新分区都可以查,旧数据为null而已。
4、parquet格式:修改字段名后,无法从旧数据解析原字段内容,相当于新旧字段名没有印射关系。因为旧数据中只有旧字段名没有新字段名,而且新字段名无法印射旧字段名,所以select不能解析新字段名。
5、orc格式:修改字段名后,可以从旧数据解析原字段内容,相当于新旧字段名有印射关系。虽然旧数据中只有旧字段名没有新字段名,但是新字段名印射了旧字段名,相当于一个别名,所以select可以解析新字段名。