This is an automated email from the ASF dual-hosted git repository. morrysnow pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/doris.git
The following commit(s) were added to refs/heads/master by this push: new 8a144548a3f [fix](nereids) SqlParser support load data into temporary partition (#45025) 8a144548a3f is described below commit 8a144548a3f8dcada0c3d9e31f60513b84da1bb2 Author: zxealous <zhouchang...@baidu.com> AuthorDate: Tue Dec 10 13:52:26 2024 +0800 [fix](nereids) SqlParser support load data into temporary partition (#45025) Problem Summary: Currently, use DorisParser.g4 parse sql, and not support specified temporary partition while load data now. LOAD LABEL db.label ( DATA INFILE("hdfs://hdfs:9000/user/partition/partition_type") INTO TABLE `tb` TEMPORARY PARTITION (partition_g)) WITH BROKER "ahdfs" ( "username" = "xxx", "password" = "") PROPERTIES( "max_filter_ratio"="0.5" ) ERROR 1105 (HY000): errCode = 2, detailMessage = mismatched input 'TEMPORARY' expecting {')', ','}(line 1, pos xxx) This pr support specified temporary partition while load data, because load operation will backup to old optimizer, so only support specified temporary partition while load data can successfully load data into the temporary partition. TODO: 1.nereids need support load data to temporary partition. 2.nereids need support export data from temporary partition. --- .../antlr4/org/apache/doris/nereids/DorisParser.g4 | 2 +- .../test_broker_load_with_partition.out | 8 ++++ .../test_broker_load_with_partition.groovy | 54 ++++++++++++++++++++++ 3 files changed, 63 insertions(+), 1 deletion(-) diff --git a/fe/fe-core/src/main/antlr4/org/apache/doris/nereids/DorisParser.g4 b/fe/fe-core/src/main/antlr4/org/apache/doris/nereids/DorisParser.g4 index 1b01cad4293..fdad6baccdc 100644 --- a/fe/fe-core/src/main/antlr4/org/apache/doris/nereids/DorisParser.g4 +++ b/fe/fe-core/src/main/antlr4/org/apache/doris/nereids/DorisParser.g4 @@ -929,7 +929,7 @@ identityOrFunction dataDesc : ((WITH)? mergeType)? DATA INFILE LEFT_PAREN filePaths+=STRING_LITERAL (COMMA filePath+=STRING_LITERAL)* RIGHT_PAREN INTO TABLE targetTableName=identifier - (PARTITION partition=identifierList)? + (partitionSpec)? (COLUMNS TERMINATED BY comma=STRING_LITERAL)? (LINES TERMINATED BY separator=STRING_LITERAL)? (FORMAT AS format=identifierOrText)? diff --git a/regression-test/data/load_p0/broker_load/test_broker_load_with_partition.out b/regression-test/data/load_p0/broker_load/test_broker_load_with_partition.out index 5015d98b3c9..e62e67a8f74 100644 --- a/regression-test/data/load_p0/broker_load/test_broker_load_with_partition.out +++ b/regression-test/data/load_p0/broker_load/test_broker_load_with_partition.out @@ -8,3 +8,11 @@ 11001 2023-09-01 1 2 10 11001 2023-09-01 2 1 10 +-- !select -- +11001 2023-10-01 1 3 10 +11001 2023-10-01 2 2 10 +11001 2023-10-01 2 3 10 +11001 2023-09-01 1 1 10 +11001 2023-09-01 1 2 10 +11001 2023-09-01 2 1 10 + diff --git a/regression-test/suites/load_p0/broker_load/test_broker_load_with_partition.groovy b/regression-test/suites/load_p0/broker_load/test_broker_load_with_partition.groovy index 032b48baf60..45f0cc50be7 100644 --- a/regression-test/suites/load_p0/broker_load/test_broker_load_with_partition.groovy +++ b/regression-test/suites/load_p0/broker_load/test_broker_load_with_partition.groovy @@ -54,6 +54,13 @@ suite("test_broker_load_with_partition", "load_p0") { assertTrue(result2.size() == 1) assertTrue(result2[0].size() == 1) assertTrue(result2[0][0] == 1, "Insert should update 1 rows") + + def result3 = sql """ + ALTER TABLE ${testTable} ADD TEMPORARY PARTITION partition_t VALUES LESS THAN ("2023-11-01"); + """ + assertTrue(result3.size() == 1) + assertTrue(result3[0].size() == 1) + assertTrue(result3[0][0] == 0, "Alter table should update 0 rows") } def load_from_hdfs_partition = {testTablex, label, hdfsFilePath, format, brokerName, hdfsUser, hdfsPasswd -> @@ -78,6 +85,28 @@ suite("test_broker_load_with_partition", "load_p0") { assertTrue(result1[0][0] == 0, "Query OK, 0 rows affected") } + def load_from_hdfs_tmp_partition = {testTablex, label, hdfsFilePath, format, brokerName, hdfsUser, hdfsPasswd -> + def result1= sql """ + LOAD LABEL ${label} ( + DATA INFILE("${hdfsFilePath}") + INTO TABLE ${testTablex} + TEMPORARY PARTITION (`partition_t`) + COLUMNS TERMINATED BY "," + FORMAT as "${format}" + ) + with BROKER "${brokerName}" ( + "username"="${hdfsUser}", + "password"="${hdfsPasswd}") + PROPERTIES ( + "timeout"="1200", + "max_filter_ratio"="0.1"); + """ + + assertTrue(result1.size() == 1) + assertTrue(result1[0].size() == 1) + assertTrue(result1[0][0] == 0, "Query OK, 0 rows affected") + } + def check_load_result = {checklabel, testTablex -> max_try_milli_secs = 10000 while(max_try_milli_secs) { @@ -97,6 +126,25 @@ suite("test_broker_load_with_partition", "load_p0") { } } + def check_load_tmp_partition_result = {checklabel, testTablex -> + max_try_milli_secs = 10000 + while(max_try_milli_secs) { + result = sql "show load where label = '${checklabel}'" + log.info("result: ${result}") + if(result[0][2] == "FINISHED") { + //sql "sync" + qt_select "select * from ${testTablex} TEMPORARY PARTITION (`partition_t`) order by k1" + break + } else { + sleep(1000) // wait 1 second every time + max_try_milli_secs -= 1000 + if(max_try_milli_secs <= 0) { + assertEquals(1, 2) + } + } + } + } + // if 'enableHdfs' in regression-conf.groovy has been set to true, // the test will run these case as below. if (enableHdfs()) { @@ -115,6 +163,12 @@ suite("test_broker_load_with_partition", "load_p0") { brokerName, hdfsUser, hdfsPasswd) check_load_result.call(test_load_label, testTable) + + def test_tmp_partition_load_label = UUID.randomUUID().toString().replaceAll("-", "") + load_from_hdfs_tmp_partition.call(testTable, test_tmp_partition_load_label, + hdfs_csv_file_path, "csv", brokerName, hdfsUser, hdfsPasswd) + + check_load_tmp_partition_result.call(test_tmp_partition_load_label, testTable) } finally { try_sql("DROP TABLE IF EXISTS ${testTable}") } --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@doris.apache.org For additional commands, e-mail: commits-h...@doris.apache.org