This is an automated email from the ASF dual-hosted git repository.

hellostephen pushed a commit to branch branch-2.1
in repository https://gitbox.apache.org/repos/asf/doris.git


The following commit(s) were added to refs/heads/branch-2.1 by this push:
     new 9eac4f27978 [regression](case) fix typo (#38142)
9eac4f27978 is described below

commit 9eac4f279789dbeb7d3aefff02277ee5722b8502
Author: Dongyang Li <hello_step...@qq.com>
AuthorDate: Mon Jul 22 14:26:30 2024 +0800

    [regression](case) fix typo (#38142)
    
    ## Proposed changes
    
    Issue Number: close #xxx
    
    <!--Describe your changes.-->
    
    Co-authored-by: stephen <hello-step...@qq.com>
---
 .../suites/load_p0/tvf/test_s3_tvf.groovy          | 88 +++++++++++-----------
 1 file changed, 44 insertions(+), 44 deletions(-)

diff --git a/regression-test/suites/load_p0/tvf/test_s3_tvf.groovy 
b/regression-test/suites/load_p0/tvf/test_s3_tvf.groovy
index 6bf6c8f7e73..609abd37329 100644
--- a/regression-test/suites/load_p0/tvf/test_s3_tvf.groovy
+++ b/regression-test/suites/load_p0/tvf/test_s3_tvf.groovy
@@ -56,7 +56,7 @@ suite("test_s3_tvf", "load_p0") {
     /* normal */
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, ["K00", "K01", "K02", "K03", 
"K04", "K05", "K06", "K07", "K08", "K09", "K10", "K11", "K12", "K13", "K14", 
"K15", "K16", "K17", "K18"], "", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -64,14 +64,14 @@ suite("test_s3_tvf", "load_p0") {
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as 
k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 
as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as 
k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, 
TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, 
TO_QUANTILE_STATE(c5, 1.0) as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv")
+            .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv")
             .addProperty("format", "csv")
             .addProperty("column_separator", "|")
             .addProperty("force_parsing_by_standard_uri", "true"))
 
     for(String table : arrayTables) {
-        attributeList.add(new TvfAttribute(table, ["K00", "K01", "K02", "K03", 
"K04", "K05", "K06", "K07", "K08", "K09", "K10", "K11", "K12", "K13", "K14", 
"K15", "K16", "K17"], "", "").addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv")
+        attributeList.add(new TvfAttribute(table, ["K00", "K01", "K02", "K03", 
"K04", "K05", "K06", "K07", "K08", "K09", "K10", "K11", "K12", "K13", "K14", 
"K15", "K16", "K17"], "", "").addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -79,7 +79,7 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17", "k18"], "", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -87,14 +87,14 @@ suite("test_s3_tvf", "load_p0") {
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as 
k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 
as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as 
k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, 
TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, 
TO_QUANTILE_STATE(c5, 1.0) as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv")
+            .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv")
             .addProperty("format", "csv")
             .addProperty("column_separator", "|")
             .addProperty("force_parsing_by_standard_uri", "true"))
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17"], "", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -102,27 +102,27 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17", "k18"], "", "")
-                .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv";)
+                .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv";)
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|"))
     }
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as 
k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 
as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as 
k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, 
TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, 
TO_QUANTILE_STATE(c5, 1.0) as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv";)
+            .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv";)
             .addProperty("format", "csv")
             .addProperty("column_separator", "|"))
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17"], "", "")
-                .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv";)
+                .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv";)
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|"))
     }
 
 //    for(String table : basicTables) {
 //        attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", 
"k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", 
"k14", "k15", "k16", "k17", "k18"], "", "")
-//                .addProperty("uri", 
"https://{$s3Endpoint}/{$s3BucketName}/regression/load/data/basic_data.csv";)
+//                .addProperty("uri", 
"https://${s3Endpoint}/${s3BucketName}/regression/load/data/basic_data.csv";)
 //                .addProperty("format", "csv")
 //                .addProperty("column_separator", "|")
 //                .addProperty("use_path_style", "true"))
@@ -130,14 +130,14 @@ suite("test_s3_tvf", "load_p0") {
 //
 //    attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as 
k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 
as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as 
k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, 
TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, 
TO_QUANTILE_STATE(c5, 1.0) as kd21",
 //            
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-//            .addProperty("uri", 
"https://{$s3Endpoint}/{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv";)
+//            .addProperty("uri", 
"https://${s3Endpoint}/${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv";)
 //            .addProperty("format", "csv")
 //            .addProperty("column_separator", "|")
 //            .addProperty("use_path_style", "true"))
 //
 //    for(String table : arrayTables) {
 //        attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", 
"k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", 
"k14", "k15", "k16", "k17"], "", "")
-//                .addProperty("uri", 
"https://{$s3Endpoint}/{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv";)
+//                .addProperty("uri", 
"https://${s3Endpoint}/${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv";)
 //                .addProperty("format", "csv")
 //                .addProperty("column_separator", "|")
 //                .addProperty("use_path_style", "true"))
@@ -146,7 +146,7 @@ suite("test_s3_tvf", "load_p0") {
     /* error */
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17", "k18"], "", "", true)
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_with_errors.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_with_errors.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -154,14 +154,14 @@ suite("test_s3_tvf", "load_p0") {
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as 
k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 
as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as 
k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, 
TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, 
TO_QUANTILE_STATE(c5, 1.0) as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "", true)
-            .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_with_errors.csv")
+            .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_with_errors.csv")
             .addProperty("format", "csv")
             .addProperty("column_separator", "|")
             .addProperty("force_parsing_by_standard_uri", "true"))
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17", "kd01", "kd02", "kd03", "kd04", "kd05", "kd06", "kd07", 
"kd08", "kd09", "kd10", "kd11", "kd12", "kd13", "kd14", "kd15", "kd16"], "", 
"", true)
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data_with_errors.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data_with_errors.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -170,7 +170,7 @@ suite("test_s3_tvf", "load_p0") {
     /* skip lines */
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17", "k18"], "", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_with_errors.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_with_errors.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("skip_lines", "10")
@@ -179,7 +179,7 @@ suite("test_s3_tvf", "load_p0") {
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as 
k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 
as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as 
k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, 
TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, 
TO_QUANTILE_STATE(c5, 1.0) as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_with_errors.csv")
+            .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_with_errors.csv")
             .addProperty("format", "csv")
             .addProperty("column_separator", "|")
             .addProperty("skip_lines", "10")
@@ -187,7 +187,7 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17"], "", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data_with_errors.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data_with_errors.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("skip_lines", "10")
@@ -197,21 +197,21 @@ suite("test_s3_tvf", "load_p0") {
     /* compress type */
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17", "k18"], "", "")
-                .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.gz";)
+                .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.gz";)
                 .addProperty("format", "csv")
                 .addProperty("column_separator", 
"|").addProperty("compress_type", "GZ"))
     }
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as 
k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 
as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as 
k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, 
TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, 
TO_QUANTILE_STATE(c5, 1.0) as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.gz";)
+            .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.gz";)
             .addProperty("format", "csv")
             .addProperty("column_separator", "|")
             .addProperty("compress_type", "GZ"))
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17"], "", "")
-                .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv.gz";)
+                .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv.gz";)
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("compress_type", "GZ"))
@@ -219,21 +219,21 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17", "k18"], "", "")
-                .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.bz2";)
+                .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.bz2";)
                 .addProperty("format", "csv")
                 .addProperty("column_separator", 
"|").addProperty("compress_type", "BZ2"))
     }
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as 
k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 
as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as 
k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, 
TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, 
TO_QUANTILE_STATE(c5, 1.0) as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.bz2";)
+            .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.bz2";)
             .addProperty("format", "csv")
             .addProperty("column_separator", "|")
             .addProperty("compress_type", "BZ2"))
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17"], "", "")
-                .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv.bz2";)
+                .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv.bz2";)
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("compress_type", "BZ2"))
@@ -241,21 +241,21 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17", "k18"], "", "")
-                .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.lz4";)
+                .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.lz4";)
                 .addProperty("format", "csv")
                 .addProperty("column_separator", 
"|").addProperty("compress_type", "LZ4FRAME"))
     }
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as 
k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 
as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as 
k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, 
TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, 
TO_QUANTILE_STATE(c5, 1.0) as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.lz4";)
+            .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.lz4";)
             .addProperty("format", "csv")
             .addProperty("column_separator", "|")
             .addProperty("compress_type", "LZ4FRAME"))
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17"], "", "")
-                .addProperty("uri", 
"https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv.lz4";)
+                .addProperty("uri", 
"https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv.lz4";)
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("compress_type", "LZ4FRAME"))
@@ -263,7 +263,7 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17", "k18"], "WHERE c1 > 50", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -271,14 +271,14 @@ suite("test_s3_tvf", "load_p0") {
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as 
k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 
as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as 
k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, 
TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, 
TO_QUANTILE_STATE(c5, 1.0) as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"WHERE c1 > 50", "")
-            .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv")
+            .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv")
             .addProperty("format", "csv")
             .addProperty("column_separator", "|")
             .addProperty("force_parsing_by_standard_uri", "true"))
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17"], "WHERE c1 > 50", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -286,7 +286,7 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : uniqTable) {
         attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", 
"k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", 
"k15", "k16", "k17", "k18"], "", "ORDER BY c1")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv")
                 .addProperty("format", "csv")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -294,7 +294,7 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18","k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18",
 "", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.parq")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.parq")
                 .addProperty("format", "parquet")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -302,14 +302,14 @@ suite("test_s3_tvf", "load_p0") {
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18, 
to_bitmap(k05) as k19, HLL_HASH(k05) as k20, TO_QUANTILE_STATE(k04, 1.0) as 
k21, to_bitmap(k05) as kd19, HLL_HASH(k05) as kd20, TO_QUANTILE_STATE(k04, 1.0) 
as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.parq")
+            .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.parq")
             .addProperty("format", "parquet")
             .addProperty("column_separator", "|")
             .addProperty("force_parsing_by_standard_uri", "true"))
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "", 
"")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.parq")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.parq")
                 .addProperty("format", "parquet")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -317,7 +317,7 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18","k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18",
 "", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.orc")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.orc")
                 .addProperty("format", "orc")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -325,14 +325,14 @@ suite("test_s3_tvf", "load_p0") {
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18, 
to_bitmap(k05) as k19, HLL_HASH(k05) as k20, TO_QUANTILE_STATE(k04, 1.0) as 
k21, to_bitmap(k05) as kd19, HLL_HASH(k05) as kd20, TO_QUANTILE_STATE(k04, 1.0) 
as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.orc")
+            .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.orc")
             .addProperty("format", "orc")
             .addProperty("column_separator", "|")
             .addProperty("force_parsing_by_standard_uri", "true"))
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "", 
"")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.orc")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.orc")
                 .addProperty("format", "orc")
                 .addProperty("column_separator", "|")
                 .addProperty("force_parsing_by_standard_uri", "true"))
@@ -340,7 +340,7 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18","k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18",
 "", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.json")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.json")
                 .addProperty("format", "json")
                 .addProperty("read_json_by_line", "false")
                 .addProperty("strip_outer_array", "true")
@@ -350,7 +350,7 @@ suite("test_s3_tvf", "load_p0") {
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18, 
to_bitmap(k05) as k19, HLL_HASH(k05) as k20, TO_QUANTILE_STATE(k04, 1.0) as 
k21, to_bitmap(k05) as kd19, HLL_HASH(k05) as kd20, TO_QUANTILE_STATE(k04, 1.0) 
as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.json")
+            .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.json")
             .addProperty("format", "json")
             .addProperty("read_json_by_line", "false")
             .addProperty("strip_outer_array", "true")
@@ -359,7 +359,7 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "", 
"")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.json")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.json")
                 .addProperty("format", "json")
                 .addProperty("read_json_by_line", "false")
                 .addProperty("strip_outer_array", "true")
@@ -369,7 +369,7 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : basicTables) {
         attributeList.add(new TvfAttribute(table, 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18","k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18",
 "", "")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_by_line.json")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_by_line.json")
                 .addProperty("format", "json")
                 .addProperty("read_json_by_line", "true")
                 .addProperty("strip_outer_array", "false")
@@ -379,7 +379,7 @@ suite("test_s3_tvf", "load_p0") {
 
     attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18, 
to_bitmap(k05) as k19, HLL_HASH(k05) as k20, TO_QUANTILE_STATE(k04, 1.0) as 
k21, to_bitmap(k05) as kd19, HLL_HASH(k05) as kd20, TO_QUANTILE_STATE(k04, 1.0) 
as kd21",
             
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21"
 ,"", "")
-            .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_by_line.json")
+            .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_by_line.json")
             .addProperty("format", "json")
             .addProperty("read_json_by_line", "true")
             .addProperty("strip_outer_array", "false")
@@ -388,7 +388,7 @@ suite("test_s3_tvf", "load_p0") {
 
     for(String table : arrayTables) {
         attributeList.add(new TvfAttribute(table, 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", 
"k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "", 
"")
-                .addProperty("uri", 
"s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data_by_line.json")
+                .addProperty("uri", 
"s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data_by_line.json")
                 .addProperty("format", "json")
                 .addProperty("read_json_by_line", "true")
                 .addProperty("strip_outer_array", "false")


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@doris.apache.org
For additional commands, e-mail: commits-h...@doris.apache.org

Reply via email to