This is an automated email from the ASF dual-hosted git repository. kxiao pushed a commit to branch branch-2.0-beta in repository https://gitbox.apache.org/repos/asf/doris.git
commit f20d3c4ee9d911052ae1cec4ac1d1e7e5d3ae9c9 Author: zclllyybb <zhaochan...@selectdb.com> AuthorDate: Tue Jun 6 09:57:38 2023 +0800 [fix](regression-test) Add lost ddl file for tpcds_sf1_p2 #20288 --- .../suites/tpcds_sf1_p2/ddl/catalog_page.sql | 17 +++++++++++ regression-test/suites/tpcds_sf1_p2/load.groovy | 33 +--------------------- 2 files changed, 18 insertions(+), 32 deletions(-) diff --git a/regression-test/suites/tpcds_sf1_p2/ddl/catalog_page.sql b/regression-test/suites/tpcds_sf1_p2/ddl/catalog_page.sql new file mode 100644 index 0000000000..7b06b67fa5 --- /dev/null +++ b/regression-test/suites/tpcds_sf1_p2/ddl/catalog_page.sql @@ -0,0 +1,17 @@ +CREATE TABLE IF NOT EXISTS catalog_page ( + cp_catalog_page_sk bigint, + cp_catalog_page_id char(16), + cp_start_date_sk integer, + cp_end_date_sk integer, + cp_department varchar(50), + cp_catalog_number integer, + cp_catalog_page_number integer, + cp_description varchar(100), + cp_type varchar(100) +) +DUPLICATE KEY(cp_catalog_page_sk, cp_catalog_page_id) +DISTRIBUTED BY HASH(cp_catalog_page_sk) BUCKETS 3 +PROPERTIES ( + "replication_num" = "1" +) + diff --git a/regression-test/suites/tpcds_sf1_p2/load.groovy b/regression-test/suites/tpcds_sf1_p2/load.groovy index 1b527dcc81..b8e5e059d3 100644 --- a/regression-test/suites/tpcds_sf1_p2/load.groovy +++ b/regression-test/suites/tpcds_sf1_p2/load.groovy @@ -19,33 +19,7 @@ // /testing/trino-product-tests/src/main/resources/sql-tests/testcases/tpcds // and modified by Doris. suite("load") { - def tables=["store", "store_returns", "customer", "date_dim", "web_sales", - "catalog_sales", "store_sales", "item", "web_returns", "catalog_returns", - "catalog_page", "web_site", "customer_address", "customer_demographics", - "ship_mode", "promotion", "inventory", "time_dim", "income_band", - "call_center", "reason", "household_demographics", "warehouse", "web_page"] - def columnsMap = [ - "item": """tmp_item_sk, tmp_item_id, tmp_rec_start_date, tmp_rec_end_date, tmp_item_desc, - tmp_current_price, tmp_wholesale_cost, tmp_brand_id, tmp_brand, tmp_class_id, tmp_class, - tmp_category_id, tmp_category, tmp_manufact_id, tmp_manufact, tmp_size, tmp_formulation, - tmp_color, tmp_units, tmp_container, tmp_manager_id, tmp_product_name, - i_item_sk=tmp_item_sk, i_item_id=tmp_item_id, i_rec_start_date=tmp_rec_start_date, - i_rec_end_date=tmp_rec_end_date, i_item_desc=tmp_item_desc, i_current_price=tmp_current_price, - i_wholesale_cost=tmp_wholesale_cost, i_brand_id=tmp_brand_id, i_brand=tmp_brand, - i_class_id=tmp_class_id, i_class=tmp_class, i_category_id=tmp_category_id, - i_category=nullif(tmp_category, ''), i_manufact_id=tmp_manufact_id, i_manufact=tmp_manufact, - i_size=tmp_size, i_formulation=tmp_formulation, i_color=tmp_color, i_units=tmp_units, - i_container=tmp_container, i_manager_id=tmp_manager_id, i_product_name=tmp_product_name""", - - "customer_address": """tmp_address_sk, tmp_address_id, tmp_street_number, tmp_street_name, tmp_street_type, tmp_suite_number, - tmp_city, tmp_county, tmp_state, tmp_zip, tmp_country, tmp_gmt_offset, tmp_location_type, - ca_address_sk=tmp_address_sk, ca_address_id=tmp_address_id, ca_street_number=tmp_street_number, - ca_street_name=tmp_street_name, ca_street_type=tmp_street_type, ca_suite_number=tmp_suite_number, ca_city=tmp_city, - ca_county=nullif(tmp_county, ''), ca_state=tmp_state, ca_zip=tmp_zip, ca_country=tmp_country, - ca_gmt_offset=tmp_gmt_offset, ca_location_type=tmp_location_type""", - ] - - def specialTables = ["item", "customer_address"] + def tables=["catalog_page"] for (String table in tables) { sql """ DROP TABLE IF EXISTS $table """ @@ -72,11 +46,6 @@ suite("load") { set 'column_separator', '|' set 'compress_type', 'GZ' - if (specialTables.contains(tableName)) { - set "columns", columnsMap[tableName] - } - - // relate to ${DORIS_HOME}/regression-test/data/demo/streamload_input.csv. // also, you can stream load a http stream, e.g. http://xxx/some.csv file """${getS3Url()}/regression/tpcds/sf1/${tableName}.dat.gz""" --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@doris.apache.org For additional commands, e-mail: commits-h...@doris.apache.org