This is an automated email from the ASF dual-hosted git repository.
dataroaring pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/doris.git
The following commit(s) were added to refs/heads/branch-3.0 by this push:
new 07ec37d8e17 branch-3.0: [test](warmup) fix
test_warm_up_cluster_event_rename #55145 (#55196)
07ec37d8e17 is described below
commit 07ec37d8e178bbbb49b5a9389b005b471cc4e32d
Author: github-actions[bot]
<41898282+github-actions[bot]@users.noreply.github.com>
AuthorDate: Sun Aug 24 15:26:28 2025 +0800
branch-3.0: [test](warmup) fix test_warm_up_cluster_event_rename #55145
(#55196)
Cherry-picked from #55145
Co-authored-by: Kaijie Chen <[email protected]>
---
.../test_warm_up_cluster_event_rename.groovy | 21 +++++++++++++++++----
1 file changed, 17 insertions(+), 4 deletions(-)
diff --git
a/regression-test/suites/cloud_p0/cache/multi_cluster/warm_up/cluster/test_warm_up_cluster_event_rename.groovy
b/regression-test/suites/cloud_p0/cache/multi_cluster/warm_up/cluster/test_warm_up_cluster_event_rename.groovy
index fc3ac52ae91..857eac205f2 100644
---
a/regression-test/suites/cloud_p0/cache/multi_cluster/warm_up/cluster/test_warm_up_cluster_event_rename.groovy
+++
b/regression-test/suites/cloud_p0/cache/multi_cluster/warm_up/cluster/test_warm_up_cluster_event_rename.groovy
@@ -89,14 +89,26 @@ suite('test_warm_up_cluster_event_rename', 'docker') {
def submitted_segment = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_submitted_segment_num")
def finished_segment = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_finished_segment_num")
def failed_segment = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_failed_segment_num")
+ def submitted_segment_size = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_submitted_segment_size")
+ def finished_segment_size = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_finished_segment_size")
+ def failed_segment_size = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_failed_segment_size")
def submitted_index = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_submitted_index_num")
def finished_index = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_finished_index_num")
def failed_index = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_failed_index_num")
+ def submitted_index_size = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_submitted_index_size")
+ def finished_index_size = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_finished_index_size")
+ def failed_index_size = getBrpcMetrics(ip, port,
"file_cache_event_driven_warm_up_failed_index_size")
logger.info("${cluster} be ${ip}:${port},
submitted_segment=${submitted_segment}"
+ ", finished_segment=${finished_segment},
failed_segment=${failed_segment}"
+ + ", submitted_segment_size=${submitted_segment_size}"
+ + ", finished_segment_size=${finished_segment_size}"
+ + ", failed_segment_size=${failed_segment_size}"
+ ", submitted_index=${submitted_index}"
+ ", finished_index=${finished_index}"
- + ", failed_index=${failed_index}")
+ + ", failed_index=${failed_index}"
+ + ", submitted_index_size=${submitted_index_size}"
+ + ", finished_index_size=${finished_index_size}"
+ + ", failed_index_size=${failed_index_size}")
}
}
@@ -174,7 +186,8 @@ suite('test_warm_up_cluster_event_rename', 'docker') {
sql """use @${clusterName1}"""
// Simple setup to simulate data load and access
- sql """CREATE TABLE IF NOT EXISTS customer (id INT, name STRING)
DUPLICATE KEY(id) DISTRIBUTED BY HASH(id) BUCKETS 3 PROPERTIES
("file_cache_ttl_seconds" = "3600")"""
+ sql """CREATE TABLE IF NOT EXISTS customer (id INT, name STRING)
DUPLICATE KEY(id) DISTRIBUTED BY HASH(id) BUCKETS 3
+ PROPERTIES ("file_cache_ttl_seconds" = "3600",
"disable_auto_compaction" = "true")"""
// Start warm up job
def jobId_ = sql """
@@ -199,8 +212,8 @@ suite('test_warm_up_cluster_event_rename', 'docker') {
logFileCacheDownloadMetrics(clusterName2)
checkTTLCacheSizeSumEqual(clusterName1, clusterName2)
- srcSumOld = getClusterTTLCacheSizeSum(clusterName1)
- dstSumOld = getClusterTTLCacheSizeSum(clusterName2)
+ def srcSumOld = getClusterTTLCacheSizeSum(clusterName1)
+ def dstSumOld = getClusterTTLCacheSizeSum(clusterName2)
// rename
sql """ALTER SYSTEM RENAME COMPUTE GROUP ${clusterName2}
${clusterName3}"""
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]