This is an automated email from the ASF dual-hosted git repository.

dataroaring pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/doris.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
     new 07ec37d8e17 branch-3.0: [test](warmup) fix 
test_warm_up_cluster_event_rename #55145 (#55196)
07ec37d8e17 is described below

commit 07ec37d8e178bbbb49b5a9389b005b471cc4e32d
Author: github-actions[bot] 
<41898282+github-actions[bot]@users.noreply.github.com>
AuthorDate: Sun Aug 24 15:26:28 2025 +0800

    branch-3.0: [test](warmup) fix test_warm_up_cluster_event_rename #55145 
(#55196)
    
    Cherry-picked from #55145
    
    Co-authored-by: Kaijie Chen <[email protected]>
---
 .../test_warm_up_cluster_event_rename.groovy        | 21 +++++++++++++++++----
 1 file changed, 17 insertions(+), 4 deletions(-)

diff --git 
a/regression-test/suites/cloud_p0/cache/multi_cluster/warm_up/cluster/test_warm_up_cluster_event_rename.groovy
 
b/regression-test/suites/cloud_p0/cache/multi_cluster/warm_up/cluster/test_warm_up_cluster_event_rename.groovy
index fc3ac52ae91..857eac205f2 100644
--- 
a/regression-test/suites/cloud_p0/cache/multi_cluster/warm_up/cluster/test_warm_up_cluster_event_rename.groovy
+++ 
b/regression-test/suites/cloud_p0/cache/multi_cluster/warm_up/cluster/test_warm_up_cluster_event_rename.groovy
@@ -89,14 +89,26 @@ suite('test_warm_up_cluster_event_rename', 'docker') {
             def submitted_segment = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_submitted_segment_num")
             def finished_segment = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_finished_segment_num")
             def failed_segment = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_failed_segment_num")
+            def submitted_segment_size = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_submitted_segment_size")
+            def finished_segment_size = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_finished_segment_size")
+            def failed_segment_size = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_failed_segment_size")
             def submitted_index = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_submitted_index_num")
             def finished_index = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_finished_index_num")
             def failed_index = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_failed_index_num")
+            def submitted_index_size = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_submitted_index_size")
+            def finished_index_size = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_finished_index_size")
+            def failed_index_size = getBrpcMetrics(ip, port, 
"file_cache_event_driven_warm_up_failed_index_size")
             logger.info("${cluster} be ${ip}:${port}, 
submitted_segment=${submitted_segment}"
                     + ", finished_segment=${finished_segment}, 
failed_segment=${failed_segment}"
+                    + ", submitted_segment_size=${submitted_segment_size}"
+                    + ", finished_segment_size=${finished_segment_size}"
+                    + ", failed_segment_size=${failed_segment_size}"
                     + ", submitted_index=${submitted_index}"
                     + ", finished_index=${finished_index}"
-                    + ", failed_index=${failed_index}")
+                    + ", failed_index=${failed_index}"
+                    + ", submitted_index_size=${submitted_index_size}"
+                    + ", finished_index_size=${finished_index_size}"
+                    + ", failed_index_size=${failed_index_size}")
         }
     }
 
@@ -174,7 +186,8 @@ suite('test_warm_up_cluster_event_rename', 'docker') {
         sql """use @${clusterName1}"""
 
         // Simple setup to simulate data load and access
-        sql """CREATE TABLE IF NOT EXISTS customer (id INT, name STRING) 
DUPLICATE KEY(id) DISTRIBUTED BY HASH(id) BUCKETS 3 PROPERTIES 
("file_cache_ttl_seconds" = "3600")"""
+        sql """CREATE TABLE IF NOT EXISTS customer (id INT, name STRING) 
DUPLICATE KEY(id) DISTRIBUTED BY HASH(id) BUCKETS 3
+        PROPERTIES ("file_cache_ttl_seconds" = "3600", 
"disable_auto_compaction" = "true")"""
 
         // Start warm up job
         def jobId_ = sql """
@@ -199,8 +212,8 @@ suite('test_warm_up_cluster_event_rename', 'docker') {
         logFileCacheDownloadMetrics(clusterName2)
         checkTTLCacheSizeSumEqual(clusterName1, clusterName2)
 
-        srcSumOld = getClusterTTLCacheSizeSum(clusterName1)
-        dstSumOld = getClusterTTLCacheSizeSum(clusterName2)
+        def srcSumOld = getClusterTTLCacheSizeSum(clusterName1)
+        def dstSumOld = getClusterTTLCacheSizeSum(clusterName2)
 
         // rename
         sql """ALTER SYSTEM RENAME COMPUTE GROUP ${clusterName2} 
${clusterName3}"""


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to