This is an automated email from the ASF dual-hosted git repository.

dataroaring pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/doris.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
     new 53d0aebcdda [fix](cases) fix some failure cases (#55740)
53d0aebcdda is described below

commit 53d0aebcdda6b53e0fb6e3f3434a729ec717c37c
Author: Yongqiang YANG <[email protected]>
AuthorDate: Sat Sep 6 22:58:18 2025 +0800

    [fix](cases) fix some failure cases (#55740)
---
 ...t_cloud_mow_stream_load_with_commit_fail.groovy | 22 +++++++++++-----------
 .../cloud/test_cloud_version_already_merged.groovy |  2 +-
 2 files changed, 12 insertions(+), 12 deletions(-)

diff --git 
a/regression-test/suites/fault_injection_p0/cloud/test_cloud_mow_stream_load_with_commit_fail.groovy
 
b/regression-test/suites/fault_injection_p0/cloud/test_cloud_mow_stream_load_with_commit_fail.groovy
index 77555980dba..a5c713591d0 100644
--- 
a/regression-test/suites/fault_injection_p0/cloud/test_cloud_mow_stream_load_with_commit_fail.groovy
+++ 
b/regression-test/suites/fault_injection_p0/cloud/test_cloud_mow_stream_load_with_commit_fail.groovy
@@ -168,7 +168,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
 
             time 10000 // limit inflight 10s
 
-            directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+            directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
             check { result, exception, startTime, endTime ->
                 log.info("Stream load result: ${result}")
@@ -193,7 +193,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
             file "test_stream_load2.csv"
 
             time 10000 // limit inflight 10s
-            directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+            directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
             check { result, exception, startTime, endTime ->
                 log.info("Stream load result: ${result}")
@@ -218,7 +218,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
             file "test_stream_load2.csv"
 
             time 10000 // limit inflight 10s
-            directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+            directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
             check { result, exception, startTime, endTime ->
                 log.info("Stream load result: ${result}")
@@ -244,7 +244,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
                 file "test_stream_load3.csv"
 
                 time 10000 // limit inflight 10s
-                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
                 check { result, exception, startTime, endTime ->
                     log.info("Stream load result: ${result}")
@@ -268,7 +268,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
                 file "test_stream_load3.csv"
 
                 time 10000 // limit inflight 10s
-                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
                 check { result, exception, startTime, endTime ->
                     log.info("Stream load result: ${result}")
@@ -296,7 +296,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
                 file "test_stream_load4.csv"
 
                 time 10000 // limit inflight 10s
-                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
                 check { result, exception, startTime, endTime ->
                     log.info("Stream load result: ${result}")
@@ -321,7 +321,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
                 file "test_stream_load4.csv"
 
                 time 10000 // limit inflight 10s
-                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
                 check { result, exception, startTime, endTime ->
                     log.info("Stream load result: ${result}")
@@ -346,7 +346,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
                 file "test_stream_load5.csv"
 
                 time 10000 // limit inflight 10s
-                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
                 check { result, exception, startTime, endTime ->
                     log.info("Stream load result: ${result}")
@@ -370,7 +370,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
                 file "test_stream_load5.csv"
 
                 time 10000 // limit inflight 10s
-                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
                 check { result, exception, startTime, endTime ->
                     log.info("Stream load result: ${result}")
@@ -397,7 +397,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
                 file "test_stream_load6.csv"
 
                 time 10000 // limit inflight 10s
-                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
                 check { result, exception, startTime, endTime ->
                     log.info("Stream load result: ${result}")
@@ -421,7 +421,7 @@ suite("test_cloud_mow_stream_load_with_commit_fail", 
"nonConcurrent") {
                 file "test_stream_load.csv"
 
                 time 10000 // limit inflight 10s
-                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId)
+                directToBe backendIdToBackendIP.get(backendId), 
backendId_to_backendHttpPort.get(backendId).toInteger()
 
                 check { result, exception, startTime, endTime ->
                     log.info("Stream load result: ${result}")
diff --git 
a/regression-test/suites/fault_injection_p0/cloud/test_cloud_version_already_merged.groovy
 
b/regression-test/suites/fault_injection_p0/cloud/test_cloud_version_already_merged.groovy
index 111fe556dfc..b6e106c7600 100644
--- 
a/regression-test/suites/fault_injection_p0/cloud/test_cloud_version_already_merged.groovy
+++ 
b/regression-test/suites/fault_injection_p0/cloud/test_cloud_version_already_merged.groovy
@@ -86,7 +86,7 @@ suite("test_cloud_version_already_merged", "nonConcurrent") {
 
         test {
             sql """ SELECT * from ${tblName} ORDER BY k1 """
-            exception "version already merged, meet error during remote 
capturing rowsets"
+            exception "versions are already compacted"
         }
 
     } finally {


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to