This is an automated email from the ASF dual-hosted git repository.

yiguolei pushed a commit to branch branch-4.0
in repository https://gitbox.apache.org/repos/asf/doris.git


The following commit(s) were added to refs/heads/branch-4.0 by this push:
     new cea96b8aeff branch-4.0: [Fix](regress-test) fix streaming job for s3 
docker case #60053 (#60086)
cea96b8aeff is described below

commit cea96b8aeff3ee8d992e64392120833553a9fc61
Author: github-actions[bot] 
<41898282+github-actions[bot]@users.noreply.github.com>
AuthorDate: Wed Jan 21 17:47:24 2026 +0800

    branch-4.0: [Fix](regress-test) fix streaming job for s3 docker case #60053 
(#60086)
    
    Cherry-picked from #60053
    
    Co-authored-by: wudi <[email protected]>
---
 .../test_streaming_job_alter_offset_restart_fe.groovy             | 8 ++++----
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git 
a/regression-test/suites/job_p0/streaming_job/test_streaming_job_alter_offset_restart_fe.groovy
 
b/regression-test/suites/job_p0/streaming_job/test_streaming_job_alter_offset_restart_fe.groovy
index 66e18248774..02d44f80a97 100644
--- 
a/regression-test/suites/job_p0/streaming_job/test_streaming_job_alter_offset_restart_fe.groovy
+++ 
b/regression-test/suites/job_p0/streaming_job/test_streaming_job_alter_offset_restart_fe.groovy
@@ -93,7 +93,7 @@ suite("test_streaming_job_alter_offset_restart_fe", "docker") 
{
         log.info("jobInfo: " + jobInfo)
         assert jobInfo.get(0).get(0) == 
"{\"fileName\":\"regression/load/data/example_1.csv\"}";
         assert jobInfo.get(0).get(1) == 
"{\"fileName\":\"regression/load/data/example_1.csv\"}";
-        assert jobInfo.get(0).get(2) == 
"{\"scannedRows\":10,\"loadBytes\":218,\"fileNumber\":0,\"fileSize\":0}"
+        assert jobInfo.get(0).get(2) == 
"{\"scannedRows\":10,\"loadBytes\":218,\"fileNumber\":1,\"fileSize\":138}"
 
         sql """
             PAUSE JOB where jobname =  '${jobName}'
@@ -112,7 +112,7 @@ suite("test_streaming_job_alter_offset_restart_fe", 
"docker") {
         """
         log.info("jobInfo: " + jobInfo)
         assert jobInfo.get(0).get(0) == 
"{\"fileName\":\"regression/load/data/anoexist1234.csv\"}";
-        assert jobInfo.get(0).get(1) == 
"{\"scannedRows\":10,\"loadBytes\":218,\"fileNumber\":0,\"fileSize\":0}"
+        assert jobInfo.get(0).get(1) == 
"{\"scannedRows\":10,\"loadBytes\":218,\"fileNumber\":1,\"fileSize\":138}"
         assert jobInfo.get(0).get(2) == 
"{\"offset\":\"{\\\"fileName\\\":\\\"regression/load/data/anoexist1234.csv\\\"}\"}"
 
         // Restart FE
@@ -131,7 +131,7 @@ suite("test_streaming_job_alter_offset_restart_fe", 
"docker") {
         """
         log.info("jobInfo: " + jobInfo)
         assert jobInfo.get(0).get(0) == 
"{\"fileName\":\"regression/load/data/anoexist1234.csv\"}";
-        assert jobInfo.get(0).get(1) == 
"{\"scannedRows\":10,\"loadBytes\":218,\"fileNumber\":0,\"fileSize\":0}"
+        assert jobInfo.get(0).get(1) == 
"{\"scannedRows\":10,\"loadBytes\":218,\"fileNumber\":1,\"fileSize\":138}"
         assert jobInfo.get(0).get(2) == 
"{\"offset\":\"{\\\"fileName\\\":\\\"regression/load/data/anoexist1234.csv\\\"}\"}"
 
         // resume to check whether consumption will resume
@@ -163,7 +163,7 @@ suite("test_streaming_job_alter_offset_restart_fe", 
"docker") {
         log.info("jobInfo: " + jobInfo)
         assert jobInfo.get(0).get(0) == 
"{\"fileName\":\"regression/load/data/example_1.csv\"}";
         assert jobInfo.get(0).get(1) == 
"{\"fileName\":\"regression/load/data/example_1.csv\"}";
-        assert jobInfo.get(0).get(2) == 
"{\"scannedRows\":30,\"loadBytes\":643,\"fileNumber\":0,\"fileSize\":0}"
+        assert jobInfo.get(0).get(2) == 
"{\"scannedRows\":30,\"loadBytes\":643,\"fileNumber\":3,\"fileSize\":394}"
         assert jobInfo.get(0).get(3) == 
"{\"offset\":\"{\\\"fileName\\\":\\\"regression/load/data/anoexist1234.csv\\\"}\"}"
 
         sql """ DROP JOB IF EXISTS where jobname =  '${jobName}' """


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to