This is an automated email from the ASF dual-hosted git repository.

kassiez pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/doris-website.git


The following commit(s) were added to refs/heads/master by this push:
     new c85b4bd04de [doc](flink) Update flink doc (#2286)
c85b4bd04de is described below

commit c85b4bd04de83bd77c8b572a2ce1d0ce006388c5
Author: wudi <w...@selectdb.com>
AuthorDate: Tue Apr 15 14:16:26 2025 +0800

    [doc](flink) Update flink doc (#2286)
    
    ## Versions
    
    - [x] dev
    - [x] 3.0
    - [x] 2.1
    - [ ] 2.0
    
    ## Languages
    
    - [x] Chinese
    - [x] English
    
    ## Docs Checklist
    
    - [ ] Checked by AI
    - [ ] Test Cases Built
---
 docs/ecosystem/flink-doris-connector.md            | 22 +++++++++++-----------
 .../current/ecosystem/flink-doris-connector.md     | 22 +++++++++++-----------
 .../version-2.1/ecosystem/flink-doris-connector.md | 22 +++++++++++-----------
 .../version-3.0/ecosystem/flink-doris-connector.md | 22 +++++++++++-----------
 .../version-2.1/ecosystem/flink-doris-connector.md | 22 +++++++++++-----------
 .../version-3.0/ecosystem/flink-doris-connector.md | 22 +++++++++++-----------
 6 files changed, 66 insertions(+), 66 deletions(-)

diff --git a/docs/ecosystem/flink-doris-connector.md 
b/docs/ecosystem/flink-doris-connector.md
index c8ef9ed37bf..3ed37f5f540 100644
--- a/docs/ecosystem/flink-doris-connector.md
+++ b/docs/ecosystem/flink-doris-connector.md
@@ -220,7 +220,7 @@ When Flink reads data from Doris, the Doris Source is 
currently a bounded stream
 ##### Thrift Method
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -228,18 +228,18 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '127.0.0.1:8030',  -- Fe的host:HttpPort
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 ##### ArrowFlightSQL
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -247,14 +247,14 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '{fe.conf:http_port}', 
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'source.use-flight-sql' = 'true',
       'source.flight-sql-port' = '{fe.conf:arrow_flight_sql_port}',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 #### Using DataStream API to Read Data
@@ -265,7 +265,7 @@ When using the DataStream API to read data, you need to 
include the dependencies
 final StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment();
 DorisOptions option = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("")
         .build();
@@ -324,7 +324,7 @@ CREATE TABLE student_sink (
     WITH (
       'connector' = 'doris',
       'fenodes' = '10.16.10.6:28737',
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = 'password',
       'sink.label-prefix' = 'doris_label'
@@ -406,7 +406,7 @@ properties.setProperty("format", "csv");
 DorisOptions.Builder dorisBuilder = DorisOptions.builder();
 dorisBuilder
         .setFenodes("10.16.10.6:28737")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("");
 DorisExecutionOptions.Builder executionBuilder = 
DorisExecutionOptions.builder();
@@ -467,7 +467,7 @@ props.setProperty("format", "json");
 props.setProperty("read_json_by_line", "true");
 DorisOptions dorisOptions = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("").build();
 
@@ -578,7 +578,7 @@ The Flink Doris Connector integrates **Flink CDC** ([Flink 
CDC Documentation](ht
 
 :::info Note
 
-1. When using full database synchronization, you need to add the corresponding 
Flink CDC dependencies in the `$FLINK_HOME/lib` directory, such as 
**flink-sql-connector-mysql-cdc-${version}.jar**, 
**flink-sql-connector-oracle-cdc-${version}.jar**. FlinkCDC version 3.1 and 
later is not compatible with previous versions. You can download the 
dependencies from the following links: [FlinkCDC 
3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),
 [FlinkCDC  [...]
+1. When using full database synchronization, you need to add the corresponding 
Flink CDC dependencies in the `$FLINK_HOME/lib` directory (Fat Jar), such as 
**flink-sql-connector-mysql-cdc-${version}.jar**, 
**flink-sql-connector-oracle-cdc-${version}.jar**. FlinkCDC version 3.1 and 
later is not compatible with previous versions. You can download the 
dependencies from the following links: [FlinkCDC 
3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),
  [...]
 2. For versions after Connector 24.0.0, the required Flink CDC version must be 
3.1 or higher. You can download it 
[here](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/).
 If Flink CDC is used to synchronize MySQL and Oracle, you must also add the 
relevant JDBC drivers under `$FLINK_HOME/lib`.
 
 #### MySQL Whole Database Synchronization
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/ecosystem/flink-doris-connector.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/ecosystem/flink-doris-connector.md
index 914c5def6a8..7c2e806947d 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/ecosystem/flink-doris-connector.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/ecosystem/flink-doris-connector.md
@@ -223,7 +223,7 @@ Flink 读取 Doris 中数据时,目前 Doris Source 是有界流,不支持
 ##### Thrift 方式
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -231,18 +231,18 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '127.0.0.1:8030',  -- Fe的host:HttpPort
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 ##### ArrowFlightSQL 方式
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -250,14 +250,14 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '{fe.conf:http_port}', 
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'source.use-flight-sql' = 'true',
       'source.flight-sql-port' = '{fe.conf:arrow_flight_sql_port}',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 #### 使用 DataStream API 读取数据
@@ -268,7 +268,7 @@ SELECT * FROM students;
 final StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment();
 DorisOptions option = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("")
         .build();
@@ -327,7 +327,7 @@ CREATE TABLE student_sink (
     WITH (
       'connector' = 'doris',
       'fenodes' = '10.16.10.6:28737',
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = 'password',
       'sink.label-prefix' = 'doris_label'
@@ -409,7 +409,7 @@ properties.setProperty("format", "csv");
 DorisOptions.Builder dorisBuilder = DorisOptions.builder();
 dorisBuilder
         .setFenodes("10.16.10.6:28737")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("");
 DorisExecutionOptions.Builder executionBuilder = 
DorisExecutionOptions.builder();
@@ -470,7 +470,7 @@ props.setProperty("format", "json");
 props.setProperty("read_json_by_line", "true");
 DorisOptions dorisOptions = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("").build();
 
@@ -581,7 +581,7 @@ Flink Doris Connector 中集成了[Flink 
CDC](https://nightlies.apache.org/flink
 
 :::info 注意
 
-1. 使用整库同步时需要在 `$FLINK_HOME/lib` 目录下添加对应的 Flink CDC 依赖,比如 
**`flink-sql-connector-mysql-cdc-${version}.jar`**,**`flink-sql-connector-oracle-cdc-${version}.jar`**,FlinkCDC
 从 3.1 
版本与之前版本不兼容,下载地址分别为为[FlinkCDC3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),[FlinkCDC
 
2.x](https://repo.maven.apache.org/maven2/com/ververica/flink-sql-connector-mysql-cdc/)。
+1. 使用整库同步时需要在 `$FLINK_HOME/lib` 目录下添加对应的 Flink CDC 依赖 (Fat Jar),比如 
**`flink-sql-connector-mysql-cdc-${version}.jar`**,**`flink-sql-connector-oracle-cdc-${version}.jar`**,FlinkCDC
 从 3.1 
版本与之前版本不兼容,下载地址分别为为[FlinkCDC3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),[FlinkCDC
 
2.x](https://repo.maven.apache.org/maven2/com/ververica/flink-sql-connector-mysql-cdc/)。
 2. Connector 24.0.0 之后依赖的 Flink CDC 版本需要在 3.1 
以上,下载地址见[这里](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),FlinkCDC
 如果需使用 Flink CDC 同步 MySQL 和 Oracle,还需要在 `$FLINK_HOME/lib` 下增加相关的 JDBC 驱动。
 
 :::
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/ecosystem/flink-doris-connector.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/ecosystem/flink-doris-connector.md
index 914c5def6a8..7c2e806947d 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/ecosystem/flink-doris-connector.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/ecosystem/flink-doris-connector.md
@@ -223,7 +223,7 @@ Flink 读取 Doris 中数据时,目前 Doris Source 是有界流,不支持
 ##### Thrift 方式
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -231,18 +231,18 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '127.0.0.1:8030',  -- Fe的host:HttpPort
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 ##### ArrowFlightSQL 方式
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -250,14 +250,14 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '{fe.conf:http_port}', 
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'source.use-flight-sql' = 'true',
       'source.flight-sql-port' = '{fe.conf:arrow_flight_sql_port}',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 #### 使用 DataStream API 读取数据
@@ -268,7 +268,7 @@ SELECT * FROM students;
 final StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment();
 DorisOptions option = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("")
         .build();
@@ -327,7 +327,7 @@ CREATE TABLE student_sink (
     WITH (
       'connector' = 'doris',
       'fenodes' = '10.16.10.6:28737',
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = 'password',
       'sink.label-prefix' = 'doris_label'
@@ -409,7 +409,7 @@ properties.setProperty("format", "csv");
 DorisOptions.Builder dorisBuilder = DorisOptions.builder();
 dorisBuilder
         .setFenodes("10.16.10.6:28737")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("");
 DorisExecutionOptions.Builder executionBuilder = 
DorisExecutionOptions.builder();
@@ -470,7 +470,7 @@ props.setProperty("format", "json");
 props.setProperty("read_json_by_line", "true");
 DorisOptions dorisOptions = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("").build();
 
@@ -581,7 +581,7 @@ Flink Doris Connector 中集成了[Flink 
CDC](https://nightlies.apache.org/flink
 
 :::info 注意
 
-1. 使用整库同步时需要在 `$FLINK_HOME/lib` 目录下添加对应的 Flink CDC 依赖,比如 
**`flink-sql-connector-mysql-cdc-${version}.jar`**,**`flink-sql-connector-oracle-cdc-${version}.jar`**,FlinkCDC
 从 3.1 
版本与之前版本不兼容,下载地址分别为为[FlinkCDC3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),[FlinkCDC
 
2.x](https://repo.maven.apache.org/maven2/com/ververica/flink-sql-connector-mysql-cdc/)。
+1. 使用整库同步时需要在 `$FLINK_HOME/lib` 目录下添加对应的 Flink CDC 依赖 (Fat Jar),比如 
**`flink-sql-connector-mysql-cdc-${version}.jar`**,**`flink-sql-connector-oracle-cdc-${version}.jar`**,FlinkCDC
 从 3.1 
版本与之前版本不兼容,下载地址分别为为[FlinkCDC3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),[FlinkCDC
 
2.x](https://repo.maven.apache.org/maven2/com/ververica/flink-sql-connector-mysql-cdc/)。
 2. Connector 24.0.0 之后依赖的 Flink CDC 版本需要在 3.1 
以上,下载地址见[这里](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),FlinkCDC
 如果需使用 Flink CDC 同步 MySQL 和 Oracle,还需要在 `$FLINK_HOME/lib` 下增加相关的 JDBC 驱动。
 
 :::
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/ecosystem/flink-doris-connector.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/ecosystem/flink-doris-connector.md
index ac945ce24b4..ebd79b091e8 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/ecosystem/flink-doris-connector.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/ecosystem/flink-doris-connector.md
@@ -223,7 +223,7 @@ Flink 读取 Doris 中数据时,目前 Doris Source 是有界流,不支持
 ##### Thrift 方式
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -231,18 +231,18 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '127.0.0.1:8030',  -- Fe的host:HttpPort
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 ##### ArrowFlightSQL 方式
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -250,14 +250,14 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '{fe.conf:http_port}', 
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'source.use-flight-sql' = 'true',
       'source.flight-sql-port' = '{fe.conf:arrow_flight_sql_port}',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 #### 使用 DataStream API 读取数据
@@ -268,7 +268,7 @@ SELECT * FROM students;
 final StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment();
 DorisOptions option = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("")
         .build();
@@ -327,7 +327,7 @@ CREATE TABLE student_sink (
     WITH (
       'connector' = 'doris',
       'fenodes' = '10.16.10.6:28737',
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = 'password',
       'sink.label-prefix' = 'doris_label'
@@ -409,7 +409,7 @@ properties.setProperty("format", "csv");
 DorisOptions.Builder dorisBuilder = DorisOptions.builder();
 dorisBuilder
         .setFenodes("10.16.10.6:28737")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("");
 DorisExecutionOptions.Builder executionBuilder = 
DorisExecutionOptions.builder();
@@ -470,7 +470,7 @@ props.setProperty("format", "json");
 props.setProperty("read_json_by_line", "true");
 DorisOptions dorisOptions = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("").build();
 
@@ -581,7 +581,7 @@ Flink Doris Connector 中集成了[Flink 
CDC](https://nightlies.apache.org/flink
 
 :::info 注意
 
-1. 使用整库同步时需要在 `$FLINK_HOME/lib` 目录下添加对应的 Flink CDC 依赖,比如 
**`flink-sql-connector-mysql-cdc-${version}.jar`**,**`flink-sql-connector-oracle-cdc-${version}.jar`**,FlinkCDC
 从 3.1 
版本与之前版本不兼容,下载地址分别为为[FlinkCDC3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),[FlinkCDC
 
2.x](https://repo.maven.apache.org/maven2/com/ververica/flink-sql-connector-mysql-cdc/)。
+1. 使用整库同步时需要在 `$FLINK_HOME/lib` 目录下添加对应的 Flink CDC 包依赖 (Fat Jar),比如 
**`flink-sql-connector-mysql-cdc-${version}.jar`**,**`flink-sql-connector-oracle-cdc-${version}.jar`**,FlinkCDC
 从 3.1 
版本与之前版本不兼容,下载地址分别为为[FlinkCDC3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),[FlinkCDC
 
2.x](https://repo.maven.apache.org/maven2/com/ververica/flink-sql-connector-mysql-cdc/)。
 2. Connector 24.0.0 之后依赖的 Flink CDC 版本需要在 3.1 
以上,下载地址见[这里](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),FlinkCDC
 如果需使用 Flink CDC 同步 MySQL 和 Oracle,还需要在 `$FLINK_HOME/lib` 下增加相关的 JDBC 驱动。
 
 :::
diff --git a/versioned_docs/version-2.1/ecosystem/flink-doris-connector.md 
b/versioned_docs/version-2.1/ecosystem/flink-doris-connector.md
index 66610f08a3f..98112b21774 100644
--- a/versioned_docs/version-2.1/ecosystem/flink-doris-connector.md
+++ b/versioned_docs/version-2.1/ecosystem/flink-doris-connector.md
@@ -220,7 +220,7 @@ When Flink reads data from Doris, the Doris Source is 
currently a bounded stream
 ##### Thrift Method
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -228,18 +228,18 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '127.0.0.1:8030',  -- Fe的host:HttpPort
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 ##### ArrowFlightSQL
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -247,14 +247,14 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '{fe.conf:http_port}', 
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'source.use-flight-sql' = 'true',
       'source.flight-sql-port' = '{fe.conf:arrow_flight_sql_port}',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 #### Using DataStream API to Read Data
@@ -265,7 +265,7 @@ When using the DataStream API to read data, you need to 
include the dependencies
 final StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment();
 DorisOptions option = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("")
         .build();
@@ -324,7 +324,7 @@ CREATE TABLE student_sink (
     WITH (
       'connector' = 'doris',
       'fenodes' = '10.16.10.6:28737',
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = 'password',
       'sink.label-prefix' = 'doris_label'
@@ -406,7 +406,7 @@ properties.setProperty("format", "csv");
 DorisOptions.Builder dorisBuilder = DorisOptions.builder();
 dorisBuilder
         .setFenodes("10.16.10.6:28737")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("");
 DorisExecutionOptions.Builder executionBuilder = 
DorisExecutionOptions.builder();
@@ -467,7 +467,7 @@ props.setProperty("format", "json");
 props.setProperty("read_json_by_line", "true");
 DorisOptions dorisOptions = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("").build();
 
@@ -578,7 +578,7 @@ The Flink Doris Connector integrates **Flink CDC** ([Flink 
CDC Documentation](ht
 
 :::info Note
 
-1. When using full database synchronization, you need to add the corresponding 
Flink CDC dependencies in the `$FLINK_HOME/lib` directory, such as 
**flink-sql-connector-mysql-cdc-${version}.jar**, 
**flink-sql-connector-oracle-cdc-${version}.jar**. FlinkCDC version 3.1 and 
later is not compatible with previous versions. You can download the 
dependencies from the following links: [FlinkCDC 
3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),
 [FlinkCDC  [...]
+1. When using full database synchronization, you need to add the corresponding 
Flink CDC dependencies in the `$FLINK_HOME/lib` directory (Fat Jar), such as 
**flink-sql-connector-mysql-cdc-${version}.jar**, 
**flink-sql-connector-oracle-cdc-${version}.jar**. FlinkCDC version 3.1 and 
later is not compatible with previous versions. You can download the 
dependencies from the following links: [FlinkCDC 
3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),
  [...]
 2. For versions after Connector 24.0.0, the required Flink CDC version must be 
3.1 or higher. You can download it 
[here](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/).
 If Flink CDC is used to synchronize MySQL and Oracle, you must also add the 
relevant JDBC drivers under `$FLINK_HOME/lib`.
 
 #### MySQL Whole Database Synchronization
diff --git a/versioned_docs/version-3.0/ecosystem/flink-doris-connector.md 
b/versioned_docs/version-3.0/ecosystem/flink-doris-connector.md
index 66610f08a3f..98112b21774 100644
--- a/versioned_docs/version-3.0/ecosystem/flink-doris-connector.md
+++ b/versioned_docs/version-3.0/ecosystem/flink-doris-connector.md
@@ -220,7 +220,7 @@ When Flink reads data from Doris, the Doris Source is 
currently a bounded stream
 ##### Thrift Method
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -228,18 +228,18 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '127.0.0.1:8030',  -- Fe的host:HttpPort
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 ##### ArrowFlightSQL
 
 ```SQL
-CREATE TABLE students (
+CREATE TABLE student (
     id INT,
     name STRING,
     age INT
@@ -247,14 +247,14 @@ CREATE TABLE students (
     WITH (
       'connector' = 'doris',
       'fenodes' = '{fe.conf:http_port}', 
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'source.use-flight-sql' = 'true',
       'source.flight-sql-port' = '{fe.conf:arrow_flight_sql_port}',
       'username' = 'root',
       'password' = ''
 );
 
-SELECT * FROM students;
+SELECT * FROM student;
 ```
 
 #### Using DataStream API to Read Data
@@ -265,7 +265,7 @@ When using the DataStream API to read data, you need to 
include the dependencies
 final StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment();
 DorisOptions option = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("")
         .build();
@@ -324,7 +324,7 @@ CREATE TABLE student_sink (
     WITH (
       'connector' = 'doris',
       'fenodes' = '10.16.10.6:28737',
-      'table.identifier' = 'test.students',
+      'table.identifier' = 'test.student',
       'username' = 'root',
       'password' = 'password',
       'sink.label-prefix' = 'doris_label'
@@ -406,7 +406,7 @@ properties.setProperty("format", "csv");
 DorisOptions.Builder dorisBuilder = DorisOptions.builder();
 dorisBuilder
         .setFenodes("10.16.10.6:28737")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("");
 DorisExecutionOptions.Builder executionBuilder = 
DorisExecutionOptions.builder();
@@ -467,7 +467,7 @@ props.setProperty("format", "json");
 props.setProperty("read_json_by_line", "true");
 DorisOptions dorisOptions = DorisOptions.builder()
         .setFenodes("127.0.0.1:8030")
-        .setTableIdentifier("test.students")
+        .setTableIdentifier("test.student")
         .setUsername("root")
         .setPassword("").build();
 
@@ -578,7 +578,7 @@ The Flink Doris Connector integrates **Flink CDC** ([Flink 
CDC Documentation](ht
 
 :::info Note
 
-1. When using full database synchronization, you need to add the corresponding 
Flink CDC dependencies in the `$FLINK_HOME/lib` directory, such as 
**flink-sql-connector-mysql-cdc-${version}.jar**, 
**flink-sql-connector-oracle-cdc-${version}.jar**. FlinkCDC version 3.1 and 
later is not compatible with previous versions. You can download the 
dependencies from the following links: [FlinkCDC 
3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),
 [FlinkCDC  [...]
+1. When using full database synchronization, you need to add the corresponding 
Flink CDC dependencies in the `$FLINK_HOME/lib` directory (Fat Jar), such as 
**flink-sql-connector-mysql-cdc-${version}.jar**, 
**flink-sql-connector-oracle-cdc-${version}.jar**. FlinkCDC version 3.1 and 
later is not compatible with previous versions. You can download the 
dependencies from the following links: [FlinkCDC 
3.x](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/),
  [...]
 2. For versions after Connector 24.0.0, the required Flink CDC version must be 
3.1 or higher. You can download it 
[here](https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-mysql-cdc/).
 If Flink CDC is used to synchronize MySQL and Oracle, you must also add the 
relevant JDBC drivers under `$FLINK_HOME/lib`.
 
 #### MySQL Whole Database Synchronization


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@doris.apache.org
For additional commands, e-mail: commits-h...@doris.apache.org

Reply via email to