This is an automated email from the ASF dual-hosted git repository.

liaoxin pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/doris-website.git


The following commit(s) were added to refs/heads/master by this push:
     new a665b9d5b6 Update load-manual.md (#1339)
a665b9d5b6 is described below

commit a665b9d5b63eb470ba6117c237d13c2747b6cdcc
Author: Ruffian Jiang <jiangyj0...@outlook.com>
AuthorDate: Tue Nov 19 18:45:34 2024 +0800

    Update load-manual.md (#1339)
---
 docs/data-operate/import/load-manual.md                             | 6 +++---
 .../current/data-operate/import/load-manual.md                      | 4 ++--
 .../version-2.1/data-operate/import/load-manual.md                  | 4 ++--
 .../version-3.0/data-operate/import/load-manual.md                  | 4 ++--
 versioned_docs/version-2.1/data-operate/import/load-manual.md       | 6 +++---
 versioned_docs/version-3.0/data-operate/import/load-manual.md       | 6 +++---
 6 files changed, 15 insertions(+), 15 deletions(-)

diff --git a/docs/data-operate/import/load-manual.md 
b/docs/data-operate/import/load-manual.md
index 6dcf57e666..c7c34cdda8 100644
--- a/docs/data-operate/import/load-manual.md
+++ b/docs/data-operate/import/load-manual.md
@@ -32,7 +32,7 @@ Apache Doris offers various methods for importing and 
integrating data, allowing
 
     - For higher concurrency or frequency (more than 20 concurrent writes or 
multiple writes per minute), you can enable enable [Group 
Commit](./import-way/group-commit-manual.md) and use JDBC INSERT or Stream Load.
 
-    - For high throughput, you can use [Stream 
Load](./import-way/stream-load-manua) via HTTP.
+    - For high throughput, you can use [Stream 
Load](./import-way/stream-load-manual) via HTTP.
 
 - **Streaming Synchronization**: Real-time data streams (e.g., Flink, Kafka, 
transactional databases) are imported into Doris tables, ideal for real-time 
analysis and querying.
 
@@ -48,7 +48,7 @@ Apache Doris offers various methods for importing and 
integrating data, allowing
 
     - You can use [INSERT INTO SELECT](./import-way/insert-into-manual.md) to 
synchronously load files from S3, HDFS, and NAS into Doris, and you can perform 
the operation asynchronously using a [JOB](../scheduler/job-scheduler.md).
 
-    - You can use [Stream Load](./import-way/stream-load-manua) or [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) to write local files into 
Doris.
+    - You can use [Stream Load](./import-way/stream-load-manual) or [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) to write local files into 
Doris.
 
 - **External Data Source Integration**: Query and partially import data from 
external sources (e.g., Hive, JDBC, Iceberg) into Doris tables.
 
@@ -70,4 +70,4 @@ Doris's import process mainly involves various aspects such 
as data sources, dat
 | [INSERT INTO SELECT](./import-way/insert-into-manual.md) | Importing from an 
external source like a table in a catalog or files in s3.      | SQL            
         | Depending on memory size  | Synchronous, Asynchronous via Job     |
 | [Routine Load](./import-way/routine-load-manual.md)      | Real-time import 
from Kafka                            | csv, json               | Micro-batch 
import MB to GB | Asynchronous     |
 | [MySQL Load](./import-way/mysql-load-manual.md)          | Importing from 
local files.                             | csv                     | Less than 
1GB          | Synchronous     |
-| [Group Commit](./import-way/group-commit-manual.md)          | Writing with 
high frequency.                            | Depending on the import method 
used                     |  Micro-batch import KB         | -     |
\ No newline at end of file
+| [Group Commit](./import-way/group-commit-manual.md)          | Writing with 
high frequency.                            | Depending on the import method 
used                     |  Micro-batch import KB         | -     |
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/load-manual.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/load-manual.md
index 7e056884b3..9cd401158e 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/load-manual.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/load-manual.md
@@ -32,7 +32,7 @@ Apache Doris 提供了多种导入和集成数据的方法,您可以使用合
 
     - 并发较高或者频次较高(大于 20 并发或者 1 分钟写入多次)时建议打开 [Group 
Commit](./import-way/group-commit-manual.md),使用 JDBC INSERT 或者 Stream Load 写入数据。
 
-    - 吞吐较高时推荐使用 [Stream Load](./import-way/stream-load-manua) 通过 HTTP 写入数据。
+    - 吞吐较高时推荐使用 [Stream Load](./import-way/stream-load-manual) 通过 HTTP 写入数据。
 
 - **流式同步**:通过实时数据流(如 Flink、Kafka、事务数据库)将数据实时导入到 Doris 表中,适用于需要实时分析和查询的场景。
 
@@ -47,7 +47,7 @@ Apache Doris 提供了多种导入和集成数据的方法,您可以使用合
 
     - 可以使用 [INSERT INTO SELECT](./import-way/insert-into-manual.md) 将 S3、HDFS 
和 NAS 中的文件同步写入到 Doris 中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。
 
-    - 可以使用 [Stream Load](./import-way/stream-load-manua) 或者 [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) 将本地文件写入 Doris 中。
+    - 可以使用 [Stream Load](./import-way/stream-load-manual) 或者 [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) 将本地文件写入 Doris 中。
 
 - **外部数据源集成**:通过与外部数据源(如 Hive、JDBC、Iceberg 等)的集成,实现对外部数据的查询和部分数据导入到 Doris 表中。
     - 可以创建 [Catalog](../../lakehouse/lakehouse-overview.md) 读取外部数据源中的数据,使用 
[INSERT INTO SELECT](./import-way/insert-into-manual.md) 将外部数据源中的数据同步写入到 Doris 
中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/load-manual.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/load-manual.md
index 7e056884b3..9cd401158e 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/load-manual.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/load-manual.md
@@ -32,7 +32,7 @@ Apache Doris 提供了多种导入和集成数据的方法,您可以使用合
 
     - 并发较高或者频次较高(大于 20 并发或者 1 分钟写入多次)时建议打开 [Group 
Commit](./import-way/group-commit-manual.md),使用 JDBC INSERT 或者 Stream Load 写入数据。
 
-    - 吞吐较高时推荐使用 [Stream Load](./import-way/stream-load-manua) 通过 HTTP 写入数据。
+    - 吞吐较高时推荐使用 [Stream Load](./import-way/stream-load-manual) 通过 HTTP 写入数据。
 
 - **流式同步**:通过实时数据流(如 Flink、Kafka、事务数据库)将数据实时导入到 Doris 表中,适用于需要实时分析和查询的场景。
 
@@ -47,7 +47,7 @@ Apache Doris 提供了多种导入和集成数据的方法,您可以使用合
 
     - 可以使用 [INSERT INTO SELECT](./import-way/insert-into-manual.md) 将 S3、HDFS 
和 NAS 中的文件同步写入到 Doris 中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。
 
-    - 可以使用 [Stream Load](./import-way/stream-load-manua) 或者 [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) 将本地文件写入 Doris 中。
+    - 可以使用 [Stream Load](./import-way/stream-load-manual) 或者 [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) 将本地文件写入 Doris 中。
 
 - **外部数据源集成**:通过与外部数据源(如 Hive、JDBC、Iceberg 等)的集成,实现对外部数据的查询和部分数据导入到 Doris 表中。
     - 可以创建 [Catalog](../../lakehouse/lakehouse-overview.md) 读取外部数据源中的数据,使用 
[INSERT INTO SELECT](./import-way/insert-into-manual.md) 将外部数据源中的数据同步写入到 Doris 
中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。
diff --git 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/load-manual.md
 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/load-manual.md
index 7e056884b3..9cd401158e 100644
--- 
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/load-manual.md
+++ 
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/load-manual.md
@@ -32,7 +32,7 @@ Apache Doris 提供了多种导入和集成数据的方法,您可以使用合
 
     - 并发较高或者频次较高(大于 20 并发或者 1 分钟写入多次)时建议打开 [Group 
Commit](./import-way/group-commit-manual.md),使用 JDBC INSERT 或者 Stream Load 写入数据。
 
-    - 吞吐较高时推荐使用 [Stream Load](./import-way/stream-load-manua) 通过 HTTP 写入数据。
+    - 吞吐较高时推荐使用 [Stream Load](./import-way/stream-load-manual) 通过 HTTP 写入数据。
 
 - **流式同步**:通过实时数据流(如 Flink、Kafka、事务数据库)将数据实时导入到 Doris 表中,适用于需要实时分析和查询的场景。
 
@@ -47,7 +47,7 @@ Apache Doris 提供了多种导入和集成数据的方法,您可以使用合
 
     - 可以使用 [INSERT INTO SELECT](./import-way/insert-into-manual.md) 将 S3、HDFS 
和 NAS 中的文件同步写入到 Doris 中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。
 
-    - 可以使用 [Stream Load](./import-way/stream-load-manua) 或者 [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) 将本地文件写入 Doris 中。
+    - 可以使用 [Stream Load](./import-way/stream-load-manual) 或者 [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) 将本地文件写入 Doris 中。
 
 - **外部数据源集成**:通过与外部数据源(如 Hive、JDBC、Iceberg 等)的集成,实现对外部数据的查询和部分数据导入到 Doris 表中。
     - 可以创建 [Catalog](../../lakehouse/lakehouse-overview.md) 读取外部数据源中的数据,使用 
[INSERT INTO SELECT](./import-way/insert-into-manual.md) 将外部数据源中的数据同步写入到 Doris 
中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。
diff --git a/versioned_docs/version-2.1/data-operate/import/load-manual.md 
b/versioned_docs/version-2.1/data-operate/import/load-manual.md
index 6dcf57e666..c7c34cdda8 100644
--- a/versioned_docs/version-2.1/data-operate/import/load-manual.md
+++ b/versioned_docs/version-2.1/data-operate/import/load-manual.md
@@ -32,7 +32,7 @@ Apache Doris offers various methods for importing and 
integrating data, allowing
 
     - For higher concurrency or frequency (more than 20 concurrent writes or 
multiple writes per minute), you can enable enable [Group 
Commit](./import-way/group-commit-manual.md) and use JDBC INSERT or Stream Load.
 
-    - For high throughput, you can use [Stream 
Load](./import-way/stream-load-manua) via HTTP.
+    - For high throughput, you can use [Stream 
Load](./import-way/stream-load-manual) via HTTP.
 
 - **Streaming Synchronization**: Real-time data streams (e.g., Flink, Kafka, 
transactional databases) are imported into Doris tables, ideal for real-time 
analysis and querying.
 
@@ -48,7 +48,7 @@ Apache Doris offers various methods for importing and 
integrating data, allowing
 
     - You can use [INSERT INTO SELECT](./import-way/insert-into-manual.md) to 
synchronously load files from S3, HDFS, and NAS into Doris, and you can perform 
the operation asynchronously using a [JOB](../scheduler/job-scheduler.md).
 
-    - You can use [Stream Load](./import-way/stream-load-manua) or [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) to write local files into 
Doris.
+    - You can use [Stream Load](./import-way/stream-load-manual) or [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) to write local files into 
Doris.
 
 - **External Data Source Integration**: Query and partially import data from 
external sources (e.g., Hive, JDBC, Iceberg) into Doris tables.
 
@@ -70,4 +70,4 @@ Doris's import process mainly involves various aspects such 
as data sources, dat
 | [INSERT INTO SELECT](./import-way/insert-into-manual.md) | Importing from an 
external source like a table in a catalog or files in s3.      | SQL            
         | Depending on memory size  | Synchronous, Asynchronous via Job     |
 | [Routine Load](./import-way/routine-load-manual.md)      | Real-time import 
from Kafka                            | csv, json               | Micro-batch 
import MB to GB | Asynchronous     |
 | [MySQL Load](./import-way/mysql-load-manual.md)          | Importing from 
local files.                             | csv                     | Less than 
1GB          | Synchronous     |
-| [Group Commit](./import-way/group-commit-manual.md)          | Writing with 
high frequency.                            | Depending on the import method 
used                     |  Micro-batch import KB         | -     |
\ No newline at end of file
+| [Group Commit](./import-way/group-commit-manual.md)          | Writing with 
high frequency.                            | Depending on the import method 
used                     |  Micro-batch import KB         | -     |
diff --git a/versioned_docs/version-3.0/data-operate/import/load-manual.md 
b/versioned_docs/version-3.0/data-operate/import/load-manual.md
index 6dcf57e666..c7c34cdda8 100644
--- a/versioned_docs/version-3.0/data-operate/import/load-manual.md
+++ b/versioned_docs/version-3.0/data-operate/import/load-manual.md
@@ -32,7 +32,7 @@ Apache Doris offers various methods for importing and 
integrating data, allowing
 
     - For higher concurrency or frequency (more than 20 concurrent writes or 
multiple writes per minute), you can enable enable [Group 
Commit](./import-way/group-commit-manual.md) and use JDBC INSERT or Stream Load.
 
-    - For high throughput, you can use [Stream 
Load](./import-way/stream-load-manua) via HTTP.
+    - For high throughput, you can use [Stream 
Load](./import-way/stream-load-manual) via HTTP.
 
 - **Streaming Synchronization**: Real-time data streams (e.g., Flink, Kafka, 
transactional databases) are imported into Doris tables, ideal for real-time 
analysis and querying.
 
@@ -48,7 +48,7 @@ Apache Doris offers various methods for importing and 
integrating data, allowing
 
     - You can use [INSERT INTO SELECT](./import-way/insert-into-manual.md) to 
synchronously load files from S3, HDFS, and NAS into Doris, and you can perform 
the operation asynchronously using a [JOB](../scheduler/job-scheduler.md).
 
-    - You can use [Stream Load](./import-way/stream-load-manua) or [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) to write local files into 
Doris.
+    - You can use [Stream Load](./import-way/stream-load-manual) or [Doris 
Streamloader](../../ecosystem/doris-streamloader.md) to write local files into 
Doris.
 
 - **External Data Source Integration**: Query and partially import data from 
external sources (e.g., Hive, JDBC, Iceberg) into Doris tables.
 
@@ -70,4 +70,4 @@ Doris's import process mainly involves various aspects such 
as data sources, dat
 | [INSERT INTO SELECT](./import-way/insert-into-manual.md) | Importing from an 
external source like a table in a catalog or files in s3.      | SQL            
         | Depending on memory size  | Synchronous, Asynchronous via Job     |
 | [Routine Load](./import-way/routine-load-manual.md)      | Real-time import 
from Kafka                            | csv, json               | Micro-batch 
import MB to GB | Asynchronous     |
 | [MySQL Load](./import-way/mysql-load-manual.md)          | Importing from 
local files.                             | csv                     | Less than 
1GB          | Synchronous     |
-| [Group Commit](./import-way/group-commit-manual.md)          | Writing with 
high frequency.                            | Depending on the import method 
used                     |  Micro-batch import KB         | -     |
\ No newline at end of file
+| [Group Commit](./import-way/group-commit-manual.md)          | Writing with 
high frequency.                            | Depending on the import method 
used                     |  Micro-batch import KB         | -     |


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@doris.apache.org
For additional commands, e-mail: commits-h...@doris.apache.org

Reply via email to