This is an automated email from the ASF dual-hosted git repository. liaoxin pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/doris-website.git
The following commit(s) were added to refs/heads/master by this push: new fc217733022 [doc](load) fix load doc error (#1634) fc217733022 is described below commit fc2177330224a506928e4d3f508114b49cb87eed Author: hui lai <1353307...@qq.com> AuthorDate: Fri Dec 27 15:05:57 2024 +0800 [doc](load) fix load doc error (#1634) --- docs/data-operate/import/complex-types/bitmap.md | 1 + docs/data-operate/import/complex-types/hll.md | 1 + docs/data-operate/import/complex-types/variant.md | 1 + docs/data-operate/import/data-source/kafka.md | 17 ++++++++--------- docs/data-operate/import/data-source/local-file.md | 1 + .../zh-CN/docusaurus-plugin-content-docs/current.json | 4 ---- .../data-operate/import/complex-types/bitmap.md | 1 + .../current/data-operate/import/complex-types/hll.md | 1 + .../data-operate/import/complex-types/variant.md | 1 + .../current/data-operate/import/data-source/kafka.md | 19 +++++++++---------- .../data-operate/import/data-source/local-file.md | 3 ++- .../docusaurus-plugin-content-docs/version-2.1.json | 4 ---- .../data-operate/import/complex-types/bitmap.md | 1 + .../data-operate/import/complex-types/hll.md | 1 + .../data-operate/import/complex-types/variant.md | 1 + .../data-operate/import/data-source/kafka.md | 19 +++++++++---------- .../data-operate/import/data-source/local-file.md | 3 ++- .../docusaurus-plugin-content-docs/version-3.0.json | 4 ---- .../data-operate/import/complex-types/bitmap.md | 1 + .../data-operate/import/complex-types/hll.md | 1 + .../data-operate/import/complex-types/variant.md | 1 + .../data-operate/import/data-source/kafka.md | 19 +++++++++---------- .../data-operate/import/data-source/local-file.md | 3 ++- sidebars.json | 16 +++++----------- .../data-operate/import/complex-types/bitmap.md | 1 + .../data-operate/import/complex-types/hll.md | 1 + .../data-operate/import/complex-types/variant.md | 1 + .../data-operate/import/data-source/kafka.md | 17 ++++++++--------- .../data-operate/import/data-source/local-file.md | 1 + .../data-operate/import/complex-types/bitmap.md | 1 + .../data-operate/import/complex-types/hll.md | 1 + .../data-operate/import/complex-types/variant.md | 1 + .../data-operate/import/data-source/kafka.md | 17 ++++++++--------- .../data-operate/import/data-source/local-file.md | 1 + versioned_sidebars/version-2.1-sidebars.json | 9 +++++++-- versioned_sidebars/version-3.0-sidebars.json | 9 +++++++-- 36 files changed, 97 insertions(+), 87 deletions(-) diff --git a/docs/data-operate/import/complex-types/bitmap.md b/docs/data-operate/import/complex-types/bitmap.md index 71610345f9f..e60058bf773 100644 --- a/docs/data-operate/import/complex-types/bitmap.md +++ b/docs/data-operate/import/complex-types/bitmap.md @@ -1,3 +1,4 @@ +--- { "title": "Bitmap", "language": "en" diff --git a/docs/data-operate/import/complex-types/hll.md b/docs/data-operate/import/complex-types/hll.md index 5979826a9dd..a0c3300c13b 100644 --- a/docs/data-operate/import/complex-types/hll.md +++ b/docs/data-operate/import/complex-types/hll.md @@ -1,3 +1,4 @@ +--- { "title": "HLL", "language": "en" diff --git a/docs/data-operate/import/complex-types/variant.md b/docs/data-operate/import/complex-types/variant.md index 3eac77b15b1..62ad04021d3 100644 --- a/docs/data-operate/import/complex-types/variant.md +++ b/docs/data-operate/import/complex-types/variant.md @@ -1,3 +1,4 @@ +--- { "title": "Variant", "language": "en" diff --git a/docs/data-operate/import/data-source/kafka.md b/docs/data-operate/import/data-source/kafka.md index 1254ab8bef4..c84d72d8758 100644 --- a/docs/data-operate/import/data-source/kafka.md +++ b/docs/data-operate/import/data-source/kafka.md @@ -1,3 +1,4 @@ +--- { "title": "Kafka", "language": "en" @@ -27,9 +28,7 @@ Doris provides the following methods to load data from Kafka: - **Using Routine Load to consume Kafka data** -Doris continuously consumes data from Kafka Topics through Routine Load. After submitting a Routine Load job, Doris generates load tasks in real-time to consume messages from the specified Topic in the Kafka cluster. Routine Load supports CSV and JSON formats, with Exactly-Once semantics, ensuring that data is neither lost nor duplicated. - -For more documentation, please refer to [Routine Load](../import-way/routine-load-manual.md). +Doris continuously consumes data from Kafka Topics through Routine Load. After submitting a Routine Load job, Doris generates load tasks in real-time to consume messages from the specified Topic in the Kafka cluster. Routine Load supports CSV and JSON formats, with Exactly-Once semantics, ensuring that data is neither lost nor duplicated.For more documentation, please refer to [Routine Load](../import-way/routine-load-manual.md). - **Doris Kafka Connector to consume Kafka data** @@ -178,11 +177,11 @@ mysql> select * from test_multi_table_load2; For methods of configuring Kafka with authentication, please refer to [Kafka Security Authentication](../import-way/routine-load-manual.md#kafka-security-authentication). -# Using Doris Kafka Connector to consume Kafka data +## Using Doris Kafka Connector to consume Kafka data The Doris Kafka Connector is a tool for loading Kafka data streams into the Doris database. Users can easily load various serialization formats (such as JSON, Avro, Protobuf) through the Kafka Connect plugin, and it supports parsing data formats from the Debezium component. -## Start in Distributed Mode +### Start in Distributed Mode [Distributed](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers) mode provides scalability and automatic fault tolerance for Kafka Connect. In this mode, multiple worker processes can be started using the same `group.id`, which will coordinate the execution of connectors and tasks across all available worker processes. @@ -253,7 +252,7 @@ curl -i http://127.0.0.1:8083/connectors/test-doris-sink-cluster/tasks/0/restart For an introduction to Distributed mode, please refer to [Distributed Workers](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers). -## Load Ordinary Data +### Load Ordinary Data 1. Load sample data: @@ -314,7 +313,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## Load Data Collected by Debezium Component +### Load Data Collected by Debezium Component 1. The MySQL database has the following table: @@ -374,7 +373,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## Load Data in AVRO Serialization Format +### Load Data in AVRO Serialization Format ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ @@ -402,7 +401,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## Load Data in Protobuf Serialization Format +### Load Data in Protobuf Serialization Format ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ diff --git a/docs/data-operate/import/data-source/local-file.md b/docs/data-operate/import/data-source/local-file.md index cb644207019..b62818d7868 100644 --- a/docs/data-operate/import/data-source/local-file.md +++ b/docs/data-operate/import/data-source/local-file.md @@ -1,3 +1,4 @@ +--- { "title": "local file", "language": "en" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current.json b/i18n/zh-CN/docusaurus-plugin-content-docs/current.json index 2a05c19351b..a8c5e3c014b 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/current.json +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current.json @@ -99,10 +99,6 @@ "message": "数据导入", "description": "The label for category Loading Data in sidebar docs" }, - "sidebar.docs.category.Complex Type": { - "message": "复杂类型", - "description": "The label for category Complex Type in sidebar docs" - }, "sidebar.docs.category.Data Source": { "message": "数据源", "description": "The label for category Data Source in sidebar docs" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/bitmap.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/bitmap.md index f62d4fb2201..040898c7fb0 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/bitmap.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/bitmap.md @@ -1,3 +1,4 @@ +--- { "title": "Bitmap", "language": "zh-CN" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/hll.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/hll.md index 82985722989..2dd6cebb03b 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/hll.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/hll.md @@ -1,3 +1,4 @@ +--- { "title": "HLL", "language": "zh-CN" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/variant.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/variant.md index 23a26467ccb..4fec36539e8 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/variant.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/complex-types/variant.md @@ -1,3 +1,4 @@ +--- { "title": "Variant", "language": "zh-CN" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/data-source/kafka.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/data-source/kafka.md index 01ad2ae1bfe..5cc00328e08 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/data-source/kafka.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/data-source/kafka.md @@ -1,3 +1,4 @@ +--- { "title": "Kafka", "language": "zh-CN" @@ -27,9 +28,7 @@ Doris 提供以下方式从 Kafka 导入数据: - **使用 Routine Load 消费 Kafka 数据** -Doris 通过 Routine Load 持续消费 Kafka Topic 中的数据。提交 Routine Load 作业后,Doris 会实时生成导入任务,消费 Kafka 集群中指定 Topic 的消息。Routine Load 支持 CSV 和 JSON 格式,具备 Exactly-Once 语义,确保数据不丢失且不重复。 - -更多文档请参考 [Routine Load](../import-way/routine-load-manual.md)。 +Doris 通过 Routine Load 持续消费 Kafka Topic 中的数据。提交 Routine Load 作业后,Doris 会实时生成导入任务,消费 Kafka 集群中指定 Topic 的消息。Routine Load 支持 CSV 和 JSON 格式,具备 Exactly-Once 语义,确保数据不丢失且不重复。更多文档请参考 [Routine Load](../import-way/routine-load-manual.md)。 - **Doris Kafka Connector 消费 Kafka 数据** @@ -178,11 +177,11 @@ mysql> select * from test_multi_table_load2; 有关带有认证的 Kafka 配置方法,请参见 [Kafka 安全认证](../import-way/routine-load-manual.md#kafka-安全认证)。 -# 使用 Doris Kafka Connector 消费 Kafka 数据 +## 使用 Doris Kafka Connector 消费 Kafka 数据 Doris Kafka Connector 是将 Kafka 数据流导入 Doris 数据库的工具。用户可通过 Kafka Connect 插件轻松导入多种序列化格式(如 JSON、Avro、Protobuf),并支持解析 Debezium 组件的数据格式。 -## 以 Distributed 模式启动 +### 以 Distributed 模式启动 [Distributed](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers) 模式为 Kafka Connect 提供可扩展性和自动容错功能。在此模式下,可以使用相同的 `group.id` 启动多个工作进程,它们会协调在所有可用工作进程中安排连接器和任务的执行。 @@ -253,7 +252,7 @@ curl -i http://127.0.0.1:8083/connectors/test-doris-sink-cluster/tasks/0/restart 关于 Distributed 模式的介绍请参见 [Distributed Workers](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers)。 -## 消费普通数据 +### 消费普通数据 1. 导入数据样本: @@ -314,7 +313,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## 消费 Debezium 组件采集的数据 +### 消费 Debezium 组件采集的数据 1. MySQL 数据库中有如下表: @@ -374,7 +373,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## 消费 AVRO 序列化格式数据 +### 消费 AVRO 序列化格式数据 ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ @@ -402,7 +401,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## 消费 Protobuf 序列化格式数据 +### 消费 Protobuf 序列化格式数据 ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ @@ -428,4 +427,4 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X "value.converter.schema.registry.url":"http://127.0.0.1:8081" } }' -``` \ No newline at end of file +``` diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/data-source/local-file.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/data-source/local-file.md index ac0dbe44399..dfd7a6b2c6d 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/data-source/local-file.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/data-source/local-file.md @@ -1,5 +1,6 @@ +--- { - "title": "local file", + "title": "本地文件", "language": "zh-CN" } --- diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1.json b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1.json index 3bb2abeaad1..e79a42af59f 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1.json +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1.json @@ -99,10 +99,6 @@ "message": "数据导入", "description": "The label for category Loading Data in sidebar docs" }, - "sidebar.docs.category.Complex Type": { - "message": "复杂类型", - "description": "The label for category Complex Type in sidebar docs" - }, "sidebar.docs.category.Data Source": { "message": "数据源", "description": "The label for category Data Source in sidebar docs" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/bitmap.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/bitmap.md index f62d4fb2201..040898c7fb0 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/bitmap.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/bitmap.md @@ -1,3 +1,4 @@ +--- { "title": "Bitmap", "language": "zh-CN" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/hll.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/hll.md index 82985722989..2dd6cebb03b 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/hll.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/hll.md @@ -1,3 +1,4 @@ +--- { "title": "HLL", "language": "zh-CN" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/variant.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/variant.md index 23a26467ccb..4fec36539e8 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/variant.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/complex-types/variant.md @@ -1,3 +1,4 @@ +--- { "title": "Variant", "language": "zh-CN" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/data-source/kafka.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/data-source/kafka.md index 01ad2ae1bfe..5cc00328e08 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/data-source/kafka.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/data-source/kafka.md @@ -1,3 +1,4 @@ +--- { "title": "Kafka", "language": "zh-CN" @@ -27,9 +28,7 @@ Doris 提供以下方式从 Kafka 导入数据: - **使用 Routine Load 消费 Kafka 数据** -Doris 通过 Routine Load 持续消费 Kafka Topic 中的数据。提交 Routine Load 作业后,Doris 会实时生成导入任务,消费 Kafka 集群中指定 Topic 的消息。Routine Load 支持 CSV 和 JSON 格式,具备 Exactly-Once 语义,确保数据不丢失且不重复。 - -更多文档请参考 [Routine Load](../import-way/routine-load-manual.md)。 +Doris 通过 Routine Load 持续消费 Kafka Topic 中的数据。提交 Routine Load 作业后,Doris 会实时生成导入任务,消费 Kafka 集群中指定 Topic 的消息。Routine Load 支持 CSV 和 JSON 格式,具备 Exactly-Once 语义,确保数据不丢失且不重复。更多文档请参考 [Routine Load](../import-way/routine-load-manual.md)。 - **Doris Kafka Connector 消费 Kafka 数据** @@ -178,11 +177,11 @@ mysql> select * from test_multi_table_load2; 有关带有认证的 Kafka 配置方法,请参见 [Kafka 安全认证](../import-way/routine-load-manual.md#kafka-安全认证)。 -# 使用 Doris Kafka Connector 消费 Kafka 数据 +## 使用 Doris Kafka Connector 消费 Kafka 数据 Doris Kafka Connector 是将 Kafka 数据流导入 Doris 数据库的工具。用户可通过 Kafka Connect 插件轻松导入多种序列化格式(如 JSON、Avro、Protobuf),并支持解析 Debezium 组件的数据格式。 -## 以 Distributed 模式启动 +### 以 Distributed 模式启动 [Distributed](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers) 模式为 Kafka Connect 提供可扩展性和自动容错功能。在此模式下,可以使用相同的 `group.id` 启动多个工作进程,它们会协调在所有可用工作进程中安排连接器和任务的执行。 @@ -253,7 +252,7 @@ curl -i http://127.0.0.1:8083/connectors/test-doris-sink-cluster/tasks/0/restart 关于 Distributed 模式的介绍请参见 [Distributed Workers](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers)。 -## 消费普通数据 +### 消费普通数据 1. 导入数据样本: @@ -314,7 +313,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## 消费 Debezium 组件采集的数据 +### 消费 Debezium 组件采集的数据 1. MySQL 数据库中有如下表: @@ -374,7 +373,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## 消费 AVRO 序列化格式数据 +### 消费 AVRO 序列化格式数据 ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ @@ -402,7 +401,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## 消费 Protobuf 序列化格式数据 +### 消费 Protobuf 序列化格式数据 ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ @@ -428,4 +427,4 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X "value.converter.schema.registry.url":"http://127.0.0.1:8081" } }' -``` \ No newline at end of file +``` diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/data-source/local-file.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/data-source/local-file.md index ac0dbe44399..dfd7a6b2c6d 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/data-source/local-file.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/data-source/local-file.md @@ -1,5 +1,6 @@ +--- { - "title": "local file", + "title": "本地文件", "language": "zh-CN" } --- diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0.json b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0.json index 5b954ab391a..283c6ffdcff 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0.json +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0.json @@ -99,10 +99,6 @@ "message": "数据导入", "description": "The label for category Loading Data in sidebar docs" }, - "sidebar.docs.category.Complex Type": { - "message": "复杂类型", - "description": "The label for category Complex Type in sidebar docs" - }, "sidebar.docs.category.Data Source": { "message": "数据源", "description": "The label for category Data Source in sidebar docs" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/bitmap.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/bitmap.md index f62d4fb2201..040898c7fb0 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/bitmap.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/bitmap.md @@ -1,3 +1,4 @@ +--- { "title": "Bitmap", "language": "zh-CN" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/hll.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/hll.md index 82985722989..2dd6cebb03b 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/hll.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/hll.md @@ -1,3 +1,4 @@ +--- { "title": "HLL", "language": "zh-CN" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/variant.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/variant.md index 23a26467ccb..4fec36539e8 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/variant.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/complex-types/variant.md @@ -1,3 +1,4 @@ +--- { "title": "Variant", "language": "zh-CN" diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/data-source/kafka.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/data-source/kafka.md index 01ad2ae1bfe..5cc00328e08 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/data-source/kafka.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/data-source/kafka.md @@ -1,3 +1,4 @@ +--- { "title": "Kafka", "language": "zh-CN" @@ -27,9 +28,7 @@ Doris 提供以下方式从 Kafka 导入数据: - **使用 Routine Load 消费 Kafka 数据** -Doris 通过 Routine Load 持续消费 Kafka Topic 中的数据。提交 Routine Load 作业后,Doris 会实时生成导入任务,消费 Kafka 集群中指定 Topic 的消息。Routine Load 支持 CSV 和 JSON 格式,具备 Exactly-Once 语义,确保数据不丢失且不重复。 - -更多文档请参考 [Routine Load](../import-way/routine-load-manual.md)。 +Doris 通过 Routine Load 持续消费 Kafka Topic 中的数据。提交 Routine Load 作业后,Doris 会实时生成导入任务,消费 Kafka 集群中指定 Topic 的消息。Routine Load 支持 CSV 和 JSON 格式,具备 Exactly-Once 语义,确保数据不丢失且不重复。更多文档请参考 [Routine Load](../import-way/routine-load-manual.md)。 - **Doris Kafka Connector 消费 Kafka 数据** @@ -178,11 +177,11 @@ mysql> select * from test_multi_table_load2; 有关带有认证的 Kafka 配置方法,请参见 [Kafka 安全认证](../import-way/routine-load-manual.md#kafka-安全认证)。 -# 使用 Doris Kafka Connector 消费 Kafka 数据 +## 使用 Doris Kafka Connector 消费 Kafka 数据 Doris Kafka Connector 是将 Kafka 数据流导入 Doris 数据库的工具。用户可通过 Kafka Connect 插件轻松导入多种序列化格式(如 JSON、Avro、Protobuf),并支持解析 Debezium 组件的数据格式。 -## 以 Distributed 模式启动 +### 以 Distributed 模式启动 [Distributed](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers) 模式为 Kafka Connect 提供可扩展性和自动容错功能。在此模式下,可以使用相同的 `group.id` 启动多个工作进程,它们会协调在所有可用工作进程中安排连接器和任务的执行。 @@ -253,7 +252,7 @@ curl -i http://127.0.0.1:8083/connectors/test-doris-sink-cluster/tasks/0/restart 关于 Distributed 模式的介绍请参见 [Distributed Workers](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers)。 -## 消费普通数据 +### 消费普通数据 1. 导入数据样本: @@ -314,7 +313,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## 消费 Debezium 组件采集的数据 +### 消费 Debezium 组件采集的数据 1. MySQL 数据库中有如下表: @@ -374,7 +373,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## 消费 AVRO 序列化格式数据 +### 消费 AVRO 序列化格式数据 ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ @@ -402,7 +401,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## 消费 Protobuf 序列化格式数据 +### 消费 Protobuf 序列化格式数据 ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ @@ -428,4 +427,4 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X "value.converter.schema.registry.url":"http://127.0.0.1:8081" } }' -``` \ No newline at end of file +``` diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/data-source/local-file.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/data-source/local-file.md index ac0dbe44399..dfd7a6b2c6d 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/data-source/local-file.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/data-source/local-file.md @@ -1,5 +1,6 @@ +--- { - "title": "local file", + "title": "本地文件", "language": "zh-CN" } --- diff --git a/sidebars.json b/sidebars.json index 11c49e7c948..0b4aa01c0b5 100644 --- a/sidebars.json +++ b/sidebars.json @@ -176,15 +176,6 @@ "label": "Loading Data", "items": [ "data-operate/import/load-manual", - { - "type": "category", - "label": "Complex Type", - "items": [ - "data-operate/import/complex-types/bitmap", - "data-operate/import/complex-types/hll", - "data-operate/import/complex-types/variant" - ] - }, { "type": "category", "label": "Data Source", @@ -214,6 +205,7 @@ "data-operate/import/import-way/mysql-load-manual" ] }, + "data-operate/import/load-data-format", { "type": "category", "label": "Complex Data Types", @@ -221,10 +213,12 @@ "data-operate/import/complex-types/array", "data-operate/import/complex-types/map", "data-operate/import/complex-types/struct", - "data-operate/import/complex-types/json" + "data-operate/import/complex-types/json", + "data-operate/import/complex-types/bitmap", + "data-operate/import/complex-types/hll", + "data-operate/import/complex-types/variant" ] }, - "data-operate/import/load-data-format", "data-operate/import/error-data-handling", "data-operate/import/load-data-convert", "data-operate/import/load-high-availability", diff --git a/versioned_docs/version-2.1/data-operate/import/complex-types/bitmap.md b/versioned_docs/version-2.1/data-operate/import/complex-types/bitmap.md index 71610345f9f..e60058bf773 100644 --- a/versioned_docs/version-2.1/data-operate/import/complex-types/bitmap.md +++ b/versioned_docs/version-2.1/data-operate/import/complex-types/bitmap.md @@ -1,3 +1,4 @@ +--- { "title": "Bitmap", "language": "en" diff --git a/versioned_docs/version-2.1/data-operate/import/complex-types/hll.md b/versioned_docs/version-2.1/data-operate/import/complex-types/hll.md index 5979826a9dd..a0c3300c13b 100644 --- a/versioned_docs/version-2.1/data-operate/import/complex-types/hll.md +++ b/versioned_docs/version-2.1/data-operate/import/complex-types/hll.md @@ -1,3 +1,4 @@ +--- { "title": "HLL", "language": "en" diff --git a/versioned_docs/version-2.1/data-operate/import/complex-types/variant.md b/versioned_docs/version-2.1/data-operate/import/complex-types/variant.md index 3eac77b15b1..62ad04021d3 100644 --- a/versioned_docs/version-2.1/data-operate/import/complex-types/variant.md +++ b/versioned_docs/version-2.1/data-operate/import/complex-types/variant.md @@ -1,3 +1,4 @@ +--- { "title": "Variant", "language": "en" diff --git a/versioned_docs/version-2.1/data-operate/import/data-source/kafka.md b/versioned_docs/version-2.1/data-operate/import/data-source/kafka.md index 1254ab8bef4..c84d72d8758 100644 --- a/versioned_docs/version-2.1/data-operate/import/data-source/kafka.md +++ b/versioned_docs/version-2.1/data-operate/import/data-source/kafka.md @@ -1,3 +1,4 @@ +--- { "title": "Kafka", "language": "en" @@ -27,9 +28,7 @@ Doris provides the following methods to load data from Kafka: - **Using Routine Load to consume Kafka data** -Doris continuously consumes data from Kafka Topics through Routine Load. After submitting a Routine Load job, Doris generates load tasks in real-time to consume messages from the specified Topic in the Kafka cluster. Routine Load supports CSV and JSON formats, with Exactly-Once semantics, ensuring that data is neither lost nor duplicated. - -For more documentation, please refer to [Routine Load](../import-way/routine-load-manual.md). +Doris continuously consumes data from Kafka Topics through Routine Load. After submitting a Routine Load job, Doris generates load tasks in real-time to consume messages from the specified Topic in the Kafka cluster. Routine Load supports CSV and JSON formats, with Exactly-Once semantics, ensuring that data is neither lost nor duplicated.For more documentation, please refer to [Routine Load](../import-way/routine-load-manual.md). - **Doris Kafka Connector to consume Kafka data** @@ -178,11 +177,11 @@ mysql> select * from test_multi_table_load2; For methods of configuring Kafka with authentication, please refer to [Kafka Security Authentication](../import-way/routine-load-manual.md#kafka-security-authentication). -# Using Doris Kafka Connector to consume Kafka data +## Using Doris Kafka Connector to consume Kafka data The Doris Kafka Connector is a tool for loading Kafka data streams into the Doris database. Users can easily load various serialization formats (such as JSON, Avro, Protobuf) through the Kafka Connect plugin, and it supports parsing data formats from the Debezium component. -## Start in Distributed Mode +### Start in Distributed Mode [Distributed](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers) mode provides scalability and automatic fault tolerance for Kafka Connect. In this mode, multiple worker processes can be started using the same `group.id`, which will coordinate the execution of connectors and tasks across all available worker processes. @@ -253,7 +252,7 @@ curl -i http://127.0.0.1:8083/connectors/test-doris-sink-cluster/tasks/0/restart For an introduction to Distributed mode, please refer to [Distributed Workers](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers). -## Load Ordinary Data +### Load Ordinary Data 1. Load sample data: @@ -314,7 +313,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## Load Data Collected by Debezium Component +### Load Data Collected by Debezium Component 1. The MySQL database has the following table: @@ -374,7 +373,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## Load Data in AVRO Serialization Format +### Load Data in AVRO Serialization Format ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ @@ -402,7 +401,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## Load Data in Protobuf Serialization Format +### Load Data in Protobuf Serialization Format ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ diff --git a/versioned_docs/version-2.1/data-operate/import/data-source/local-file.md b/versioned_docs/version-2.1/data-operate/import/data-source/local-file.md index cb644207019..b62818d7868 100644 --- a/versioned_docs/version-2.1/data-operate/import/data-source/local-file.md +++ b/versioned_docs/version-2.1/data-operate/import/data-source/local-file.md @@ -1,3 +1,4 @@ +--- { "title": "local file", "language": "en" diff --git a/versioned_docs/version-3.0/data-operate/import/complex-types/bitmap.md b/versioned_docs/version-3.0/data-operate/import/complex-types/bitmap.md index 71610345f9f..e60058bf773 100644 --- a/versioned_docs/version-3.0/data-operate/import/complex-types/bitmap.md +++ b/versioned_docs/version-3.0/data-operate/import/complex-types/bitmap.md @@ -1,3 +1,4 @@ +--- { "title": "Bitmap", "language": "en" diff --git a/versioned_docs/version-3.0/data-operate/import/complex-types/hll.md b/versioned_docs/version-3.0/data-operate/import/complex-types/hll.md index 5979826a9dd..a0c3300c13b 100644 --- a/versioned_docs/version-3.0/data-operate/import/complex-types/hll.md +++ b/versioned_docs/version-3.0/data-operate/import/complex-types/hll.md @@ -1,3 +1,4 @@ +--- { "title": "HLL", "language": "en" diff --git a/versioned_docs/version-3.0/data-operate/import/complex-types/variant.md b/versioned_docs/version-3.0/data-operate/import/complex-types/variant.md index 3eac77b15b1..62ad04021d3 100644 --- a/versioned_docs/version-3.0/data-operate/import/complex-types/variant.md +++ b/versioned_docs/version-3.0/data-operate/import/complex-types/variant.md @@ -1,3 +1,4 @@ +--- { "title": "Variant", "language": "en" diff --git a/versioned_docs/version-3.0/data-operate/import/data-source/kafka.md b/versioned_docs/version-3.0/data-operate/import/data-source/kafka.md index 1254ab8bef4..c84d72d8758 100644 --- a/versioned_docs/version-3.0/data-operate/import/data-source/kafka.md +++ b/versioned_docs/version-3.0/data-operate/import/data-source/kafka.md @@ -1,3 +1,4 @@ +--- { "title": "Kafka", "language": "en" @@ -27,9 +28,7 @@ Doris provides the following methods to load data from Kafka: - **Using Routine Load to consume Kafka data** -Doris continuously consumes data from Kafka Topics through Routine Load. After submitting a Routine Load job, Doris generates load tasks in real-time to consume messages from the specified Topic in the Kafka cluster. Routine Load supports CSV and JSON formats, with Exactly-Once semantics, ensuring that data is neither lost nor duplicated. - -For more documentation, please refer to [Routine Load](../import-way/routine-load-manual.md). +Doris continuously consumes data from Kafka Topics through Routine Load. After submitting a Routine Load job, Doris generates load tasks in real-time to consume messages from the specified Topic in the Kafka cluster. Routine Load supports CSV and JSON formats, with Exactly-Once semantics, ensuring that data is neither lost nor duplicated.For more documentation, please refer to [Routine Load](../import-way/routine-load-manual.md). - **Doris Kafka Connector to consume Kafka data** @@ -178,11 +177,11 @@ mysql> select * from test_multi_table_load2; For methods of configuring Kafka with authentication, please refer to [Kafka Security Authentication](../import-way/routine-load-manual.md#kafka-security-authentication). -# Using Doris Kafka Connector to consume Kafka data +## Using Doris Kafka Connector to consume Kafka data The Doris Kafka Connector is a tool for loading Kafka data streams into the Doris database. Users can easily load various serialization formats (such as JSON, Avro, Protobuf) through the Kafka Connect plugin, and it supports parsing data formats from the Debezium component. -## Start in Distributed Mode +### Start in Distributed Mode [Distributed](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers) mode provides scalability and automatic fault tolerance for Kafka Connect. In this mode, multiple worker processes can be started using the same `group.id`, which will coordinate the execution of connectors and tasks across all available worker processes. @@ -253,7 +252,7 @@ curl -i http://127.0.0.1:8083/connectors/test-doris-sink-cluster/tasks/0/restart For an introduction to Distributed mode, please refer to [Distributed Workers](https://docs.confluent.io/platform/current/connect/index.html#distributed-workers). -## Load Ordinary Data +### Load Ordinary Data 1. Load sample data: @@ -314,7 +313,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## Load Data Collected by Debezium Component +### Load Data Collected by Debezium Component 1. The MySQL database has the following table: @@ -374,7 +373,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## Load Data in AVRO Serialization Format +### Load Data in AVRO Serialization Format ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ @@ -402,7 +401,7 @@ curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X }' ``` -## Load Data in Protobuf Serialization Format +### Load Data in Protobuf Serialization Format ```Bash curl -i http://127.0.0.1:8083/connectors -H "Content-Type: application/json" -X POST -d '{ diff --git a/versioned_docs/version-3.0/data-operate/import/data-source/local-file.md b/versioned_docs/version-3.0/data-operate/import/data-source/local-file.md index cb644207019..b62818d7868 100644 --- a/versioned_docs/version-3.0/data-operate/import/data-source/local-file.md +++ b/versioned_docs/version-3.0/data-operate/import/data-source/local-file.md @@ -1,3 +1,4 @@ +--- { "title": "local file", "language": "en" diff --git a/versioned_sidebars/version-2.1-sidebars.json b/versioned_sidebars/version-2.1-sidebars.json index 4b860211e7c..89ba292c82e 100644 --- a/versioned_sidebars/version-2.1-sidebars.json +++ b/versioned_sidebars/version-2.1-sidebars.json @@ -166,6 +166,8 @@ "type": "category", "label": "Data Source", "items": [ + "data-operate/import/data-source/local-file", + "data-operate/import/data-source/kafka", "data-operate/import/data-source/hdfs", "data-operate/import/data-source/amazon-s3", "data-operate/import/data-source/google-cloud-storage", @@ -188,6 +190,7 @@ "data-operate/import/import-way/mysql-load-manual" ] }, + "data-operate/import/load-data-format", { "type": "category", "label": "Complex Data Types", @@ -195,10 +198,12 @@ "data-operate/import/complex-types/array", "data-operate/import/complex-types/map", "data-operate/import/complex-types/struct", - "data-operate/import/complex-types/json" + "data-operate/import/complex-types/json", + "data-operate/import/complex-types/bitmap", + "data-operate/import/complex-types/hll", + "data-operate/import/complex-types/variant" ] }, - "data-operate/import/load-data-format", "data-operate/import/error-data-handling", "data-operate/import/load-data-convert", "data-operate/import/load-high-availability", diff --git a/versioned_sidebars/version-3.0-sidebars.json b/versioned_sidebars/version-3.0-sidebars.json index c19272f7781..3052ef32c3d 100644 --- a/versioned_sidebars/version-3.0-sidebars.json +++ b/versioned_sidebars/version-3.0-sidebars.json @@ -178,6 +178,8 @@ "type": "category", "label": "Data Source", "items": [ + "data-operate/import/data-source/local-file", + "data-operate/import/data-source/kafka", "data-operate/import/data-source/hdfs", "data-operate/import/data-source/amazon-s3", "data-operate/import/data-source/google-cloud-storage", @@ -201,6 +203,7 @@ "data-operate/import/import-way/mysql-load-manual" ] }, + "data-operate/import/load-data-format", { "type": "category", "label": "Complex Data Types", @@ -208,10 +211,12 @@ "data-operate/import/complex-types/array", "data-operate/import/complex-types/map", "data-operate/import/complex-types/struct", - "data-operate/import/complex-types/json" + "data-operate/import/complex-types/json", + "data-operate/import/complex-types/bitmap", + "data-operate/import/complex-types/hll", + "data-operate/import/complex-types/variant" ] }, - "data-operate/import/load-data-format", "data-operate/import/error-data-handling", "data-operate/import/load-data-convert", "data-operate/import/load-high-availability", --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@doris.apache.org For additional commands, e-mail: commits-h...@doris.apache.org