This is an automated email from the ASF dual-hosted git repository. liaoxin pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/doris-website.git
The following commit(s) were added to refs/heads/master by this push: new a948a4646bb [doc](load) fix load manual (#1763) a948a4646bb is described below commit a948a4646bbd7937f5cfc402dbc76a7285c491fa Author: Xin Liao <liao...@selectdb.com> AuthorDate: Sun Jan 12 22:28:16 2025 +0800 [doc](load) fix load manual (#1763) --- docs/data-operate/import/load-manual.md | 24 ++++++++++----------- .../current/data-operate/import/load-manual.md | 25 +++++++++++----------- .../version-2.1/data-operate/import/load-manual.md | 25 +++++++++++----------- .../version-3.0/data-operate/import/load-manual.md | 25 +++++++++++----------- .../version-2.1/data-operate/import/load-manual.md | 24 ++++++++++----------- .../version-3.0/data-operate/import/load-manual.md | 24 ++++++++++----------- 6 files changed, 72 insertions(+), 75 deletions(-) diff --git a/docs/data-operate/import/load-manual.md b/docs/data-operate/import/load-manual.md index 11642ade9bb..d86544cd12d 100644 --- a/docs/data-operate/import/load-manual.md +++ b/docs/data-operate/import/load-manual.md @@ -42,11 +42,11 @@ Apache Doris offers various methods for importing and integrating data, allowing - You can use [Flink CDC](../../ecosystem/flink-doris-connector.md) or [Datax](../../ecosystem/datax.md) to write transactional database CDC data streams into Doris. -- **Batch Import**: Data is batch-loaded from external storage systems (e.g., S3, HDFS, local files, NAS) into Doris tables, suitable for non-real-time data import needs. +- **Batch Import**: Data is batch-loaded from external storage systems (e.g., Object Storage, HDFS, local files, NAS) into Doris tables, suitable for non-real-time data import needs. - - You can use [Broker Load](./import-way/broker-load-manual.md) to write files from S3 and HDFS into Doris. + - You can use [Broker Load](./import-way/broker-load-manual.md) to write files from Object Storage and HDFS into Doris. - - You can use [INSERT INTO SELECT](./import-way/insert-into-manual.md) to synchronously load files from S3, HDFS, and NAS into Doris, and you can perform the operation asynchronously using a [JOB](../scheduler/job-scheduler.md). + - You can use [INSERT INTO SELECT](./import-way/insert-into-manual.md) to synchronously load files from Object Storage, HDFS, and NAS into Doris, and you can perform the operation asynchronously using a [JOB](../scheduler/job-scheduler.md). - You can use [Stream Load](./import-way/stream-load-manual) or [Doris Streamloader](../../ecosystem/doris-streamloader.md) to write local files into Doris. @@ -62,12 +62,12 @@ Each import method in Doris is an implicit transaction by default. For more info Doris's import process mainly involves various aspects such as data sources, data formats, import methods, error handling, data transformation, and transactions. You can quickly browse the scenarios suitable for each import method and the supported file formats in the table below. -| Import Method | Use Case | Supported File Formats | Single Import Volume | Import Mode | -| :-------------------------------------------- | :----------------------------------------- | ----------------------- | ----------------- | -------- | -| [Stream Load](./import-way/stream-load-manual) | Importing local files or push data in applications via http. | csv, json, parquet, orc | Less than 10GB | Synchronous | -| [Broker Load](./import-way/broker-load-manual.md) | Importing from object storage, HDFS, etc. | csv, json, parquet, orc | Tens of GB to hundreds of GB | Asynchronous | -| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | Writing data via JDBC. | SQL | Simple testing | Synchronous | -| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | Importing from an external source like a table in a catalog or files in s3. | SQL | Depending on memory size | Synchronous, Asynchronous via Job | -| [Routine Load](./import-way/routine-load-manual.md) | Real-time import from Kafka | csv, json | Micro-batch import MB to GB | Asynchronous | -| [MySQL Load](./import-way/mysql-load-manual.md) | Importing from local files. | csv | Less than 1GB | Synchronous | -| [Group Commit](./group-commit-manual.md) | Writing with high frequency. | Depending on the import method used | Micro-batch import KB | - | +| Import Method | Use Case | Supported File Formats | Import Mode | +| :-------------------------------------------- | :----------------------------------------- | ----------------------- | -------- | +| [Stream Load](./import-way/stream-load-manual) | Importing local files or push data in applications via http. | csv, json, parquet, orc | Synchronous | +| [Broker Load](./import-way/broker-load-manual.md) | Importing from object storage, HDFS, etc. | csv, json, parquet, orc | Asynchronous | +| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | Writing data via JDBC. | SQL | Synchronous | +| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | Importing from an external source like a table in a catalog or files in Object Storage, HDFS. | SQL | Synchronous, Asynchronous via Job | +| [Routine Load](./import-way/routine-load-manual.md) | Real-time import from Kafka | csv, json | Asynchronous | +| [MySQL Load](./import-way/mysql-load-manual.md) | Importing from local files. | csv | Synchronous | +| [Group Commit](./group-commit-manual.md) | Writing with high frequency. | Depending on the import method used | - | diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/load-manual.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/load-manual.md index baa7f297759..2abf083868f 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/load-manual.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/data-operate/import/load-manual.md @@ -42,10 +42,10 @@ Apache Doris 提供了多种导入和集成数据的方法,您可以使用合 - 可以使用 [Flink CDC](../../ecosystem/flink-doris-connector.md) 或 [ Datax](../../ecosystem/datax.md) 将事务数据库的 CDC 数据流写入到 Doris 中。 -- **批量导入**:将数据从外部存储系统(如 S3、HDFS、本地文件、NAS)批量加载到 Doris 表中,适用于非实时数据导入的需求。 - - 可以使用 [Broker Load](./import-way/broker-load-manual.md) 将 S3 和 HDFS 中的文件写入到 Doris 中。 +- **批量导入**:将数据从外部存储系统(如对象存储、HDFS、本地文件、NAS)批量加载到 Doris 表中,适用于非实时数据导入的需求。 + - 可以使用 [Broker Load](./import-way/broker-load-manual.md) 将对象存储和 HDFS 中的文件写入到 Doris 中。 - - 可以使用 [INSERT INTO SELECT](./import-way/insert-into-manual.md) 将 S3、HDFS 和 NAS 中的文件同步写入到 Doris 中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。 + - 可以使用 [INSERT INTO SELECT](./import-way/insert-into-manual.md) 将对象存储、HDFS 和 NAS 中的文件同步写入到 Doris 中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。 - 可以使用 [Stream Load](./import-way/stream-load-manual) 或者 [Doris Streamloader](../../ecosystem/doris-streamloader.md) 将本地文件写入 Doris 中。 @@ -60,13 +60,12 @@ Doris 的每个导入默认都是一个隐式事务,事务相关的更多信 Doris 的导入主要涉及数据源、数据格式、导入方式、错误数据处理、数据转换、事务多个方面。您可以在如下表格中快速浏览各导入方式适合的场景和支持的文件格式。 -| 导入方式 | 使用场景 | 支持的文件格式 | 单次导入数据量 | 导入模式 | -| :-------------------------------------------- | :----------------------------------------- | ----------------------- | ----------------- | -------- | -| [Stream Load](./import-way/stream-load-manual) | 导入本地文件或者应用程序写入 | csv、json、parquet、orc | 小于10GB | 同步 | -| [Broker Load](./import-way/broker-load-manual.md) | 从对象存储、HDFS等导入 | csv、json、parquet、orc | 数十GB到数百 GB | 异步 | -| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | 通过JDBC等接口导入 | SQL | 简单测试用 | 同步 | -| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | 可以导入外部表或者对象存储、HDFS中的文件 | SQL | 根据内存大小而定 | 同步 | -| [Routine Load](./import-way/routine-load-manual.md) | 从kakfa实时导入 | csv、json | 微批导入 MB 到 GB | 异步 | -| [MySQL Load](./import-way/mysql-load-manual.md) | 从本地数据导入 | csv | 小于10GB | 同步 | -| [Group Commit](./group-commit-manual.md) | 高频小批量导入 | 根据使用的导入方式而定 | 微批导入KB | - | - +| 导入方式 | 使用场景 | 支持的文件格式 | 导入模式 | +| :-------------------------------------------- | :----------------------------------------- | ----------------------- | -------- | +| [Stream Load](./import-way/stream-load-manual) | 导入本地文件或者应用程序写入 | csv、json、parquet、orc | 同步 | +| [Broker Load](./import-way/broker-load-manual.md) | 从对象存储、HDFS等导入 | csv、json、parquet、orc | 异步 | +| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | 通过JDBC等接口导入 | SQL | 同步 | +| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | 可以导入外部表或者对象存储、HDFS中的文件 | SQL | 同步 | +| [Routine Load](./import-way/routine-load-manual.md) | 从kakfa实时导入 | csv、json | 异步 | +| [MySQL Load](./import-way/mysql-load-manual.md) | 从本地数据导入 | csv | 同步 | +| [Group Commit](./group-commit-manual.md) | 高频小批量导入 | 根据使用的导入方式而定 | - | diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/load-manual.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/load-manual.md index baa7f297759..2abf083868f 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/load-manual.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/data-operate/import/load-manual.md @@ -42,10 +42,10 @@ Apache Doris 提供了多种导入和集成数据的方法,您可以使用合 - 可以使用 [Flink CDC](../../ecosystem/flink-doris-connector.md) 或 [ Datax](../../ecosystem/datax.md) 将事务数据库的 CDC 数据流写入到 Doris 中。 -- **批量导入**:将数据从外部存储系统(如 S3、HDFS、本地文件、NAS)批量加载到 Doris 表中,适用于非实时数据导入的需求。 - - 可以使用 [Broker Load](./import-way/broker-load-manual.md) 将 S3 和 HDFS 中的文件写入到 Doris 中。 +- **批量导入**:将数据从外部存储系统(如对象存储、HDFS、本地文件、NAS)批量加载到 Doris 表中,适用于非实时数据导入的需求。 + - 可以使用 [Broker Load](./import-way/broker-load-manual.md) 将对象存储和 HDFS 中的文件写入到 Doris 中。 - - 可以使用 [INSERT INTO SELECT](./import-way/insert-into-manual.md) 将 S3、HDFS 和 NAS 中的文件同步写入到 Doris 中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。 + - 可以使用 [INSERT INTO SELECT](./import-way/insert-into-manual.md) 将对象存储、HDFS 和 NAS 中的文件同步写入到 Doris 中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。 - 可以使用 [Stream Load](./import-way/stream-load-manual) 或者 [Doris Streamloader](../../ecosystem/doris-streamloader.md) 将本地文件写入 Doris 中。 @@ -60,13 +60,12 @@ Doris 的每个导入默认都是一个隐式事务,事务相关的更多信 Doris 的导入主要涉及数据源、数据格式、导入方式、错误数据处理、数据转换、事务多个方面。您可以在如下表格中快速浏览各导入方式适合的场景和支持的文件格式。 -| 导入方式 | 使用场景 | 支持的文件格式 | 单次导入数据量 | 导入模式 | -| :-------------------------------------------- | :----------------------------------------- | ----------------------- | ----------------- | -------- | -| [Stream Load](./import-way/stream-load-manual) | 导入本地文件或者应用程序写入 | csv、json、parquet、orc | 小于10GB | 同步 | -| [Broker Load](./import-way/broker-load-manual.md) | 从对象存储、HDFS等导入 | csv、json、parquet、orc | 数十GB到数百 GB | 异步 | -| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | 通过JDBC等接口导入 | SQL | 简单测试用 | 同步 | -| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | 可以导入外部表或者对象存储、HDFS中的文件 | SQL | 根据内存大小而定 | 同步 | -| [Routine Load](./import-way/routine-load-manual.md) | 从kakfa实时导入 | csv、json | 微批导入 MB 到 GB | 异步 | -| [MySQL Load](./import-way/mysql-load-manual.md) | 从本地数据导入 | csv | 小于10GB | 同步 | -| [Group Commit](./group-commit-manual.md) | 高频小批量导入 | 根据使用的导入方式而定 | 微批导入KB | - | - +| 导入方式 | 使用场景 | 支持的文件格式 | 导入模式 | +| :-------------------------------------------- | :----------------------------------------- | ----------------------- | -------- | +| [Stream Load](./import-way/stream-load-manual) | 导入本地文件或者应用程序写入 | csv、json、parquet、orc | 同步 | +| [Broker Load](./import-way/broker-load-manual.md) | 从对象存储、HDFS等导入 | csv、json、parquet、orc | 异步 | +| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | 通过JDBC等接口导入 | SQL | 同步 | +| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | 可以导入外部表或者对象存储、HDFS中的文件 | SQL | 同步 | +| [Routine Load](./import-way/routine-load-manual.md) | 从kakfa实时导入 | csv、json | 异步 | +| [MySQL Load](./import-way/mysql-load-manual.md) | 从本地数据导入 | csv | 同步 | +| [Group Commit](./group-commit-manual.md) | 高频小批量导入 | 根据使用的导入方式而定 | - | diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/load-manual.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/load-manual.md index baa7f297759..2abf083868f 100644 --- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/load-manual.md +++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/data-operate/import/load-manual.md @@ -42,10 +42,10 @@ Apache Doris 提供了多种导入和集成数据的方法,您可以使用合 - 可以使用 [Flink CDC](../../ecosystem/flink-doris-connector.md) 或 [ Datax](../../ecosystem/datax.md) 将事务数据库的 CDC 数据流写入到 Doris 中。 -- **批量导入**:将数据从外部存储系统(如 S3、HDFS、本地文件、NAS)批量加载到 Doris 表中,适用于非实时数据导入的需求。 - - 可以使用 [Broker Load](./import-way/broker-load-manual.md) 将 S3 和 HDFS 中的文件写入到 Doris 中。 +- **批量导入**:将数据从外部存储系统(如对象存储、HDFS、本地文件、NAS)批量加载到 Doris 表中,适用于非实时数据导入的需求。 + - 可以使用 [Broker Load](./import-way/broker-load-manual.md) 将对象存储和 HDFS 中的文件写入到 Doris 中。 - - 可以使用 [INSERT INTO SELECT](./import-way/insert-into-manual.md) 将 S3、HDFS 和 NAS 中的文件同步写入到 Doris 中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。 + - 可以使用 [INSERT INTO SELECT](./import-way/insert-into-manual.md) 将对象存储、HDFS 和 NAS 中的文件同步写入到 Doris 中,配合 [JOB](../scheduler/job-scheduler.md) 可以异步写入。 - 可以使用 [Stream Load](./import-way/stream-load-manual) 或者 [Doris Streamloader](../../ecosystem/doris-streamloader.md) 将本地文件写入 Doris 中。 @@ -60,13 +60,12 @@ Doris 的每个导入默认都是一个隐式事务,事务相关的更多信 Doris 的导入主要涉及数据源、数据格式、导入方式、错误数据处理、数据转换、事务多个方面。您可以在如下表格中快速浏览各导入方式适合的场景和支持的文件格式。 -| 导入方式 | 使用场景 | 支持的文件格式 | 单次导入数据量 | 导入模式 | -| :-------------------------------------------- | :----------------------------------------- | ----------------------- | ----------------- | -------- | -| [Stream Load](./import-way/stream-load-manual) | 导入本地文件或者应用程序写入 | csv、json、parquet、orc | 小于10GB | 同步 | -| [Broker Load](./import-way/broker-load-manual.md) | 从对象存储、HDFS等导入 | csv、json、parquet、orc | 数十GB到数百 GB | 异步 | -| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | 通过JDBC等接口导入 | SQL | 简单测试用 | 同步 | -| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | 可以导入外部表或者对象存储、HDFS中的文件 | SQL | 根据内存大小而定 | 同步 | -| [Routine Load](./import-way/routine-load-manual.md) | 从kakfa实时导入 | csv、json | 微批导入 MB 到 GB | 异步 | -| [MySQL Load](./import-way/mysql-load-manual.md) | 从本地数据导入 | csv | 小于10GB | 同步 | -| [Group Commit](./group-commit-manual.md) | 高频小批量导入 | 根据使用的导入方式而定 | 微批导入KB | - | - +| 导入方式 | 使用场景 | 支持的文件格式 | 导入模式 | +| :-------------------------------------------- | :----------------------------------------- | ----------------------- | -------- | +| [Stream Load](./import-way/stream-load-manual) | 导入本地文件或者应用程序写入 | csv、json、parquet、orc | 同步 | +| [Broker Load](./import-way/broker-load-manual.md) | 从对象存储、HDFS等导入 | csv、json、parquet、orc | 异步 | +| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | 通过JDBC等接口导入 | SQL | 同步 | +| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | 可以导入外部表或者对象存储、HDFS中的文件 | SQL | 同步 | +| [Routine Load](./import-way/routine-load-manual.md) | 从kakfa实时导入 | csv、json | 异步 | +| [MySQL Load](./import-way/mysql-load-manual.md) | 从本地数据导入 | csv | 同步 | +| [Group Commit](./group-commit-manual.md) | 高频小批量导入 | 根据使用的导入方式而定 | - | diff --git a/versioned_docs/version-2.1/data-operate/import/load-manual.md b/versioned_docs/version-2.1/data-operate/import/load-manual.md index 11642ade9bb..d86544cd12d 100644 --- a/versioned_docs/version-2.1/data-operate/import/load-manual.md +++ b/versioned_docs/version-2.1/data-operate/import/load-manual.md @@ -42,11 +42,11 @@ Apache Doris offers various methods for importing and integrating data, allowing - You can use [Flink CDC](../../ecosystem/flink-doris-connector.md) or [Datax](../../ecosystem/datax.md) to write transactional database CDC data streams into Doris. -- **Batch Import**: Data is batch-loaded from external storage systems (e.g., S3, HDFS, local files, NAS) into Doris tables, suitable for non-real-time data import needs. +- **Batch Import**: Data is batch-loaded from external storage systems (e.g., Object Storage, HDFS, local files, NAS) into Doris tables, suitable for non-real-time data import needs. - - You can use [Broker Load](./import-way/broker-load-manual.md) to write files from S3 and HDFS into Doris. + - You can use [Broker Load](./import-way/broker-load-manual.md) to write files from Object Storage and HDFS into Doris. - - You can use [INSERT INTO SELECT](./import-way/insert-into-manual.md) to synchronously load files from S3, HDFS, and NAS into Doris, and you can perform the operation asynchronously using a [JOB](../scheduler/job-scheduler.md). + - You can use [INSERT INTO SELECT](./import-way/insert-into-manual.md) to synchronously load files from Object Storage, HDFS, and NAS into Doris, and you can perform the operation asynchronously using a [JOB](../scheduler/job-scheduler.md). - You can use [Stream Load](./import-way/stream-load-manual) or [Doris Streamloader](../../ecosystem/doris-streamloader.md) to write local files into Doris. @@ -62,12 +62,12 @@ Each import method in Doris is an implicit transaction by default. For more info Doris's import process mainly involves various aspects such as data sources, data formats, import methods, error handling, data transformation, and transactions. You can quickly browse the scenarios suitable for each import method and the supported file formats in the table below. -| Import Method | Use Case | Supported File Formats | Single Import Volume | Import Mode | -| :-------------------------------------------- | :----------------------------------------- | ----------------------- | ----------------- | -------- | -| [Stream Load](./import-way/stream-load-manual) | Importing local files or push data in applications via http. | csv, json, parquet, orc | Less than 10GB | Synchronous | -| [Broker Load](./import-way/broker-load-manual.md) | Importing from object storage, HDFS, etc. | csv, json, parquet, orc | Tens of GB to hundreds of GB | Asynchronous | -| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | Writing data via JDBC. | SQL | Simple testing | Synchronous | -| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | Importing from an external source like a table in a catalog or files in s3. | SQL | Depending on memory size | Synchronous, Asynchronous via Job | -| [Routine Load](./import-way/routine-load-manual.md) | Real-time import from Kafka | csv, json | Micro-batch import MB to GB | Asynchronous | -| [MySQL Load](./import-way/mysql-load-manual.md) | Importing from local files. | csv | Less than 1GB | Synchronous | -| [Group Commit](./group-commit-manual.md) | Writing with high frequency. | Depending on the import method used | Micro-batch import KB | - | +| Import Method | Use Case | Supported File Formats | Import Mode | +| :-------------------------------------------- | :----------------------------------------- | ----------------------- | -------- | +| [Stream Load](./import-way/stream-load-manual) | Importing local files or push data in applications via http. | csv, json, parquet, orc | Synchronous | +| [Broker Load](./import-way/broker-load-manual.md) | Importing from object storage, HDFS, etc. | csv, json, parquet, orc | Asynchronous | +| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | Writing data via JDBC. | SQL | Synchronous | +| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | Importing from an external source like a table in a catalog or files in Object Storage, HDFS. | SQL | Synchronous, Asynchronous via Job | +| [Routine Load](./import-way/routine-load-manual.md) | Real-time import from Kafka | csv, json | Asynchronous | +| [MySQL Load](./import-way/mysql-load-manual.md) | Importing from local files. | csv | Synchronous | +| [Group Commit](./group-commit-manual.md) | Writing with high frequency. | Depending on the import method used | - | diff --git a/versioned_docs/version-3.0/data-operate/import/load-manual.md b/versioned_docs/version-3.0/data-operate/import/load-manual.md index 11642ade9bb..d86544cd12d 100644 --- a/versioned_docs/version-3.0/data-operate/import/load-manual.md +++ b/versioned_docs/version-3.0/data-operate/import/load-manual.md @@ -42,11 +42,11 @@ Apache Doris offers various methods for importing and integrating data, allowing - You can use [Flink CDC](../../ecosystem/flink-doris-connector.md) or [Datax](../../ecosystem/datax.md) to write transactional database CDC data streams into Doris. -- **Batch Import**: Data is batch-loaded from external storage systems (e.g., S3, HDFS, local files, NAS) into Doris tables, suitable for non-real-time data import needs. +- **Batch Import**: Data is batch-loaded from external storage systems (e.g., Object Storage, HDFS, local files, NAS) into Doris tables, suitable for non-real-time data import needs. - - You can use [Broker Load](./import-way/broker-load-manual.md) to write files from S3 and HDFS into Doris. + - You can use [Broker Load](./import-way/broker-load-manual.md) to write files from Object Storage and HDFS into Doris. - - You can use [INSERT INTO SELECT](./import-way/insert-into-manual.md) to synchronously load files from S3, HDFS, and NAS into Doris, and you can perform the operation asynchronously using a [JOB](../scheduler/job-scheduler.md). + - You can use [INSERT INTO SELECT](./import-way/insert-into-manual.md) to synchronously load files from Object Storage, HDFS, and NAS into Doris, and you can perform the operation asynchronously using a [JOB](../scheduler/job-scheduler.md). - You can use [Stream Load](./import-way/stream-load-manual) or [Doris Streamloader](../../ecosystem/doris-streamloader.md) to write local files into Doris. @@ -62,12 +62,12 @@ Each import method in Doris is an implicit transaction by default. For more info Doris's import process mainly involves various aspects such as data sources, data formats, import methods, error handling, data transformation, and transactions. You can quickly browse the scenarios suitable for each import method and the supported file formats in the table below. -| Import Method | Use Case | Supported File Formats | Single Import Volume | Import Mode | -| :-------------------------------------------- | :----------------------------------------- | ----------------------- | ----------------- | -------- | -| [Stream Load](./import-way/stream-load-manual) | Importing local files or push data in applications via http. | csv, json, parquet, orc | Less than 10GB | Synchronous | -| [Broker Load](./import-way/broker-load-manual.md) | Importing from object storage, HDFS, etc. | csv, json, parquet, orc | Tens of GB to hundreds of GB | Asynchronous | -| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | Writing data via JDBC. | SQL | Simple testing | Synchronous | -| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | Importing from an external source like a table in a catalog or files in s3. | SQL | Depending on memory size | Synchronous, Asynchronous via Job | -| [Routine Load](./import-way/routine-load-manual.md) | Real-time import from Kafka | csv, json | Micro-batch import MB to GB | Asynchronous | -| [MySQL Load](./import-way/mysql-load-manual.md) | Importing from local files. | csv | Less than 1GB | Synchronous | -| [Group Commit](./group-commit-manual.md) | Writing with high frequency. | Depending on the import method used | Micro-batch import KB | - | +| Import Method | Use Case | Supported File Formats | Import Mode | +| :-------------------------------------------- | :----------------------------------------- | ----------------------- | -------- | +| [Stream Load](./import-way/stream-load-manual) | Importing local files or push data in applications via http. | csv, json, parquet, orc | Synchronous | +| [Broker Load](./import-way/broker-load-manual.md) | Importing from object storage, HDFS, etc. | csv, json, parquet, orc | Asynchronous | +| [INSERT INTO VALUES](./import-way/insert-into-manual.md) | Writing data via JDBC. | SQL | Synchronous | +| [INSERT INTO SELECT](./import-way/insert-into-manual.md) | Importing from an external source like a table in a catalog or files in Object Storage, HDFS. | SQL | Synchronous, Asynchronous via Job | +| [Routine Load](./import-way/routine-load-manual.md) | Real-time import from Kafka | csv, json | Asynchronous | +| [MySQL Load](./import-way/mysql-load-manual.md) | Importing from local files. | csv | Synchronous | +| [Group Commit](./group-commit-manual.md) | Writing with high frequency. | Depending on the import method used | - | --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@doris.apache.org For additional commands, e-mail: commits-h...@doris.apache.org