From faa717f0a5784193707151b58d5eef673642ad95 Mon Sep 17 00:00:00 2001 From: yangxin Date: Wed, 9 Jul 2025 15:18:27 +0800 Subject: [PATCH 01/17] Cloud: Import ux optimization --- tidb-cloud/import-csv-files-serverless.md | 192 ++++++++++++---------- 1 file changed, 106 insertions(+), 86 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index a38102b427262..e7738f74960f5 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -99,43 +99,48 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 2. Click the name of your target cluster to go to its overview page, and then click **Data** > **Import** in the left navigation pane. -2. Select **Import data from Cloud Storage**, and then click **Amazon S3**. +2. Click **Import data from Cloud Storage**. -3. On the **Import Data from Amazon S3** page, provide the following information for the source CSV files: +3. On the **Import Data from Cloud Storage** page, provide the following information: - - **Import File Count**: select **One file** or **Multiple files** as needed. - - **Included Schema Files**: this field is only visible when importing multiple files. If the source folder contains the target table schemas, select **Yes**. Otherwise, select **No**. - - **Data Format**: select **CSV**. - - **File URI** or **Folder URI**: + - Select **Amazon S3** as the storage provider. + - enter the **Source Files URI**. - When importing one file, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `s3://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - - **Bucket Access**: you can use either an AWS Role ARN or an AWS access key to access your bucket. For more information, see [Configure Amazon S3 access](/tidb-cloud/serverless-external-storage.md#configure-amazon-s3-access). + - select the **Credential** - **AWS Role ARN**: enter the AWS Role ARN value. - **AWS Access Key**: enter the AWS access key ID and AWS secret access key. -4. Click **Connect**. +4. Click **Next**. -5. In the **Destination** section, select the target database and table. +5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When importing multiple files, you can use **Advanced Settings** > **Mapping Settings** to define a custom mapping rule for each target table and its corresponding CSV file. After that, the data source files will be re-scanned using the provided custom mapping rule. + * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. - When you enter the source file URI and name in **Source File URIs and Names**, make sure it is in the following format `s3://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `s3://sampledata/ingest/TableName.01.csv`. + After setting the custom mapping rules, the system will re-scan the source files accordingly. - You can also use wildcards to match the source files. For example: + * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. + * You can also use wildcards to match multiple files: - - `s3://[bucket_name]/[data_source_folder]/my-data?.csv`: all CSV files starting with `my-data` followed by one character (such as `my-data1.csv` and `my-data2.csv`) in that folder will be imported into the same target table. + * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (e.g., `my-data1.csv`, `my-data2.csv`). - - `s3://[bucket_name]/[data_source_folder]/my-data*.csv`: all CSV files in the folder starting with `my-data` will be imported into the same target table. + * `my-data*.csv`: Matches all CSV files that start with `my-data` (e.g., `my-data-2023.csv`, `my-data-final.csv`). - Note that only `?` and `*` are supported. + > Only `*` and `?` wildcards are supported. - > **Note:** - > - > The URI must contain the data source folder. + * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + + * Select **CSV** as the data format. + + > **Note:** + > + > When importing one file, you can only use mapping settings and the source file cannot be modified. + +6. Click **Next**. -6. Click **Start Import**. +7. Wait the scan result, check the data files and target tables, and then click **Start Import**. -7. When the import progress shows **Completed**, check the imported tables. +8. When the import progress shows **Completed**, check the imported tables. @@ -151,41 +156,46 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 2. Click the name of your target cluster to go to its overview page, and then click **Data** > **Import** in the left navigation pane. -2. Select **Import data from Cloud Storage**, and then click **Google Cloud Storage**. +2. Click **Import data from Cloud Storage**. -3. On the **Import Data from Google Cloud Storage** page, provide the following information for the source CSV files: +3. On the **Import Data from Cloud Storage** page, provide the following information: - - **Import File Count**: select **One file** or **Multiple files** as needed. - - **Included Schema Files**: this field is only visible when importing multiple files. If the source folder contains the target table schemas, select **Yes**. Otherwise, select **No**. - - **Data Format**: select **CSV**. - - **File URI** or **Folder URI**: - - When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.csv`. - - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `[gcs|gs]://sampledata/ingest/`. - - **Bucket Access**: you can use a service account key to access your bucket. For more information, see [Configure GCS access](/tidb-cloud/serverless-external-storage.md#configure-gcs-access). + - Select **Google Cloud Storage** as the storage provider. + - **Source Files URI**. + - When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `s3://sampledata/ingest/TableName.01.csv`. + - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. + - **Credential**: you can use a GCS IAM Role Service Account key to access your bucket. For more information, see [Configure GCS access](/tidb-cloud/serverless-external-storage.md#configure-gcs-access). -4. Click **Connect**. +4. Click **Next**. -5. In the **Destination** section, select the target database and table. +5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When importing multiple files, you can use **Advanced Settings** > **Mapping Settings** to define a custom mapping rule for each target table and its corresponding CSV file. After that, the data source files will be re-scanned using the provided custom mapping rule. + * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. - When you enter the source file URI and name in **Source File URIs and Names**, make sure it is in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.csv`. + After setting the custom mapping rules, the system will re-scan the source files accordingly. - You can also use wildcards to match the source files. For example: + * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. + * You can also use wildcards to match multiple files: - - `[gcs|gs]://[bucket_name]/[data_source_folder]/my-data?.csv`: all CSV files starting with `my-data` followed by one character (such as `my-data1.csv` and `my-data2.csv`) in that folder will be imported into the same target table. + * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (e.g., `my-data1.csv`, `my-data2.csv`). - - `[gcs|gs]://[bucket_name]/[data_source_folder]/my-data*.csv`: all CSV files in the folder starting with `my-data` will be imported into the same target table. + * `my-data*.csv`: Matches all CSV files that start with `my-data` (e.g., `my-data-2023.csv`, `my-data-final.csv`). - Note that only `?` and `*` are supported. + > Only `*` and `?` wildcards are supported. - > **Note:** - > - > The URI must contain the data source folder. + * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + + * Select **CSV** as the data format. + + > **Note:** + > + > When importing one file, you can only use mapping settings and the source file cannot be modified. + +6. Click **Next**. -6. Click **Start Import**. +7. Wait the scan result, check the data files and target tables, and then click **Start Import**. -7. When the import progress shows **Completed**, check the imported tables. +8. When the import progress shows **Completed**, check the imported tables. @@ -201,41 +211,46 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 2. Click the name of your target cluster to go to its overview page, and then click **Data** > **Import** in the left navigation pane. -2. Select **Import data from Cloud Storage**, and then click **Azure Blob Storage**. +2. Click **Import data from Cloud Storage**. -3. On the **Import Data from Azure Blob Storage** page, provide the following information for the source CSV files: +3. On the **Import Data from Cloud Storage** page, provide the following information: - - **Import File Count**: select **One file** or **Multiple files** as needed. - - **Included Schema Files**: this field is only visible when importing multiple files. If the source folder contains the target table schemas, select **Yes**. Otherwise, select **No**. - - **Data Format**: select **CSV**. - - **File URI** or **Folder URI**: - - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[azure|https]://sampledata/ingest/TableName.01.csv`. - - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `[azure|https]://sampledata/ingest/`. - - **Bucket Access**: you can use a shared access signature (SAS) token to access your bucket. For more information, see [Configure Azure Blob Storage access](/tidb-cloud/serverless-external-storage.md#configure-azure-blob-storage-access). + - Select **Azure Blob Storage** as the storage provider. + - **Source Files URI**. + - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `s3://sampledata/ingest/TableName.01.csv`. + - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. + - **Credential**: you can use a shared access signature (SAS) token to access your bucket. For more information, see [Configure Azure Blob Storage access](/tidb-cloud/serverless-external-storage.md#configure-azure-blob-storage-access) -4. Click **Connect**. +4. Click **Next**. -5. In the **Destination** section, select the target database and table. +5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When importing multiple files, you can use **Advanced Settings** > **Mapping Settings** to define a custom mapping rule for each target table and its corresponding CSV file. After that, the data source files will be re-scanned using the provided custom mapping rule. + * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. - When you enter the source file URI and name in **Source File URIs and Names**, make sure it is in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[azure|https]://sampledata/ingest/TableName.01.csv`. + After setting the custom mapping rules, the system will re-scan the source files accordingly. - You can also use wildcards to match the source files. For example: + * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. + * You can also use wildcards to match multiple files: - - `[azure|https]://[bucket_name]/[data_source_folder]/my-data?.csv`: all CSV files starting with `my-data` followed by one character (such as `my-data1.csv` and `my-data2.csv`) in that folder will be imported into the same target table. + * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (e.g., `my-data1.csv`, `my-data2.csv`). - - `[azure|https]://[bucket_name]/[data_source_folder]/my-data*.csv`: all CSV files in the folder starting with `my-data` will be imported into the same target table. + * `my-data*.csv`: Matches all CSV files that start with `my-data` (e.g., `my-data-2023.csv`, `my-data-final.csv`). - Note that only `?` and `*` are supported. + > Only `*` and `?` wildcards are supported. - > **Note:** - > - > The URI must contain the data source folder. + * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + + * Select **CSV** as the data format. + + > **Note:** + > + > When importing one file, you can only use mapping settings and the source file cannot be modified. + +6. Click **Next**. -6. Click **Start Import**. +7. Wait the scan result, check the data files and target tables, and then click **Start Import**. -7. When the import progress shows **Completed**, check the imported tables. +8. When the import progress shows **Completed**, check the imported tables. @@ -251,41 +266,46 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 2. Click the name of your target cluster to go to its overview page, and then click **Data** > **Import** in the left navigation pane. -2. Select **Import data from Cloud Storage**, and then click **Alibaba Cloud OSS**. +2. Click **Import data from Cloud Storage**. -3. On the **Import Data from Alibaba Cloud OSS** page, provide the following information for the source CSV files: +3. On the **Import Data from Cloud Storage** page, provide the following information: - - **Import File Count**: select **One file** or **Multiple files** as needed. - - **Included Schema Files**: this field is only visible when importing multiple files. If the source folder contains the target table schemas, select **Yes**. Otherwise, select **No**. - - **Data Format**: select **CSV**. - - **File URI** or **Folder URI**: - - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `oss://sampledata/ingest/TableName.01.csv`. - - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `oss://sampledata/ingest/`. - - **Bucket Access**: you can use an AccessKey pair to access your bucket. For more information, see [Configure Alibaba Cloud Object Storage Service (OSS) access](/tidb-cloud/serverless-external-storage.md#configure-alibaba-cloud-object-storage-service-oss-access). + - Select **Alibaba Cloud OSS** as the storage provider. + - **Source Files URI**. + - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `s3://sampledata/ingest/TableName.01.csv`. + - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. + - **Credential**: you can use an AccessKey pair to access your bucket. For more information, see [Configure Alibaba Cloud Object Storage Service (OSS) access](/tidb-cloud/serverless-external-storage.md#configure-alibaba-cloud-object-storage-service-oss-access). -4. Click **Connect**. +4. Click **Next**. -5. In the **Destination** section, select the target database and table. +5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When importing multiple files, you can use **Advanced Settings** > **Mapping Settings** to define a custom mapping rule for each target table and its corresponding CSV file. After that, the data source files will be re-scanned using the provided custom mapping rule. + * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. - When you enter the source file URI and name in **Source File URIs and Names**, make sure it is in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `oss://sampledata/ingest/TableName.01.csv`. + After setting the custom mapping rules, the system will re-scan the source files accordingly. - You can also use wildcards to match the source files. For example: + * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. + * You can also use wildcards to match multiple files: - - `oss://[bucket_name]/[data_source_folder]/my-data?.csv`: all CSV files starting with `my-data` followed by one character (such as `my-data1.csv` and `my-data2.csv`) in that folder will be imported into the same target table. + * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (e.g., `my-data1.csv`, `my-data2.csv`). - - `oss://[bucket_name]/[data_source_folder]/my-data*.csv`: all CSV files in the folder starting with `my-data` will be imported into the same target table. + * `my-data*.csv`: Matches all CSV files that start with `my-data` (e.g., `my-data-2023.csv`, `my-data-final.csv`). - Note that only `?` and `*` are supported. + > Only `*` and `?` wildcards are supported. - > **Note:** - > - > The URI must contain the data source folder. + * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + + * Select **CSV** as the data format. + + > **Note:** + > + > When importing one file, you can only use mapping settings and the source file cannot be modified. + +6. Click **Next**. -6. Click **Start Import**. +7. Wait the scan result, check the data files and target tables, and then click **Start Import**. -7. When the import progress shows **Completed**, check the imported tables. +8. When the import progress shows **Completed**, check the imported tables. From edc41a01d6e87657efd35e0497f365298dbdab5c Mon Sep 17 00:00:00 2001 From: Leon Yang Date: Wed, 9 Jul 2025 15:21:42 +0800 Subject: [PATCH 02/17] Apply suggestions from code review Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> --- tidb-cloud/import-csv-files-serverless.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index e7738f74960f5..112c8ad1ad754 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -162,7 +162,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: - Select **Google Cloud Storage** as the storage provider. - **Source Files URI**. - - When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `s3://sampledata/ingest/TableName.01.csv`. +- When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use a GCS IAM Role Service Account key to access your bucket. For more information, see [Configure GCS access](/tidb-cloud/serverless-external-storage.md#configure-gcs-access). @@ -217,7 +217,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: - Select **Azure Blob Storage** as the storage provider. - **Source Files URI**. - - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `s3://sampledata/ingest/TableName.01.csv`. +- When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[azure|https]://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use a shared access signature (SAS) token to access your bucket. For more information, see [Configure Azure Blob Storage access](/tidb-cloud/serverless-external-storage.md#configure-azure-blob-storage-access) @@ -272,7 +272,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: - Select **Alibaba Cloud OSS** as the storage provider. - **Source Files URI**. - - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `s3://sampledata/ingest/TableName.01.csv`. +- When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `oss://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use an AccessKey pair to access your bucket. For more information, see [Configure Alibaba Cloud Object Storage Service (OSS) access](/tidb-cloud/serverless-external-storage.md#configure-alibaba-cloud-object-storage-service-oss-access). From 5317eb05e8f2ad689c96d583467af46569309d77 Mon Sep 17 00:00:00 2001 From: yangxin Date: Wed, 9 Jul 2025 15:26:42 +0800 Subject: [PATCH 03/17] use for example --- tidb-cloud/import-csv-files-serverless.md | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index 112c8ad1ad754..dc01a170762a5 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -122,9 +122,9 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: - * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (e.g., `my-data1.csv`, `my-data2.csv`). + * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). - * `my-data*.csv`: Matches all CSV files that start with `my-data` (e.g., `my-data-2023.csv`, `my-data-final.csv`). + * `my-data*.csv`: Matches all CSV files that start with `my-data` (for example, `my-data-2023.csv`, `my-data-final.csv`). > Only `*` and `?` wildcards are supported. @@ -177,9 +177,9 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: - * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (e.g., `my-data1.csv`, `my-data2.csv`). + * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). - * `my-data*.csv`: Matches all CSV files that start with `my-data` (e.g., `my-data-2023.csv`, `my-data-final.csv`). + * `my-data*.csv`: Matches all CSV files that start with `my-data` (for example, `my-data-2023.csv`, `my-data-final.csv`). > Only `*` and `?` wildcards are supported. @@ -232,9 +232,9 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: - * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (e.g., `my-data1.csv`, `my-data2.csv`). + * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). - * `my-data*.csv`: Matches all CSV files that start with `my-data` (e.g., `my-data-2023.csv`, `my-data-final.csv`). + * `my-data*.csv`: Matches all CSV files that start with `my-data` (for example, `my-data-2023.csv`, `my-data-final.csv`). > Only `*` and `?` wildcards are supported. @@ -287,9 +287,9 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: - * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (e.g., `my-data1.csv`, `my-data2.csv`). + * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). - * `my-data*.csv`: Matches all CSV files that start with `my-data` (e.g., `my-data-2023.csv`, `my-data-final.csv`). + * `my-data*.csv`: Matches all CSV files that start with `my-data` (for example, `my-data-2023.csv`, `my-data-final.csv`). > Only `*` and `?` wildcards are supported. From 4187e1d79bb8404cba10b6193cf7a7aba72de9eb Mon Sep 17 00:00:00 2001 From: yangxin Date: Wed, 9 Jul 2025 15:32:20 +0800 Subject: [PATCH 04/17] fix --- tidb-cloud/import-csv-files-serverless.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index dc01a170762a5..9ec23217606b5 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -162,7 +162,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: - Select **Google Cloud Storage** as the storage provider. - **Source Files URI**. -- When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.csv`. + - When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use a GCS IAM Role Service Account key to access your bucket. For more information, see [Configure GCS access](/tidb-cloud/serverless-external-storage.md#configure-gcs-access). @@ -217,7 +217,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: - Select **Azure Blob Storage** as the storage provider. - **Source Files URI**. -- When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[azure|https]://sampledata/ingest/TableName.01.csv`. + - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[azure|https]://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use a shared access signature (SAS) token to access your bucket. For more information, see [Configure Azure Blob Storage access](/tidb-cloud/serverless-external-storage.md#configure-azure-blob-storage-access) @@ -272,7 +272,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: - Select **Alibaba Cloud OSS** as the storage provider. - **Source Files URI**. -- When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `oss://sampledata/ingest/TableName.01.csv`. + - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `oss://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use an AccessKey pair to access your bucket. For more information, see [Configure Alibaba Cloud Object Storage Service (OSS) access](/tidb-cloud/serverless-external-storage.md#configure-alibaba-cloud-object-storage-service-oss-access). From 15c8bc40ddd47087df333aa0b1ceee1d860d521c Mon Sep 17 00:00:00 2001 From: yangxin Date: Wed, 9 Jul 2025 15:38:22 +0800 Subject: [PATCH 05/17] polish --- tidb-cloud/import-csv-files-serverless.md | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index 9ec23217606b5..39116052b128e 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -138,7 +138,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 6. Click **Next**. -7. Wait the scan result, check the data files and target tables, and then click **Start Import**. +7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. @@ -193,7 +193,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 6. Click **Next**. -7. Wait the scan result, check the data files and target tables, and then click **Start Import**. +7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. @@ -248,7 +248,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 6. Click **Next**. -7. Wait the scan result, check the data files and target tables, and then click **Start Import**. +7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. @@ -303,7 +303,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 6. Click **Next**. -7. Wait the scan result, check the data files and target tables, and then click **Start Import**. +7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. From f8fd35a891b6f8864eb012cb2df70e731f059119 Mon Sep 17 00:00:00 2001 From: yangxin Date: Wed, 9 Jul 2025 16:01:33 +0800 Subject: [PATCH 06/17] modify all --- tidb-cloud/import-parquet-files-serverless.md | 184 ++++++++++-------- tidb-cloud/import-sample-data-serverless.md | 26 ++- 2 files changed, 118 insertions(+), 92 deletions(-) diff --git a/tidb-cloud/import-parquet-files-serverless.md b/tidb-cloud/import-parquet-files-serverless.md index c7ba5cf642fba..927bf66587899 100644 --- a/tidb-cloud/import-parquet-files-serverless.md +++ b/tidb-cloud/import-parquet-files-serverless.md @@ -107,43 +107,48 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 2. Click the name of your target cluster to go to its overview page, and then click **Data** > **Import** in the left navigation pane. -2. Select **Import data from Cloud Storage**, and then click **Amazon S3**. +2. Click **Import data from Cloud Storage**. -3. On the **Import Data from Amazon S3** page, provide the following information for the source Parquet files: +3. On the **Import Data from Cloud Storage** page, provide the following information: - - **Import File Count**: select **One file** or **Multiple files** as needed. - - **Included Schema Files**: this field is only visible when importing multiple files. If the source folder contains the target table schemas, select **Yes**. Otherwise, select **No**. - - **Data Format**: select **Parquet**. - - **File URI** or **Folder URI**: + - Select **Amazon S3** as the storage provider. + - enter the **Source Files URI**. - When importing one file, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `s3://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - - **Bucket Access**: you can use either an AWS Role ARN or an AWS access key to access your bucket. For more information, see [Configure Amazon S3 access](/tidb-cloud/serverless-external-storage.md#configure-amazon-s3-access). + - select the **Credential** - **AWS Role ARN**: enter the AWS Role ARN value. - **AWS Access Key**: enter the AWS access key ID and AWS secret access key. -4. Click **Connect**. +4. Click **Next**. -5. In the **Destination** section, select the target database and table. +5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When importing multiple files, you can use **Advanced Settings** > **Mapping Settings** to define a custom mapping rule for each target table and its corresponding Parquet file. After that, the data source files will be re-scanned using the provided custom mapping rule. + * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. - When you enter the source file URI and name in **Source File URIs and Names**, make sure it is in the following format `s3://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `s3://sampledata/ingest/TableName.01.parquet`. + After setting the custom mapping rules, the system will re-scan the source files accordingly. - You can also use wildcards to match the source files. For example: + * In **Source File URIs and Names**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. + * You can also use wildcards to match multiple files: - - `s3://[bucket_name]/[data_source_folder]/my-data?.parquet`: all Parquet files starting with `my-data` followed by one character (such as `my-data1.parquet` and `my-data2.parquet`) in that folder will be imported into the same target table. + * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). - - `s3://[bucket_name]/[data_source_folder]/my-data*.parquet`: all Parquet files in the folder starting with `my-data` will be imported into the same target table. + * `my-data*.parquet`: Matches all Parquet files that start with `my-data` (for example, `my-data-2023.parquet`, `my-data-final.parquet`). - Note that only `?` and `*` are supported. + > Only `*` and `?` wildcards are supported. - > **Note:** - > - > The URI must contain the data source folder. + * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + + * Select **Parquet** as the data format. + + > **Note:** + > + > When importing one file, you can only use mapping settings and the source file cannot be modified. + +6. Click **Next**. -6. Click **Start Import**. +7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. -7. When the import progress shows **Completed**, check the imported tables. +8. When the import progress shows **Completed**, check the imported tables. @@ -159,41 +164,46 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 2. Click the name of your target cluster to go to its overview page, and then click **Data** > **Import** in the left navigation pane. -2. Select **Import data from Cloud Storage**, and then click **Google Cloud Storage**. +2. Click **Import data from Cloud Storage**. -3. On the **Import Data from Google Cloud Storage** page, provide the following information for the source Parquet files: +3. On the **Import Data from Google Cloud Storage** page, provide the following information: - - **Import File Count**: select **One file** or **Multiple files** as needed. - - **Included Schema Files**: this field is only visible when importing multiple files. If the source folder contains the target table schemas, select **Yes**. Otherwise, select **No**. - - **Data Format**: select **Parquet**. - - **File URI** or **Folder URI**: + - Select **Google Cloud Storage** as the storage provider. + - **Source Files URI**. - When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `[gcs|gs]://sampledata/ingest/`. - - **Bucket Access**: you can use a GCS IAM Role to access your bucket. For more information, see [Configure GCS access](/tidb-cloud/serverless-external-storage.md#configure-gcs-access). + - **Credential**: you can use a GCS IAM Role Service Account key to access your bucket. For more information, see [Configure GCS access](/tidb-cloud/serverless-external-storage.md#configure-gcs-access). -4. Click **Connect**. +4. Click **Next**. -5. In the **Destination** section, select the target database and table. +5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When importing multiple files, you can use **Advanced Settings** > **Mapping Settings** to define a custom mapping rule for each target table and its corresponding Parquet file. After that, the data source files will be re-scanned using the provided custom mapping rule. + * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. - When you enter the source file URI and name in **Source File URIs and Names**, make sure it is in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.parquet`. + After setting the custom mapping rules, the system will re-scan the source files accordingly. - You can also use wildcards to match the source files. For example: + * In **Source File URIs and Names**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. + * You can also use wildcards to match multiple files: - - `[gcs|gs]://[bucket_name]/[data_source_folder]/my-data?.parquet`: all Parquet files starting with `my-data` followed by one character (such as `my-data1.parquet` and `my-data2.parquet`) in that folder will be imported into the same target table. + * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). - - `[gcs|gs]://[bucket_name]/[data_source_folder]/my-data*.parquet`: all Parquet files in the folder starting with `my-data` will be imported into the same target table. + * `my-data*.parquet`: Matches all Parquet files that start with `my-data` (for example, `my-data-2023.parquet`, `my-data-final.parquet`). - Note that only `?` and `*` are supported. + > Only `*` and `?` wildcards are supported. - > **Note:** - > - > The URI must contain the data source folder. + * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + + * Select **Parquet** as the data format. + + > **Note:** + > + > When importing one file, you can only use mapping settings and the source file cannot be modified. + +6. Click **Next**. -6. Click **Start Import**. +7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. -7. When the import progress shows **Completed**, check the imported tables. +8. When the import progress shows **Completed**, check the imported tables. @@ -209,41 +219,46 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 2. Click the name of your target cluster to go to its overview page, and then click **Data** > **Import** in the left navigation pane. -2. Select **Import data from Cloud Storage**, and then click **Azure Blob Storage**. +2. Click **Import data from Cloud Storage**. -3. On the **Import Data from Azure Blob Storage** page, provide the following information for the source Parquet files: +3. On the **Import Data from Cloud Storage** page, provide the following information: - - **Import File Count**: select **One file** or **Multiple files** as needed. - - **Included Schema Files**: this field is only visible when importing multiple files. If the source folder contains the target table schemas, select **Yes**. Otherwise, select **No**. - - **Data Format**: select **Parquet**. - - **File URI** or **Folder URI**: + - Select **Azure Blob Storage** as the storage provider. + - **Source Files URI**. - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `[azure|https]://sampledata/ingest/TableName.01.parquet`. - - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `[azure|https]://sampledata/ingest/`. - - **Bucket Access**: you can use a shared access signature (SAS) token to access your bucket. For more information, see [Configure Azure Blob Storage access](/tidb-cloud/serverless-external-storage.md#configure-azure-blob-storage-access). + - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. + - **Credential**: you can use a shared access signature (SAS) token to access your bucket. For more information, see [Configure Azure Blob Storage access](/tidb-cloud/serverless-external-storage.md#configure-azure-blob-storage-access) -4. Click **Connect**. +4. Click **Next**. -5. In the **Destination** section, select the target database and table. +5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When importing multiple files, you can use **Advanced Settings** > **Mapping Settings** to define a custom mapping rule for each target table and its corresponding Parquet file. After that, the data source files will be re-scanned using the provided custom mapping rule. + * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. - When you enter the source file URI and name in **Source File URIs and Names**, make sure it is in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `[azure|https]://sampledata/ingest/TableName.01.parquet`. + After setting the custom mapping rules, the system will re-scan the source files accordingly. - You can also use wildcards to match the source files. For example: + * In **Source File URIs and Names**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. + * You can also use wildcards to match multiple files: - - `[azure|https]://[bucket_name]/[data_source_folder]/my-data?.parquet`: all Parquet files starting with `my-data` followed by one character (such as `my-data1.parquet` and `my-data2.parquet`) in that folder will be imported into the same target table. + * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). - - `[azure|https]://[bucket_name]/[data_source_folder]/my-data*.parquet`: all Parquet files in the folder starting with `my-data` will be imported into the same target table. + * `my-data*.parquet`: Matches all Parquet files that start with `my-data` (for example, `my-data-2023.parquet`, `my-data-final.parquet`). - Note that only `?` and `*` are supported. + > Only `*` and `?` wildcards are supported. - > **Note:** - > - > The URI must contain the data source folder. + * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + + * Select **Parquet** as the data format. + + > **Note:** + > + > When importing one file, you can only use mapping settings and the source file cannot be modified. + +6. Click **Next**. -6. Click **Start Import**. +7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. -7. When the import progress shows **Completed**, check the imported tables. +8. When the import progress shows **Completed**, check the imported tables. @@ -259,41 +274,46 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 2. Click the name of your target cluster to go to its overview page, and then click **Data** > **Import** in the left navigation pane. -2. Select **Import data from Cloud Storage**, and then click **Alibaba Cloud OSS**. +2. Click **Import data from Cloud Storage**. -3. On the **Import Data from Alibaba Cloud OSS** page, provide the following information for the source Parquet files: +3. On the **Import Data from Cloud Storage** page, provide the following information: - - **Import File Count**: select **One file** or **Multiple files** as needed. - - **Included Schema Files**: this field is only visible when importing multiple files. If the source folder contains the target table schemas, select **Yes**. Otherwise, select **No**. - - **Data Format**: select **Parquet**. - - **File URI** or **Folder URI**: + - Select **Alibaba Cloud OSS** as the storage provider. + - **Source Files URI**. - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `oss://sampledata/ingest/TableName.01.parquet`. - - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `oss://sampledata/ingest/`. - - **Bucket Access**: you can use an AccessKey pair to access your bucket. For more information, see [Configure Alibaba Cloud Object Storage Service (OSS) access](/tidb-cloud/serverless-external-storage.md#configure-alibaba-cloud-object-storage-service-oss-access). + - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. + - **Credential**: you can use an AccessKey pair to access your bucket. For more information, see [Configure Alibaba Cloud Object Storage Service (OSS) access](/tidb-cloud/serverless-external-storage.md#configure-alibaba-cloud-object-storage-service-oss-access). -4. Click **Connect**. +4. Click **Next**. -5. In the **Destination** section, select the target database and table. +5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When importing multiple files, you can use **Advanced Settings** > **Mapping Settings** to define a custom mapping rule for each target table and its corresponding Parquet file. After that, the data source files will be re-scanned using the provided custom mapping rule. + * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. - When you enter the source file URI and name in **Source File URIs and Names**, make sure it is in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `oss://sampledata/ingest/TableName.01.parquet`. + After setting the custom mapping rules, the system will re-scan the source files accordingly. - You can also use wildcards to match the source files. For example: + * In **Source File URIs and Names**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. + * You can also use wildcards to match multiple files: - - `oss://[bucket_name]/[data_source_folder]/my-data?.parquet`: all Parquet files starting with `my-data` followed by one character (such as `my-data1.parquet` and `my-data2.parquet`) in that folder will be imported into the same target table. + * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). - - `oss://[bucket_name]/[data_source_folder]/my-data*.parquet`: all Parquet files in the folder starting with `my-data` will be imported into the same target table. + * `my-data*.parquet`: Matches all Parquet files that start with `my-data` (for example, `my-data-2023.parquet`, `my-data-final.parquet`). - Note that only `?` and `*` are supported. + > Only `*` and `?` wildcards are supported. - > **Note:** - > - > The URI must contain the data source folder. + * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + + * Select **Parquet** as the data format. + + > **Note:** + > + > When importing one file, you can only use mapping settings and the source file cannot be modified. + +6. Click **Next**. -6. Click **Start Import**. +7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. -7. When the import progress shows **Completed**, check the imported tables. +8. When the import progress shows **Completed**, check the imported tables. diff --git a/tidb-cloud/import-sample-data-serverless.md b/tidb-cloud/import-sample-data-serverless.md index ebd2cc08c2b1a..13102785d553f 100644 --- a/tidb-cloud/import-sample-data-serverless.md +++ b/tidb-cloud/import-sample-data-serverless.md @@ -21,19 +21,25 @@ This document describes how to import the sample data into TiDB Cloud Serverless 2. Click the name of your target cluster to go to its overview page, and then click **Data** > **Import** in the left navigation pane. -2. Select **Import data from Cloud Storage**, and then click **Amazon S3**. +2. Click **Import data from Cloud Storage**. -3. On the **Import Data from Amazon S3** page, configure the following source data information: +3. On the **Import Data from Cloud Storage** page, provide the following information: - - **Import File Count**: for the sample data, select **Multiple files**. - - **Included Schema Files**: for the sample data, select **Yes**. - - **Data Format**: select **SQL**. - - **Folder URI** or **File URI**: enter the sample data URI `s3://tidbcloud-sample-data/data-ingestion/`. - - **Bucket Access**: for the sample data, you can only use a Role ARN to access its bucket. For your own data, you can use either an AWS access key or a Role ARN to access your bucket. - - **AWS Role ARN**: enter `arn:aws:iam::801626783489:role/import-sample-access`. - - **AWS Access Key**: skip this option for the sample data. + - Select **Amazon S3** as the storage provider. + - **Source Files URI**: enter the sample data URI `s3://tidbcloud-sample-data/data-ingestion/`. + - select the **Credential** + - **AWS Role ARN**: enter `arn:aws:iam::801626783489:role/import-sample-access`. + - **AWS Access Key**: skip this option for the sample data. -4. Click **Connect** > **Start Import**. +4. Click **Next**. + +5. In the **Destination Mapping** section, use the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) and select **SQL** as the data format + +6. Click **Next**. + +7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. + +8. When the import progress shows **Completed**, check the imported tables. When the data import progress shows **Completed**, you have successfully imported the sample data and the database schema to your database in TiDB Cloud Serverless. From 74f1a4b821d8dceb9c722addca8e7a3064dbe986 Mon Sep 17 00:00:00 2001 From: yangxin Date: Wed, 9 Jul 2025 16:04:28 +0800 Subject: [PATCH 07/17] polish --- tidb-cloud/import-csv-files-serverless.md | 10 +++++----- tidb-cloud/import-parquet-files-serverless.md | 10 +++++----- tidb-cloud/import-sample-data-serverless.md | 2 +- 3 files changed, 11 insertions(+), 11 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index 39116052b128e..0a00039741274 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -104,10 +104,10 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - Select **Amazon S3** as the storage provider. - - enter the **Source Files URI**. + - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `s3://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - - select the **Credential** + - **Credential**: - **AWS Role ARN**: enter the AWS Role ARN value. - **AWS Access Key**: enter the AWS access key ID and AWS secret access key. @@ -161,7 +161,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - Select **Google Cloud Storage** as the storage provider. - - **Source Files URI**. + - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use a GCS IAM Role Service Account key to access your bucket. For more information, see [Configure GCS access](/tidb-cloud/serverless-external-storage.md#configure-gcs-access). @@ -216,7 +216,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - Select **Azure Blob Storage** as the storage provider. - - **Source Files URI**. + - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[azure|https]://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use a shared access signature (SAS) token to access your bucket. For more information, see [Configure Azure Blob Storage access](/tidb-cloud/serverless-external-storage.md#configure-azure-blob-storage-access) @@ -271,7 +271,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - Select **Alibaba Cloud OSS** as the storage provider. - - **Source Files URI**. + - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `oss://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use an AccessKey pair to access your bucket. For more information, see [Configure Alibaba Cloud Object Storage Service (OSS) access](/tidb-cloud/serverless-external-storage.md#configure-alibaba-cloud-object-storage-service-oss-access). diff --git a/tidb-cloud/import-parquet-files-serverless.md b/tidb-cloud/import-parquet-files-serverless.md index 927bf66587899..5caff376ec410 100644 --- a/tidb-cloud/import-parquet-files-serverless.md +++ b/tidb-cloud/import-parquet-files-serverless.md @@ -112,10 +112,10 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - Select **Amazon S3** as the storage provider. - - enter the **Source Files URI**. + - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `s3://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - - select the **Credential** + - **Credential**: - **AWS Role ARN**: enter the AWS Role ARN value. - **AWS Access Key**: enter the AWS access key ID and AWS secret access key. @@ -169,7 +169,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Google Cloud Storage** page, provide the following information: - Select **Google Cloud Storage** as the storage provider. - - **Source Files URI**. + - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `[gcs|gs]://sampledata/ingest/`. - **Credential**: you can use a GCS IAM Role Service Account key to access your bucket. For more information, see [Configure GCS access](/tidb-cloud/serverless-external-storage.md#configure-gcs-access). @@ -224,7 +224,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - Select **Azure Blob Storage** as the storage provider. - - **Source Files URI**. + - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `[azure|https]://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use a shared access signature (SAS) token to access your bucket. For more information, see [Configure Azure Blob Storage access](/tidb-cloud/serverless-external-storage.md#configure-azure-blob-storage-access) @@ -279,7 +279,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - Select **Alibaba Cloud OSS** as the storage provider. - - **Source Files URI**. + - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `oss://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - **Credential**: you can use an AccessKey pair to access your bucket. For more information, see [Configure Alibaba Cloud Object Storage Service (OSS) access](/tidb-cloud/serverless-external-storage.md#configure-alibaba-cloud-object-storage-service-oss-access). diff --git a/tidb-cloud/import-sample-data-serverless.md b/tidb-cloud/import-sample-data-serverless.md index 13102785d553f..8e609d056d292 100644 --- a/tidb-cloud/import-sample-data-serverless.md +++ b/tidb-cloud/import-sample-data-serverless.md @@ -27,7 +27,7 @@ This document describes how to import the sample data into TiDB Cloud Serverless - Select **Amazon S3** as the storage provider. - **Source Files URI**: enter the sample data URI `s3://tidbcloud-sample-data/data-ingestion/`. - - select the **Credential** + - **Credential**: - **AWS Role ARN**: enter `arn:aws:iam::801626783489:role/import-sample-access`. - **AWS Access Key**: skip this option for the sample data. From c8c09c483c76dd93cf87426f2b6adf242c29f950 Mon Sep 17 00:00:00 2001 From: yangxin Date: Thu, 10 Jul 2025 17:23:37 +0800 Subject: [PATCH 08/17] polish --- tidb-cloud/import-csv-files-serverless.md | 24 +++++++------------ tidb-cloud/import-parquet-files-serverless.md | 24 +++++++------------ 2 files changed, 16 insertions(+), 32 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index 0a00039741274..e841b4aeb2194 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -117,9 +117,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. - After setting the custom mapping rules, the system will re-scan the source files accordingly. - - * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. + * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). @@ -136,7 +134,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: > > When importing one file, you can only use mapping settings and the source file cannot be modified. -6. Click **Next**. +6. Click **Next**, the system will scan the source files accordingly. 7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. @@ -172,9 +170,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. - After setting the custom mapping rules, the system will re-scan the source files accordingly. - - * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. + * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). @@ -191,7 +187,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: > > When importing one file, you can only use mapping settings and the source file cannot be modified. -6. Click **Next**. +6. Click **Next**, the system will scan the source files accordingly.. 7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. @@ -227,9 +223,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. - After setting the custom mapping rules, the system will re-scan the source files accordingly. - - * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. + * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). @@ -246,7 +240,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: > > When importing one file, you can only use mapping settings and the source file cannot be modified. -6. Click **Next**. +6. Click **Next**, the system will scan the source files accordingly.. 7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. @@ -282,9 +276,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. - After setting the custom mapping rules, the system will re-scan the source files accordingly. - - * In **Source File URIs and Names**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. + * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). @@ -301,7 +293,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: > > When importing one file, you can only use mapping settings and the source file cannot be modified. -6. Click **Next**. +6. Click **Next**, the system will scan the source files accordingly.. 7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. diff --git a/tidb-cloud/import-parquet-files-serverless.md b/tidb-cloud/import-parquet-files-serverless.md index 5caff376ec410..28fcdf5fe1b6b 100644 --- a/tidb-cloud/import-parquet-files-serverless.md +++ b/tidb-cloud/import-parquet-files-serverless.md @@ -125,9 +125,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. - After setting the custom mapping rules, the system will re-scan the source files accordingly. - - * In **Source File URIs and Names**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. + * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. * You can also use wildcards to match multiple files: * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). @@ -144,7 +142,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: > > When importing one file, you can only use mapping settings and the source file cannot be modified. -6. Click **Next**. +6. Click **Next**, the system will scan the source files accordingly.. 7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. @@ -180,9 +178,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. - After setting the custom mapping rules, the system will re-scan the source files accordingly. - - * In **Source File URIs and Names**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. + * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. * You can also use wildcards to match multiple files: * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). @@ -199,7 +195,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: > > When importing one file, you can only use mapping settings and the source file cannot be modified. -6. Click **Next**. +6. Click **Next**, the system will scan the source files accordingly.. 7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. @@ -235,9 +231,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. - After setting the custom mapping rules, the system will re-scan the source files accordingly. - - * In **Source File URIs and Names**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. + * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. * You can also use wildcards to match multiple files: * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). @@ -254,7 +248,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: > > When importing one file, you can only use mapping settings and the source file cannot be modified. -6. Click **Next**. +6. Click **Next**, the system will scan the source files accordingly.. 7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. @@ -290,9 +284,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. - After setting the custom mapping rules, the system will re-scan the source files accordingly. - - * In **Source File URIs and Names**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. + * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. * You can also use wildcards to match multiple files: * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). @@ -309,7 +301,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: > > When importing one file, you can only use mapping settings and the source file cannot be modified. -6. Click **Next**. +6. Click **Next**, the system will scan the source files accordingly. 7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. From 20acb01f239e4bb4c0c11d89f6f3958ae2fa110c Mon Sep 17 00:00:00 2001 From: Leon Yang Date: Fri, 11 Jul 2025 10:42:44 +0800 Subject: [PATCH 09/17] Apply suggestions from code review Co-authored-by: Grace Cai --- tidb-cloud/import-csv-files-serverless.md | 12 ++++++------ tidb-cloud/import-parquet-files-serverless.md | 8 ++++---- tidb-cloud/import-sample-data-serverless.md | 2 +- 3 files changed, 11 insertions(+), 11 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index e841b4aeb2194..446e2a254d99a 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -103,11 +103,11 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - - Select **Amazon S3** as the storage provider. + - **Storage Provider**: select **Amazon S3**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `s3://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - - **Credential**: + - **Credential**: you can use either an AWS Role ARN or an AWS access key to access your bucket. For more information, see [Configure Amazon S3 access](/tidb-cloud/serverless-external-storage.md#configure-amazon-s3-access). - **AWS Role ARN**: enter the AWS Role ARN value. - **AWS Access Key**: enter the AWS access key ID and AWS secret access key. @@ -115,7 +115,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. + * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: @@ -158,7 +158,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - - Select **Google Cloud Storage** as the storage provider. + - **Storage Provider**: select **Google Cloud Storage**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. @@ -211,7 +211,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - - Select **Azure Blob Storage** as the storage provider. + - **Storage Provider**: select **Azure Blob Storage**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[azure|https]://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. @@ -264,7 +264,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - - Select **Alibaba Cloud OSS** as the storage provider. + - **Storage Provider**: select **Alibaba Cloud OSS**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `oss://sampledata/ingest/TableName.01.csv`. - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. diff --git a/tidb-cloud/import-parquet-files-serverless.md b/tidb-cloud/import-parquet-files-serverless.md index 28fcdf5fe1b6b..3f4e60cb25375 100644 --- a/tidb-cloud/import-parquet-files-serverless.md +++ b/tidb-cloud/import-parquet-files-serverless.md @@ -111,7 +111,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - - Select **Amazon S3** as the storage provider. + - **Storage Provider**: select **Amazon S3**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `s3://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. @@ -166,7 +166,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Google Cloud Storage** page, provide the following information: - - Select **Google Cloud Storage** as the storage provider. + - **Storage Provider**: select **Google Cloud Storage**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `[gcs|gs]://sampledata/ingest/`. @@ -219,7 +219,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - - Select **Azure Blob Storage** as the storage provider. + - **Storage Provider**: select **Azure Blob Storage**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `[azure|https]://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. @@ -272,7 +272,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 3. On the **Import Data from Cloud Storage** page, provide the following information: - - Select **Alibaba Cloud OSS** as the storage provider. + - **Storage Provider**: select **Alibaba Cloud OSS**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `oss://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. diff --git a/tidb-cloud/import-sample-data-serverless.md b/tidb-cloud/import-sample-data-serverless.md index 8e609d056d292..6092a617f37eb 100644 --- a/tidb-cloud/import-sample-data-serverless.md +++ b/tidb-cloud/import-sample-data-serverless.md @@ -25,7 +25,7 @@ This document describes how to import the sample data into TiDB Cloud Serverless 3. On the **Import Data from Cloud Storage** page, provide the following information: - - Select **Amazon S3** as the storage provider. + - **Storage Provider**: select **Amazon S3**. - **Source Files URI**: enter the sample data URI `s3://tidbcloud-sample-data/data-ingestion/`. - **Credential**: - **AWS Role ARN**: enter `arn:aws:iam::801626783489:role/import-sample-access`. From 24360c03d871f2bf11e9ff13da235b64d7587631 Mon Sep 17 00:00:00 2001 From: yangxin Date: Fri, 11 Jul 2025 10:46:10 +0800 Subject: [PATCH 10/17] polish --- tidb-cloud/import-csv-files-serverless.md | 6 +++--- tidb-cloud/import-parquet-files-serverless.md | 10 +++++----- 2 files changed, 8 insertions(+), 8 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index 446e2a254d99a..e505790d109be 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -168,7 +168,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. + * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: @@ -221,7 +221,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. + * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: @@ -274,7 +274,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source CSV file with its corresponding target table. + * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. * You can also use wildcards to match multiple files: diff --git a/tidb-cloud/import-parquet-files-serverless.md b/tidb-cloud/import-parquet-files-serverless.md index 3f4e60cb25375..7c1b1fde62f36 100644 --- a/tidb-cloud/import-parquet-files-serverless.md +++ b/tidb-cloud/import-parquet-files-serverless.md @@ -115,7 +115,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `s3://sampledata/ingest/TableName.01.parquet`. - When importing multiple files, enter the source file URI and name in the following format `s3://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. - - **Credential**: + - **Credential**: you can use either an AWS Role ARN or an AWS access key to access your bucket. For more information, see [Configure Amazon S3 access](/tidb-cloud/serverless-external-storage.md#configure-amazon-s3-access). - **AWS Role ARN**: enter the AWS Role ARN value. - **AWS Access Key**: enter the AWS access key ID and AWS secret access key. @@ -123,7 +123,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. + * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. * You can also use wildcards to match multiple files: @@ -176,7 +176,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. + * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. * You can also use wildcards to match multiple files: @@ -229,7 +229,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. + * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. * You can also use wildcards to match multiple files: @@ -282,7 +282,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * If you **do not use** the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md), you need to manually configure the mapping settings. This allows you to define custom rules to associate each source Parquet file with its corresponding target table. + * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. * You can also use wildcards to match multiple files: From 27eb027637aca6c38a038f708dc4f623e679eeab Mon Sep 17 00:00:00 2001 From: Test User Date: Fri, 11 Jul 2025 14:17:37 +0800 Subject: [PATCH 11/17] csv: update the steps for mapping source files --- tidb-cloud/import-csv-files-serverless.md | 26 +++++++++++------------ 1 file changed, 12 insertions(+), 14 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index e505790d109be..b885b6683bb20 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -115,28 +115,26 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. + The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. - * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. - * You can also use wildcards to match multiple files: + > **Note:** + > + > When importing a single file, you can only use the default mapping settings, and the **Source** field cannot be modified. - * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). + - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. - * `my-data*.csv`: Matches all CSV files that start with `my-data` (for example, `my-data-2023.csv`, `my-data-final.csv`). + - To manually configure the mapping rules to associate your source CSV files with the target database and table, unselect this option, and then fill in the following fields: - > Only `*` and `?` wildcards are supported. + - **Source**: enter the file name pattern in the `[file_name].csv` format. For example: `TableName.01.csv`. You can also use wildcards to match multiple files. Only `*` and `?` wildcards are supported. - * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + - `my-data?.csv`: matches all CSV files that start with `my-data` followed by a single character, such as `my-data1.csv` and `my-data2.csv`. + - `my-data*.csv`: matches all CSV files that start with `my-data`, such as `my-data-2023.csv` and `my-data-final.csv`. - * Select **CSV** as the data format. + - **Target Database** and **Target Table**: select the target database and table to import the data to. - > **Note:** - > - > When importing one file, you can only use mapping settings and the source file cannot be modified. - -6. Click **Next**, the system will scan the source files accordingly. +6. Click **Next**. TiDB Cloud scans the source files accordingly. -7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. +7. Review the scan results, check the data files found and corresponding target tables, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. From c961f2114319a1ee93c9e13b9b1dd62e47574ba6 Mon Sep 17 00:00:00 2001 From: Test User Date: Fri, 11 Jul 2025 14:37:48 +0800 Subject: [PATCH 12/17] csv: update step 5 to step 8 --- tidb-cloud/import-csv-files-serverless.md | 82 +++++++++++------------ 1 file changed, 38 insertions(+), 44 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index b885b6683bb20..3eb628dee0745 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -115,11 +115,11 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. + When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. > **Note:** > - > When importing a single file, you can only use the default mapping settings, and the **Source** field cannot be modified. + > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. @@ -166,28 +166,26 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. + When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. - * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. - * You can also use wildcards to match multiple files: - - * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). + > **Note:** + > + > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - * `my-data*.csv`: Matches all CSV files that start with `my-data` (for example, `my-data-2023.csv`, `my-data-final.csv`). + - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. - > Only `*` and `?` wildcards are supported. + - To manually configure the mapping rules to associate your source CSV files with the target database and table, unselect this option, and then fill in the following fields: - * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + - **Source**: enter the file name pattern in the `[file_name].csv` format. For example: `TableName.01.csv`. You can also use wildcards to match multiple files. Only `*` and `?` wildcards are supported. - * Select **CSV** as the data format. + - `my-data?.csv`: matches all CSV files that start with `my-data` followed by a single character, such as `my-data1.csv` and `my-data2.csv`. + - `my-data*.csv`: matches all CSV files that start with `my-data`, such as `my-data-2023.csv` and `my-data-final.csv`. - > **Note:** - > - > When importing one file, you can only use mapping settings and the source file cannot be modified. + - **Target Database** and **Target Table**: select the target database and table to import the data to. -6. Click **Next**, the system will scan the source files accordingly.. +6. Click **Next**. TiDB Cloud scans the source files accordingly. -7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. +7. Review the scan results, check the data files found and corresponding target tables, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. @@ -219,28 +217,26 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. + When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. - * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. - * You can also use wildcards to match multiple files: - - * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). + > **Note:** + > + > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - * `my-data*.csv`: Matches all CSV files that start with `my-data` (for example, `my-data-2023.csv`, `my-data-final.csv`). + - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. - > Only `*` and `?` wildcards are supported. + - To manually configure the mapping rules to associate your source CSV files with the target database and table, unselect this option, and then fill in the following fields: - * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + - **Source**: enter the file name pattern in the `[file_name].csv` format. For example: `TableName.01.csv`. You can also use wildcards to match multiple files. Only `*` and `?` wildcards are supported. - * Select **CSV** as the data format. + - `my-data?.csv`: matches all CSV files that start with `my-data` followed by a single character, such as `my-data1.csv` and `my-data2.csv`. + - `my-data*.csv`: matches all CSV files that start with `my-data`, such as `my-data-2023.csv` and `my-data-final.csv`. - > **Note:** - > - > When importing one file, you can only use mapping settings and the source file cannot be modified. + - **Target Database** and **Target Table**: select the target database and table to import the data to. -6. Click **Next**, the system will scan the source files accordingly.. +6. Click **Next**. TiDB Cloud scans the source files accordingly. -7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. +7. Review the scan results, check the data files found and corresponding target tables, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. @@ -272,28 +268,26 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. - - * In **Source**, enter file names in the format `[file_name].csv`. For example: `TableName.01.csv`. - * You can also use wildcards to match multiple files: + When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. - * `my-data?.csv`: Matches all CSV files that start with `my-data` followed by a single character (for example, `my-data1.csv`, `my-data2.csv`). + > **Note:** + > + > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - * `my-data*.csv`: Matches all CSV files that start with `my-data` (for example, `my-data-2023.csv`, `my-data-final.csv`). + - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. - > Only `*` and `?` wildcards are supported. + - To manually configure the mapping rules to associate your source CSV files with the target database and table, unselect this option, and then fill in the following fields: - * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + - **Source**: enter the file name pattern in the `[file_name].csv` format. For example: `TableName.01.csv`. You can also use wildcards to match multiple files. Only `*` and `?` wildcards are supported. - * Select **CSV** as the data format. + - `my-data?.csv`: matches all CSV files that start with `my-data` followed by a single character, such as `my-data1.csv` and `my-data2.csv`. + - `my-data*.csv`: matches all CSV files that start with `my-data`, such as `my-data-2023.csv` and `my-data-final.csv`. - > **Note:** - > - > When importing one file, you can only use mapping settings and the source file cannot be modified. + - **Target Database** and **Target Table**: select the target database and table to import the data to. -6. Click **Next**, the system will scan the source files accordingly.. +6. Click **Next**. TiDB Cloud scans the source files accordingly. -7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. +7. Review the scan results, check the data files found and corresponding target tables, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. From 69e6f59dbfd137d33a3ab3ff1439bc3ec3bee3f1 Mon Sep 17 00:00:00 2001 From: Test User Date: Fri, 11 Jul 2025 15:02:25 +0800 Subject: [PATCH 13/17] parquet: update step5 to step8 --- tidb-cloud/import-parquet-files-serverless.md | 104 ++++++++---------- 1 file changed, 48 insertions(+), 56 deletions(-) diff --git a/tidb-cloud/import-parquet-files-serverless.md b/tidb-cloud/import-parquet-files-serverless.md index 7c1b1fde62f36..254198cf8fd24 100644 --- a/tidb-cloud/import-parquet-files-serverless.md +++ b/tidb-cloud/import-parquet-files-serverless.md @@ -123,28 +123,26 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. + When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. - * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. - * You can also use wildcards to match multiple files: - - * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). + > **Note:** + > + > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - * `my-data*.parquet`: Matches all Parquet files that start with `my-data` (for example, `my-data-2023.parquet`, `my-data-final.parquet`). + - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. - > Only `*` and `?` wildcards are supported. + - To manually configure the mapping rules to associate your source Parquet files with the target database and table, unselect this option, and then fill in the following fields: - * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + - **Source**: enter the file name pattern in the `[file_name].parquet` format. For example: `TableName.01.parquet`. You can also use wildcards to match multiple files. Only `*` and `?` wildcards are supported. - * Select **Parquet** as the data format. + - `my-data?.parquet`: matches all Parquet files that start with `my-data` followed by a single character, such as `my-data1.parquet` and `my-data2.parquet`. + - `my-data*.parquet`: matches all Parquet files that start with `my-data`, such as `my-data-2023.parquet` and `my-data-final.parquet`. - > **Note:** - > - > When importing one file, you can only use mapping settings and the source file cannot be modified. + - **Target Database** and **Target Table**: select the target database and table to import the data to. -6. Click **Next**, the system will scan the source files accordingly.. +6. Click **Next**. TiDB Cloud scans the source files accordingly. -7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. +7. Review the scan results, check the data files found and corresponding target tables, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. @@ -176,28 +174,26 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. + When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. - * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. - * You can also use wildcards to match multiple files: - - * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). + > **Note:** + > + > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - * `my-data*.parquet`: Matches all Parquet files that start with `my-data` (for example, `my-data-2023.parquet`, `my-data-final.parquet`). + - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. - > Only `*` and `?` wildcards are supported. + - To manually configure the mapping rules to associate your source Parquet files with the target database and table, unselect this option, and then fill in the following fields: - * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + - **Source**: enter the file name pattern in the `[file_name].parquet` format. For example: `TableName.01.parquet`. You can also use wildcards to match multiple files. Only `*` and `?` wildcards are supported. - * Select **Parquet** as the data format. + - `my-data?.parquet`: matches all Parquet files that start with `my-data` followed by a single character, such as `my-data1.parquet` and `my-data2.parquet`. + - `my-data*.parquet`: matches all Parquet files that start with `my-data`, such as `my-data-2023.parquet` and `my-data-final.parquet`. - > **Note:** - > - > When importing one file, you can only use mapping settings and the source file cannot be modified. + - **Target Database** and **Target Table**: select the target database and table to import the data to. -6. Click **Next**, the system will scan the source files accordingly.. +6. Click **Next**. TiDB Cloud scans the source files accordingly. -7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. +7. Review the scan results, check the data files found and corresponding target tables, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. @@ -229,28 +225,26 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. + When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. - * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. - * You can also use wildcards to match multiple files: - - * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). + > **Note:** + > + > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - * `my-data*.parquet`: Matches all Parquet files that start with `my-data` (for example, `my-data-2023.parquet`, `my-data-final.parquet`). + - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. - > Only `*` and `?` wildcards are supported. + - To manually configure the mapping rules to associate your source Parquet files with the target database and table, unselect this option, and then fill in the following fields: - * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + - **Source**: enter the file name pattern in the `[file_name].parquet` format. For example: `TableName.01.parquet`. You can also use wildcards to match multiple files. Only `*` and `?` wildcards are supported. - * Select **Parquet** as the data format. + - `my-data?.parquet`: matches all Parquet files that start with `my-data` followed by a single character, such as `my-data1.parquet` and `my-data2.parquet`. + - `my-data*.parquet`: matches all Parquet files that start with `my-data`, such as `my-data-2023.parquet` and `my-data-final.parquet`. - > **Note:** - > - > When importing one file, you can only use mapping settings and the source file cannot be modified. + - **Target Database** and **Target Table**: select the target database and table to import the data to. -6. Click **Next**, the system will scan the source files accordingly.. +6. Click **Next**. TiDB Cloud scans the source files accordingly. -7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. +7. Review the scan results, check the data files found and corresponding target tables, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. @@ -282,28 +276,26 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - * The **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. If you unselect it, you need to manually configure the mapping rules to associate your source CSV files with the target database and table. + When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. - * In **Source**, enter file names in the format `[file_name].parquet`. For example: `TableName.01.parquet`. - * You can also use wildcards to match multiple files: - - * `my-data?.parquet`: Matches all Parquet files that start with `my-data` followed by a single character (for example, `my-data1.parquet`, `my-data2.parquet`). + > **Note:** + > + > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - * `my-data*.parquet`: Matches all Parquet files that start with `my-data` (for example, `my-data-2023.parquet`, `my-data-final.parquet`). + - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. - > Only `*` and `?` wildcards are supported. + - To manually configure the mapping rules to associate your source Parquet files with the target database and table, unselect this option, and then fill in the following fields: - * If you **use** the naming conventions, all source files that conform to the format will be automatically mapped to their corresponding tables. + - **Source**: enter the file name pattern in the `[file_name].parquet` format. For example: `TableName.01.parquet`. You can also use wildcards to match multiple files. Only `*` and `?` wildcards are supported. - * Select **Parquet** as the data format. + - `my-data?.parquet`: matches all Parquet files that start with `my-data` followed by a single character, such as `my-data1.parquet` and `my-data2.parquet`. + - `my-data*.parquet`: matches all Parquet files that start with `my-data`, such as `my-data-2023.parquet` and `my-data-final.parquet`. - > **Note:** - > - > When importing one file, you can only use mapping settings and the source file cannot be modified. + - **Target Database** and **Target Table**: select the target database and table to import the data to. -6. Click **Next**, the system will scan the source files accordingly. +6. Click **Next**. TiDB Cloud scans the source files accordingly. -7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. +7. Review the scan results, check the data files found and corresponding target tables, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. From 845ea1c635bc1daf02d48112383e3991d08f8260 Mon Sep 17 00:00:00 2001 From: Grace Cai Date: Fri, 11 Jul 2025 15:08:58 +0800 Subject: [PATCH 14/17] minor wording updates --- tidb-cloud/import-sample-data-serverless.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/tidb-cloud/import-sample-data-serverless.md b/tidb-cloud/import-sample-data-serverless.md index 6092a617f37eb..4f3e0c0fb9469 100644 --- a/tidb-cloud/import-sample-data-serverless.md +++ b/tidb-cloud/import-sample-data-serverless.md @@ -33,11 +33,11 @@ This document describes how to import the sample data into TiDB Cloud Serverless 4. Click **Next**. -5. In the **Destination Mapping** section, use the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) and select **SQL** as the data format +5. In the **Destination Mapping** section, keep the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option selected and select **SQL** as the data format. 6. Click **Next**. -7. Review the scan result, check data files found and target tables to import, and then click **Start Import**. +7. Review the scan results, check the data files found and corresponding target tables, and then click **Start Import**. 8. When the import progress shows **Completed**, check the imported tables. From 51b5e0037f5abebbbc816f689dc22e53be24ac02 Mon Sep 17 00:00:00 2001 From: Grace Cai Date: Fri, 11 Jul 2025 15:09:38 +0800 Subject: [PATCH 15/17] correct the sample url --- tidb-cloud/import-csv-files-serverless.md | 6 +++--- tidb-cloud/import-parquet-files-serverless.md | 2 +- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index 3eb628dee0745..0aabd70c47483 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -159,7 +159,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: - **Storage Provider**: select **Google Cloud Storage**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[gcs|gs]://sampledata/ingest/TableName.01.csv`. - - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. + - When importing multiple files, enter the source file URI and name in the following format `[gcs|gs]://[bucket_name]/[data_source_folder]/`. For example, `[gcs|gs]://sampledata/ingest/`. - **Credential**: you can use a GCS IAM Role Service Account key to access your bucket. For more information, see [Configure GCS access](/tidb-cloud/serverless-external-storage.md#configure-gcs-access). 4. Click **Next**. @@ -210,7 +210,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: - **Storage Provider**: select **Azure Blob Storage**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `[azure|https]://sampledata/ingest/TableName.01.csv`. - - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. + - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `[azure|https]://sampledata/ingest/`. - **Credential**: you can use a shared access signature (SAS) token to access your bucket. For more information, see [Configure Azure Blob Storage access](/tidb-cloud/serverless-external-storage.md#configure-azure-blob-storage-access) 4. Click **Next**. @@ -261,7 +261,7 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: - **Storage Provider**: select **Alibaba Cloud OSS**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].csv`. For example, `oss://sampledata/ingest/TableName.01.csv`. - - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. + - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `oss://sampledata/ingest/`. - **Credential**: you can use an AccessKey pair to access your bucket. For more information, see [Configure Alibaba Cloud Object Storage Service (OSS) access](/tidb-cloud/serverless-external-storage.md#configure-alibaba-cloud-object-storage-service-oss-access). 4. Click **Next**. diff --git a/tidb-cloud/import-parquet-files-serverless.md b/tidb-cloud/import-parquet-files-serverless.md index 254198cf8fd24..36cff36a76aa8 100644 --- a/tidb-cloud/import-parquet-files-serverless.md +++ b/tidb-cloud/import-parquet-files-serverless.md @@ -218,7 +218,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: - **Storage Provider**: select **Azure Blob Storage**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `[azure|https]://sampledata/ingest/TableName.01.parquet`. - - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. + - When importing multiple files, enter the source file URI and name in the following format `[azure|https]://[bucket_name]/[data_source_folder]/`. For example, `[azure|https]://sampledata/ingest/`. - **Credential**: you can use a shared access signature (SAS) token to access your bucket. For more information, see [Configure Azure Blob Storage access](/tidb-cloud/serverless-external-storage.md#configure-azure-blob-storage-access) 4. Click **Next**. From ba9c74a6db5fa7837e134c886cf7764ec56c2234 Mon Sep 17 00:00:00 2001 From: Grace Cai Date: Fri, 11 Jul 2025 15:09:52 +0800 Subject: [PATCH 16/17] Update tidb-cloud/import-parquet-files-serverless.md --- tidb-cloud/import-parquet-files-serverless.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tidb-cloud/import-parquet-files-serverless.md b/tidb-cloud/import-parquet-files-serverless.md index 36cff36a76aa8..7922b3077eab7 100644 --- a/tidb-cloud/import-parquet-files-serverless.md +++ b/tidb-cloud/import-parquet-files-serverless.md @@ -269,7 +269,7 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: - **Storage Provider**: select **Alibaba Cloud OSS**. - **Source Files URI**: - When importing one file, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/[file_name].parquet`. For example, `oss://sampledata/ingest/TableName.01.parquet`. - - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `s3://sampledata/ingest/`. + - When importing multiple files, enter the source file URI and name in the following format `oss://[bucket_name]/[data_source_folder]/`. For example, `oss://sampledata/ingest/`. - **Credential**: you can use an AccessKey pair to access your bucket. For more information, see [Configure Alibaba Cloud Object Storage Service (OSS) access](/tidb-cloud/serverless-external-storage.md#configure-alibaba-cloud-object-storage-service-oss-access). 4. Click **Next**. From 7f416e2e3d8e5b1af1360502bb83a6a1f5110654 Mon Sep 17 00:00:00 2001 From: yangxin Date: Mon, 14 Jul 2025 17:16:12 +0800 Subject: [PATCH 17/17] rename to File naming conventions --- tidb-cloud/import-csv-files-serverless.md | 24 +++++++++---------- tidb-cloud/import-parquet-files-serverless.md | 24 +++++++++---------- tidb-cloud/import-sample-data-serverless.md | 2 +- 3 files changed, 25 insertions(+), 25 deletions(-) diff --git a/tidb-cloud/import-csv-files-serverless.md b/tidb-cloud/import-csv-files-serverless.md index 0aabd70c47483..bc75aae79bc01 100644 --- a/tidb-cloud/import-csv-files-serverless.md +++ b/tidb-cloud/import-csv-files-serverless.md @@ -115,13 +115,13 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. + When a directory is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. > **Note:** > - > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. + > When a single file is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. + - To let TiDB Cloud automatically map all source files that follow the [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. - To manually configure the mapping rules to associate your source CSV files with the target database and table, unselect this option, and then fill in the following fields: @@ -166,13 +166,13 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. + When a directory is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. > **Note:** > - > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. + > When a single file is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. + - To let TiDB Cloud automatically map all source files that follow the [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. - To manually configure the mapping rules to associate your source CSV files with the target database and table, unselect this option, and then fill in the following fields: @@ -217,13 +217,13 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. + When a directory is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. > **Note:** > - > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. + > When a single file is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. + - To let TiDB Cloud automatically map all source files that follow the [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. - To manually configure the mapping rules to associate your source CSV files with the target database and table, unselect this option, and then fill in the following fields: @@ -268,13 +268,13 @@ To import the CSV files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. + When a directory is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. > **Note:** > - > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. + > When a single file is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. + - To let TiDB Cloud automatically map all source files that follow the [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **CSV** as the data format. - To manually configure the mapping rules to associate your source CSV files with the target database and table, unselect this option, and then fill in the following fields: diff --git a/tidb-cloud/import-parquet-files-serverless.md b/tidb-cloud/import-parquet-files-serverless.md index 7922b3077eab7..e56be8b584b0c 100644 --- a/tidb-cloud/import-parquet-files-serverless.md +++ b/tidb-cloud/import-parquet-files-serverless.md @@ -123,13 +123,13 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. + When a directory is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. > **Note:** > - > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. + > When a single file is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. + - To let TiDB Cloud automatically map all source files that follow the [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. - To manually configure the mapping rules to associate your source Parquet files with the target database and table, unselect this option, and then fill in the following fields: @@ -174,13 +174,13 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. + When a directory is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. > **Note:** > - > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. + > When a single file is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. + - To let TiDB Cloud automatically map all source files that follow the [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. - To manually configure the mapping rules to associate your source Parquet files with the target database and table, unselect this option, and then fill in the following fields: @@ -225,13 +225,13 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. + When a directory is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. > **Note:** > - > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. + > When a single file is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. + - To let TiDB Cloud automatically map all source files that follow the [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. - To manually configure the mapping rules to associate your source Parquet files with the target database and table, unselect this option, and then fill in the following fields: @@ -276,13 +276,13 @@ To import the Parquet files to TiDB Cloud Serverless, take the following steps: 5. In the **Destination Mapping** section, specify how source files are mapped to target tables. - When a directory is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. + When a directory is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is selected by default. > **Note:** > - > When a single file is specified in **Source Files URI**, the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. + > When a single file is specified in **Source Files URI**, the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option is not displayed, and TiDB Cloud automatically populates the **Source** field with the file name. In this case, you only need to select the target database and table for data import. - - To let TiDB Cloud automatically map all source files that follow the [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. + - To let TiDB Cloud automatically map all source files that follow the [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) to their corresponding tables, keep this option selected and select **Parquet** as the data format. - To manually configure the mapping rules to associate your source Parquet files with the target database and table, unselect this option, and then fill in the following fields: diff --git a/tidb-cloud/import-sample-data-serverless.md b/tidb-cloud/import-sample-data-serverless.md index 4f3e0c0fb9469..b1a506766a673 100644 --- a/tidb-cloud/import-sample-data-serverless.md +++ b/tidb-cloud/import-sample-data-serverless.md @@ -33,7 +33,7 @@ This document describes how to import the sample data into TiDB Cloud Serverless 4. Click **Next**. -5. In the **Destination Mapping** section, keep the **Use [TiDB Dumpling file naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option selected and select **SQL** as the data format. +5. In the **Destination Mapping** section, keep the **Use [File naming conventions](/tidb-cloud/naming-conventions-for-data-import.md) for automatic mapping** option selected and select **SQL** as the data format. 6. Click **Next**.