Updated on 2026-03-20 GMT+08:00

To Doris

Table 1 Parameter description

Type

Parameter

Description

Example Value

Basic parameters

Schema/Table Space

Name of the database to which data will be written. The schema can be automatically created. Click the icon next to the text box to select a schema or tablespace.

schema

Table Name

Name of the table to which data will be written. Click the icon next to the text box. The dialog box for selecting the table is displayed.

This parameter can be configured as a macro variable of date and time and a path name can contain multiple macro variables. When the macro variable of date and time works with a scheduled job, the incremental data can be synchronized periodically. For details, see Incremental Synchronization Using the Macro Variables of Date and Time.

NOTE:

If you have configured a macro variable of date and time and schedule a CDM job through DataArts Studio DataArts Factory, the system replaces the macro variable of date and time with (Planned start time of the data development jobOffset) rather than (Actual start time of the CDM jobOffset).

table

Clear Data Before Import

Whether to clear the data in the destination table before data import. The options are as follows:
  • Do not clear: The data in the destination table is not cleared before data import. The imported data is just added to the table.
  • Clear all data: All data is cleared from the destination table before data import.
  • Clear part of data: Part of the data in the destination table is cleared before data import. If you select Clear part of data, you must configure WHERE Clause to specify which part will be deleted.

Clear part of data

Where Clause

This parameter is displayed when Clear Data Before Import is set to Clear part of data.

It specifies the data to be deleted from the destination table before data import.

age > 18 and age <=60

Write Mode

STREAM_LOAD is recommended.

  • JDBC: JDBC is used to execute SQL statements to write data.
  • STREAM_LOAD: STREAM_LOAD is used to import data.

STREAM_LOAD

stream load config properties

Set stream load parameters when you set Write Mode to STREAM_LOAD.

The parameters are included in the request header for writing data using stream load. The following parameters are supported:

  • Public parameters:
    • format: format of the data to be imported. CSV and JSON are supported. CSV is recommended because it delivers better performance.
    • partitions: partitions to which data is to be imported. If this parameter is not specified, data is imported to all partitions of the Doris table by default. Use commas (,) to separate partitions, such as p1, p2.
    • strict_mode: whether to enable the strict mode. The default value is true. If this parameter is set to false, the field conversion fails and the value is set to NULL. This will cause data loss.
    • timeout: timeout interval for the data import job. The value ranges from 1 to 259200. The unit is second. The default value is 600.
    • compress_type: format that CSV data is compressed into by Doris. The value can be gz, lz4, or none.
  • Parameters supported for the JSON format:

    strip_outer_array: whether to strip the outermost array structure. The default value is true.

  • Parameters supported for the CSV format:
    • line_separator: line separator in the Doris data source. The default value is \n for the CSV format.
    • row_delimiter: row delimiter for the StarRocks data source. The default value is \n for the CSV format.
    • column_separator: field separator. The default value is \t for the CSV format.

-

Advanced attributes

Pre-import SQL Statement

The SQL statement that is first executed before a task is executed. Currently, only one SQL statement can be executed in wizard mode.

create temp table

Post-import SQL Statement

The SQL statement that is executed after a task is executed. Currently, only one SQL statement can be executed.

merge into

Number of Loader Threads

Number of threads started in each loader. A larger number allows more concurrent write operations. Conflict handling policies do not support "replace into" or "on duplicate key update".

1