Doris
Last updated
Last updated
Doris sink connector
exactly-once & cdc supported Doris version is >= 1.1.x
Array data type supported Doris version is >= 1.2.x
Map data type will be support in Doris version is 2.x
Used to send data to Doris. Both support streaming and batch mode. The internal implementation of Doris sink connector is cached and imported by stream load in batches.
fenodes
String
Yes
-
Doris
cluster fenodes address, the format is "fe_ip:fe_http_port, ..."
query-port
int
No
9030
Doris
Fenodes query_port
username
String
Yes
-
Doris
user username
password
String
Yes
-
Doris
user password
database
String
Yes
-
The database name of Doris
table, use ${database_name}
to represent the upstream table name
table
String
Yes
-
The table name of Doris
table, use ${table_name}
to represent the upstream table name
table.identifier
String
Yes
-
The name of Doris
table, it will deprecate after version 2.3.5, please use database
and table
instead.
sink.label-prefix
String
Yes
-
The label prefix used by stream load imports. In the 2pc scenario, global uniqueness is required to ensure the EOS semantics of Nexus.
sink.enable-2pc
bool
No
false
sink.enable-delete
bool
No
-
sink.check-interval
int
No
10000
check exception with the interval while loading
sink.max-retries
int
No
3
the max retry times if writing records to database failed
sink.buffer-size
int
No
256 * 1024
the buffer size to cache data for stream load.
sink.buffer-count
int
No
3
the buffer count to cache data for stream load.
doris.batch.size
int
No
1024
the batch size of the write to doris each http request, when the row reaches the size or checkpoint is executed, the data of cached will write to server.
needs_unsupported_type_casting
boolean
No
false
Whether to enable the unsupported type casting, such as Decimal64 to Double
schema_save_mode
Enum
no
CREATE_SCHEMA_WHEN_NOT_EXIST
the schema save mode, please refer to schema_save_mode
below
data_save_mode
Enum
no
APPEND_DATA
the data save mode, please refer to data_save_mode
below
save_mode_create_template
string
no
see below
see below
custom_sql
String
no
-
When data_save_mode selects CUSTOM_PROCESSING, you should fill in the CUSTOM_SQL parameter. This parameter usually fills in a SQL that can be executed. SQL will be executed before synchronization tasks.
doris.config
map
yes
-
This option is used to support operations such as insert
, delete
, and update
when automatically generate sql,and supported formats.
Before the synchronous task is turned on, different treatment schemes are selected for the existing surface structure of the target side.
Option introduction:
RECREATE_SCHEMA
:Will create when the table does not exist, delete and rebuild when the table is saved
CREATE_SCHEMA_WHEN_NOT_EXIST
:Will Created when the table does not exist, skipped when the table is saved
ERROR_WHEN_SCHEMA_NOT_EXIST
:Error will be reported when the table does not exist
Before the synchronous task is turned on, different processing schemes are selected for data existing data on the target side.
Option introduction:
DROP_DATA
: Preserve database structure and delete data
APPEND_DATA
:Preserve database structure, preserve data
CUSTOM_PROCESSING
:User defined processing
ERROR_WHEN_DATA_EXISTS
:When there is data, an error is reported
We use templates to automatically create Doris tables, which will create corresponding table creation statements based on the type of upstream data and schema type, and the default template can be modified according to the situation.
Default template:
If a custom field is filled in the template, such as adding an id
field
The connector will automatically obtain the corresponding type from the upstream to complete the filling, and remove the id field from rowtype_fields
. This method can be used to customize the modification of field types and attributes.
You can use the following placeholders
database: Used to get the database in the upstream schema
table_name: Used to get the table name in the upstream schema
rowtype_fields: Used to get all the fields in the upstream schema, we will automatically map to the field description of Doris
rowtype_primary_key: Used to get the primary key in the upstream schema (maybe a list)
rowtype_unique_key: Used to get the unique key in the upstream schema (maybe a list)
rowtype_duplicate_key: Used to get the duplicate key in the upstream schema (only for doris source, maybe a list)
BOOLEAN
BOOLEAN
TINYINT
TINYINT
SMALLINT
SMALLINT TINYINT
INT
INT SMALLINT TINYINT
BIGINT
BIGINT INT SMALLINT TINYINT
LARGEINT
BIGINT INT SMALLINT TINYINT
FLOAT
FLOAT
DOUBLE
DOUBLE FLOAT
DECIMAL
DECIMAL DOUBLE FLOAT
DATE
DATE
DATETIME
TIMESTAMP
CHAR
STRING
VARCHAR
STRING
STRING
STRING
ARRAY
ARRAY
MAP
MAP
JSON
STRING
HLL
Not supported yet
BITMAP
Not supported yet
QUANTILE_STATE
Not supported yet
STRUCT
Not supported yet
The supported formats include CSV and JSON
The following example describes writing multiple data types to Doris, and users need to create corresponding tables downstream
This example defines a Nexus synchronization task that automatically generates data through FakeSource and sends it to Doris Sink,FakeSource simulates CDC data with schema, score (int type),Doris needs to create a table sink named test.e2e_table_sink and a corresponding table for it.
Whether to enable two-phase commit (2pc), the default is false. For two-phase commit, please refer to .
Whether to enable deletion. This option requires Doris table to enable batch delete function (0.15+ version is enabled by default), and only supports Unique model. you can get more detail at this
Supported import data formats
example1
example2