Development Rules
The following table describes the parameter specifications you need to comply with to read Hudi tables on Flink streams.
Parameter |
Mandatory |
Description |
Example |
---|---|---|---|
Connector |
Yes |
Type of the table to be read |
hudi |
Path |
Yes |
Path for storing the table |
Set this parameter based on site requirements. |
table.type |
Yes |
Hudi table type. The default value is COPY_ON_WRITE. |
MERGE_ON_READ |
hoodie.datasource.write.recordkey.field |
Yes |
Primary key of the table |
Set this parameter as needed. |
write.precombine.field |
Yes |
Data combination field |
Set this parameter as needed. |
read.tasks |
No |
Hudi table read parallelism. The default value is 4. |
4 |
read.streaming.enabled |
Yes |
|
Set this parameter based on the site requirements. For streaming read, set this parameter to true. |
read.streaming.start-commit |
No |
Start commit (closed interval) in the yyyyMMddHHmmss format. By default, the latest commit is used. |
- |
hoodie.datasource.write.keygenerator.type |
No |
Primary key generation type of the upstream table |
COMPLEX |
read.streaming.check-interval |
No |
Check interval for finding new source commits. The default value is 1 minute. |
5 (The default value is recommended for heavy traffic.) |
read.end-commit |
No |
|
- |
changelog.enabled |
No |
Whether to write changelog messages. The default value is false. Set this parameter to true for CDC. |
false |
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.See the reply and handling status in My Cloud VOC.
For any further questions, feel free to contact us through the chatbot.
Chatbot