Recording the Time When Data Is Written to the Database
When you create a job on the CDM console to migrate tables or files of a relational database, you can add a field to record the time when they were written to the database.
Prerequisites
A link has been created, and the source end of the connector is a relational database.
Creating a Table/File Migration Job
- Create a table/file migration job, and select the created source connector and destination connector.
Figure 1 Configuring the job
- Click Next to go to the Map Field page and click .
Figure 2 Configuring field mapping
- Click the Custom Fields tab, set the field name and value, and click OK.
Name: Enter InputTime.
Value: Enter ${timestamp()}. For more time macro variables, see Table 1.Figure 3 Add Field
Table 1 Macro variable definition of time and date Macro Variable
Description
Display Effect
${dateformat(yyyy-MM-dd)}
Returns the current date in yyyy-MM-dd format.
2017-10-16
${dateformat(yyyy/MM/dd)}
Returns the current date in yyyy/MM/dd format.
2017/10/16
${dateformat(yyyy_MM_dd HH:mm:ss)}
Returns the current time in yyyy_MM_dd HH:mm:ss format.
2017_10_16 09:00:00
${dateformat(yyyy-MM-dd HH:mm:ss, -1, DAY)}
Returns the current time in yyyy-MM-dd HH:mm:ss format. The date is one day before the current day.
2017-10-15 09:00:00
${timestamp()}
Returns the timestamp of the current time, that is, the number of milliseconds that have elapsed since 00:00:00 on January 1, 1970.
1508115600000
${timestamp(-10, MINUTE)}
Returns the timestamp generated 10 minutes before the current time point.
1508115000000
${timestamp(dateformat(yyyyMMdd))}
Returns the timestamp of 00:00:00 of the current day.
1508083200000
${timestamp(dateformat(yyyyMMdd,-1,DAY))}
Returns the timestamp of 00:00:00 of the previous day.
1507996800000
${timestamp(dateformat(yyyyMMddHH))}
Returns the timestamp of the current hour.
1508115600000
- After a field is added, its sample value is not displayed on the console. This does not affect the field value transmission. CDM directly writes the field value to the destination end.
- The Custom Fields tab is available only when the source connector is JDBC, HBase, MongoDB, Elasticsearch, or Kafka, or the destination connector is HBase.
- Click Next and set task parameters. Generally, retain the default values of all parameters.
In this step, you can configure the following optional functions:
- Retry Upon Failure: If the job fails to be executed, you can determine whether to automatically retry. Retain the default value Never.
- Group: Select the group to which the job belongs. The default group is DEFAULT. On the Job Management page, jobs can be displayed, started, or exported by group.
- Schedule Execution: If you want the job to be automatically executed at a scheduled time, retain the default value No.
- Concurrent Extractors: Enter the number of extractors to be concurrently executed. Retain the default value 1.
- Write Dirty Data: Specify this parameter if data that fails to be processed or filtered out during job execution needs to be written to OBS for future viewing. Before writing dirty data, create an OBS link. Retain the default value No so that dirty data is not recorded.
- Delete Job After Completion: Retain the default value Do not delete.
- Click Save and Run. On the Table/File Migration page, you can view the job execution progress and result.
- After the job is successfully executed, in the Operation column of the job, click Historical Record to view the job's historical execution records and read/write statistics.
On the Historical Record page, click Log to view the job log.
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.