Updated on 2024-08-19 GMT+08:00

Type Conversion Functions

Syntax

CAST(value AS type)

Description

This function is used to forcibly convert types.

Precautions

  • If the input is NULL, NULL is returned.
  • The cast function does not support converting a string to the JSON format.

Example 1: Convert the amount value to an integer.

The following example converts the amount value to an integer.

insert into temp select cast(amount as INT) from source_stream;
Table 1 Examples of type conversion functions

Example

Description

Example

cast(v1 as string)

Converts v1 to a string. The value of v1 can be of the numeric type or of the timestamp, date, or time type.

Table T1:

| content (INT)           |
| -------------           |
| 5                       |

Statement:

SELECT
  cast(content as varchar)
FROM
  T1;

Result:

"5"

cast (v1 as int)

Converts v1 to the int type. The value of v1 can be a number or a character.

Table T1:

| content  (STRING)           |
| -------------               |
| "5"                         |

Statement:

SELECT
  cast(content as int)
FROM
  T1;

Result:

5

cast(v1 as timestamp)

Converts v1 to the timestamp type. The value of v1 can be of the string, date, or time type.

Table T1:

| content  (STRING)          |
| -------------              |
| "2018-01-01 00:00:01"     |

Statement:

SELECT
  cast(content as timestamp)
FROM
  T1;

Result:

1514736001000

cast(v1 as date)

Converts v1 to the date type. The value of v1 can be of the string or timestamp type.

Table T1:

| content  (TIMESTAMP)     |
| -------------            |
| 1514736001000            |

Statement:

SELECT
  cast(content as date)
FROM
  T1;

Result:

"2018-01-01"

Flink jobs do not support the conversion of bigint to timestamp using CAST. You can convert it using to_timestamp.

Example 2:

  1. Create a Flink opensource SQL job by referring to Kafka Source Table and Print Result Table, enter the following job running script, and submit the job.
    Note: When creating a job, set Flink Version to 1.12 in the Running Parameters area on the job editing page, select Save Job Log, and set the OBS bucket for saving job logs to facilitate subsequent job log viewing. Change the values of the parameters in bold in the following script according to the actual situation.
    CREATE TABLE kafkaSource (
      cast_int_to_string int, 
      cast_String_to_int string,
      case_string_to_timestamp string, 
      case_timestamp_to_date timestamp
    ) WITH (
      'connector' = 'kafka',
      'topic' = 'KafkaTopic',
      'properties.bootstrap.servers' = 'KafkaAddress1:KafkaPort,KafkaAddress2:KafkaPort',
      'properties.group.id' = 'GroupId',
      'scan.startup.mode' = 'latest-offset',
      "format" = "json"
    );
    
    CREATE TABLE printSink (
      cast_int_to_string string, 
      cast_String_to_int int, 
      case_string_to_timestamp timestamp, 
      case_timestamp_to_date date
    ) WITH (
      'connector' = 'print'
    );
    
    insert into printSink select 
      cast(cast_int_to_string as string),
      cast(cast_String_to_int as int),
      cast(case_string_to_timestamp as timestamp),  
      cast(case_timestamp_to_date as date)
    from kafkaSource;
  2. Connect to the Kafka cluster and send the following test data to the Kafka topic:
    {"cast_int_to_string":"1", "cast_String_to_int": "1", "case_string_to_timestamp": "2022-04-02 15:00:00", "case_timestamp_to_date": "2022-04-02 15:00:00"}
  3. View output.
    • Method 1:
      1. Log in to the DLI management console and choose Job Management > Flink Streaming Jobs.
      2. Locate the row that contains the target Flink job, and choose More & > FlinkUI in the Operation column.
      3. On the Flink UI, choose Task Managers, click the task name, and select Stdout to view the job run logs.
    • Method 2: If you select Save Job Log for Running Parameters before submitting the job, perform the following operations:
      1. Log in to the DLI management console and choose Job Management > Flink Streaming Jobs.
      2. Click the name of the corresponding Flink job, choose Run Log, click OBS Bucket, and locate the folder of the corresponding log based on the job running date.
      3. Go to the folder of the corresponding date, find the folder whose name contains taskmanager, download the taskmanager.out file, and view the result log.
    The query result is as follows:
    +I(1,1,2022-04-02T15:00,2022-04-02)