- Product Bulletin
- Service Overview
- Getting Started
-
User Guide
-
Working with FlexusRDS for MySQL
- Using IAM to Grant Access to FlexusRDS
- Buying a FlexusRDS for MySQL Instance
- Connecting to a FlexusRDS for MySQL Instance
- Suggestions on Using FlexusRDS for MySQL
- Data Migration
-
Instance Management
- Upgrading the Minor Version of a FlexusRDS for MySQL Instance
- Upgrading a FlexusRDS for MySQL Instance to an RDS DB Instance
- Changing the Name of a FlexusRDS for MySQL Instance
- Rebooting FlexusRDS for MySQL Instances
- Resetting the Administrator Password of a FlexusRDS for MySQL Instance
- Enabling Storage Autoscaling for a FlexusRDS for MySQL Instance
- Binding an EIP to a FlexusRDS for MySQL Instance or Unbinding an EIP from a FlexusRDS for MySQL Instance
- Changing the VPC and Subnet of a FlexusRDS for MySQL Instance
- Renewing FlexusRDS for MySQL Instances
- Unsubscribing from a FlexusRDS for MySQL Instance
- Backups and Restorations
- Parameters
- Monitoring
- Logs
- Interconnection with CTS
- FlexusRDS for MySQL Tags
- FlexusRDS for MySQL Quotas
-
Working with FlexusRDS for PostgreSQL
- Permissions Management
- Buying a FlexusRDS for PostgreSQL Instance
- Connecting to a FlexusRDS for PostgreSQL Instance
- Suggestions on Using FlexusRDS for PostgreSQL
- Data Migration
-
Instance Management
- Changing the Name of a FlexusRDS for PostgreSQL Instance
- Rebooting FlexusRDS for PostgreSQL Instances
- Resetting the Administrator Password of a FlexusRDS for PostgreSQL Instance
- Enabling Storage Autoscaling for a FlexusRDS for PostgreSQL Instance
- Binding an EIP to a FlexusRDS for PostgreSQL Instance or Unbinding an EIP from a FlexusRDS for PostgreSQL Instance
- Renewing FlexusRDS for PostgreSQL Instances
- Unsubscribing from a FlexusRDS for PostgreSQL Instance
-
Backups and Restorations
- Creating a Manual Backup for a FlexusRDS for PostgreSQL Instance
- Deleting a Manual Backup of a FlexusRDS for PostgreSQL Instance
- Downloading a Full Backup of a FlexusRDS for PostgreSQL Instance
- Checking and Exporting Backup Information of a FlexusRDS for PostgreSQL Instance
- Restoring a FlexusRDS for PostgreSQL Instance
- Parameters
- Monitoring Management
- Logs
- Interconnection with CTS
- FlexusRDS for PostgreSQL Tags
- FlexusRDS for PostgreSQL Quotas
-
Working with FlexusRDS for MySQL
- API Reference
-
FAQs
- What Is FlexusRDS?
- Can FlexusRDS for MySQL and RDS for MySQL Instances Access Each Other?
- Can I Change the Instance Class of My FlexusRDS Instance?
- Does FlexusRDS Support Cross-AZ HA?
- What Is the Backup Policy of FlexusRDS?
- How Are FlexusRDS Backups Billed?
- How Do I Download FlexusRDS Backups to Restore Data Locally?
- Why Does the Root User of My FlexusRDS for MySQL Instance Not Have Super Permissions?
- Which Storage Engine Does FlexusRDS for MySQL Provide?
- Why Does FlexusRDS for MySQL Not Support the MyISAM Storage Engine?
- Does FlexusRDS for MySQL Support Read Replicas?
- Why Can't I Ping the EIP After It Is Bound to My FlexusRDS for MySQL Instance?
- How Do I Create Accounts and Databases for My FlexusRDS Instance?
- What Major Versions and Minor Versions Does FlexusRDS for MySQL Support?
- Why Are Pay-per-Use Instances Not Provided by FlexusRDS?
- Can I Scale Up the Storage Space of My FlexusRDS for MySQL Instance?
- General Reference
Copied.
Migrating Data to FlexusRDS for PostgreSQL Using the Export and Import Functions of DAS
Scenarios
To back up or migrate data, you can use Data Admin Service (DAS) to export data from the source database first and then import the data to from your local PC or OBS bucket to the destination database.
For more information, see Data Import and Export.
Constraints
- Only one file that is no larger than 1 GB can be imported at a time.
- Binary fields such as BINARY, VARBINARY, TINYBLOB, BLOB, MEDIUMBLOB, and LONGBLOB are not supported.
- If there are more than 100,000 tables in a FlexusRDS for PostgreSQL instance, an error will be reported when you export data using the Export Database function of DAS. In this case, use the Export SQL Result function instead.
Exporting Data
- On the Instances page, locate the DB instance and click Log In in the Operation column.
- On the displayed login page, enter the username and password and click Log In.
- On the top menu bar, choose Import and Export > Export.
- On the displayed page, click Create Task and choose Export Database or Export SQL Result as required. The following takes database export as an example.
Alternatively, click Quick Export and select the target database. On the displayed page, select a storage path and click OK.
Figure 1 Quick export - On the displayed page, set parameters as required in areas Basic Information and Advanced Settings. Then, select the tables to be exported on the right.
Figure 2 Creating an export task
NOTE:
- In a SQL result export task, the executed SQL statements cannot exceed 5 MB.
Figure 3 Export SQL Result
- Databases are classified into user databases and system databases. System databases cannot be exported. If system database data is required, deploy system database services in a created user database, so that you can export the system database data from the user database.
- DAS connects to your standby database to export data. This prevents the primary database from being affected by data export. However, if the standby database has a high replication delay, the exported data may not be the latest.
- In a SQL result export task, the executed SQL statements cannot exceed 5 MB.
- After settings are complete, click OK.
- In the task list, view the task ID, type, status, and progress.
- Click Details in the Operation column to view task details.
Figure 4 Task list
Importing Data
- On the top menu bar, choose Import and Export > Import.
- Import a file from your local PC or an OBS bucket.
Figure 5 Creating an import task
- From your local PC
In the upper left corner, click Create Task. On the displayed page, select an import type, select Upload file for File Source, set the attachment storage, and upload the file. Then, set other parameters as required.
For security purposes, imported files are stored in OBS buckets.
NOTE:
- To keep your data secure, provide your own OBS bucket to store the attachments you upload. In this way, DAS automatically connects to your OBS bucket for in-memory reading.
- If you select Delete the uploaded file upon an import success., the file you uploaded will be automatically deleted from the OBS bucket after being imported to the destination database.
- From an OBS bucket
In the upper left corner, click Create Task. On the displayed page, select an import type, select Choose from OBS for File Source, and select a file from the bucket. Then, set other parameters as required.
NOTE:
The file uploaded from an OBS bucket will not be deleted upon an import success.
- After setting the import parameters, click Create. Confirm the information again before you click OK because original data may be overwritten after data import.
- View the import progress in the task list or check task details.
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.See the reply and handling status in My Cloud VOC.
For any further questions, feel free to contact us through the chatbot.
Chatbot