- What's New
- Service Overview
- Getting Started
-
User Guide
- Function Overview
- Permissions Management
-
Instance Management
- Buying a DDM Instance
- Splitting Read-only and Read-Write Services
- Changing Class of a DDM Node
- Scaling Out a DDM Instance
- Scaling In a DDM Instance
- Changing Billing Mode of a DDM Instance
- Renewing a DDM Instance
- Restarting a DDM Instance
- Unsubscribing from a DDM Instance
- Deleting a DDM Instance
- Modifying Parameters of a DDM Instance
- Splitting Read and Write Requests
- Configuring a Parameter Template
- Connection Management
-
Parameter Template Management
- Creating a Parameter Template
- Editing a Parameter Template
- Comparing Two Parameter Templates
- Viewing Parameter Change History
- Replicating a Parameter Template
- Applying a Parameter Template
- Viewing Application Records of a Parameter Template
- Modifying the Description of a Parameter Template
- Deleting a Parameter Template
- Task Center
- Schema Management
- Shard Configuration
- Data Node Management
- Account Management
- Backups and Restorations
-
Data Migration
- Overview
- Migration Evaluation
- Scenario 1: Migrating Data from Huawei Cloud RDS to DDM
- Scenario 2: Migrating Data from an On-Premises RDS Instance for MySQL to DDM
- Scenario 3: Migrating Data from a Third-Party RDS for MySQL Instance to DDM
- Scenario 4: Migrating Data from a Self-Built MySQL Instance to DDM
- Scenario 5: Migrating Data from Heterogeneous Databases to DDM
- Scenario 6: Exporting Data from a DDM Instance
- Slow Queries
- Monitoring Management
- Auditing
-
SQL Syntax
- Introduction
- DDL
- DML
- Functions
- Use Constraints
- Supported SQL Statements
- Global Sequence
- Database Management Syntax
- Advanced SQL Functions
- Quotas
- Change History
-
API Reference
- Before You Start
- API Overview
- Calling APIs
-
APIs (Recommended)
-
DDM Instances
- Buying a DDM instance
- Querying DDM Instances
- Querying Details of a DDM Instance
- Modifying the Name of a DDM Instance
- Modifying the Security Group of a DDM Instance
- Deleting a DDM Instance
- Restarting a DDM Instance
- Reloading Table Data
- Scaling out a DDM instance
- Scaling in a DDM instance
- Modifying the Read Policy of the Associated DB Instance
- Synchronizing Data Node Information
- Querying Nodes of a DDM Instance
- Querying Details of a DDM Instance Node
- Querying Parameters of a Specified DDM Instance
- Modifying Parameters of a DDM Instance
- Querying DDM Engine Information
- Querying DDM Node Classes Available in an AZ
- Changing the Node Class of a DDM Instance
- Schemas
- DDM Accounts
- Monitoring
-
DDM Instances
- APIs (Unavailable Soon)
- Appendix
- Change History
- SDK Reference
-
Best Practices
- Overview
- Formulating Sharding Rules
- Determining the Number of Shards in a Schema
- Using Broadcast and Unsharded Tables
- Transaction Models
- SQL Standards
- Migrating an Entire RDS Database to DDM
- Migrating an Entire MyCat Database to DDM
- Accessing DDM Using a JDBC Connection Pool
- Logging In to a DDM Instance Using Navicat
- Migrating Data from RDS for MySQL to DDM Using DRS
- Performance White Paper
-
FAQs
- General Questions
-
DDM Usage
- How Does DDM Perform Sharding?
- What Do I Do If I Fail to Connect to a DDM Instance Using the JDBC Driver?
- What Version and Parameters Should I Select?
- Why It Takes So Long Time to Export Data from MySQL Using mysqldump?
- What Do I Do If a Duplicate Primary Key Error Occurs When Data Is Imported into DDM?
- What Should I Do If an Error Message Is Returned When I Specify an Auto-Increment Primary Key During Migration?
- What Do I Do If an Error Is Reported When Parameter Configuration Does Not Time Out?
- Which Should I Delete First, a Schema or its Associated RDS Instances?
- Can I Manually Delete Databases and Accounts Remained in Data Nodes After a Schema Is Deleted?
- SQL Syntax
- RDS-related Questions
- Connection Management
- Resource Freezing, Release, Deletion, and Unsubscription
- Change History
- Videos
Show all
Using Broadcast and Unsharded Tables
Unsharded Table
The DDM console does not allow creation of unsharded tables. To create unsharded tables, you can log in to your DDM instance using a MySQL client or an application.
If a table contains fewer than 10 million data records and has no requests for JOIN operations with other sharded tables for query, set the table type to Unsharded Table to store data in the default shard.
The following is an example SQL statement for creating an unsharded table:
CREATE TABLE single( id int NOT NULL AUTO_INCREMENT COMMENT 'Primary key ID', name varchar(128), PRIMARY KEY(id) );
Broadcast Table
In business databases, some tables have a small amount of data and are not frequently updated, but they are often used for JOIN operations.
- The broadcast table stores the same data in each shard. Data insertion, update, and deletion in broadcast tables are executed in each shard in real time.
- Queries of broadcast tables are only executed in one shard.
- Any tables can JOIN with broadcast tables.
- Before using any broadcast hint, ensure that there are available tables.
Example:
When the order management system of an e-commerce enterprise needs to query and collect statistics on order data in Guangdong province (China), JOIN query is performed for a regional table and an order flow table. The order data volume is large and the order flow table needs to be sharded. In this scenario, you can set the regional table as a broadcast table to avoid cross-database JOIN.
The following is an example SQL statement for creating a broadcast table:
CREATE TABLE broadcast_tbl ( id int NOT NULL AUTO_INCREMENT COMMENT 'Primary key ID', name varchar(128), PRIMARY KEY(id) ) ENGINE=InnoDB DEFAULT CHARSET=utf8 COLLATE=utf8_unicode_ci BROADCAST;
Operate broadcast tables or perform full table scanning operations during off-peak hours, such as executing SQL statements that do not contain sharding conditions. Otherwise, an error may occur, indicating that the number of backend RDS connections is insufficient.
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.