DM 分库分表合并场景
本文介绍如何在分库分表合并场景中使用 Data Migration (DM)。使用场景中,三个上游 MySQL 实例的分库和分表数据需要同步至下游 TiDB 集群。
上游实例
假设上游库结构如下:
- 实例 1
Schema | Tables |
---|---|
user | information, log_north, log_bak |
store_01 | sale_01, sale_02 |
store_02 | sale_01, sale_02 |
- 实例 2
Schema | Tables |
---|---|
user | information, log_east, log_bak |
store_01 | sale_01, sale_02 |
store_02 | sale_01, sale_02 |
- 实例 3
Schema | Tables |
---|---|
user | information, log_south, log_bak |
store_01 | sale_01, sale_02 |
store_02 | sale_01, sale_02 |
同步需求
- 合并三个实例中的
user
.information
表至下游 TiDB 中的user
.information
表。 - 合并三个实例中的
user
.log_{north|south|east}
表至下游TiDB中的user
.log_{north|south|east}
表。 - 合并三个实例中的
store_{01|02}
.sale_{01|02}
表至下游TiDB中的store
.sale
表。 - 过滤掉三个实例的
user
.log_{north|south|east}
表的所有删除操作。 - 过滤掉三个实例的
user
.information
表的所有删除操作。 - 过滤掉三个实例的
store_{01|02}
.sale_{01|02}
表的所有删除操作。 - 过滤掉三个实例的
user
.log_bak
表。 - 因为
store_{01|02}
.sale_{01|02}
表带有 bigint 型的自增主键,将其合并至 TiDB 时会引发冲突。你需要有相应的方案来避免冲突。
下游实例
假设同步后下游库结构如下:
Schema | Tables |
---|---|
user | information, log_north, log_east, log_south |
store | sale |
同步方案
要满足同步需求 #1 和 #2,配置 Table routing 规则 如下:
routes:
...
user-route-rule:
schema-pattern: "user"
target-schema: "user"
要满足同步需求 #3,配置 table routing 规则 如下:
routes:
...
store-route-rule:
schema-pattern: "store_*"
target-schema: "store"
sale-route-rule:
schema-pattern: "store_*"
table-pattern: "sale_*"
target-schema: "store"
target-table: "sale"
要满足同步需求 #4 和 #5,配置 Binlog event filter 规则 如下:
filters:
...
user-filter-rule:
schema-pattern: "user"
events: ["truncate table", "drop table", "delete", "drop database"]
action: Ignore
注意:
同步需求 #4、#5 和 #7 的操作意味着过滤掉所有对
user
库的删除操作,所以此处配置了库级别的过滤规则。但是user
库以后加入表的删除操作也都会被过滤。要满足同步需求 #6,配置 Binlog event filter 规则 如下:
filters:
...
sale-filter-rule:
schema-pattern: "store_*"
table-pattern: "sale_*"
events: ["truncate table", "drop table", "delete"]
action: Ignore
store-filter-rule:
schema-pattern: "store_*"
events: ["drop database"]
action: Ignore
要满足同步需求 #7,配置 Black & white table lists 如下:
black-white-list:
log-bak-ignored:
ignore-tables:
- db-name: "user"
tbl-name: "log_bak"
要满足同步需求 #8,首先参考自增主键冲突处理来解决冲突,保证在同步到下游时不会因为分表中有相同的主键值而使同步出现异常;然后需要配置
ignore-checking-items
来跳过自增主键冲突的检查:ignore-checking-items: ["auto_increment_ID"]
同步任务配置
同步任务的完整配置如下。详情请参阅 Data Migration 任务配置文件。
name: "shard_merge"
task-mode: all
meta-schema: "dm_meta"
remove-meta: false
ignore-checking-items: ["auto_increment_ID"]
target-database:
host: "192.168.0.1"
port: 4000
user: "root"
password: ""
mysql-instances:
-
source-id: "instance-1"
route-rules: ["user-route-rule", "store-route-rule", "sale-route-rule"]
filter-rules: ["user-filter-rule", "store-filter-rule", "sale-filter-rule"]
black-white-list: "log-bak-ignored"
mydumper-config-name: "global"
loader-config-name: "global"
syncer-config-name: "global"
-
source-id: "instance-2"
route-rules: ["user-route-rule", "store-route-rule", "sale-route-rule"]
filter-rules: ["user-filter-rule", "store-filter-rule", "sale-filter-rule"]
black-white-list: "log-bak-ignored"
mydumper-config-name: "global"
loader-config-name: "global"
syncer-config-name: "global"
-
source-id: "instance-3"
route-rules: ["user-route-rule", "store-route-rule", "sale-route-rule"]
filter-rules: ["user-filter-rule", "store-filter-rule", "sale-filter-rule"]
black-white-list: "log-bak-ignored"
mydumper-config-name: "global"
loader-config-name: "global"
syncer-config-name: "global"
# 所有实例共享的其他通用配置
routes:
user-route-rule:
schema-pattern: "user"
target-schema: "user"
store-route-rule:
schema-pattern: "store_*"
target-schema: "store"
sale-route-rule:
schema-pattern: "store_*"
table-pattern: "sale_*"
target-schema: "store"
target-table: "sale"
filters:
user-filter-rule:
schema-pattern: "user"
events: ["truncate table", "drop table", "delete", "drop database"]
action: Ignore
sale-filter-rule:
schema-pattern: "store_*"
table-pattern: "sale_*"
events: ["truncate table", "drop table", "delete"]
action: Ignore
store-filter-rule:
schema-pattern: "store_*"
events: ["drop database"]
action: Ignore
black-white-list:
log-bak-ignored:
ignore-tables:
- db-name: "user"
tbl-name: "log_bak"
mydumpers:
global:
threads: 4
chunk-filesize: 64
skip-tz-utc: true
loaders:
global:
pool-size: 16
dir: "./dumped_data"
syncers:
global:
worker-count: 16
batch: 100
max-retry: 100