怎么部署MySQL Group Replication

48次阅读
没有评论

共计 9224 个字符,预计需要花费 24 分钟才能阅读完成。

这篇文章主要介绍“怎么部署 MySQL Group Replication”,在日常操作中,相信很多人在怎么部署 MySQL Group Replication 问题上存在疑惑,丸趣 TV 小编查阅了各式资料,整理出简单好用的操作方法,希望对大家解答”怎么部署 MySQL Group Replication”的疑惑有所帮助!接下来,请跟着丸趣 TV 小编一起来学习吧!

一、环境准备

名称  
  版本  
  备注
操作系统  
 RHEL6.5_X86_64 
 
数据库  
 5.7.18-15 
Percona 二进制版本
复制节点  10.26.7.129
node1
 
 10.26.7.142
node2
 
 10.26.7.166
node3

二、MGR 详细部署步骤
1、MYSQL5.7 安装
本次学习实验采用的是 Percona-Server-5.7.18-15-Linux.x86_64.ssl101 二进制版本,具体安装过程略
2、节点 1my.cnf 参数配置(主写节点)
#replicate
server-id=1001
skip-slave-start = false
read-only = false
expire_logs_days = 2
max_binlog_size = 1G
max_binlog_cache_size = 2G
log-bin = /home/mysql/mysql-bin
log-bin-index = /home/mysql/bin-index
binlog_format = row
log-slave-updates = 1
sync_binlog = 1
log-slow-slave-statements = 1
max-relay-log-size = 1G
relay-log = /home/mysql/mysql-relay
relay-log-index = /home/mysql/relay-index
gtid_mode=ON
enforce_gtid_consistency=ON
master_info_repository=TABLE
relay_log_info_repository=TABLE
binlog_checksum=NONE
#group replication
transaction_write_set_extraction  =XXHASH64   
loose-group_replication_group_name  = 2faa74b8-66bb-11e7-ab7e-b8ca3a6a61ec   #务必以 uuid 形式配置
loose-group_replication_start_on_boot  =off
loose-group_replication_local_address  = 10.26.7.129:24001   #不同节点配置不同节点本身的 IP 地址和端口,区分 MYSQL 自身的 3306 端口
loose-group_replication_group_seeds  = 10.26.7.129:24001,10.26.7.142:24001,10.26.7.166:24001
loose-group_replication_bootstrap_group  =off
3、创建复制账户(主写节点)
set sql_log_bin=0;
create user rpl_user@ %
grant replication slave on *.* to rpl_user@ %   identified by rpl_pass
flush privileges;
set sql_log_bin=1;
change master to master_user= rpl_user ,master_password= rpl_pass for channel group_replication_recovery
4、安装组复制插件并启动组复制(主写节点)
安装插件
install plugin group_replication soname group_replication.so
检查插件是否正确安装
show plugins
+—————————–+———-+——————–+———————-+———+
| Name  | Status  | Type  | Library  | License |
+—————————–+———-+——————–+———————-+———+
| group_replication  | ACTIVE  | GROUP REPLICATION  | group_replication.so | GPL  |
+—————————–+———-+——————–+———————-+———+
启动组复制
set global group_replication_bootstrap_group=ON;
start group_replication;
set global group_replication_bootstrap_group=OFF;
检查组复制成员及状态
select * from performance_schema.replication_group_members;
+—————————+————————————–+————-+————-+————–+
| CHANNEL_NAME  | MEMBER_ID  | MEMBER_HOST | MEMBER_PORT | MEMBER_STATE |
+—————————+————————————–+————-+————-+————–+
| group_replication_applier | 2faa74b8-66bb-11e7-ab7e-b8ca3a6a61ec | node1  |  3306 | ONLINE  |
5、添加组复制成员实例 node2 和 node3
*** 添加节点前,务必做 DNS 解析,如果没有配置 DNS 解析服务器,需要在每个节点配置 hosts 解析   /etc/hosts***
10.26.7.166 node3
10.26.7.142 node2
10.26.7.129 node1

*** 节点 my.cnf 参数文件 server-id 和 loose-group_replication_local_address = node2:24001 需要分别更改 ***

node2
set sql_log_bin=0;
create user rpl_user@ %
grant replication slave on *.* to rpl_user@ %   identified by rpl_pass
flush privileges;

set sql_log_bin=1;
change master to master_user= rpl_user ,master_password= rpl_pass for channel group_replication_recovery

install plugin group_replication soname group_replication.so
show plugins

set global group_replication_allow_local_disjoint_gtids_join=ON;
start group_replication;

node3 同样执行上述命令

然后检查组复制信息
select * from performance_schema.replication_group_members;   
+—————————+————————————–+————-+————-+————–+
| CHANNEL_NAME  | MEMBER_ID  | MEMBER_HOST | MEMBER_PORT | MEMBER_STATE |
+—————————+————————————–+————-+————-+————–+
| group_replication_applier | 2faa74b8-66bb-11e7-ab7e-b8ca3a6a61ec | node1  |  3306 | ONLINE  |
| group_replication_applier | 35e38786-66bb-11e7-bcc3-b8ca3a6a61a4 | node2  |  3306 | ONLINE  |
| group_replication_applier | 3bbedb1e-66bb-11e7-8fc0-b8ca3a6a7c48 | node3  |  3306 | ONLINE  |
+—————————+————————————–+————-+————-+————–+

测试组复制是否正常:
(root:localhost:Sat Jul 15 13:26:33 2017)[(none)] create database dbtest;
Query OK, 1 row affected (0.01 sec)
(root:localhost:Sat Jul 15 13:26:40 2017)[(none)] use dbtest;
Database changed
(root:localhost:Sat Jul 15 13:26:45 2017)[dbtest] create table t1(id int primary key);
Query OK, 0 rows affected (0.01 sec)
(root:localhost:Sat Jul 15 13:26:54 2017)[dbtest] insert into t1 values(1);
Query OK, 1 row affected (0.00 sec)
node2 和 node3 执行查询
(root:localhost:Sat Jul 15 12:57:32 2017)[db01] use dbtest;
Database changed
(root:localhost:Sat Jul 15 13:27:26 2017)[dbtest] select * from t1;
+—-+
| id |
+—-+
|  1 |
+—-+
1 row in set (0.00 sec)
三、错误问题以及汇总:
1、错误案例 01
错误信息:2017-07-15T01:36:06.929941Z 4 [ERROR] Plugin group_replication reported: The group name group-replication-test is not a valid UUID
错误原因:loose-group_replication_group_name 参数没有按照 UUID 格式指定,被认为设置该参数无效
解决方案:更改 loose-group_replication_group_name 参数值为,loose-group_replication_group_name  = 2faa74b8-66bb-11e7-ab7e-b8ca3a6a61ec
2、错误案例 02
错误信息:
2017-07-15T01:29:27.271909Z 0 [Warning] unknown variable loose-group_replication_group_name=group-replication-test
2017-07-15T01:29:27.271926Z 0 [Warning] unknown variable loose-group_replication_start_on_boot=off
2017-07-15T01:29:27.271930Z 0 [Warning] unknown variable loose-group_replication_local_address=10.26.7.129:3306
2017-07-15T01:29:27.271935Z 0 [Warning] unknown variable loose-group_replication_group_seeds=10.26.7.129:3306,10.26.7.142:3306,10.26.7.166:3306
2017-07-15T01:29:27.271939Z 0 [Warning] unknown variable loose-group_replication_bootstrap_group=off
错误原因:因为先设置了这些参数,而没有装 group_replication 插件,导致数据库实例无法识别这些参数
解决方案:安装 group replication 插件,install plugin group_replication soname group_replication.so(uninstall plugin group_replication 卸载,show plugins 查看)
3、错误案例 03
错误信息:
2017-07-15T01:54:54.447829Z 0 [Note] Plugin group_replication reported: Unable to bind to 0.0.0.0:3306 (socket=60, errno=98)!
2017-07-15T01:54:54.447948Z 0 [ERROR] Plugin group_replication reported: Unable to announce tcp port 3306. Port already in use?
2017-07-15T01:54:54.448101Z 0 [ERROR] Plugin group_replication reported: [GCS] Error joining the group while waiting for the network layer to become ready.
错误原因:配置的组复制监听端口和 MYSQL 实例端口冲突
解决方案:调整下面参数
loose-group_replication_local_address 
= 10.26.7.129:24001  
#不同节点配置不同节点本身的 IP 地址和端口,区分 MYSQL 自身的 3306 端口
loose-group_replication_group_seeds  = 10.26.7.129:24001,10.26.7.142:24001,10.26.7.166:24001
4、错误案例 04
错误信息:
2017-07-15T04:20:01.249529Z 21 [ERROR] Slave I/O for channel group_replication_recovery : error connecting to master rpl_user@node2:3306 – retry-time: 60  retries: 1, Error_code: 2005
错误原因:没有配置 DNS 解析或者 hosts 解析,节点无法连接其他数据库
解决方案:配置 hosts 解析,每个节点 /etc/hosts 添加如下内容
10.26.7.166 node3
10.26.7.142 node2
10.26.7.129 node1
5、错误案例 05
错误信息
2017-07-15T03:42:45.395407Z 288 [ERROR] Slave SQL for channel group_replication_recovery : Error Can t create database db01 database exists on query. Default database: db01 . Query: create database db01 , Error_code: 1007
2017-07-15T03:42:45.395472Z 288 [Warning] Slave: Can t create database db01 database exists Error_code: 1007
2017-07-15T03:42:45.395503Z 288 [ERROR] Error running query, slave SQL thread aborted. Fix the problem, and restart the slave SQL thread with SLAVE START . We stopped at log mysql-bin.000005 position 434
错误原因:这个错误是由于节点由于各种原因退出 mgr 组,后面又加入了 mgr 组,但之前存在的数据依旧存在
解决方案:删除要加入组复制节点存在的数据库即可,但其他节点不是主写节点,需要先调整参数 set global super_read_only=0; 然后执行 drop database db01; 再重新加入组
set global group_replication_allow_local_disjoint_gtids_join=ON;
start group_replication;
6、错误案例 06
错误信息:
2017-07-15T03:44:09.982428Z 18 [ERROR] Slave SQL for channel group_replication_recovery : Error Can t create database db01 database exists on query. Default database: db01 . Query: create database db01 , Error_code: 1007
2017-07-15T03:44:09.982493Z 18 [Warning] Slave: Can t create database db01 database exists Error_code: 1007
2017-07-15T03:44:09.982522Z 18 [ERROR] Error running query, slave SQL thread aborted. Fix the problem, and restart the slave SQL thread with SLAVE START . We stopped at log mysql-bin.000005 position 434
错误原因:同上错误案例 05
解决方案:同上错误案例 05
7、错误案例 07
错误信息:
2017-07-15T03:49:10.370846Z 0 [ERROR] Plugin group_replication reported: This member has more executed transactions than those present in the group. Local transactions: 2faa74b8-66bb-11e7-ab7e-b8ca3a6a61ec:1-4,35e38786-66bb-11e7-bcc3-b8ca3a6a61a4:1 Group transactions: 2faa74b8-66bb-11e7-ab7e-b8ca3a6a61ec:1-16
错误原因:同上错误案例 05,在从库执行了多余的事务
解决方案:同上错误案例 05,直接重新加入节点即可
set global group_replication_allow_local_disjoint_gtids_join=ON;
start group_replication;
8、错误案例 08
错误信息
ERROR 3098 (HY000): The table does not comply with the requirements by an external plugin.
错误原因:由于主节点创建了表 t1,但没有指定主键(此时表结构可以复制到各节点,一旦插入数据 DML 操作即会报错)
解决方案:为表增加主键,然后做 DML 操作(MGR 需要各表都有主键)
alter table t1 add primary key(id);
insert into t1 values(1),(2);   
9、错误案例 09
错误信息:
mysqldump -R -E –triggers –single-transaction  –master-data=2 -B db01 db01.sql
Warning: A partial dump from a server that has GTIDs will by default include the GTIDs of all transactions, even those that changed suppressed parts of the database. If you don t want to restore GTIDs, pass –set-gtid-purged=OFF. To make a complete dump, pass –all-databases –triggers –routines –events.
mysqldump: Couldn t execute SAVEPOINT sp : The MySQL server is running with the –transaction-write-set-extraction!=OFF option so it cannot execute this statement (1290)
错误原因:mgr 不支持 mysqldump 的事务一致性备份,因为其不支持 savepoint
解决方案:通过 xtrabackup 或者不加 –single-transaction 备份
10、错误案例 10
错误信息:
create table t2 as select * from t1;
ERROR 1786 (HY000): Statement violates GTID consistency: CREATE TABLE … SELECT.
错误原因:配置 MGR,开启了 GTID,所有 GTID 不支持的操作,MGR 也不支持
解决方案:使用 create table t2 like t1; insert into t2 select * from t; 分开两个事务执行

到此,关于“怎么部署 MySQL Group Replication”的学习就结束了,希望能够解决大家的疑惑。理论与实践的搭配能更好的帮助大家学习,快去试试吧!若想继续学习更多相关知识,请继续关注丸趣 TV 网站,丸趣 TV 小编会继续努力为大家带来更多实用的文章!

正文完
 
丸趣
版权声明:本站原创文章,由 丸趣 2023-07-19发表,共计9224字。
转载说明:除特殊说明外本站除技术相关以外文章皆由网络搜集发布,转载请注明出处。
评论(没有评论)