利用maxwell组件监听mysql之binlog日志进行实时同步数据

本文涉及的产品
云数据库 RDS MySQL,集群版 2核4GB 100GB
推荐场景:
搭建个人博客
RDS MySQL Serverless 基础系列,0.5-2RCU 50GB
日志服务 SLS,月写入数据量 50GB 1个月
简介: 利用maxwell组件监听mysql之binlog日志进行实时同步数据

一 maxwell组件介绍



Maxwell是一个守护进程,它能监听并读取MySQL的binlog,然后解析输出为json,支持将数据输出到Kafka、Kinesis或其他流媒体平台,支持表和库过滤。


注意:对于增删改都有输出,但对于truncate操作,没有输出。


源码地址:https://github.com/zendesk/maxwell


下载地址:https://github.com/zendesk/maxwell/releases/download/v1.21.1/maxwell-1.21.1.tar.gz


示意如下:


mysql> insert into `test`.`maxwell` set id = 1, daemon = 'Stanislaw Lem';
  maxwell: {
    "database": "test",
    "table": "maxwell",
    "type": "insert",
    "ts": 1449786310,
    "xid": 940752,
    "commit": true,
    "data": { "id":1, "daemon": "Stanislaw Lem" }
  }
  mysql> update test.maxwell set daemon = 'firebus!  firebus!' where id = 1;
  maxwell: {
    "database": "test",
    "table": "maxwell",
    "type": "update",
    "ts": 1449786341,
    "xid": 940786,
    "commit": true,
    "data": {"id":1, "daemon": "Firebus!  Firebus!"},
    "old":  {"daemon": "Stanislaw Lem"}
  }


二 设备与组件版本梳理:


1.  linux内核版本(CentOS Linux 7):(命令:uname -a)


Linux slave1 3.10.0-693.el7.x86_64 #1 SMP Tue Aug 22 21:09:27 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux


2.  mysql版本:(SQL命令:select version(); 或 status)


Server version: 5.6.43-log MySQL Community Server (GPL)


3.  maxwell版本:maxwell-1.21.1


4.  kafka版本:kafka_2.11-2.2.0


5.  zookeeper版本:zookeeper-3.4.5-cdh5.7.0


三 设备介绍


二台虚拟机,分别为salve1(192.168.175.21)和slave2(192.168.175.22),slave1上安装mysql和kafka,slave2上启动maxwell守护进程和zookeeper。


四 简要过程梳理



网络异常,图片无法展示
|


主流程示意图


如上图:这次主要介绍从binlog > maxwell > kafka的过程,而kafka后面的过程,就可以有很多种了,比如:


(1)binlog > maxwell > kafka > spark streaming > hdfs、kudu;

(2)binlog > maxwell > kafka > flume > hdfs;

(3)binlog > maxwell > kafka > es > kibana;


第一种spark streaming+hdfs、kudu,是目前我所在公司中使用的场景。简要流程梳理如下:


1.  在slave1上安装mysql;
2.  在slave2上启动maxwell,测试是否可以正常读取binlog;
3.  maxwell初步测试ok;
4.  在slave2上安装并启动zk;
5.  在slave1上安装并启动kafka server;
6.  通过kafka producer和consumer测试启动是否成功;
7.  启动maxwell将解析后的json数据发送到kafka;
8.  启动kafka consumer测试数据是否成功发送。


五 详细过程


1 在slave1(192.168.175.21)上安装mysql


详细过程,可参考笔记: https://www.jianshu.com/p/09936d9c7bf2


(1)在创建root账号并设置远程访问之后,接着创建maxwell账号并设置远程访问和权限:


mysql> CREATE USER 'maxwell'@'%' IDENTIFIED BY 'XXXXXX';
mysql> GRANT ALL ON maxwell.* TO 'maxwell'@'%';
mysql> GRANT SELECT, REPLICATION CLIENT, REPLICATION SLAVE ON *.* TO 'maxwell'@'%';


权限:Maxwell需要权限将状态存储在schema_database选项(默认Maxwell)指定的数据库中。


(2)针对maxwell配置mysql: 确保配置了server_id,并打开了基于行的复制。

可参考maxwell快速入门文档:


https://github.com/zendesk/maxwell/blob/master/docs/docs/quickstart.md


$ vi my.cnf
[mysqld]
server_id=1
log-bin=master
binlog_format=row


注意:  binlog_format是一个基于会话的属性需要关闭所有活动连接才能完全转换为基于行的复制。


mysql配置文件my.cnf查找方式:


$ whereis my
my:/etc/my.cnf


配置完成后,要重启mysql服务,方可生效


2 在slave2上启动maxwell,测试是否可以正常读取binlog


(1)在slave2上测试是否可以进行远程访问数据


mysql -h 192.168.175.21 -P 3306 - u root -proot   #root登录成功后,窗口无需关闭,后面还要接着测试用
mysql -h 192.168.175.21 -P 3306 -u maxwell -p111111


(2)下载并解压maxwell


#解压到指定的文件夹
tar xzvf maxwell-1.21.1.tar.gz -C /usr/loca/hadoop/app


(3)命令行启动maxwell,将解析后的日志输出到控制台进行测试


[root@slave2 maxwell-1.21.1]# pwd
/usr/local/hadoop/app/maxwell-1.21.1
[root@slave2 maxwell-1.21.1]# bin/maxwell --user='maxwell' --password='111111' --host='192.168.175.21' --producer=stdout


启动成功后,会展示如下日志内容:


[root@slave2 maxwell-1.21.1]# bin/maxwell --user='maxwell' --password='111111' --host='192.168.175.21' --producer=stdout
Using kafka version: 1.0.0
15:49:04,967 WARN  MaxwellMetrics - Metrics will not be exposed: metricsReportingType not configured.
15:49:08,334 INFO  Maxwell - Maxwell v1.21.1 is booting (StdoutProducer), starting at Position[BinlogPosition[master.000002:984013], lastHeartbeat=1555141484690]
15:49:10,071 INFO  MysqlSavedSchema - Restoring schema id 2 (last modified at Position[BinlogPosition[master.000001:40754], lastHeartbeat=1555112822550])
15:49:12,759 INFO  MysqlSavedSchema - Restoring schema id 1 (last modified at Position[BinlogPosition[master.000001:3816], lastHeartbeat=0])
15:49:13,122 INFO  MysqlSavedSchema - beginning to play deltas...
15:49:13,138 INFO  MysqlSavedSchema - played 1 deltas in 12ms
15:49:13,426 INFO  BinlogConnectorReplicator - Setting initial binlog pos to: master.000002:984013
15:49:13,997 INFO  BinaryLogClient - Connected to 192.168.175.21:3306 at master.000002/984013 (sid:6379, cid:59)
15:49:14,003 INFO  BinlogConnectorLifecycleListener - Binlog connected.


在slave2上root连接mysql的窗口中,执行insert,delete,update操作:


MySQL [mysql]> insert into tb_dept (id,name,description) values(16,'xiaoman','manger');
Query OK, 1 row affected (0.02 sec)
MySQL [mysql]> delete from tb_dept where id = 16;
Query OK, 1 row affected (0.01 sec)
MySQL [mysql]> update tb_dept set name='xiaofei' where id=14;
Query OK, 1 row affected (0.02 sec)
Rows matched: 1  Changed: 1  Warnings: 0


在maxwell的stdout窗口中会产生如下日志:


{"database":"mysql","table":"tb_dept","type":"insert","ts":1555142065,"xid":6349,"commit":true,"data":{"Id":16,"Name":"xiaoman","description":"manger"}}
{"database":"mysql","table":"tb_dept","type":"delete","ts":1555142096,"xid":6361,"commit":true,"data":{"Id":16,"Name":"xiaoman","description":"manger"}}
{"database":"mysql","table":"tb_dept","type":"update","ts":1555142157,"xid":6383,"commit":true,"data":{"Id":14,"Name":"xiaofei","description":"sales"},"old":{"Name":"xiaoming1"}}


3 在slave2上安装并启动zk


具体安装方式,可参考笔记:https://www.jianshu.com/p/10d5a20ab9b7


注意启动完成后,要检查zk是否安装成功


#查看状态
zkServer.sh status


4 在slave1上安装并启动kafka server:


(1)具体启动安装方式,可参考笔记:https://www.jianshu.com/p/3d017bdbfb3c

修改kafka配置文件 $KAFKA_HOME/config/server.properties:


broker.id=1
listeners=PLAINTEXT://slave1:9092
log.dirs=/usr/local/app/tmp/kafka-logs
zookeeper.connect=slave2:2181


(2)通过kafka producer和consumer测试启动是否成功


5 在slave1的kafka上创建名为maxwell的topic


(1)创建topic


kafka-topics.sh --create --zookeeper slave2:2181 --replication-factor 1 --partitions 1 --topic maxwell


(2)检查topic是否创建成功


#查看topic列表:
kafka-topics.sh --list --zookeeper slave2:2181
#查看topic具体描述:
kafka-topics.sh --describe --zookeeper slave2:2181  --topic maxwell


6 在slave1上启动消费topic名称为maxwell的kafka consumer:


kafka-console-consumer.sh --zookeeper slave2:2181 --topic maxwell --from-beginning


7 在slave2上启动maxwell进程,将解析后的json数据输出到kafka:


bin/maxwell --user='maxwell' --password='111111' --host='192.168.175.21' --producer=kafka --kafka.bootstrap.servers=192.168.175.21:9092 --kafka_topic=maxwell


注意:启动之前,需要将输出到stdout上的maxwell进程停掉,否则会报错。


同样,启动成功后,会输出Binlog连接成功的日志信息。如下:


16:22:19,127 INFO  AppInfoParser - Kafka version : 1.0.0
16:22:19,129 INFO  AppInfoParser - Kafka commitId : aaa7af6d4a11b29d
16:22:19,391 INFO  Maxwell - Maxwell v1.21.1 is booting (MaxwellKafkaProducer), starting at Position[BinlogPosition[master.000002:1084309], lastHeartbeat=1555143612778]
16:22:20,916 INFO  MysqlSavedSchema - Restoring schema id 2 (last modified at Position[BinlogPosition[master.000001:40754], lastHeartbeat=1555112822550])
16:22:22,861 INFO  MysqlSavedSchema - Restoring schema id 1 (last modified at Position[BinlogPosition[master.000001:3816], lastHeartbeat=0])
16:22:23,178 INFO  MysqlSavedSchema - beginning to play deltas...
16:22:23,192 INFO  MysqlSavedSchema - played 1 deltas in 7ms
16:22:23,482 INFO  BinlogConnectorReplicator - Setting initial binlog pos to: master.000002:1084309
16:22:23,887 INFO  BinaryLogClient - Connected to 192.168.175.21:3306 at master.000002/1084309 (sid:6379, cid:64)
16:22:23,894 INFO  BinlogConnectorLifecycleListener - Binlog connected.


8 验证


(1)在slave1中的root账号登录的mysql窗口中,执行一条更新操作:


MySQL [mysql]> update tb_dept set name='xiaofei123' where id=14;
Query OK, 1 row affected (0.02 sec)
Rows matched: 1  Changed: 1  Warnings: 0


(2)随机在消费topic=maxwell的kafka consumer中输出日志如下:


{"database":"mysql","table":"tb_dept","type":"update","ts":1555143780,"xid":6888,"commit":true,"data":{"Id":14,"Name":"xiaofei123","description":"sales"},"old":{"Name":"xiaofei"}}


到此,流程梳理完毕!


六 遗留问题:


1 mysql数据库


设置远程访问之后,在本地一直访问不了,尝试修改密码同样访问不了。猜测是由于设置远程访问时的%百分号影响的。故而这里访问数据库用的都是远程访问的,其实正常生产环境,我们也都是远程进行访问的。


ERROR 1045 (28000): Access denied for user 'root'@'localhost' (using password: YES)


2 kafka:


在将maxwell进程和kafka server运行在同一台虚拟机上时,启动kafka consumer时,经常报ConsumerRebalanceFailedException的异常,如下:没有找到具体原因,最终通过kafka server和maxwell不运行在一起将问题解决。其实生产环境,一般kafka都是有专用的集群,也都不会和maxwell运行在一起。


[2019-04-13 13:12:03,599] ERROR Unknown error when running consumer:  (kafka.tools.ConsoleConsumer$)
kafka.common.ConsumerRebalanceFailedException: console-consumer-72779_slave2-1555132312263-9d33f2d8 can't rebalance after 4 retries
    at kafka.consumer.ZookeeperConsumerConnector$ZKRebalancerListener.syncedRebalance(ZookeeperConsumerConnector.scala:660)
    at kafka.consumer.ZookeeperConsumerConnector.kafka$consumer$ZookeeperConsumerConnector$$reinitializeConsumer(ZookeeperConsumerConnector.scala:967)
    at kafka.consumer.ZookeeperConsumerConnector$WildcardStreamsHandler.<init>(ZookeeperConsumerConnector.scala:1001)
    at kafka.consumer.ZookeeperConsumerConnector.createMessageStreamsByFilter(ZookeeperConsumerConnector.scala:163)
    at kafka.consumer.OldConsumer.<init>(BaseConsumer.scala:75)
    at kafka.tools.ConsoleConsumer$.run(ConsoleConsumer.scala:63)
    at kafka.tools.ConsoleConsumer$.main(ConsoleConsumer.scala:47)
    at kafka.tools.ConsoleConsumer.main(ConsoleConsumer.scala)


相关文章
|
17天前
|
存储 关系型数据库 MySQL
|
17天前
|
存储 SQL 关系型数据库
|
12天前
|
SQL 安全 关系型数据库
MySQL的binlog日志的简介与查看
MySQL的binlog日志的简介与查看
26 4
|
13天前
|
关系型数据库 MySQL 数据库
MySQL 启动日志报错: File /mysql-bin.index not found (Errcode: 13 - Permission denied)
MySQL 启动日志报错: File /mysql-bin.index not found (Errcode: 13 - Permission denied)
31 2
|
14天前
|
存储 关系型数据库 MySQL
|
16天前
|
监控 数据库
neo4j数据插入操作有日志吗
【6月更文挑战第29天】neo4j数据插入操作有日志吗
22 1
|
18天前
|
SQL 运维 关系型数据库
|
11天前
|
存储 JavaScript 容器
TS,添加注释,//,ctrl + /,shift + alt + a,输出语句,console.log(“Hello Ts‘),变量和数据类型导读,变量就是用来存储数据的容器,变量的使用,TS
TS,添加注释,//,ctrl + /,shift + alt + a,输出语句,console.log(“Hello Ts‘),变量和数据类型导读,变量就是用来存储数据的容器,变量的使用,TS
|
2月前
|
存储 安全 关系型数据库
Mysql 的binlog日志的优缺点
MySQL的binlog(二进制日志)是一个记录数据库更改的日志文件,它包含了所有对数据库执行的更改操作,如INSERT、UPDATE和DELETE等。binlog的主要目的是复制和恢复。以下是binlog日志的优缺点: ### 优点: 1. **数据恢复**:当数据库出现意外故障或数据丢失时,可以利用binlog进行点恢复(point-in-time recovery),将数据恢复到某一特定时间点。 2. **主从复制**:binlog是实现MySQL主从复制功能的核心组件。主服务器将binlog中的事件发送到从服务器,从服务器再重放这些事件,从而实现数据的同步。 3. **审计**:b
101 1
|
2月前
|
SQL 关系型数据库 MySQL
mysql的binlog恢复数据
mysql的binlog恢复数据
62 0