主库:

[mysqld]

server-id = 2233
port = 13306
basedir = /usr/local/mysql
datadir = /usr/local/mysql/data
socket = /usr/local/mysql/mysql.sock
pid-file=/usr/local/mysql/data/mysql.pid
skip-external-locking
skip_name_resolve = 0
skip-slave-start
transaction_isolation = REPEATABLE-READ
character-set-server = utf8
collation-server = utf8_general_ci
init_connect='SET NAMES utf8'
lower_case_table_names = 1
max_connections = 2000
max_connect_errors = 1000
max_allowed_packet = 24M
open_files_limit=65536
max_prepared_stmt_count=1048576
tmp_table_size = 256M
max_heap_table_size = 256M
query_cache_size = 128M
query_cache_type = 1
query_cache_limit=4M
sort_buffer_size = 2M    #线程缓存
binlog_cache_size = 5242880
back_log = 130
log_timestamps=SYSTEM

log-error=/usr/local/mysql/data/mysqld.log
log-bin=mysql-bin
binlog_rows_query_log_events=1
log_bin_trust_function_creators=1

table_open_cache=2000
innodb_open_files=1024
binlog_format = row

gtid_mode=ON
enforce-gtid-consistency

innodb_flush_log_at_trx_commit = 1
sync_binlog = 0
innodb_buffer_pool_size = 6G
innodb_file_per_table=1
innodb_data_file_path = ibdata1:12M;ibdata2:1G:autoextend
innodb_flush_method = O_DIRECT
innodb_log_file_size=512M
innodb_log_files_in_group=3
innodb_log_buffer_size = 128M
innodb_io_capacity=800
innodb_adaptive_flushing=1
innodb_support_xa = ON       
innodb_read_io_threads=12
innodb_write_io_threads=12
#auto_increment_offset=1
#auto_increment_increment=2

innodb_lock_wait_timeout = 100
innodb_large_prefix = 1
innodb_purge_threads = 4
#innodb_print_all_deadlocks = 1
innodb_buffer_pool_dump_at_shutdown=on
innodb_buffer_pool_load_at_startup=ON
innodb_buffer_pool_dump_now=ON
innodb_buffer_pool_load_now=ON
innodb_thread_concurrency=0
innodb_read_ahead_threshold=42
innodb_change_buffering=all
innodb_change_buffer_max_size=30

secure-file-priv=''

[mysqldump]
quick
max_allowed_packet = 24M

[mysql]
auto-rehash #使用TAB键自动补全

[mysqld_safe]
open-files-limit = 65536

从库:

[mysqld]

server-id = 2233
port = 13306
basedir = /usr/local/mysql
datadir = /usr/local/mysql/data
socket = /usr/local/mysql/mysql.sock
pid-file=/usr/local/mysql/data/mysql.pid
skip-external-locking
skip_name_resolve = 0
skip-slave-start
transaction_isolation = REPEATABLE-READ
character-set-server = utf8
collation-server = utf8_general_ci
init_connect='SET NAMES utf8'
lower_case_table_names = 1
max_connections = 2000
max_connect_errors = 1000
max_allowed_packet = 24M
open_files_limit=65536
max_prepared_stmt_count=1048576
tmp_table_size = 256M
max_heap_table_size = 256M
query_cache_size = 128M
query_cache_type = 1

query_cache_limit=4M
sort_buffer_size = 2M    #线程缓存
binlog_cache_size = 5242880
back_log = 130

log_timestamps=SYSTEM

log-error=/usr/local/mysql/data/mysql-err.log
log-bin=mysql-bin

binlog_rows_query_log_events=1

log_bin_trust_function_creators=1

table_open_cache=2000
innodb_open_files=1024
binlog_format = row

relay_log_recovery = 1
relay-log=mysql-relay
slave_skip_errors = all
slave_net_timeout=60
#replicate-do-db =
replicate-ignore-db = mysql
#并行复制参数
slave-parallel-type=LOGICAL_CLOCK

#slave_preserve_commit_order开启时slave-parallel-typ必须为LOGICAL_CLOCK,保证顺序执行语句。

slave_preserve_commit_order=1
slave-parallel-workers=8
master_info_repository = TABLE
relay_log_info_repository = TABLE
gtid_mode=ON

#将mster的记录到slave的二进制日志文件中
log-slave-updates=ON
enforce-gtid-consistency

innodb_flush_log_at_trx_commit = 1
sync_binlog = 0
innodb_buffer_pool_size = 6G
innodb_file_per_table=1
innodb_data_file_path = ibdata1:12M;ibdata2:1G:autoextend
innodb_flush_method = O_DIRECT
innodb_log_file_size=512M
innodb_log_files_in_group=3
innodb_log_buffer_size = 128M
innodb_io_capacity=800
innodb_adaptive_flushing=1
innodb_support_xa = ON       
innodb_read_io_threads=12
innodb_write_io_threads=12
#auto_increment_offset=1
#auto_increment_increment=2

innodb_lock_wait_timeout = 100
innodb_large_prefix = 1
innodb_purge_threads = 4
#innodb_print_all_deadlocks = 1
innodb_buffer_pool_dump_at_shutdown=on
innodb_buffer_pool_load_at_startup=ON
innodb_buffer_pool_dump_now=ON
innodb_buffer_pool_load_now=ON
innodb_thread_concurrency=0
innodb_read_ahead_threshold=42
innodb_change_buffering=all
innodb_change_buffer_max_size=30

secure-file-priv=''

[mysqldump]
quick
max_allowed_packet = 24M

[mysql]
auto-rehash #使用TAB键自动补全

[mysqld_safe]
open-files-limit = 65536

mysql主库与从库配置(并行复制配置)的更多相关文章

  1. MySQL5.7 并行复制配置

    转自:https://www.cnblogs.com/langdashu/p/6125621.html [MySQL] 号称永久解决了复制延迟问题的并行复制,MySQL5.7 一.缘由: 某天看到主从 ...

  2. 5.7 并行复制配置 基于GTID 搭建中从 基于GTID的备份与恢复,同步中断处理

    5.7 并行复制配置 基于GTID 搭建中从 基于GTID的备份与恢复,同步中断处理 这个文章包含三个部分 1:gtid的多线程复制2:同步中断处理3:GTID的备份与恢复 下面文字相关的东西 大部分 ...

  3. MySQL 5.7 基于GTID主从复制+并行复制+半同步复制

    环境准备 IP HOSTNAME SERVICE SYSTEM 192.168.131.129 mysql-master1 mysql CentOS7.6 192.168.131.130 mysql- ...

  4. MySQL实现主从库,AB复制配置

    AB复制是一种数据复制技术,是myslq数据库提供的一种高可用.高性能的解决方案. AB复制的模式:一主一从 .一主多从.双主.多主多从 复制的工作原理:要想实现ab复制,那么前提是master上必须 ...

  5. 【MySQL】你以为设置了并行复制就降低延迟了?这个你绝对想不到!

    在MySQL官方版本中,为了保证其的高可用性,一般情况我们会采用主从复制的方式来解决.当然,方法很多.而我们今天所要处理的是采用GTID方式并且开了多线程复制后,仍然延迟的情况,糟糕的是,延迟还在不断 ...

  6. MySQL Bug剖析之Slave节点并行复制死锁

    此文已由作者温正湖授权网易云社区发布. 欢迎访问网易云社区,了解更多网易技术产品运营经验. 有天一早,DBA同学就找上来了,说有个DDB集群下的RDS实例Slave节点(从库)死锁了,请求支援.说实话 ...

  7. MySQL的并行复制多线程复制MTS(Multi-Threaded Slaves)

    MySQL的并行复制多线程复制MTS(Multi-Threaded Slaves) http://www.tuicool.com/articles/m2Unmeq 姜承饶 简称MTS:基于binlog ...

  8. MySQL 5.7基于组提交的并行复制

    参考链接: http://mysql.taobao.org/monthly/2016/08/01/ https://www.kancloud.cn/thinkphp/mysql-parallel-ap ...

  9. MySQL并行复制(MTS)原理(完整版)

    目录 MySQL 5.6并行复制架构 MySQL 5.7并行复制原理 Master 组提交(group commit) 支持并行复制的GTID slave LOGICAL_CLOCK(由order c ...

随机推荐

  1. 本地测试出现:Call to undefined function curl_init()

    网上搜索Call to undefined function curl_init(),清一色的以下解决办法: 1.在php.ini中找到extension=php_curl.dll,去掉前面的,php ...

  2. HDU1452:Happy 2004(积性函数)(因子和)

    题意 给出\(x\),求\(2004^x\)的所有因子和 分析 \(2004=2*2*3*167\) 则\(2004^x\)=\(2^{2x}*3^x*167^x\) s[\(2004^x\)]=s[ ...

  3. hdu 1348【凸包模板】

    #include<iostream> #include<iostream> #include<algorithm> #include<cmath> us ...

  4. Pycharm安装及第一次使用导航

    Pycharm:Pycharm是一种Python IDE,带有一整套可以帮助用户在使用Python语言开发时提高其效率的工具,比如调试.语法高亮.Project管理.代码跳转.智能提示.自动完成.单元 ...

  5. mybatis-plus 获取新增id

    <insert id="insert" parameterType="com.xxx.xxxx.pojo.User"> insert into t_ ...

  6. 2017 JUST Programming Contest 3.0 E. The Architect Omar

    E. The Architect Omar time limit per test 1.0 s memory limit per test 256 MB input standard input ou ...

  7. codevs 3044 矩形面积求并 || hdu 1542

    这个线段树的作用其实是维护一组(1维 平面(?) 上的)线段覆盖的区域的总长度,支持加入/删除一条线段. 线段树只能维护整数下标,因此要离散化. 也可以理解为将每一条处理的线段分解为一些小线段,要求每 ...

  8. 从0开始的hbase

    2016马上要结束了,回顾一下这一年对hbase的学习历程. 1,年初hbase的状态 使用场景:主要是用来存储业务线的mysql表,增量同步到hbase,然后每天晚上全量导入hdfs做离线计算. h ...

  9. pandas 选择某几列

    转自:https://blog.csdn.net/aaa_aaa1sdf/article/details/77414387 col_n = ['名称','收盘价','日期'] a = pd.DataF ...

  10. 阿里云虚拟主机的域名添加https的方法

    第一步:购买CDN套餐,阿里云虚拟主机目前是不支持https的,不过可以通过阿里云的CDN服务来跳转一下实现部署https 静态HTTPS请求数根据你的网站访问量来选择 第二步:申请SSL证书服务,有 ...