Redis主从及哨兵
Redis主从用法
像MySQL一样,redis是支持主从同步的,而且也支持一主多从以及多级从结构。
主从结构,一是为了纯粹的冗余备份,二是为了提升读性能,比如很消耗性能的SORT就可以由从服务器来承担。redis的主从同步是异步进行的,这意味着主从同步不会影响主逻辑,也不会降低redis的处理性能。
主从架构中,可以考虑关闭主服务器的数据持久化功能,只让从服务器进行持久化,这样可以提高主服务器的处理性能。
在主从架构中,从服务器通常被设置为只读模式,这样可以避免从服务器的数据被误修改。但是从服务器仍然可以接受CONFIG等指令,所以还是不应该将从服务器直接暴露到不安全的网络环境中。如果必须如此,那可以考虑给重要指令进行重命名,来避免命令被外人误执行。
Redis同步原理
从服务器会向主服务器发出SYNC指令,当主服务器接到此命令后,就会调用BGSAVE指令来创建一个子进程专门进行数据持久化工作,也就是将主服务器的数据写入RDB文件中。在数据持久化期间,主服务器将执行的写指令都缓存在内存中。
在BGSAVE指令执行完成后,主服务器会将持久化好的RDB文件发送给从服务器,从服务器接到此文件后会将其存储到磁盘上,然后再将其读取到内存中。这个动作完成后,主服务器会将这段时间缓存的写指令再以redis协议的格式发送给从服务器。
另外,要说的一点是,即使有多个从服务器同时发来SYNC指令,主服务器也只会执行一次BGSAVE,然后把持久化好的RDB文件发给多个下游。在redis2.8版本之前,如果从服务器与主服务器因某些原因断开连接的话,都会进行一次主从之间的全量的数据同步;而在2.8版本之后,redis支持了效率更高的增量同步策略,这大大降低了连接断开的恢复成本。
主服务器会在内存中维护一个缓冲区,缓冲区中存储着将要发给从服务器的内容。从服务器在与主服务器出现网络瞬断之后,从服务器会尝试再次与主服务器连接,一旦连接成功,从服务器就会把“希望同步的主服务器ID”和“希望请求的数据的偏移位置(replication offset)”发送出去。主服务器接收到这样的同步请求后,首先会验证主服务器ID是否和自己的ID匹配,其次会检查“请求的偏移位置”是否存在于自己的缓冲区中,如果两者都满足的话,主服务器就会向从服务器发送增量内容。
增量同步功能,需要服务器端支持全新的PSYNC指令。这个指令,只有在redis-2.8之后才具有。
建立复制:
slaveof {masterHost} {masterPort}
从节点查看复制状态信息:
info replication
断开复制:
slaveof no one
安全性:
主节点使用requirepass参数进行密码验证,客户端使用auth命令校验,从节点的masterauth参数与主节点密码保持一致
只读:
默认情况下,从节点使用slave-read-only=yes配置为只读模式
复制支持树状结构,从节点可以复制另一个从节点,实现一层层向下的复制流。
复制原理:
1)执行slaveof后从节点只保存主节点的地址信息便直接返回,保存主节点信息
2)从节点内部通过每秒运行的定时任务维护复制相关逻辑,与主节点建立socket连接
3)从节点发送ping命令请求进行首次通信
4)权限验证
5)同步数据集,2.8以后采用新复制命令psync进行数据同步
6)命令持续复制
psync命令需要以下组件支持:
主从节点各自复制偏移量:master_repl_offset,slave_repl_offset
主节点复制积压缓冲区
主节点运行id
全量复制时间开销:
主节点bgsave时间
RDB文件网络传输时间
从节点清空数据时间
从节点加载RDB时间
可能的AOF重写时间
心跳:
主从节点在建立复制后,它们之间维护着长连接并彼此发送心跳命令
异步复制
复制可能遇到的问题:
复制数据延迟
读到过期数据
从节点故障
Redis配置主从备份及主备切换方案配置
192.168.1.101 master 6379
192.168.1.102 slave1 6379
192.168.1.103 slave2 6379
修改master文件:/etc/redis/6379.conf
pidfile /var/run/redis_6379.pid
logfile /data/logs/redis.master.log
protected-mode yes
masterauth "mypass"
requirepass "mypass"
daemonize yes
tcp-backlog 511
timeout 0
tcp-keepalive 60
loglevel notice
databases 16
dir /data
stop-writes-on-bgsave-error no
repl-timeout 60
repl-ping-slave-period 10
repl-disable-tcp-nodelay no
repl-backlog-size 10M
repl-backlog-ttl 7200
slave-serve-stale-data yes
slave-read-only yes
slave-priority 100
lua-time-limit 5000
slowlog-log-slower-than 10000
slowlog-max-len 128
hash-max-ziplist-entries 512
hash-max-ziplist-value 64
list-max-ziplist-entries 512
list-max-ziplist-value 64
set-max-intset-entries 512
zset-max-ziplist-entries 128
zset-max-ziplist-value 64
activerehashing yes
client-output-buffer-limit normal 0 0 0
client-output-buffer-limit slave 512mb 128mb 60
client-output-buffer-limit pubsub 32mb 8mb 60
hz 10
port 6379
maxmemory 512mb
maxmemory-policy volatile-lru
appendonly yes
appendfsync everysec
appendfilename appendonly-6379.aof
dbfilename dump-6379.rdb
aof-rewrite-incremental-fsync yes
no-appendfsync-on-rewrite yes
auto-aof-rewrite-min-size 64m
auto-aof-rewrite-percentage 89
rdbcompression yes
rdbchecksum yes
repl-diskless-sync no
repl-diskless-sync-delay 5
maxclients 10000
hll-sparse-max-bytes 3000
min-slaves-to-write 0
min-slaves-max-lag 10
aof-load-truncated yes
notify-keyspace-events ""
修改slave1文件:/etc/redis/6379.conf
pidfile /var/run/redis_6379.pid
logfile /data/logs/redis.slave1.log
protected-mode yes
masterauth "mypass"
requirepass "mypass"
daemonize yes
tcp-backlog 511
timeout 0
tcp-keepalive 60
loglevel notice
databases 16
dir /data
stop-writes-on-bgsave-error no
repl-timeout 60
repl-ping-slave-period 10
repl-disable-tcp-nodelay no
repl-backlog-size 10000000
repl-backlog-ttl 7200
slave-serve-stale-data yes
slave-read-only yes
slave-priority 100
lua-time-limit 5000
slowlog-log-slower-than 10000
slowlog-max-len 128
hash-max-ziplist-entries 512
hash-max-ziplist-value 64
list-max-ziplist-entries 512
list-max-ziplist-value 64
set-max-intset-entries 512
zset-max-ziplist-entries 128
zset-max-ziplist-value 64
activerehashing yes
client-output-buffer-limit normal 0 0 0
client-output-buffer-limit slave 512mb 128mb 60
client-output-buffer-limit pubsub 32mb 8mb 60
hz 10
port 6379
maxmemory 512mb
maxmemory-policy volatile-lru
appendonly yes
appendfsync everysec
appendfilename "appendonly-6379.aof"
dbfilename "dump-6379.rdb"
aof-rewrite-incremental-fsync yes
no-appendfsync-on-rewrite yes
auto-aof-rewrite-min-size 62500kb
auto-aof-rewrite-percentage 81
rdbcompression yes
rdbchecksum yes
repl-diskless-sync no
repl-diskless-sync-delay 5
maxclients 4064
hll-sparse-max-bytes 3000
min-slaves-to-write 0
min-slaves-max-lag 10
aof-load-truncated yes
notify-keyspace-events ""
slaveof 192.168.1.101 6379
修改slave2文件:/etc/redis/6379.conf
pidfile /var/run/redis_6379.pid
logfile /data/logs/redis.slave2.log
protected-mode yes
masterauth "mypass"
requirepass "mypass"
daemonize yes
tcp-backlog 511
timeout 0
tcp-keepalive 60
loglevel notice
databases 16
dir /data
stop-writes-on-bgsave-error no
repl-timeout 60
repl-ping-slave-period 10
repl-disable-tcp-nodelay no
repl-backlog-size 10000000
repl-backlog-ttl 7200
slave-serve-stale-data yes
slave-read-only yes
slave-priority 100
lua-time-limit 5000
slowlog-log-slower-than 10000
slowlog-max-len 128
hash-max-ziplist-entries 512
hash-max-ziplist-value 64
list-max-ziplist-entries 512
list-max-ziplist-value 64
set-max-intset-entries 512
zset-max-ziplist-entries 128
zset-max-ziplist-value 64
activerehashing yes
client-output-buffer-limit normal 0 0 0
client-output-buffer-limit slave 512mb 128mb 60
client-output-buffer-limit pubsub 32mb 8mb 60
hz 10
port 6379
maxmemory 512mb
maxmemory-policy volatile-lru
appendonly yes
appendfsync everysec
appendfilename "appendonly-6379.aof"
dbfilename "dump-6379.rdb"
aof-rewrite-incremental-fsync yes
no-appendfsync-on-rewrite yes
auto-aof-rewrite-min-size 62500kb
auto-aof-rewrite-percentage 81
rdbcompression yes
rdbchecksum yes
repl-diskless-sync no
repl-diskless-sync-delay 5
maxclients 4064
hll-sparse-max-bytes 3000
min-slaves-to-write 0
min-slaves-max-lag 10
aof-load-truncated yes
notify-keyspace-events ""
slaveof 192.168.1.101 6379
启动master
[root@mydb1 bin]# pwd
/usr/local/redis/bin
[root@mydb1 bin]# ./redis-server /etc/redis/6379.conf &
启动slave1
[root@mydb2 bin]# ./redis-server /etc/redis/6379.conf &
启动slave2
[root@mydb3 bin]# ./redis-server /etc/redis/6379.conf &
查看master信息
[root@mydb1 bin]# ./redis-cli -h 192.168.1.101 -p 6379 -a "mypass"
192.168.1.101:6379> info replication
# Replication
role:master
connected_slaves:2
slave0:ip=192.168.1.102,port=6379,state=online,offset=2507,lag=0
slave1:ip=192.168.1.103,port=6379,state=online,offset=2507,lag=0
master_repl_offset:2507
repl_backlog_active:1
repl_backlog_size:1048576
repl_backlog_first_byte_offset:2
repl_backlog_histlen:2506
查看slave1信息
[root@mydb2 bin]# ./redis-cli -h 192.168.1.102 -p 6379 -a "mypass"
192.168.1.102:6379> info replication
# Replication
role:slave
master_host:192.168.1.101
master_port:6379
master_link_status:up
master_last_io_seconds_ago:4
master_sync_in_progress:0
slave_repl_offset:2619
slave_priority:100
slave_read_only:1
connected_slaves:0
master_repl_offset:0
repl_backlog_active:0
repl_backlog_size:1048576
repl_backlog_first_byte_offset:0
repl_backlog_histlen:0
查看slave2信息
[root@mydb3 bin]# ./redis-cli -h 192.168.1.103 -p 6379 -a "mypass"
192.168.1.103:6379> info replication
# Replication
role:slave
master_host:192.168.1.101
master_port:6379
master_link_status:up
master_last_io_seconds_ago:6
master_sync_in_progress:0
slave_repl_offset:2703
slave_priority:100
slave_read_only:1
connected_slaves:0
master_repl_offset:0
repl_backlog_active:0
repl_backlog_size:1048576
repl_backlog_first_byte_offset:0
repl_backlog_histlen:0
测试数据同步
master上执行
192.168.1.101:6379> set name allen
OK
192.168.1.101:6379> set age 32
OK
192.168.1.101:6379> set sex male
OK
192.168.1.101:6379> set phone 13718097805
OK
192.168.1.101:6379> keys *
1) "age"
2) "name"
3) "phone"
4) "sex"
slave1上查看
192.168.1.102:6379> keys *
1) "sex"
2) "phone"
3) "age"
4) "name"
192.168.1.102:6379> get name
"allen"
slave2上查看
192.168.1.103:6379> keys *
1) "name"
2) "phone"
3) "sex"
4) "age"
192.168.1.103:6379> get name
"allen"
使用Redis Sentinel实现Redis HA
只需要设置一个sentinel配置文件如下:
/etc/redis/sentinel.conf
port 26379
bind 192.168.1.101 需要根据情况变化
protected-mode no
dir /data
logfile /data/logs/sentinel.log
sentinel monitor mymaster 192.168.1.101 6379 1
sentinel down-after-milliseconds mymaster 30000
sentinel parallel-syncs mymaster 1
sentinel failover-timeout mymaster 18000
sentinel auth-pass mymaster mypass
启动Redis Sentinel
[root@mydb1 bin]# pwd
/usr/local/redis/bin
[root@mydb1 bin]# ./redis-sentinel /etc/redis/sentinel.conf &
[root@mydb2 bin]# ./redis-sentinel /etc/redis/sentinel.conf &
[root@mydb3 bin]# ./redis-sentinel /etc/redis/sentinel.conf &
测试:场景一:slave1宕机
192.168.1.102:6379> shutdown
not connected>
192.168.1.101:6379> info Replication
# Replication
role:master
connected_slaves:1
slave0:ip=192.168.1.103,port=6379,state=online,offset=110139,lag=0
master_repl_offset:110562
repl_backlog_active:1
repl_backlog_size:1048576
repl_backlog_first_byte_offset:2
repl_backlog_histlen:110561
192.168.1.103:6379> info Replication
# Replication
role:slave
master_host:192.168.1.101
master_port:6379
master_link_status:up
master_last_io_seconds_ago:2
master_sync_in_progress:0
slave_repl_offset:113128
slave_priority:100
slave_read_only:1
connected_slaves:0
master_repl_offset:0
repl_backlog_active:0
repl_backlog_size:1048576
repl_backlog_first_byte_offset:0
repl_backlog_histlen:0
测试:场景二:slave1恢复
[root@mydb2 bin]# ./redis-server /etc/redis/6379.conf &
192.168.1.101:6379> info Replication
# Replication
role:master
connected_slaves:2
slave0:ip=192.168.1.103,port=6379,state=online,offset=139827,lag=1
slave1:ip=192.168.1.102,port=6379,state=online,offset=139968,lag=0
master_repl_offset:139968
repl_backlog_active:1
repl_backlog_size:1048576
repl_backlog_first_byte_offset:2
repl_backlog_histlen:139967
192.168.1.102:6379> info Replication
# Replication
role:slave
master_host:192.168.1.101
master_port:6379
master_link_status:up
master_last_io_seconds_ago:0
master_sync_in_progress:0
slave_repl_offset:147215
slave_priority:100
slave_read_only:1
connected_slaves:0
master_repl_offset:0
repl_backlog_active:0
repl_backlog_size:1048576
repl_backlog_first_byte_offset:0
repl_backlog_histlen:0
192.168.1.103:6379> info Replication
# Replication
role:slave
master_host:192.168.1.101
master_port:6379
master_link_status:up
master_last_io_seconds_ago:1
master_sync_in_progress:0
slave_repl_offset:150627
slave_priority:100
slave_read_only:1
connected_slaves:0
master_repl_offset:0
repl_backlog_active:0
repl_backlog_size:1048576
repl_backlog_first_byte_offset:0
repl_backlog_histlen:0
测试:场景三:master宕机
192.168.1.101:6379> shutdown
not connected>
经过一段时候后
观察slave1的日志
6423:S 12 Aug 10:32:21.683 # Error condition on socket for SYNC: Connection refused
6423:S 12 Aug 10:32:22.692 * Connecting to MASTER 192.168.1.101:6379
6423:S 12 Aug 10:32:22.692 * MASTER <-> SLAVE sync started
6423:S 12 Aug 10:32:22.693 # Error condition on socket for SYNC: Connection refused
6423:M 12 Aug 10:32:22.841 * Discarding previously cached master state.
6423:M 12 Aug 10:32:22.842 * MASTER MODE enabled (user request from 'id=3 addr=192.168.1.103:41106 fd=6 name=sentinel-a409bbba-cmd age=175 idle=0 flags=x db=0 sub=0 psub=0 multi=3 qbuf=0 qbuf-free=32768 obl=36 oll=0 omem=0 events=r cmd=exec')
6423:M 12 Aug 10:32:22.843 # CONFIG REWRITE executed with success.
6423:M 12 Aug 10:32:23.857 * Slave 192.168.1.103:6379 asks for synchronization
6423:M 12 Aug 10:32:23.857 * Full resync requested by slave 192.168.1.103:6379
6423:M 12 Aug 10:32:23.857 * Starting BGSAVE for SYNC with target: disk
6423:M 12 Aug 10:32:23.858 * Background saving started by pid 6434
6434:C 12 Aug 10:32:23.903 * DB saved on disk
6434:C 12 Aug 10:32:23.903 * RDB: 6 MB of memory used by copy-on-write
6423:M 12 Aug 10:32:24.002 * Background saving terminated with success
6423:M 12 Aug 10:32:24.002 * Synchronization with slave 192.168.1.103:6379 succeeded
观察slave2的日志
6134:S 12 Aug 10:32:50.887 # Error condition on socket for SYNC: Connection refused
6134:S 12 Aug 10:32:51.895 * Connecting to MASTER 192.168.1.101:6379
6134:S 12 Aug 10:32:51.895 * MASTER <-> SLAVE sync started
6134:S 12 Aug 10:32:51.896 # Error condition on socket for SYNC: Connection refused
6134:S 12 Aug 10:32:52.665 * Discarding previously cached master state.
6134:S 12 Aug 10:32:52.665 * SLAVE OF 192.168.1.102:6379 enabled (user request from 'id=4 addr=192.168.1.103:42804 fd=7 name=sentinel-a409bbba-cmd age=264 idle=0 flags=x db=0 sub=0 psub=0 multi=3 qbuf=141 qbuf-free=32627 obl=36 oll=0 omem=0 events=r cmd=exec')
6134:S 12 Aug 10:32:52.668 # CONFIG REWRITE executed with success.
6134:S 12 Aug 10:32:52.906 * Connecting to MASTER 192.168.1.102:6379
6134:S 12 Aug 10:32:52.906 * MASTER <-> SLAVE sync started
6134:S 12 Aug 10:32:52.906 * Non blocking connect for SYNC fired the event.
6134:S 12 Aug 10:32:52.907 * Master replied to PING, replication can continue...
6134:S 12 Aug 10:32:52.908 * Partial resynchronization not possible (no cached master)
6134:S 12 Aug 10:32:52.911 * Full resync from master: 018886996edab0be5e5be9e458f1debb32b83263:1
6134:S 12 Aug 10:32:53.054 * MASTER <-> SLAVE sync: receiving 129 bytes from master
6134:S 12 Aug 10:32:53.055 * MASTER <-> SLAVE sync: Flushing old data
6134:S 12 Aug 10:32:53.055 * MASTER <-> SLAVE sync: Loading DB in memory
6134:S 12 Aug 10:32:53.055 * MASTER <-> SLAVE sync: Finished with success
观察sentinel的日志
6157:X 12 Aug 10:32:51.778 # +sdown master mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:51.778 # +odown master mymaster 192.168.1.101 6379 #quorum 1/1
6157:X 12 Aug 10:32:51.778 # +new-epoch 1
6157:X 12 Aug 10:32:51.778 # +try-failover master mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:51.782 # +vote-for-leader a409bbbafcd5dd0da5639afb4485d228aac95b78 1
6157:X 12 Aug 10:32:51.783 # +elected-leader master mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:51.783 # +failover-state-select-slave master mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:51.836 # +selected-slave slave 192.168.1.102:6379 192.168.1.102 6379 @ mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:51.836 * +failover-state-send-slaveof-noone slave 192.168.1.102:6379 192.168.1.102 6379 @ mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:51.893 * +failover-state-wait-promotion slave 192.168.1.102:6379 192.168.1.102 6379 @ mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:52.597 # +promoted-slave slave 192.168.1.102:6379 192.168.1.102 6379 @ mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:52.597 # +failover-state-reconf-slaves master mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:52.665 * +slave-reconf-sent slave 192.168.1.103:6379 192.168.1.103 6379 @ mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:53.229 * +slave-reconf-inprog slave 192.168.1.103:6379 192.168.1.103 6379 @ mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:53.229 * +slave-reconf-done slave 192.168.1.103:6379 192.168.1.103 6379 @ mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:53.306 # +failover-end master mymaster 192.168.1.101 6379
6157:X 12 Aug 10:32:53.306 # +switch-master mymaster 192.168.1.101 6379 192.168.1.102 6379
6157:X 12 Aug 10:32:53.306 * +slave slave 192.168.1.103:6379 192.168.1.103 6379 @ mymaster 192.168.1.102 6379
已经将slave1转变为master节点
192.168.1.102:6379> info replication
# Replication
role:master
connected_slaves:1
slave0:ip=192.168.1.103,port=6379,state=online,offset=19592,lag=1
master_repl_offset:19733
repl_backlog_active:1
repl_backlog_size:1048576
repl_backlog_first_byte_offset:2
repl_backlog_histlen:19732
已经将原slave2转变为新master节点的slave
192.168.1.103:6379> info replication
# Replication
role:slave
master_host:192.168.1.102
master_port:6379
master_link_status:up
master_last_io_seconds_ago:0
master_sync_in_progress:0
slave_repl_offset:26500
slave_priority:100
slave_read_only:1
connected_slaves:0
master_repl_offset:0
repl_backlog_active:0
repl_backlog_size:1048576
repl_backlog_first_byte_offset:0
repl_backlog_histlen:0
测试:场景四:master恢复
[root@mydb1 bin]# ./redis-server /etc/redis/6379.conf &
当master挂掉时,sentinel会自动的从slave中挑选出一个作为master,并重新配置各redis实例的配置文件。当master重启后,sentinel会自动将它加入到当前的环境中,变成一个slave节点。
192.168.1.101:6379> info replication
# Replication
role:slave
master_host:192.168.1.102
master_port:6379
master_link_status:up
master_last_io_seconds_ago:1
master_sync_in_progress:0
slave_repl_offset:48093
slave_priority:100
slave_read_only:1
connected_slaves:0
master_repl_offset:0
repl_backlog_active:0
repl_backlog_size:1048576
repl_backlog_first_byte_offset:0
repl_backlog_histlen:0
192.168.1.102:6379> info replication
# Replication
role:master
connected_slaves:2
slave0:ip=192.168.1.103,port=6379,state=online,offset=49108,lag=0
slave1:ip=192.168.1.101,port=6379,state=online,offset=49108,lag=0
master_repl_offset:49108
repl_backlog_active:1
repl_backlog_size:1048576
repl_backlog_first_byte_offset:2
repl_backlog_histlen:49107
192.168.1.103:6379> info replication
# Replication
role:slave
master_host:192.168.1.102
master_port:6379
master_link_status:up
master_last_io_seconds_ago:1
master_sync_in_progress:0
slave_repl_offset:50250
slave_priority:100
slave_read_only:1
connected_slaves:0
master_repl_offset:0
repl_backlog_active:0
repl_backlog_size:1048576
repl_backlog_first_byte_offset:0
repl_backlog_histlen:0
当之前挂掉的master又重启后,如果希望还是还原到原先的主从配置
登录到sentinel,使用命令:
[root@mydb3 bin]## ./redis-cli -h 192.168.1.103 -p 26379
192.168.1.103:26379> sentinel failover mymaster
一主(Master)多从(Slave),主库不开启AOF持久化,只是每天备份一下RDB[官方给的建议是每小时备份RDB文件],而在从库上开启AOF备份,并且会用脚本将相应的备份文件推送到备份服务器。
当redis服务器挂掉时,重启时将按照以下优先级恢复数据到内存:
恢复时需要注意,要是主库挂了不能直接重启主库,否则会直接覆盖掉从库的AOF文件,一定要确保要恢复的文件都正确才能启动,否则会冲掉原来的文件。
Redis主从及哨兵的更多相关文章
- 关于redis主从|哨兵|集群模式
关于redis主从.哨兵.集群的介绍网上很多,这里就不赘述了. 一.主从 通过持久化功能,Redis保证了即使在服务器重启的情况下也不会损失(或少量损失)数据,因为持久化会把内存中数据保存到硬盘上,重 ...
- 自动配置redis主从和哨兵
redis shell 每次创建redis主从和哨兵服务,文件夹要复制好几个,配置文件改一大堆.繁琐还容易出错,就想通过shell脚本自动帮我配置好端口以及文件,下面就是脚本内容: redis-to ...
- redis主从|哨兵|集群模式
关于redis主从.哨兵.集群的介绍网上很多,这里就不赘述了. 一.主从 通过持久化功能,Redis保证了即使在服务器重启的情况下也不会损失(或少量损失)数据,因为持久化会把内存中数据保存到硬盘上,重 ...
- Redis 主从、哨兵Sentinel、Jedis
Redis 主从.哨兵Sentinel.Jedis 2017年02月15日 15:52:48 有且仅有 阅读数 6183 文章标签: redis主从sentineljedis 更多 分类专栏: 7/1 ...
- docker部署redis主从和哨兵
docker部署redis主从和哨兵 原文地址:https://www.jianshu.com/p/72ee9568c8ea 1主2从3哨兵 一.前期准备工作 1.电脑装有docker 2.假设本地i ...
- docker安装redis主从以及哨兵
docker安装redis主从以及哨兵 本文使用docker在四台机器上部署一主二从三哨兵的Redis主从结构. 服务器配置 192.168.102.128 主节点 centos7.5 192.168 ...
- redis主从配置+哨兵模式
1.搭建redis主从,一个master两个slave,加一个哨兵监听(sentinel),可以新建三个虚拟机,模拟环境,我的电脑没那么多虚拟机,就在一台虚拟机上弄的. 2.安装redis,如果是三台 ...
- redis主从,哨兵回忆手册
redis主从 持久化的开启与主从集群是否生效无关系 Slave Server同样是以非阻塞的方式完成数据同步.在同步期间,如果有客户端提交查询请求,Redis则返回同步之前的数据(注意初次同步则会阻 ...
- Redis学习之路(三)之Redis主从和哨兵模式
目录 一.Redis主从配置 1.环境说明 2.修改主从的redis配置文件 3.启动主从redis 3.数据同步验证 二.Redis哨兵模式 1.Redis sentinel介绍 2.Redis S ...
- Redis 主从同步+哨兵
简介 通过使用 Redis 自带“主从同步+哨兵守护”功能提高Redis稳定性. 主从同步:保障数据主从数据实时同步. 哨兵:实时监控主redis如果故障,将从redis作为主使用. 环境: 系统:C ...
随机推荐
- 关于keepalive
linux内核配置有一项tcp_keepalive_time,即tcp的保活定时器.当网络上两个建立连接的进程都没有数据向对方发送的时候,tcp会隔段时间发送一次保活数据,以保持连接,间隔时间就是tc ...
- LVS集群
集群: 将许多小的,性能较低的服务器做成一个大的性能高的超级服务器 集群分为负载均衡集群,高可用集群,高性能运算集群 LVS体系结构与工作原理描述 LVS集群负载均衡器接受服务的所有入站客户端计算机请 ...
- hibernate一对一映射
package loaderman.c_one2one; // 身份证 public class IdCard { // 身份证号(主键) private String cardNum;// 对象唯一 ...
- JDK8记FullGC时候Metaspace内存不会被垃圾回收
本文链接:https://blog.csdn.net/a15939557197/article/details/90635460背景前段时间有一个这样的需求:第三方调用接口,30分钟内调用120W次: ...
- 利用Spring的AbstractRoutingDataSource解决多数据源的问题【代码手动切换,非AOP】
转: 利用Spring的AbstractRoutingDataSource解决多数据源的问题 多数据源问题很常见,例如读写分离数据库配置. 原来的项目出现了新需求,局方要求新增某服务器用以提供某代码, ...
- 一百零一:CMS系统之自定义restful风格json返回格式和内容
鉴于flask-restful用起来太麻烦,这里自定义json统一返回格式和内容 from flask import jsonify class HttpCode: """ ...
- eclipse中Maven web项目的目录结构浅析
刚开始接触maven web项目的时候,相信很多人都会被它的目录结构迷惑. 为了避免初学者遇到像我一样的困扰,我就从一个纯初学者的视角,来分析一下这个东西. 1,比如说,我们拿一个常见的目录结构来看, ...
- Linux服务器端使用tcpdump抓redis报文
yum update yum install tcpdump tcpdump host -i ens192 -w /tmp/redis-zsetcapture.cap 其中ens192是网卡名, /t ...
- Linux学习—redis安装配置及远程连接
1.下载安装包并解压 进入文件夹/usr/local cd /usr/local 下载redis安装包: wget http://download.redis.io/releases/redis-.t ...
- A simple dynamic library implemented in C# 4.0 to deal with XML structure
https://github.com/cardinals/XmlToObjectParser A simple dynamic library implemented in C# 4.0 to dea ...