客户端分片 程序端实现

代理proxy,访问proxy,proxy指定redis保存位置。 Twemproxy

Redis cluster ,会造成一部分数据丢失,无中心化
1.将数据自动切分(split)到多个节点的能力
2.当集群中的一部分节点失效或者无法进行通信时,仍然可以继续处理命令请求的能力。

codis redis集群解决方案。豌豆荚开源

redis cluster 集群配置

cd /data/
mkdir `seq 7001 7008`

redis.conf 添加,端口需要变化
'''
port 7000
cluster-enabled yes
cluster-config-file nodes.conf
cluster-node-timeout 5000
appendonly yes

'''
安装ruby
yum -y install ruby rubygems

安装ruby的redis管理工具

gem install redis

1 gem installed
Installing ri documentation for redis-3.3.0...
Installing RDoc documentation for redis-3.3.0...

创建集群
redis-trib.rb create --replicas 1 172.16.230.167:7001 172.16.230.167:7002 172.16.230.167:7003 172.16.230.167:7004 172.16.230.167:7005

172.16.230.167:7006

客户端连接:

redis-cli -c -h 172.16.230.167 -p 7001
set name 111
set name1 222
会写入到不通哈希槽

cluster nodes

cluster info

添加集群节点:命令中的 add-node 表示我们要让 redis-trib 将一个节点添加到集群里面, add-node 之后跟着的是新节点的 IP 地址和端口号, 再

之后跟着的是集群中任意一个已存在节点的 IP 地址和端口号

[root@slave1 data]# redis-trib.rb add-node 172.16.230.167:7007 172.16.230.167:7001

>>> Adding node 172.16.230.167:7007 to cluster 172.16.230.167:7001
>>> Performing Cluster Check (using node 172.16.230.167:7001)
M: f2a684e0900e7b7eb91bdb77e3b559f47d0f7237 172.16.230.167:7001
slots:0-5460 (5461 slots) master
1 additional replica(s)
S: 37b6f6478ca6425022521d6a5ecf0aeedf12c601 172.16.230.167:7005
slots: (0 slots) slave
replicates 84995c68e3625c91f4c646d71566dffd709e6d01
S: d6d307c91e6ba6fbba7d898f20fd4772a3e42b9c 172.16.230.167:7004
slots: (0 slots) slave
replicates f2a684e0900e7b7eb91bdb77e3b559f47d0f7237
S: 458a01eeace9fd93544645433ad5b12587d94d3b 172.16.230.167:7006
slots: (0 slots) slave
replicates ffe722398240b59f0c993074d35dbe79386c1ec3
M: ffe722398240b59f0c993074d35dbe79386c1ec3 172.16.230.167:7003
slots:10923-16383 (5461 slots) master
1 additional replica(s)
M: 84995c68e3625c91f4c646d71566dffd709e6d01 172.16.230.167:7002
slots:5461-10922 (5462 slots) master
1 additional replica(s)
[OK] All nodes agree about slots configuration.
>>> Check for open slots...
>>> Check slots coverage...
[OK] All 16384 slots covered.
>>> Send CLUSTER MEET to node 172.16.230.167:7007 to make it join the cluster.
[OK] New node added correctly.

查询新节点
172.16.230.167:7001> cluster nodes

37b6f6478ca6425022521d6a5ecf0aeedf12c601 172.16.230.167:7005 slave 84995c68e3625c91f4c646d71566dffd709e6d01 0 1467381076060 5

connected
d6d307c91e6ba6fbba7d898f20fd4772a3e42b9c 172.16.230.167:7004 slave f2a684e0900e7b7eb91bdb77e3b559f47d0f7237 0 1467381077061 4

connected
458a01eeace9fd93544645433ad5b12587d94d3b 172.16.230.167:7006 slave ffe722398240b59f0c993074d35dbe79386c1ec3 0 1467381073056 6

connected
ffe722398240b59f0c993074d35dbe79386c1ec3 172.16.230.167:7003 master - 0 1467381074057 3 connected 10923-16383

83e5dd37a0825a4adb77d6da4738bf0c88cd6c39 172.16.230.167:7007 master - 0 1467381075059 0 connected

84995c68e3625c91f4c646d71566dffd709e6d01 172.16.230.167:7002 master - 0 1467381073556 2 connected 5461-10922
f2a684e0900e7b7eb91bdb77e3b559f47d0f7237 172.16.230.167:7001 myself,master - 0 0 1 connected 0-5460

将一个新节点转变为从节点的方法,把7008加入到集群中
[root@slave1 data]# redis-trib.rb add-node 172.16.230.167:7008 172.16.230.167:7001

进入到新加入集群的终端中
[root@slave1 data]# redis-cli -c -h 172.16.230.167 -p 7008

172.16.230.167:7008> CLUSTER nodes
84995c68e3625c91f4c646d71566dffd709e6d01 172.16.230.167:7002 master - 0 1467381351640 2 connected 5461-10922
37b6f6478ca6425022521d6a5ecf0aeedf12c601 172.16.230.167:7005 slave 84995c68e3625c91f4c646d71566dffd709e6d01 0 1467381350638 2

connected
f2a684e0900e7b7eb91bdb77e3b559f47d0f7237 172.16.230.167:7001 master - 0 1467381352142 1 connected 0-5460
83e5dd37a0825a4adb77d6da4738bf0c88cd6c39 172.16.230.167:7007 master - 0 1467381352641 0 connected
458a01eeace9fd93544645433ad5b12587d94d3b 172.16.230.167:7006 slave ffe722398240b59f0c993074d35dbe79386c1ec3 0 1467381349636 3

connected
ffe722398240b59f0c993074d35dbe79386c1ec3 172.16.230.167:7003 master - 0 1467381348635 3 connected 10923-16383
d6d307c91e6ba6fbba7d898f20fd4772a3e42b9c 172.16.230.167:7004 slave f2a684e0900e7b7eb91bdb77e3b559f47d0f7237 0 1467381348135 1

connected
6b52de99a37444a2f7cb04a65ed4239446dee048 172.16.230.167:7008 myself,master - 0 0 7 connected

172.16.230.167:7008> CLUSTER REPLICATE 83e5dd37a0825a4adb77d6da4738bf0c88cd6c39

其中命令提供的 83e5dd37a0825a4adb77d6da4738bf0c88cd6c39 就是主节点 172.16.230.167:7007 的节点 ID

172.16.230.167:7008> CLUSTER nodes
84995c68e3625c91f4c646d71566dffd709e6d01 172.16.230.167:7002 master - 0 1467381552990 2 connected 5461-10922
37b6f6478ca6425022521d6a5ecf0aeedf12c601 172.16.230.167:7005 slave 84995c68e3625c91f4c646d71566dffd709e6d01 0 1467381554995 2

connected
f2a684e0900e7b7eb91bdb77e3b559f47d0f7237 172.16.230.167:7001 master - 0 1467381553992 1 connected 0-5460
83e5dd37a0825a4adb77d6da4738bf0c88cd6c39 172.16.230.167:7007 master - 0 1467381555997 0 connected
458a01eeace9fd93544645433ad5b12587d94d3b 172.16.230.167:7006 slave ffe722398240b59f0c993074d35dbe79386c1ec3 0 1467381550487 3

connected
ffe722398240b59f0c993074d35dbe79386c1ec3 172.16.230.167:7003 master - 0 1467381550988 3 connected 10923-16383
d6d307c91e6ba6fbba7d898f20fd4772a3e42b9c 172.16.230.167:7004 slave f2a684e0900e7b7eb91bdb77e3b559f47d0f7237 0 1467381549986 1

connected
6b52de99a37444a2f7cb04a65ed4239446dee048 172.16.230.167:7008 myself,slave 83e5dd37a0825a4adb77d6da4738bf0c88cd6c39 0 0 7 connected

对集群进行重新分片:主要对172.16.230.167:7007 集群节点做重新分片

[root@slave1 ~]# redis-trib.rb reshard 172.16.230.167:7007

>>> Performing Cluster Check (using node 172.16.230.167:7007)
M: 83e5dd37a0825a4adb77d6da4738bf0c88cd6c39 172.16.230.167:7007
slots: (0 slots) master
1 additional replica(s)
S: d6d307c91e6ba6fbba7d898f20fd4772a3e42b9c 172.16.230.167:7004
slots: (0 slots) slave
replicates f2a684e0900e7b7eb91bdb77e3b559f47d0f7237
S: 6b52de99a37444a2f7cb04a65ed4239446dee048 172.16.230.167:7008
slots: (0 slots) slave
replicates 83e5dd37a0825a4adb77d6da4738bf0c88cd6c39
M: 84995c68e3625c91f4c646d71566dffd709e6d01 172.16.230.167:7002
slots:5461-10922 (5462 slots) master
1 additional replica(s)
S: 37b6f6478ca6425022521d6a5ecf0aeedf12c601 172.16.230.167:7005
slots: (0 slots) slave
replicates 84995c68e3625c91f4c646d71566dffd709e6d01
M: f2a684e0900e7b7eb91bdb77e3b559f47d0f7237 172.16.230.167:7001
slots:0-5460 (5461 slots) master
1 additional replica(s)
S: 458a01eeace9fd93544645433ad5b12587d94d3b 172.16.230.167:7006
slots: (0 slots) slave
replicates ffe722398240b59f0c993074d35dbe79386c1ec3
M: ffe722398240b59f0c993074d35dbe79386c1ec3 172.16.230.167:7003
slots:10923-16383 (5461 slots) master
1 additional replica(s)
[OK] All nodes agree about slots configuration.
>>> Check for open slots...
>>> Check slots coverage...
[OK] All 16384 slots covered.
How many slots do you want to move (from 1 to 16384)? 1000 #######分片大小
What is the receiving node ID? 83e5dd37a0825a4adb77d6da4738bf0c88cd6c39 #分配到那个集群节点的master上172.16.230.167:7007
Please enter all the source node IDs.
Type 'all' to use all the nodes as source nodes for the hash slots.
Type 'done' once you entered all the source nodes IDs.
Source node #1:all #所有

Do you want to proceed with the proposed reshard plan (yes/no)?yes

使用 consistency-test.rb检测数据一致性
下载
git clone https://github.com/antirez/redis-rb-cluster.git

[root@slave1 redis-rb-cluster]# ruby consistency-test.rb 172.16.230.167 7001
316 R (0 err) | 316 W (0 err) |
1453 R (0 err) | 1453 W (0 err) |
2708 R (0 err) | 2708 W (0 err) |
4010 R (0 err) | 4010 W (0 err) |
5309 R (0 err) | 5309 W (0 err) |
6613 R (0 err) | 6613 W (0 err) |
7915 R (0 err) | 7915 W (0 err) |
9214 R (0 err) | 9214 W (0 err) |
10517 R (0 err) | 10517 W (0 err) |
11817 R (0 err) | 11817 W (0 err) |
13121 R (0 err) | 13121 W (0 err) |
14422 R (0 err) | 14422 W (0 err) |
15724 R (0 err) | 15724 W (0 err) |
17012 R (0 err) | 17012 W (0 err) |
18312 R (0 err) | 18312 W (0 err) |
19628 R (0 err) | 19628 W (0 err) |
20949 R (0 err) | 20949 W (0 err) |
22255 R (0 err) | 22255 W (0 err) |
23294 R (0 err) | 23294 W (0 err) |
24542 R (0 err) | 24542 W (0 err) |
25852 R (0 err) | 25852 W (0 err) |

故障转移测试

查看主节点:
[root@slave1 ~]# redis-cli -c -h 172.16.230.167 -p 7001 cluster nodes | grep master
ffe722398240b59f0c993074d35dbe79386c1ec3 172.16.230.167:7003 master - 0 1467383421007 3 connected 11256-16383
83e5dd37a0825a4adb77d6da4738bf0c88cd6c39 172.16.230.167:7007 master - 0 1467383423010 8 connected 0-332 5461-5794 10923-11255
84995c68e3625c91f4c646d71566dffd709e6d01 172.16.230.167:7002 master - 0 1467383426014 2 connected 5795-10922
f2a684e0900e7b7eb91bdb77e3b559f47d0f7237 172.16.230.167:7001 myself,master - 0 0 1 connected 333-5460

使用debug segfault 让主节点崩溃
redis-cli -c -h 172.16.230.167 -p 7001 debug segfault

[root@slave1 ~]# redis-cli -c -h 172.16.230.167 -p 7002 cluster nodes | grep master
84995c68e3625c91f4c646d71566dffd709e6d01 172.16.230.167:7002 myself,master - 0 0 2 connected 5795-10922
d6d307c91e6ba6fbba7d898f20fd4772a3e42b9c 172.16.230.167:7004 master - 0 1467383553156 9 connected 333-5460
83e5dd37a0825a4adb77d6da4738bf0c88cd6c39 172.16.230.167:7007 master - 0 1467383556161 8 connected 0-332 5461-5794 10923-11255
ffe722398240b59f0c993074d35dbe79386c1ec3 172.16.230.167:7003 master - 0 1467383553657 3 connected 11256-16383
f2a684e0900e7b7eb91bdb77e3b559f47d0f7237 172.16.230.167:7001 master,fail - 1467383513361 1467383506054 1 disconnected

集群管理
yum -y install httpd php php-redis

git clone https://github.com/erikdubbelboer/phpRedisAdmin.git
cd phpRedisAdmin
git clone https://github.com/nrk/predis.git vendor

###################################

Codis 集群

Codis 由四部分组成:

Codis Proxy (codis-proxy)
Codis Dashboard (codis-config)
Codis Redis (codis-server)
ZooKeeper/Etcd

安装go
wget http://dinp.qiniudn.com/go1.7.0.linux-amd64.tar.gz
tar -zxvf go1.4.1.linux-amd64.tar.gz
mv go /usr/local/
vim /etc/profile
source /etc/profile
go version
mkdir -p $GOPATH/src/github.com/tools
go get -u github.com/tools/godep

go get github.com/golang/tools
cd /data/gopath/github.com/golang/
cp -ar tools /usr/local/go/src/golang.org/x/

go install ./
which godep
vim /etc/profile
source /etc/profile
go get -u -d github.com/CodisLabs/codis
cd CodisLabs/codis/
make
make gotest

环境变量:
export GOROOT=/usr/local/go
export GOPATH=/data/gopath
export ZOOKEEPER_HOME=/data/zookeeper
export PATH=$GOROOT/bin:$PATH
export PATH=$GOROOT/bin:$GOPATH/bin:$PATH

执行全部指令后,会在 bin 文件夹内生成 codis-config、codis-proxy、codis-server三个可执行文件。另外, bin/assets 文件夹是 codis-config

的 dashboard http 服务需要的前端资源, 需要和 codis-config 放置在同一文件夹下)
[root@slave1 bin]# ls
assets codis-config codis-proxy codis-server

编辑codis config.ini配置文件

coordinator=zookeeper
zk=172.16.230.167:2181
product=cinyi
dashboard_addr=172.16.230.167:18087
password=
backend_ping_period=5
session_max_timeout=1800
session_max_bufsize=131072
session_max_pipeline=1024
zk_session_timeout=30000
proxy_id=cinyicodis-poryx_1

配置和启动各组件
1.启动 dashboard(两台机,只需启动一台机上的dashboard)
cd /apps/svr/src/github.com/CodisLabs/codis
执行./bin/codis-config dashboard &, 该命令会启动 dashboard
以下是启动信息:
[1] 6243
[root@slave1 codis]# 2016/07/04 22:05:43 dashboard.go:160: [INFO] dashboard listening on addr: :18087
2016/07/04 22:05:43 dashboard.go:143: [INFO] dashboard node created: /zk/codis/db_cinyi/dashboard, {"addr": "172.16.230.167:18087",

"pid": 6243}
2016/07/04 22:05:43 dashboard.go:144: [WARN] ********** Attention **********
2016/07/04 22:05:43 dashboard.go:145: [WARN] You should use `kill {pid}` rather than `kill -9 {pid}` to stop me,
2016/07/04 22:05:43 dashboard.go:146: [WARN] or the node resisted on zk will not be cleaned when I'm quiting and you must remove it

manually
2016/07/04 22:05:43 dashboard.go:147: [WARN] *******************************

从上面信息可看到:
a.关闭dashboard的操作要用kil pid,不要用kill -9 pid
b.若关闭异常,下次启动时,会报错提示zk已有建立,此时需要删除 :/zk/codis/db_gls/dashboard,参考最后问题处理。
访问管理页面:http://172.16.230.167:18087,可以进行server group的增删操作、slot操作等。

初始化 slots
cd /data/gopath/src/github.com/CodisLabs/codis/bin
执行 ./codis-config -c ../config.ini slot init,该命令会在zookeeper上创建slot相关信息

配置和启动 Codis 的Redis(除配置文件scp外,其它两台机上都要操作)

cp /data/gopath/src/github.com/CodisLabs/codis/extern/redis-2.8.21 /data/redis2.8 -ar

cd /data/redis2.8/

make install

mkdir /data/{9001,9002,9003,9004}

###################################
[root@slave1 9004]# cat 9004.conf | grep -v "#" | sed '/^$/d'
daemonize yes
pidfile /var/run/redis_9004.pid
port 9004
tcp-backlog 511
timeout 0
tcp-keepalive 0
loglevel notice
logfile "/var/log/redis_9004.log"
databases 16
save 900 1
save 300 10
save 60 10000
stop-writes-on-bgsave-error yes
rdbcompression yes
rdbchecksum yes
dbfilename 9004_dump.rdb
dir /data/9004/
slave-serve-stale-data yes
slave-read-only yes
repl-diskless-sync no
repl-diskless-sync-delay 5
repl-disable-tcp-nodelay no
slave-priority 100
appendonly no
appendfilename "9004_appendonly.aof"
appendfsync everysec
no-appendfsync-on-rewrite no
auto-aof-rewrite-percentage 100
auto-aof-rewrite-min-size 64mb
aof-load-truncated yes
lua-time-limit 5000
slowlog-log-slower-than 10000
slowlog-max-len 128
latency-monitor-threshold 0
notify-keyspace-events ""
hash-max-ziplist-entries 512
hash-max-ziplist-value 64
list-max-ziplist-entries 512
list-max-ziplist-value 64
set-max-intset-entries 512
zset-max-ziplist-entries 128
zset-max-ziplist-value 64
hll-sparse-max-bytes 3000
activerehashing yes
client-output-buffer-limit normal 0 0 0
client-output-buffer-limit slave 256mb 64mb 60
client-output-buffer-limit pubsub 32mb 8mb 60
hz 10
aof-rewrite-incremental-fsync yes

###########################################

启动
/data/redis2.8/src/redis-server /data/9001/9001.conf
/data/redis2.8/src/redis-server /data/9002/9002.conf
/data/redis2.8/src/redis-server /data/9003/9003.conf
/data/redis2.8/src/redis-server /data/9004/9004.conf

添加 Redis Server Group(只在一台上操作,也可以在面板图形界面操作)
cd $CODIS_HOME/bin
./codis-config -c ../config.ini server add 1 172.16.230.167:9001 master
./codis-config -c ../config.ini server add 1 172.16.230.167:9003 slave
./codis-config -c ../config.ini server add 2 172.16.230.167:9003 master
./codis-config -c ../config.ini server add 2 172.16.230.167:9004 slave

查看

[root@slave1 bin]# ./codis-config -c ../config.ini server list
[
{
"id": 1,
"product_name": "cinyi",
"servers": [
{
"addr": "172.16.230.167:9001",
"group_id": 1,
"type": "master"
},
{
"addr": "172.16.230.167:9002",
"group_id": 1,
"type": "slave"
}
]
},
{
"id": 2,
"product_name": "cinyi",
"servers": [
{
"addr": "172.16.230.167:9003",
"group_id": 2,
"type": "master"
},
{
"addr": "172.16.230.167:9004",
"group_id": 2,
"type": "slave"
}
]
}
]

设置 server group 服务的 slot 范围(只在一台执行)

Codis 采用 Pre-sharding 的技术来实现数据的分片, 默认分成 1024 个 slots (0-1023), 对于每个key来说, 通过以下公式确定所属的 Slot Id :

SlotId = crc32(key) % 1024 每一个
slot 都会有一个且必须有一个特定的 server group id 来表示这个 slot 的数据由哪个 server group 来提供.
cd $CODIS_HOME
codis-config slot range-set <slot_from> <slot_to> <group_id> <status>

./codis-config -c ../config.ini slot range-set 0 511 1 online
./codis-config -c ../config.ini slot range-set 512 1023 2 online

启动 codis-proxy
节点一:
cd $CODIS_HOME

./codis-proxy -c ./config.ini -L /var/log/proxy.log --cpu=2 --addr=0.0.0.0:19000 --http-addr=0.0.0.0:11000 &

节点二:
cd $CODIS_HOME

./codis-proxy -c ./config.ini -L /var/log/proxy.log --cpu=2 --addr=0.0.0.0:19000 --http-addr=0.0.0.0:11000 &

刚启动的 codis-proxy 默认是处于 offline状态的, 然后设置 proxy 为 online 状态, 只有处于 online 状态的 proxy 才会对外提供服务(codis

3..0.3之前版本需要online操作)

节点一:
./codis-config -c ../config.ini proxy online cinyicodis-poryx_1

节点二:
./codis-config -c ../config.ini proxy online cinyicodis-poryx_2

五.遇到的问题
[root@mvxl2530 codis]# bin/codis-config dashboard &
[1] 14429
2016/05/11 17:45:30 dashboard.go:160: [INFO] dashboard listening on addr: :18087
[root@mvxl2530 codis]# 2016/05/11 17:45:30 dashboard.go:234: [PANIC] create zk node failed
[error]: dashboard already exists: {"addr": "172.16.230.167:18087", "pid": 13819}
[stack]:
3 /apps/svr/src/github.com/CodisLabs/codis/cmd/cconfig/dashboard.go:234
main.runDashboard
2 /apps/svr/src/github.com/CodisLabs/codis/cmd/cconfig/dashboard.go:54
main.cmdDashboard
1 /apps/svr/src/github.com/CodisLabs/codis/cmd/cconfig/main.go:85
main.runCommand
0 /apps/svr/src/github.com/CodisLabs/codis/cmd/cconfig/main.go:152
main.main
... ...
[1]+ Exit 1 bin/codis-config dashboard
原因:由于在停dashboard时,采用kill -9强杀进程方法引起,应采用kill pid方法来关闭dashboard.
临时处理:
[root@mvxl2530 zookeeper]# cd /usr/local/zookeeper
[root@mvxl2530 zookeeper]# ./bin/zkCli.sh -server 127.0.0.1:2181
[zk: 127.0.0.1:2181(CONNECTED) 1]
[zk: 127.0.0.1:2181(CONNECTED) 6] ls /zk/codis/db_zh_news/dashboard --其中zh_news为codis配置文件中的product
[fence, servers, slots, proxy, migrate_tasks, dashboard, LOCK, actions, ActionResponse]
[migrate_tasks, dashboard, actions, slots, ActionResponse]
[zk: 127.0.0.1:2181(CONNECTED) 2] rmr /zk/codis/db_zh_news/dashboard
[zk: 127.0.0.1:2181(CONNECTED) 3]

六.相关命令和参数用法:
1.codis-config使用
[root@mvxl2530 bin]# ./codis-config -h
usage: codis-config [-c <config_file>] [-L <log_file>] [--log-level=<loglevel>]
<command> [<args>...]
options:
-c set config file
-L set output log file, default is stdout
--log-level=<loglevel> set log level: info, warn, error, debug [default: info]
commands:
server
slot
dashboard
action
proxy
[root@mvxl2530 bin]# ./codis-proxy -h
2.codis proxy使用
usage: proxy [-c <config_file>] [-L <log_file>] [--log-level=<loglevel>] [--log-filesize=<filesize>] [--cpu=<cpu_num>] [--

addr=<proxy_listen_addr>] [--http-
addr=<debug_http_server_addr>]
options:
-c set config file
-L set output log file, default is stdout
--log-level=<loglevel> set log level: info, warn, error, debug [default: info]
--log-filesize=<maxsize> set max log file size, suffixes "KB", "MB", "GB" are allowed, 1KB=1024 bytes, etc. Default is 1GB.
--cpu=<cpu_num> num of cpu cores that proxy can use
--addr=<proxy_listen_addr> proxy listen address, example: 0.0.0.0:9000
--http-addr=<debug_http_server_addr> debug vars http server
3.codis-config用于增删server组
每一个 Server Group 作为一个 Redis 服务器组存在, 只允许有一个 master, 可以有多个 slave, group id 仅支持大于等于1的整数。
命令说明:
$ bin/codis-config server -h

usage:
codis-config server list
codis-config server add <group_id> <redis_addr> <role>
codis-config server remove <group_id> <redis_addr>
codis-config server promote <group_id> <redis_addr>
codis-config server add-group <group_id>
codis-config server remove-group <group_id>
如: 添加两个 server group, 每个 group 有两个 redis 实例,group的id分别为1和2, redis实例为一主一从。
添加一个group,group的id为1, 并添加一个redis master到该group

类似的,再添加group,group的id为3
$ bin/codis-config server add 3 localhost:6479 master
$ bin/codis-config server add 3 localhost:6480 slave

数据迁移,slot 从1000到1023 迁移到group3中
bin/codis-config slot migrate 1000 1023 3 --delay=10

Auto Rebalance

Codis 支持动态的根据实例内存, 自动对slot进行迁移, 以均衡数据分布.

$ bin/codis-config slot rebalance

若需要移除,参考如下:
./bin/codis-config -c ./config.ini server remove 1 172.16.230.167:9002
./bin/codis-config -c ./config.ini server remove 2 172.16.230.167:9004
./bin/codis-config -c ./config.ini server remove-group 1
./bin/codis-config -c ./config.ini server remove-group 2
redis-benchmark -h codis1 -p 19000 -n 1000000 -c 300 -r 1048576 -d 256 -t lpush ---压测

4.codis-config用于slot操作
命令说明:
$ bin/codis-config slot -h

usage:
codis-config slot init
codis-config slot info <slot_id>
codis-config slot set <slot_id> <group_id> <status>
codis-config slot range-set <slot_from> <slot_to> <group_id> <status>
codis-config slot migrate <slot_from> <slot_to> <group_id> [--delay=<delay_time_in_ms>]

数据迁移,slot 从1000到1023 迁移到group3中
bin/codis-config slot migrate 1000 1023 3 --delay=10

Auto Rebalance

Codis 支持动态的根据实例内存, 自动对slot进行迁移, 以均衡数据分布.

$ bin/codis-config slot rebalance

5.移除分片流程
---假设将分片3移除---
1. 设置codis-proxy为offline状态.
./bin/codis-config -c codis-config.ini proxy offline proxy_1

2. 迁移分片3上的数据到分片1
./bin/codis-config -c codis-config.ini slot migrate 256 511 1

3. 彻底移除分片3
./bin/codis-config -c codis-config.ini server remove-group 3

6.zk的查看操作
查看slot:
ls /zk/codis/db_gls/slots
proxy:
ls /zk/codis/db_gls/proxy
server:
ls /zk/codis/db_gls/servers
dashboard:
ls /zk/codis/db_gls/dashboard
查看迁移任务:
ls /zk/codis/db_gls/migrate_tasks

redis集群配置的更多相关文章

  1. java操作redis集群配置[可配置密码]和工具类(比较好用)

    转: java操作redis集群配置[可配置密码]和工具类 java操作redis集群配置[可配置密码]和工具类     <dependency>   <groupId>red ...

  2. Redis集群配置(linux)

     *弄了一天,有问题直接问我.qq:137416943   1.redis集群的配置和简单使用   Redis集群配置 0.首先要配置环境: 0.1 安装c++ yum install gcc-c++ ...

  3. java操作redis集群配置[可配置密码]和工具类

    java操作redis集群配置[可配置密码]和工具类     <dependency>   <groupId>redis.clients</groupId>   & ...

  4. Linux系统下安装Redis和Redis集群配置

    Linux系统下安装Redis和Redis集群配置 一. 下载.安装.配置环境: 1.1.>官网下载地址: https://redis.io/download (本人下载的是3.2.8版本:re ...

  5. centos6.5下redis集群配置(多机多节点)

    可参考官网文档:redis集群配置 需要注意的是,集群中的每个节点都会涉及到两个端口,一个是用于处理客户端操作的(如下介绍到的6379/6380),另一个是10000+{监听端口},用于集群各个节点间 ...

  6. redis简介以及redis集群配置

    简介: redis 是一个高性能的key-value数据库..它支持存储的value类型相对更多,包括string(字符串).list(链表).set(集合).zset(sorted set --有序 ...

  7. redis 集群配置实战

    文章转载自:http://hot66hot.iteye.com/blog/2050676 最近研究Redis-cluster,正好搭建了一个环境,遇到了很多坑,系统的总结下,等到redis3 rele ...

  8. spring 使用redis集群配置

    上面两篇介绍了redis集群的配置合一些基本的概念,所以接下来当然是要在项目中使用咯,redis的java支持已经做的非常好了,所以我们来试着使用这些api来进行redis的操作,首先我们需要操作re ...

  9. redis集群配置,spring整合jedis,缓存同步

    前台的商品数据(图片等加载缓慢)查询,先从redis缓存查询数据. redis是一个nosql数据库,内存版数据库,读取速度11w/s.本身具有内存淘汰机制,是单线程服务器(分时操作系统),线程安全. ...

随机推荐

  1. HTTPS 原理解析(转)

    一 前言 在说HTTPS之前先说说什么是HTTP,HTTP就是我们平时浏览网页时候使用的一种协议.HTTP协议传输的数据都是未加密的,也就是明文的,因此使用HTTP协议传输隐私信息非常不安全.为了保证 ...

  2. hdu1710 Binary Tree Traversals(二叉树的遍历)

    A binary tree is a finite set of vertices that is either empty or consists of a root r and two disjo ...

  3. HDU1518 Square(DFS,剪枝是关键呀)

    Square Time Limit : 10000/5000ms (Java/Other)   Memory Limit : 65536/32768K (Java/Other) Total Submi ...

  4. [题解+总结]动态规划大合集II

    1.前言 大合集总共14道题,出自江哥之手(这就没什么好戏了),做得让人花枝乱颤.虽说大部分是NOIP难度,也有简单的几道题目,但是还是做的很辛苦,有几道题几乎没思路,下面一道道边看边分析一下. 2. ...

  5. 彻底弄明白之java多线程中的volatile

    一. volatite 简述 Java 语言提供了一种稍弱的同步机制,即 volatile 变量.用来确保将变量的更新操作通知到其他线程,保证了新值能立即同步到主内存,以及每次使用前立即从主内存刷新. ...

  6. java8 引进lamda

    就像泛型能使开发人员对数据类型进行抽象,Lambda的目的是让程序员能够对程序行为进行抽象. 你可以这样想,它能够让程序员把一段程序代码当做数据一样使用.一个方法可以像定义和使用一个变量那样的方式被定 ...

  7. insert into select 与select into from -- sql 批量插入

    参考资料:http://www.w3school.com.cn/sql/sql_union.asp   UNION:操作符用于合并两个或多个select语句的结果集.                 ...

  8. 转载:C# this.invoke()作用 多线程操作UI 理解二

    Invoke()的作用是:在应用程序的主线程上执行指定的委托.一般应用:在辅助线程中修改UI线程( 主线程 )中对象的属性时,调用this.Invoke(); //测试的窗体 public class ...

  9. NodeJS优缺点及适用场景讨论

    概述:NodeJS宣称其目标是“旨在提供一种简单的构建可伸缩网络程序的方法”,那么它的出现是为了解决什么问题呢,它有什么优缺点以及它适用于什么场景呢? 本文就个人使用经验对这些问题进行探讨. 一. N ...

  10. Java Web include指令和动作的区别