分片是数据跨多台机器存储,MongoDB使用分片来支持具有非常大的数据集和高吞吐量操作的部署。

具有大型数据集或高吞吐量应用程序的数据库系统可能会挑战单个服务器的容量。例如,高查询率会耗尽服务器的CPU容量。工作集大小大于系统的RAM会强调磁盘驱动器的I / O容量。

有两种解决系统增长的方法:垂直和水平缩放。

垂直扩展涉及增加单个服务器的容量,例如使用更强大的CPU,添加更多RAM或增加存储空间量。可用技术的局限性可能会限制单个机器对于给定工作负载而言足够强大。此外,基于云的提供商基于可用的硬件配置具有硬性上限。结果,垂直缩放有实际的最大值。

水平扩展涉及划分系统数据集并加载多个服务器,添加其他服务器以根据需要增加容量。虽然单个机器的总体速度或容量可能不高,但每台机器处理整个工作负载的子集,可能提供比单个高速大容量服务器更高的效率。扩展部署容量只需要根据需要添加额外的服务器,这可能比单个机器的高端硬件的总体成本更低。权衡是基础架构和部署维护的复杂性增加。

MongoDB支持通过分片进行水平扩展

一、组件

  • shard:每个分片包含分片数据的子集。每个分片都可以部署为副本集(replica set)。可以分片,不分片的数据存于主分片服务器上。部署为3成员副本集
  • mongos:mongos充当查询路由器,提供客户端应用程序和分片集群之间的接口。可以部署多个mongos路由器。部署1个或者多个mongos
  • config servers:配置服务器存储群集的元数据和配置设置。从MongoDB 3.4开始,必须将配置服务器部署为3成员副本集

注意:应用程序或者客户端必须要连接mongos才能与集群的数据进行交互,永远不应连接到单个分片以执行读取或写入操作

shard的replica set的架构图:

config servers的replica set的架构图:

 分片策略

1、散列分片

  • 使用散列索引在共享群集中分区数据。散列索引计算单个字段的哈希值作为索引值; 此值用作分片键。
  • 使用散列索引解析查询时,MongoDB会自动计算哈希值。应用程序也不会需要计算哈希值。
  • 基于散列值的数据分布有助于更均匀的数据分布,尤其是在分片键单调变化的数据集中。

 2、范围分片

  • 基于分片键值将数据分成范围。然后根据分片键值为每个块分配一个范围。
  • mongos可以将操作仅路由到包含所需数据的分片。
  • 分片键的规划很重要,可能导致数据不能均匀分布。

 二、部署

1、环境说明

服务器名称 IP地址 操作系统版本 MongoDB版本 配置服务器(Config Server)端口 分片服务器1(Shard Server 1 分片服务器2(Shard Server 2) 分片服务器3(Shard Server 3) 功能
mongo1.example.net 10.10.18.10 Centos7.5 4.0 27027(Primary 27017(Primary 27018(Arbiter 27019(Secondary 配置服务器和分片服务器
mongo2.example.net 10.10.18.11 Centos7.5 4.0 27027(Secondary 27017(Secondary
27018(Primary 27019(Arbiter 配置服务器和分片服务器
mongo3.example.net 10.10.18.12 Centos7.5 4.0 27027(Secondary 27017(Arbiter 27018(Secondary 27019(Primary 配置服务器和分片服务器
mongos.example.net  192.168.11.10 Centos7.5 4.0   mongos的端口:27017     mongos

官方推荐配置中使用逻辑DNS,所以该文档中,将服务器名称和IP地址的DNS映射关系写入到各服务器的/etc/hosts文件中

2、部署MongoDB

环境中4台服务器的MongoDB的安装部署,详见:MongoDB安装

创建环境需要的目录:

mkdir -p /data/mongodb/data/{configServer,shard1,shard2,shard3}
mkdir -p /data/mongodb/{log,pid}

3、创建配置服务器(Config Server)的 Replica Set(副本集)

3台服务器上配置文件内容: /data/mongodb/configServer.conf

mongo1.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/configServer.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/configServer"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB:
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/configServer.pid"
net:
bindIp: mongo1.example.net
port:
replication:
replSetName: cs0
sharding:
clusterRole: configsvr

mongo2.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/configServer.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/configServer"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB:
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/configServer.pid"
net:
bindIp: mongo2.example.net
port:
replication:
replSetName: cs0
sharding:
clusterRole: configsvr

mongo3.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/configServer.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/configServer"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB:
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/configServer.pid"
net:
bindIp: mongo3.example.net
port:
replication:
replSetName: cs0
sharding:
clusterRole: configsvr

启动三台服务器Config Server

mongod -f /data/mongodb/configServer.conf

连接到其中一个Config Server

mongo --host mongo1.example.net --port 

结果:

 MongoDB shell version v4.0.10
connecting to: mongodb://mongo1.example.net:27027/?gssapiServiceName=mongodb
Implicit session: session { "id" : UUID("1a4d4252-11d0-40bb-90da-f144692be88d") }
MongoDB server version: 4.0.
Server has startup warnings:
--14T14::56.013+ I CONTROL [initandlisten]
--14T14::56.013+ I CONTROL [initandlisten] ** WARNING: Access control is not enabled for the database.
--14T14::56.013+ I CONTROL [initandlisten] ** Read and write access to data and configuration is unrestricted.
--14T14::56.013+ I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended.
--14T14::56.013+ I CONTROL [initandlisten]
--14T14::56.013+ I CONTROL [initandlisten]
--14T14::56.013+ I CONTROL [initandlisten] ** WARNING: /sys/kernel/mm/transparent_hugepage/enabled is 'always'.
--14T14::56.013+ I CONTROL [initandlisten] ** We suggest setting it to 'never'
--14T14::56.014+ I CONTROL [initandlisten]
--14T14::56.014+ I CONTROL [initandlisten] ** WARNING: /sys/kernel/mm/transparent_hugepage/defrag is 'always'.
--14T14::56.014+ I CONTROL [initandlisten] ** We suggest setting it to 'never'
--14T14::56.014+ I CONTROL [initandlisten]
>

配置Replica Set

rs.initiate(
{
_id: "cs0",
configsvr: true,
members: [
{ _id : , host : "mongo1.example.net:27027" },
{ _id : , host : "mongo2.example.net:27027" },
{ _id : , host : "mongo3.example.net:27027" }
]
}
)

结果:

 {
"ok" : ,
"operationTime" : Timestamp(, ),
"$gleStats" : {
"lastOpTime" : Timestamp(, ),
"electionId" : ObjectId("")
},
"lastCommittedOpTime" : Timestamp(, ),
"$clusterTime" : {
"clusterTime" : Timestamp(, ),
"signature" : {
"hash" : BinData(,"AAAAAAAAAAAAAAAAAAAAAAAAAAA="),
"keyId" : NumberLong()
}
}
}

查看Replica Set的状态

cs0:PRIMARY> rs.status()

结果:  可以看出三个服务器:1个Primary,2个Secondary

 {
"set" : "cs0",
"date" : ISODate("2019-06-14T06:33:31.348Z"),
"myState" : ,
"term" : NumberLong(),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"configsvr" : true,
"heartbeatIntervalMillis" : NumberLong(),
"optimes" : {
"lastCommittedOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"readConcernMajorityOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"appliedOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"durableOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
}
},
"lastStableCheckpointTimestamp" : Timestamp(, ),
"members" : [
{
"_id" : ,
"name" : "mongo1.example.net:27027",
"health" : ,
"state" : ,
"stateStr" : "PRIMARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-14T06:33:26Z"),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"infoMessage" : "could not find member to sync from",
"electionTime" : Timestamp(, ),
"electionDate" : ISODate("2019-06-14T06:31:59Z"),
"configVersion" : ,
"self" : true,
"lastHeartbeatMessage" : ""
},
{
"_id" : ,
"name" : "mongo2.example.net:27027",
"health" : ,
"state" : ,
"stateStr" : "SECONDARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDurable" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-14T06:33:26Z"),
"optimeDurableDate" : ISODate("2019-06-14T06:33:26Z"),
"lastHeartbeat" : ISODate("2019-06-14T06:33:29.385Z"),
"lastHeartbeatRecv" : ISODate("2019-06-14T06:33:29.988Z"),
"pingMs" : NumberLong(),
"lastHeartbeatMessage" : "",
"syncingTo" : "mongo1.example.net:27027",
"syncSourceHost" : "mongo1.example.net:27027",
"syncSourceId" : ,
"infoMessage" : "",
"configVersion" :
},
{
"_id" : ,
"name" : "mongo3.example.net:27027",
"health" : ,
"state" : ,
"stateStr" : "SECONDARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDurable" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-14T06:33:26Z"),
"optimeDurableDate" : ISODate("2019-06-14T06:33:26Z"),
"lastHeartbeat" : ISODate("2019-06-14T06:33:29.384Z"),
"lastHeartbeatRecv" : ISODate("2019-06-14T06:33:29.868Z"),
"pingMs" : NumberLong(),
"lastHeartbeatMessage" : "",
"syncingTo" : "mongo1.example.net:27027",
"syncSourceHost" : "mongo1.example.net:27027",
"syncSourceId" : ,
"infoMessage" : "",
"configVersion" :
}
],
"ok" : ,
"operationTime" : Timestamp(, ),
"$gleStats" : {
"lastOpTime" : Timestamp(, ),
"electionId" : ObjectId("7fffffff0000000000000001")
},
"lastCommittedOpTime" : Timestamp(, ),
"$clusterTime" : {
"clusterTime" : Timestamp(, ),
"signature" : {
"hash" : BinData(,"AAAAAAAAAAAAAAAAAAAAAAAAAAA="),
"keyId" : NumberLong()
}
}
}

创建管理用户

use admin
db.createUser(
{
user: "myUserAdmin",
pwd: "abc123",
roles: [{ role: "userAdminAnyDatabase", db: "admin" },"readWriteAnyDatabase"]
}
)

开启Config Server的登录验证和内部验证

使用Keyfiles进行内部认证,在其中一台服务器上创建Keyfiles

openssl rand -base64  > /data/mongodb/keyfile
chmod /data/mongodb/keyfile

将这个keyfile文件分发到其它的三台服务器上,并保证权限400

/data/mongodb/configServer.conf  配置文件中开启认证

security:
keyFile: "/data/mongodb/keyfile"
clusterAuthMode: "keyFile"
authorization: "enabled"

然后依次关闭2个Secondary,在关闭 Primary

mongod -f /data/mongodb/configServer.conf --shutdown

依次开启Primary和两个Secondary

mongod -f /data/mongodb/configServer.conf 

使用用户密码登录mongo

mongo --host mongo1.example.net --port  -u myUserAdmin --authenticationDatabase "admin" -p 'abc123'

注意:由于刚创建用户的时候没有给该用户管理集群的权限,所有此时登录后,能查看所有数据库,但是不能查看集群的状态信息。

 cs0:PRIMARY> rs.status()
{
"operationTime" : Timestamp(, ),
"ok" : ,
"errmsg" : "not authorized on admin to execute command { replSetGetStatus: 1.0, lsid: { id: UUID(\"59dd4dc0-b34f-43b9-a341-a2f43ec1dcfa\") }, $clusterTime: { clusterTime: Timestamp(1560495849, 1), signature: { hash: BinData(0, A51371EC5AA54BB1B05ED9342BFBF03CBD87F2D9), keyId: 6702270356301807629 } }, $db: \"admin\" }",
"code" : ,
"codeName" : "Unauthorized",
"$gleStats" : {
"lastOpTime" : Timestamp(, ),
"electionId" : ObjectId("7fffffff0000000000000002")
},
"lastCommittedOpTime" : Timestamp(, ),
"$clusterTime" : {
"clusterTime" : Timestamp(, ),
"signature" : {
"hash" : BinData(,"3UkTpXxyU8WI1TyS+u5vgewueGA="),
"keyId" : NumberLong("")
}
}
}
cs0:PRIMARY> show dbs
admin .000GB
config .000GB
local .000GB

赋值该用户具有集群的管理权限

use admin
db.system.users.find() #查看当前的用户信息
db.grantRolesToUser("myUserAdmin", ["clusterAdmin"])

查看集群信息

 cs0:PRIMARY> rs.status()
{
"set" : "cs0",
"date" : ISODate("2019-06-14T07:18:20.223Z"),
"myState" : ,
"term" : NumberLong(),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"configsvr" : true,
"heartbeatIntervalMillis" : NumberLong(),
"optimes" : {
"lastCommittedOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"readConcernMajorityOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"appliedOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"durableOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
}
},
"lastStableCheckpointTimestamp" : Timestamp(, ),
"members" : [
{
"_id" : ,
"name" : "mongo1.example.net:27027",
"health" : ,
"state" : ,
"stateStr" : "PRIMARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-14T07:18:10Z"),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"infoMessage" : "",
"electionTime" : Timestamp(, ),
"electionDate" : ISODate("2019-06-14T06:59:50Z"),
"configVersion" : ,
"self" : true,
"lastHeartbeatMessage" : ""
},
{
"_id" : ,
"name" : "mongo2.example.net:27027",
"health" : ,
"state" : ,
"stateStr" : "SECONDARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDurable" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-14T07:18:10Z"),
"optimeDurableDate" : ISODate("2019-06-14T07:18:10Z"),
"lastHeartbeat" : ISODate("2019-06-14T07:18:18.974Z"),
"lastHeartbeatRecv" : ISODate("2019-06-14T07:18:19.142Z"),
"pingMs" : NumberLong(),
"lastHeartbeatMessage" : "",
"syncingTo" : "mongo1.example.net:27027",
"syncSourceHost" : "mongo1.example.net:27027",
"syncSourceId" : ,
"infoMessage" : "",
"configVersion" :
},
{
"_id" : ,
"name" : "mongo3.example.net:27027",
"health" : ,
"state" : ,
"stateStr" : "SECONDARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDurable" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-14T07:18:10Z"),
"optimeDurableDate" : ISODate("2019-06-14T07:18:10Z"),
"lastHeartbeat" : ISODate("2019-06-14T07:18:18.999Z"),
"lastHeartbeatRecv" : ISODate("2019-06-14T07:18:18.998Z"),
"pingMs" : NumberLong(),
"lastHeartbeatMessage" : "",
"syncingTo" : "mongo2.example.net:27027",
"syncSourceHost" : "mongo2.example.net:27027",
"syncSourceId" : ,
"infoMessage" : "",
"configVersion" :
}
],
"ok" : ,
"operationTime" : Timestamp(, ),
"$gleStats" : {
"lastOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"electionId" : ObjectId("7fffffff0000000000000002")
},
"lastCommittedOpTime" : Timestamp(, ),
"$clusterTime" : {
"clusterTime" : Timestamp(, ),
"signature" : {
"hash" : BinData(,"lHiVw7WeO81npTi2IMW16reAN84="),
"keyId" : NumberLong("")
}
}
}

4、部署分片服务器1(Shard1)以及Replica Set(副本集)

3台服务器上配置文件内容: /data/mongodb/shard1.conf

mongo1.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/shard1.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/shard1"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB:
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/shard1.pid"
net:
bindIp: mongo1.example.net
port:
replication:
replSetName: "shard1"
sharding:
clusterRole: shardsvr

mongo2.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/shard1.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/shard1"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB:
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/shard1.pid"
net:
bindIp: mongo2.example.net
port:
replication:
replSetName: "shard1"
sharding:
clusterRole: shardsvr

mongo3.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/shard1.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/shard1"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB:
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/shard1.pid"
net:
bindIp: mongo3.example.net
port:
replication:
replSetName: "shard1"
sharding:
clusterRole: shardsvr

开启三台服务器上Shard

mongod -f /data/mongodb/shard1.conf

连接Primary服务器的Shard的副本集

mongo --host mongo1.example.net --port 

结果

 MongoDB shell version v4.0.10
connecting to: mongodb://mongo1.example.net:27017/?gssapiServiceName=mongodb
Implicit session: session { "id" : UUID("91e76384-cdae-411f-ab88-b7a8bd4555d1") }
MongoDB server version: 4.0.
Server has startup warnings:
--14T15::39.243+ I CONTROL [initandlisten]
--14T15::39.243+ I CONTROL [initandlisten] ** WARNING: Access control is not enabled for the database.
--14T15::39.243+ I CONTROL [initandlisten] ** Read and write access to data and configuration is unrestricted.
--14T15::39.243+ I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended.
--14T15::39.243+ I CONTROL [initandlisten]
--14T15::39.243+ I CONTROL [initandlisten]
--14T15::39.243+ I CONTROL [initandlisten] ** WARNING: /sys/kernel/mm/transparent_hugepage/enabled is 'always'.
--14T15::39.243+ I CONTROL [initandlisten] ** We suggest setting it to 'never'
--14T15::39.243+ I CONTROL [initandlisten]
--14T15::39.243+ I CONTROL [initandlisten] ** WARNING: /sys/kernel/mm/transparent_hugepage/defrag is 'always'.
--14T15::39.243+ I CONTROL [initandlisten] ** We suggest setting it to 'never'
--14T15::39.243+ I CONTROL [initandlisten]
>

配置Replica Set

rs.initiate(
{
_id : "shard1",
members: [
{ _id : , host : "mongo1.example.net:27017",priority: },
{ _id : , host : "mongo2.example.net:27017",priority: },
{ _id : , host : "mongo3.example.net:27017",arbiterOnly:true }
]
}
)

注意:优先级priority的值越大,越容易选举成为Primary

查看Replica Set的状态:

 shard1:PRIMARY> rs.status()
{
"set" : "shard1",
"date" : ISODate("2019-06-20T01:33:21.809Z"),
"myState" : ,
"term" : NumberLong(),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"heartbeatIntervalMillis" : NumberLong(),
"optimes" : {
"lastCommittedOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"readConcernMajorityOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"appliedOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"durableOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
}
},
"lastStableCheckpointTimestamp" : Timestamp(, ),
"members" : [
{
"_id" : ,
"name" : "mongo1.example.net:27017",
"health" : ,
"state" : ,
"stateStr" : "PRIMARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-20T01:33:13Z"),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"infoMessage" : "could not find member to sync from",
"electionTime" : Timestamp(, ),
"electionDate" : ISODate("2019-06-20T01:32:51Z"),
"configVersion" : ,
"self" : true,
"lastHeartbeatMessage" : ""
},
{
"_id" : ,
"name" : "mongo2.example.net:27017",
"health" : ,
"state" : ,
"stateStr" : "SECONDARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDurable" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-20T01:33:13Z"),
"optimeDurableDate" : ISODate("2019-06-20T01:33:13Z"),
"lastHeartbeat" : ISODate("2019-06-20T01:33:19.841Z"),
"lastHeartbeatRecv" : ISODate("2019-06-20T01:33:21.164Z"),
"pingMs" : NumberLong(),
"lastHeartbeatMessage" : "",
"syncingTo" : "mongo1.example.net:27017",
"syncSourceHost" : "mongo1.example.net:27017",
"syncSourceId" : ,
"infoMessage" : "",
"configVersion" :
},
{
"_id" : ,
"name" : "mongo3.example.net:27017",
"health" : ,
"state" : ,
"stateStr" : "ARBITER",
"uptime" : ,
"lastHeartbeat" : ISODate("2019-06-20T01:33:19.838Z"),
"lastHeartbeatRecv" : ISODate("2019-06-20T01:33:20.694Z"),
"pingMs" : NumberLong(),
"lastHeartbeatMessage" : "",
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"infoMessage" : "",
"configVersion" :
}
],
"ok" :
}

结果:  可以看出三个服务器:1个Primary,1个Secondary,1一个Arbiter

创建管理用户

use admin
db.createUser(
{
user: "myUserAdmin",
pwd: "abc123",
roles: [{ role: "userAdminAnyDatabase", db: "admin" },"readWriteAnyDatabase","clusterAdmin"]
}
)

开启Shard1的登录验证和内部验证

security:
keyFile: "/data/mongodb/keyfile"
clusterAuthMode: "keyFile"
authorization: "enabled"

然后依次关闭Arbiter、Secondary、Primary

mongod -f /data/mongodb/shard1.conf --shutdown

依次开启Primary和两个Secondary

mongod -f /data/mongodb/shard1.conf 

使用用户密码登录mongo

mongo --host mongo1.example.net --port  -u myUserAdmin --authenticationDatabase "admin" -p 'abc123'

5、部署分片服务器2(Shard2)以及Replica Set(副本集)

3台服务器上配置文件内容: /data/mongodb/shard2.conf

mongo1.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/shard2.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/shard2"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB:
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/shard2.pid"
net:
bindIp: mongo1.example.net
port:
replication:
replSetName: "shard2"
sharding:
clusterRole: shardsvr

mongo2.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/shard2.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/shard2"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB:
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/shard2.pid"
net:
bindIp: mongo2.example.net
port:
replication:
replSetName: "shard2"
sharding:
clusterRole: shardsvr

mongo3.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/shard2.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/shard2"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB:
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/shard2.pid"
net:
bindIp: mongo3.example.net
port:
replication:
replSetName: "shard2"
sharding:
clusterRole: shardsvr

开启三台服务器上Shard

mongod -f /data/mongodb/shard2.conf

连接Primary服务器的Shard的副本集

mongo --host mongo2.example.net --port 27018

配置Replica Set(注意:三个服务器的角色发生了改变)

rs.initiate(
{
_id : "shard2",
members: [
{ _id : , host : "mongo1.example.net:27018",arbiterOnly:true },
{ _id : , host : "mongo2.example.net:27018",priority: },
{ _id : , host : "mongo3.example.net:27018",priority: }
]
}
)

查看Replica Set的状态:

 shard2:PRIMARY> rs.status()
{
"set" : "shard2",
"date" : ISODate("2019-06-20T01:59:08.996Z"),
"myState" : ,
"term" : NumberLong(),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"heartbeatIntervalMillis" : NumberLong(),
"optimes" : {
"lastCommittedOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"readConcernMajorityOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"appliedOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"durableOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
}
},
"lastStableCheckpointTimestamp" : Timestamp(, ),
"members" : [
{
"_id" : ,
"name" : "mongo1.example.net:27018",
"health" : ,
"state" : ,
"stateStr" : "ARBITER",
"uptime" : ,
"lastHeartbeat" : ISODate("2019-06-20T01:59:08.221Z"),
"lastHeartbeatRecv" : ISODate("2019-06-20T01:59:07.496Z"),
"pingMs" : NumberLong(),
"lastHeartbeatMessage" : "",
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"infoMessage" : "",
"configVersion" :
},
{
"_id" : ,
"name" : "mongo2.example.net:27018",
"health" : ,
"state" : ,
"stateStr" : "PRIMARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-20T01:59:03Z"),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"infoMessage" : "could not find member to sync from",
"electionTime" : Timestamp(, ),
"electionDate" : ISODate("2019-06-20T01:57:32Z"),
"configVersion" : ,
"self" : true,
"lastHeartbeatMessage" : ""
},
{
"_id" : ,
"name" : "mongo3.example.net:27018",
"health" : ,
"state" : ,
"stateStr" : "SECONDARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDurable" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-20T01:59:03Z"),
"optimeDurableDate" : ISODate("2019-06-20T01:59:03Z"),
"lastHeartbeat" : ISODate("2019-06-20T01:59:08.220Z"),
"lastHeartbeatRecv" : ISODate("2019-06-20T01:59:08.716Z"),
"pingMs" : NumberLong(),
"lastHeartbeatMessage" : "",
"syncingTo" : "mongo2.example.net:27018",
"syncSourceHost" : "mongo2.example.net:27018",
"syncSourceId" : ,
"infoMessage" : "",
"configVersion" :
}
],
"ok" : ,
"operationTime" : Timestamp(, ),
"$clusterTime" : {
"clusterTime" : Timestamp(, ),
"signature" : {
"hash" : BinData(,"AAAAAAAAAAAAAAAAAAAAAAAAAAA="),
"keyId" : NumberLong()
}
}
}

结果:  可以看出三个服务器:1个Primary,1个Secondary,1一个Arbiter

配置登录认证的用户请按照 Shard1 的步骤

6、部署分片服务器3(Shard3)以及Replica Set(副本集)

3台服务器上配置文件内容: /data/mongodb/shard3.conf

mongo1.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/shard3.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/shard3"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB: 2
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/shard3.pid"
net:
bindIp: mongo1.example.net
port: 27019
replication:
replSetName: "shard3"
sharding:
clusterRole: shardsvr

mongo2.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/shard3.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/shard3"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB: 2
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/shard3.pid"
net:
bindIp: mongo2.example.net
port: 27019
replication:
replSetName: "shard3"
sharding:
clusterRole: shardsvr

mongo3.example.net服务器上

systemLog:
destination: file
path: "/data/mongodb/log/shard3.log"
logAppend: true
storage:
dbPath: "/data/mongodb/data/shard3"
journal:
enabled: true
wiredTiger:
engineConfig:
cacheSizeGB: 2
processManagement:
fork: true
pidFilePath: "/data/mongodb/pid/shard3.pid"
net:
bindIp: mongo3.example.net
port: 27019
replication:
replSetName: "shard3"
sharding:
clusterRole: shardsvr

开启三台服务器上Shard

mongod -f /data/mongodb/shard3.conf

连接Primary服务器的Shard的副本集

mongo --host mongo3.example.net --port 27019

配置Replica Set(注意:三个服务器的角色发生了改变)

rs.initiate(
{
_id : "shard3",
members: [
{ _id : , host : "mongo1.example.net:27019",priority: },
{ _id : , host : "mongo2.example.net:27019",arbiterOnly:true },
{ _id : , host : "mongo3.example.net:27019",priority:2 }
]
}
)

查看Replica Set的状态:

 shard3:PRIMARY> rs.status()
{
"set" : "shard3",
"date" : ISODate("2019-06-20T02:21:56.990Z"),
"myState" : ,
"term" : NumberLong(),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"heartbeatIntervalMillis" : NumberLong(),
"optimes" : {
"lastCommittedOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"readConcernMajorityOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"appliedOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"durableOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
}
},
"lastStableCheckpointTimestamp" : Timestamp(, ),
"members" : [
{
"_id" : ,
"name" : "mongo1.example.net:27019",
"health" : ,
"state" : ,
"stateStr" : "SECONDARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDurable" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-20T02:21:52Z"),
"optimeDurableDate" : ISODate("2019-06-20T02:21:52Z"),
"lastHeartbeat" : ISODate("2019-06-20T02:21:56.160Z"),
"lastHeartbeatRecv" : ISODate("2019-06-20T02:21:55.155Z"),
"pingMs" : NumberLong(),
"lastHeartbeatMessage" : "",
"syncingTo" : "mongo3.example.net:27019",
"syncSourceHost" : "mongo3.example.net:27019",
"syncSourceId" : ,
"infoMessage" : "",
"configVersion" :
},
{
"_id" : ,
"name" : "mongo2.example.net:27019",
"health" : ,
"state" : ,
"stateStr" : "ARBITER",
"uptime" : ,
"lastHeartbeat" : ISODate("2019-06-20T02:21:56.159Z"),
"lastHeartbeatRecv" : ISODate("2019-06-20T02:21:55.021Z"),
"pingMs" : NumberLong(),
"lastHeartbeatMessage" : "",
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"infoMessage" : "",
"configVersion" :
},
{
"_id" : ,
"name" : "mongo3.example.net:27019",
"health" : ,
"state" : ,
"stateStr" : "PRIMARY",
"uptime" : ,
"optime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"optimeDate" : ISODate("2019-06-20T02:21:52Z"),
"syncingTo" : "",
"syncSourceHost" : "",
"syncSourceId" : -,
"infoMessage" : "could not find member to sync from",
"electionTime" : Timestamp(, ),
"electionDate" : ISODate("2019-06-20T02:21:50Z"),
"configVersion" : ,
"self" : true,
"lastHeartbeatMessage" : ""
}
],
"ok" : ,
"operationTime" : Timestamp(, ),
"$clusterTime" : {
"clusterTime" : Timestamp(, ),
"signature" : {
"hash" : BinData(,"AAAAAAAAAAAAAAAAAAAAAAAAAAA="),
"keyId" : NumberLong()
}
}
}

结果:  可以看出三个服务器:1个Primary,1个Secondary,1一个Arbiter

配置登录认证的用户请按照 Shard1 的步骤

7、配置mongos服务器去连接分片集群

mongos.example.net 服务器上mongos的配置文件 /data/mongodb/mongos.conf

systemLog:
destination: file
path: "/data/mongodb/log/mongos.log"
logAppend: true
processManagement:
fork: true
net:
port:
bindIp: mongos.example.net
sharding:
configDB: "cs0/mongo1.example.net:27027,mongo2.example.net:27027,mongo3.example.net:27027"
security:
keyFile: "/data/mongodb/keyfile"
clusterAuthMode: "keyFile"

启动mongos服务

mongos -f /data/mongodb/mongos.conf

连接mongos

mongo --host mongos.example.net --port 27017 -u myUserAdmin --authenticationDatabase "admin" -p 'abc123'

查看当前集群结果:

mongos> sh.status()
--- Sharding Status ---
sharding version: {
"_id" : ,
"minCompatibleVersion" : ,
"currentVersion" : ,
"clusterId" : ObjectId("5d0af6ed4fa51757cd032108")
}
shards:
active mongoses:
autosplit:
Currently enabled: yes
balancer:
Currently enabled: yes
Currently running: no
Failed balancer rounds in last attempts:
Migration Results for the last hours:
No recent migrations
databases:
{ "_id" : "config", "primary" : "config", "partitioned" : true }

在集群中先加入Shard1、Shard2,剩余Shard3我们在插入数据有在进行加入(模拟实现扩容)。

sh.addShard("shard1/mongo1.example.net:27017,mongo2.example.net:27017,mongo3.example.net:27017")
sh.addShard("shard2/mongo1.example.net:27018,mongo2.example.net:27018,mongo3.example.net:27018")

结果:

 mongos> sh.addShard("shard1/mongo1.example.net:27017,mongo2.example.net:27017,mongo3.example.net:27017")
{
"shardAdded" : "shard1",
"ok" : ,
"operationTime" : Timestamp(, ),
"$clusterTime" : {
"clusterTime" : Timestamp(, ),
"signature" : {
"hash" : BinData(,"2je9FsNfMfBMHp+X/6d98B5tLH8="),
"keyId" : NumberLong("")
}
}
}
mongos> sh.addShard("shard2/mongo1.example.net:27018,mongo2.example.net:27018,mongo3.example.net:27018")
{
"shardAdded" : "shard2",
"ok" : ,
"operationTime" : Timestamp(, ),
"$clusterTime" : {
"clusterTime" : Timestamp(, ),
"signature" : {
"hash" : BinData(,"8FvJuCy8kCrMu5nB9PYILj0bzLk="),
"keyId" : NumberLong("")
}
}
}

查看集群的状态

mongos> sh.status()
--- Sharding Status ---
sharding version: {
"_id" : ,
"minCompatibleVersion" : ,
"currentVersion" : ,
"clusterId" : ObjectId("5d0af6ed4fa51757cd032108")
}
shards:
{ "_id" : "shard1", "host" : "shard1/mongo1.example.net:27017,mongo2.example.net:27017", "state" : }
{ "_id" : "shard2", "host" : "shard2/mongo2.example.net:27018,mongo3.example.net:27018", "state" : }
active mongoses:
"4.0.10" :
autosplit:
Currently enabled: yes
balancer:
Currently enabled: yes
Currently running: no
Failed balancer rounds in last attempts:
Migration Results for the last hours:
No recent migrations
databases:
{ "_id" : "config", "primary" : "config", "partitioned" : true }

8、测试

为了便于测试,设置分片chunk的大小为1M

use config
db.settings.save({"_id":"chunksize","value":1})

在连接mongos后,执行创建数据库,并启用分片存储

sh.enableSharding("user_center")

创建 "user_center"数据库,并启用分片,查看结果:

 mongos> sh.status()
--- Sharding Status ---
sharding version: {
"_id" : ,
"minCompatibleVersion" : ,
"currentVersion" : ,
"clusterId" : ObjectId("5d0af6ed4fa51757cd032108")
}
shards:
{ "_id" : "shard1", "host" : "shard1/mongo1.example.net:27017,mongo2.example.net:27017", "state" : }
{ "_id" : "shard2", "host" : "shard2/mongo2.example.net:27018,mongo3.example.net:27018", "state" : }
active mongoses:
"4.0.10" :
autosplit:
Currently enabled: yes
balancer:
Currently enabled: yes
Currently running: no
Failed balancer rounds in last attempts:
Migration Results for the last hours:
No recent migrations
databases:
{ "_id" : "config", "primary" : "config", "partitioned" : true }
config.system.sessions
shard key: { "_id" : }
unique: false
balancing: true
chunks:
shard1
{ "_id" : { "$minKey" : } } -->> { "_id" : { "$maxKey" : } } on : shard1 Timestamp(, )
{ "_id" : "user_center", "primary" : "shard1", "partitioned" : true, "version" : { "uuid" : UUID("3b05ccb5-796a-4e9e-a36e-99b860b6bee0"), "lastMod" : } }

创建 "users" 集合

sh.shardCollection("user_center.users",{"name":})   #数据库user_center中users集合使用了片键{"name":1},这个片键通过字段name的值进行数据分配

现在查看集群状态

 mongos> sh.status()
--- Sharding Status ---
sharding version: {
"_id" : 1,
"minCompatibleVersion" : 5,
"currentVersion" : 6,
"clusterId" : ObjectId("5d0af6ed4fa51757cd032108")
}
shards:
{ "_id" : "shard1", "host" : "shard1/mongo1.example.net:27017,mongo2.example.net:27017", "state" : 1 }
{ "_id" : "shard2", "host" : "shard2/mongo2.example.net:27018,mongo3.example.net:27018", "state" : 1 }
active mongoses:
"4.0.10" : 1
autosplit:
Currently enabled: yes
balancer:
Currently enabled: yes
Currently running: no
Failed balancer rounds in last 5 attempts: 0
Migration Results for the last 24 hours:
No recent migrations
databases:
{ "_id" : "config", "primary" : "config", "partitioned" : true }
config.system.sessions
shard key: { "_id" : 1 }
unique: false
balancing: true
chunks:
shard1 1
{ "_id" : { "$minKey" : 1 } } -->> { "_id" : { "$maxKey" : 1 } } on : shard1 Timestamp(1, 0)
{ "_id" : "user_center", "primary" : "shard2", "partitioned" : true, "version" : { "uuid" : UUID("33c79b3f-aa18-4755-a5e8-b8f7f3d05893"), "lastMod" : 1 } }
user_center.users
shard key: { "name" : 1 }
unique: false
balancing: true
chunks:
shard2 1
{ "name" : { "$minKey" : 1 } } -->> { "name" : { "$maxKey" : 1 } } on : shard2 Timestamp(1, 0)

写pyhton脚本插入数据

#enconding:utf8
import pymongo,string,random def random_name():
str_args = string.ascii_letters
name_list = random.sample(str_args,5)
random.shuffle(name_list)
return ''.join(name_list) def random_age():
age_args = string.digits
age_list = random.sample(age_args,2)
random.shuffle(age_list)
return int(''.join(age_list))
def insert_data_to_mongo(url,dbname,collections_name):
print(url)
client = pymongo.MongoClient(url)
db = client[dbname]
collections = db[collections_name]
for i in range(1,100000):
name = random_name()
collections.insert({"name" : name , "age" : random_age(), "status" : "pending"})
print("insert ",name) if __name__ == "__main__":
mongo_url="mongodb://myUserAdmin:abc123@192.168.11.10:27017/?maxPoolSize=100&minPoolSize=10&maxIdleTimeMS=600000"
mongo_db="user_center"
mongo_collections="users"
insert_data_to_mongo(mongo_url,mongo_db,mongo_collections)

插入数据后查看此时集群的状态:

mongos> sh.status()
--- Sharding Status ---
sharding version: {
"_id" : ,
"minCompatibleVersion" : ,
"currentVersion" : ,
"clusterId" : ObjectId("5d0af6ed4fa51757cd032108")
}
shards:
{ "_id" : "shard1", "host" : "shard1/mongo1.example.net:27017,mongo2.example.net:27017", "state" : }
{ "_id" : "shard2", "host" : "shard2/mongo2.example.net:27018,mongo3.example.net:27018", "state" : }
active mongoses:
"4.0.10" :
autosplit:
Currently enabled: yes
balancer:
Currently enabled: yes
Currently running: no
Failed balancer rounds in last attempts:
Migration Results for the last hours:
: Success
databases:
{ "_id" : "config", "primary" : "config", "partitioned" : true }
config.system.sessions
shard key: { "_id" : }
unique: false
balancing: true
chunks:
shard1
{ "_id" : { "$minKey" : } } -->> { "_id" : { "$maxKey" : } } on : shard1 Timestamp(, )
{ "_id" : "user_center", "primary" : "shard2", "partitioned" : true, "version" : { "uuid" : UUID("33c79b3f-aa18-4755-a5e8-b8f7f3d05893"), "lastMod" : } }
user_center.users
shard key: { "name" : }
unique: false
balancing: true
chunks:
shard1
shard2
{ "name" : { "$minKey" : } } -->> { "name" : "ABXEw" } on : shard1 Timestamp(, )
{ "name" : "ABXEw" } -->> { "name" : "EKdCt" } on : shard1 Timestamp(, )
{ "name" : "EKdCt" } -->> { "name" : "ITgcx" } on : shard1 Timestamp(, )
{ "name" : "ITgcx" } -->> { "name" : "JKoOz" } on : shard1 Timestamp(, )
{ "name" : "JKoOz" } -->> { "name" : "NSlcY" } on : shard1 Timestamp(, )
{ "name" : "NSlcY" } -->> { "name" : "RbrAy" } on : shard1 Timestamp(, )
{ "name" : "RbrAy" } -->> { "name" : "SQvZq" } on : shard1 Timestamp(, )
{ "name" : "SQvZq" } -->> { "name" : "TxpPM" } on : shard1 Timestamp(, )
{ "name" : "TxpPM" } -->> { "name" : "YEujn" } on : shard1 Timestamp(, )
{ "name" : "YEujn" } -->> { "name" : "cOlra" } on : shard2 Timestamp(, )
{ "name" : "cOlra" } -->> { "name" : "dFTNS" } on : shard2 Timestamp(, )
{ "name" : "dFTNS" } -->> { "name" : "hLwFZ" } on : shard2 Timestamp(, )
{ "name" : "hLwFZ" } -->> { "name" : "lVQzu" } on : shard2 Timestamp(, )
{ "name" : "lVQzu" } -->> { "name" : "mNLGP" } on : shard2 Timestamp(, )
{ "name" : "mNLGP" } -->> { "name" : "oILav" } on : shard2 Timestamp(, )
{ "name" : "oILav" } -->> { "name" : "wJWQI" } on : shard2 Timestamp(, )
{ "name" : "wJWQI" } -->> { "name" : { "$maxKey" : } } on : shard2 Timestamp(, )

可以看出,数据分别再Shard1、Shard2分片上。

将Shard3分片也加入到集群中来

mongos> sh.addShard("shard3/mongo1.example.net:27019,mongo2.example.net:27019,mongo3.example.net:27019")

在查看集群的状态:

mongos> sh.status()
--- Sharding Status ---
sharding version: {
"_id" : ,
"minCompatibleVersion" : ,
"currentVersion" : ,
"clusterId" : ObjectId("5d0af6ed4fa51757cd032108")
}
shards:
{ "_id" : "shard1", "host" : "shard1/mongo1.example.net:27017,mongo2.example.net:27017", "state" : }
{ "_id" : "shard2", "host" : "shard2/mongo2.example.net:27018,mongo3.example.net:27018", "state" : }
{ "_id" : "shard3", "host" : "shard3/mongo1.example.net:27019,mongo3.example.net:27019", "state" : }
active mongoses:
"4.0.10" :
autosplit:
Currently enabled: yes
balancer:
Currently enabled: yes
Currently running: no
Failed balancer rounds in last attempts:
Migration Results for the last hours:
: Success
databases:
{ "_id" : "config", "primary" : "config", "partitioned" : true }
config.system.sessions
shard key: { "_id" : }
unique: false
balancing: true
chunks:
shard1
{ "_id" : { "$minKey" : } } -->> { "_id" : { "$maxKey" : } } on : shard1 Timestamp(, )
{ "_id" : "user_center", "primary" : "shard2", "partitioned" : true, "version" : { "uuid" : UUID("33c79b3f-aa18-4755-a5e8-b8f7f3d05893"), "lastMod" : } }
user_center.users
shard key: { "name" : }
unique: false
balancing: true
chunks:
shard1
shard2
shard3
{ "name" : { "$minKey" : } } -->> { "name" : "ABXEw" } on : shard3 Timestamp(, )
{ "name" : "ABXEw" } -->> { "name" : "EKdCt" } on : shard3 Timestamp(, )
{ "name" : "EKdCt" } -->> { "name" : "ITgcx" } on : shard3 Timestamp(, )
{ "name" : "ITgcx" } -->> { "name" : "JKoOz" } on : shard1 Timestamp(, )
{ "name" : "JKoOz" } -->> { "name" : "NSlcY" } on : shard1 Timestamp(, )
{ "name" : "NSlcY" } -->> { "name" : "RbrAy" } on : shard1 Timestamp(, )
{ "name" : "RbrAy" } -->> { "name" : "SQvZq" } on : shard1 Timestamp(, )
{ "name" : "SQvZq" } -->> { "name" : "TxpPM" } on : shard1 Timestamp(, )
{ "name" : "TxpPM" } -->> { "name" : "YEujn" } on : shard1 Timestamp(, )
{ "name" : "YEujn" } -->> { "name" : "cOlra" } on : shard3 Timestamp(, )
{ "name" : "cOlra" } -->> { "name" : "dFTNS" } on : shard3 Timestamp(, )
{ "name" : "dFTNS" } -->> { "name" : "hLwFZ" } on : shard2 Timestamp(, )
{ "name" : "hLwFZ" } -->> { "name" : "lVQzu" } on : shard2 Timestamp(, )
{ "name" : "lVQzu" } -->> { "name" : "mNLGP" } on : shard2 Timestamp(, )
{ "name" : "mNLGP" } -->> { "name" : "oILav" } on : shard2 Timestamp(, )
{ "name" : "oILav" } -->> { "name" : "wJWQI" } on : shard2 Timestamp(, )
{ "name" : "wJWQI" } -->> { "name" : { "$maxKey" : } } on : shard2 Timestamp(, )

加入后,集群的分片数据重新平衡调整,有一部分数据分布到Shard3上。

 9、备份和恢复

 备份

备份的时候需要锁定配置服务器(ConfigServer)和分片服务器(Shard)

在备份前查看当前数据库中数据总条数

mongos> db.users.find().count()

然后启动前面的python脚本,可以在脚本中添加time.sleep来控制插入的频率。

在mongos服务器上停止平衡器。

mongos> sh.stopBalancer()

锁定配置服务器和各分片服务器,登录配置服务器和各分片服务器的Secondary执行命令

db.fsyncLock()

开始备份数据库

mongodump  -h mongo2.example.net --port  --authenticationDatabase admin -u myUserAdmin -p abc123 -o /data/backup/config
mongodump -h mongo2.example.net --port --authenticationDatabase admin -u myUserAdmin -p abc123 -o /data/backup/shard1
mongodump -h mongo3.example.net --port --authenticationDatabase admin -u myUserAdmin -p abc123 -o /data/backup/shard2
mongodump -h mongo1.example.net --port --authenticationDatabase admin -u myUserAdmin -p abc123 -o /data/backup/shard3

锁定配置服务器和各分片服务器

db.fsyncUnlock()

在mongos中开启平衡器

sh.setBalancerState(true);

在备份的过程中不会影响到数据的写入,备份后查看此时的数据

mongos> db.users.find().count()
107874

恢复

将Shard1分片服务器1中的数据库删除

shard1:PRIMARY> use user_center
switched to db user_center
shard1:PRIMARY> db.dropDatabase()
{
"dropped" : "user_center",
"ok" : ,
"operationTime" : Timestamp(, ),
"$gleStats" : {
"lastOpTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
},
"electionId" : ObjectId("7fffffff0000000000000002")
},
"lastCommittedOpTime" : Timestamp(, ),
"$configServerState" : {
"opTime" : {
"ts" : Timestamp(, ),
"t" : NumberLong()
}
},
"$clusterTime" : {
"clusterTime" : Timestamp(, ),
"signature" : {
"hash" : BinData(,"GO1yQDvdZ6oJBXdvM94noPNnJTM="),
"keyId" : NumberLong("")
}
}
}

然后使用刚备份的数据库进行恢复

mongorestore  -h mongo1.example.net --port  --authenticationDatabase admin -u myUserAdmin -p abc123 -d user_center /data/backup/shard1/user_center

2019-06-20T17:20:34.325+0800 the --db and --collection args should only be used when restoring from a BSON file. Other uses are deprecated and will not exist in the future; use --nsInclude instead
2019-06-20T17:20:34.326+0800 building a list of collections to restore from /data/backup/shard1/user_center dir
2019-06-20T17:20:34.356+0800 reading metadata for user_center.users from /data/backup/shard1/user_center/users.metadata.json
2019-06-20T17:20:34.410+0800 restoring user_center.users from /data/backup/shard1/user_center/users.bson
2019-06-20T17:20:36.836+0800 restoring indexes for collection user_center.users from metadata
2019-06-20T17:20:37.093+0800 finished restoring user_center.users (30273 documents)
2019-06-20T17:20:37.093+0800 done

根据上述步骤恢复Shard2、Shard3的数据

最后恢复的结果:

mongos> db.users.find().count()

这个应该是我在锁的时候插入的数据。

MongoDB的集群模式--Sharding(分片)的更多相关文章

  1. mongodb数据库集群及sharding分片配置

    复制集群的配置 1.安装mongodb数据库 在主节点和从节点上都安装mongodb # rpm -ivh mongo-10gen-2.4.6-mongodb_1.x86_64.rpm mongo-1 ...

  2. mongodb之集群模式

    前言 数据量大了或者并发量上来了,单机肯定是抗不住的,这个时候要开始考虑使用集群了.mongodb目前为止支持三种集群模式:主从集群,副本集集群,分片集群. 主从集群 特性 1. 一主多从2. 主负责 ...

  3. MongoDB分布式集群搭建(分片加副本集)

    # 环境准备 服务器 # 环境搭建 文件配置和目录添加 新建目录的操作要在三台机器中进行,为配置服务器新建数据目录和日志目录 mkdir -p $MONGODB_HOME/config/data mk ...

  4. MongoDB的集群模式--Replica Set

    一.Replica Set 集群分为两种架构: 奇数个节点构成Replica Set,所有节点拥有数据集.最小架构: 1个Primary节点,2个Secondary节点 偶数个节点 + 一个仲裁节点 ...

  5. MongoDB之分片集群(Sharding)

    MongoDB之分片集群(Sharding) 一.基本概念 分片(sharding)是一个通过多台机器分配数据的方法.MongoDB使用分片支持大数据集和高吞吐量的操作.大数据集和高吞吐量的数据库系统 ...

  6. MongoDB(7):集群部署实践,包含复制集,分片

    注: 刚开始学习MongoDB,写的有点麻烦了,网上教程都是很少的代码就完成了集群的部署, 纯属个人实践,错误之处望指正!有好的建议和资料请联系我QQ:1176479642 集群架构: 2mongos ...

  7. mongodb的分布式集群(4、分片和副本集的结合)

    概述 前面3篇博客讲了mongodb的分布式和集群,当中第一种的主从复制我们差点儿不用,没有什么意义,剩下的两种,我们不论单独的使用哪一个.都会出现对应的问题.比較好的一种解决方式就是.分片和副本集的 ...

  8. mongodb 搭建集群(分片+副本集)

    mongodb  搭建集群(分片+副本集) 一.搭建结构图: 二.搭建步骤:

  9. mongodb分布式集群搭建手记

    一.架构简介 目标单机搭建mongodb分布式集群(副本集 + 分片集群),演示mongodb分布式集群的安装部署.简单操作. 说明在同一个vm启动由两个分片组成的分布式集群,每个分片都是一个PSS( ...

随机推荐

  1. [考试]NOIP2015模拟题2

    // 此博文为迁移而来,写于2015年7月22日,不代表本人现在的观点与看法.原始地址:http://blog.sina.com.cn/s/blog_6022c4720102w72i.html 1.总 ...

  2. 第03组 Beta冲刺(1/4)

    队名:不等式方程组 组长博客 作业博客 团队项目进度 组员一:张逸杰(组长) 过去两天完成的任务: 文字/口头描述: 制定了初步的项目计划,并开始学习一些推荐.搜索类算法 GitHub签入纪录: 暂无 ...

  3. vue bootstrap中modal对话框不显示遮挡打不开

    使用Vue bootstrap时,点击modal却不能弹出来,被隐藏遮挡无法显示,参考下面的这个博客的说明解决了这个问题: Heap Stack Blog(pingbook.top)Vue boots ...

  4. 第01组 Alpha事后诸葛亮

    目录 一.总结思考 1.设想和目标 ①我们的软件要解决什么问题?是否定义得很清楚?是否对典型用户和典型场景有清晰的描述? ②我们达到目标了么(原计划的功能做到了几个? 按照原计划交付时间交付了么? 原 ...

  5. nmap简单使用方法

    1.作用扫描整个网络的主机服务状态和存活优点,快速,准确,效率高2.nmap 选项 Usage: nmap [Scan Type(s)] [Options] {target specification ...

  6. mysql 数据库中的每张表加同一个字段(避免重复加)

    DROP PROCEDURE IF EXISTS testEndHandle; DELIMITER $$ CREATE PROCEDURE testEndHandle() BEGIN DECLARE ...

  7. windows下Apache2.4配置php的解析运行

    1.安装php7 2.安装apache2.4 3.找到apache2.4的httpd.conf文件,配置好apache,然后通过httpd.exe -k install -n "servic ...

  8. 【操作系统之十二】分支预测、CPU亲和性(affinity)

    一.分支预测 当包含流水线技术的处理器处理分支指令时就会遇到一个问题,根据判定条件的真/假的不同,有可能会产生转跳,而这会打断流水线中指令的处理,因为处理器无法确定该指令的下一条指令,直到分支执行完毕 ...

  9. ArcGIS Server浏览地图服务无响应原因分析说明

    1.问题描述 从4月17号下午5时起,至18号晚9点,客户单位部分通过ArcGIS Server发布的地图服务(该部分地图服务的数据源为数据库SJZX)无法加载浏览,表现为长时间无响应.同时,通过Ar ...

  10. Springboot Actuator之十一:actuator PublicMetrics

    前言接下来的几篇文章我们来分析一下spring-boot-actuator 中在org.springframework.boot.actuate.metrics中的代码,如图: 这里的代码不仅多,而且 ...