zoukankan      html  css  js  c++  java
  • mongodb 分片+复制集

    mongodb版本:3.0.4

    准备:

    1、校验服务器的时间,3台同步

    2、关闭防火墙

    3、关闭selinux

    Sharding+Replica Sets 
    主机              IP                  服务及端口
    Server A      192.168.31.231    mongod shard1_1:27017
                                    mongod shard2_1:27018
                                    mongod config1:20000
                                    mongs1:30000
                                    
    Server B      192.168.31.232    mongod shard1_2:27017
                                    mongod shard2_2:27018
                                    mongod config2:20000
                                    mongs2:30000
                                    
    Server C      192.168.31.233    mongod shard1_3:27017
                                    mongod shard2_3:27018
                                    mongod config3:20000
                                    mongs3:30000> rs.initiate(config)
                                
    1.1 创建数据目录
    Server A
    mkdir /data/shard1_1
    mkdir /data/shard2_1
    mkdir /data/config 
    Server B
    mkdir /data/shard1_2
    mkdir /data/shard2_2
    mkdir /data/config
    Server C
    mkdir /data/shard1_3
    mkdir /data/shard2_3
    mkdir /data/config
    
    1.2 配置shard1的Replica Sets
    Server A 
    [root@localhost bin]# /usr/local/mongo/bin/mongod --shardsvr --replSet shard1 --port 27017 --dbpath /data/shard1_1 --logpath /data/shard1_1/shard1_1.log --logappend --fork
    
    
    Server B
    [root@localhost bin]# /usr/local/mongo/bin/mongod --shardsvr --replSet shard1 --port 27017 --dbpath /data/shard1_2 --logpath /data/shard1_2/shard1_2.log --logappend --fork
    
    Server C
    [root@localhost bin]# /usr/local/mongo/bin/mongod --shardsvr --replSet shard1 --port 27017 --dbpath /data/shard1_3 --logpath /data/shard1_3/shard1_3.log --logappend --fork
    
    用 mongo 连接其中一台机器的 27017 端口的 mongod,初始化 Replica Sets“shard1” ,执行:
    [root@localhost bin]# ./mongo --port 27017
    MongoDB shell version: 1.8.1
    connecting to: 127.0.0.1:27017/test
    > config = {_id: 'shard1', members: [
    ... {_id: 0, host: '192.168.31.231:27017'},
    ... {_id: 1, host: '192.168.31.232:27017'},
    ... {_id: 2, host: '192.168.31.233:27017',arbiterOnly:true}]
    ... }
    {
            "_id" : "shard1",
            "members" : [
                    {
                            "_id" : 0,
                            "host" : "192.168.31.231:27017"
                    },
                    {
                            "_id" : 1,
                            "host" : "192.168.31.232:27017"
                    },
                    {
                            "_id" : 2,
                            "host" : "192.168.31.233:27017",
                            "arbiterOnly":true
                    }
            ]
    }
    > rs.initiate(config)
    {
    "info" : "Config now saved locally. Should come online in about a minute.",
    "ok" : 1
    }
    
    1.3配置 shard2  所用到的 Replica Sets
    Server A 
    [root@localhost bin]# /usr/local/mongo/bin/mongod --shardsvr --replSet shard2 --port 27018 --dbpath /data/shard2_1 --logpath /data/shard2_1/shard2_1.log --logappend --fork
    
    Server B
    [root@localhost bin]# /usr/local/mongo/bin/mongod --shardsvr --replSet shard2 --port 27018 --dbpath /data/shard2_2 --logpath /data/shard2_2/shard2_2.log --logappend --fork
    
    Server C
    [root@localhost bin]# /usr/local/mongo/bin/mongod --shardsvr --replSet shard2 --port 27018 --dbpath /data/shard2_3 --logpath /data/shard2_3/shard2_3.log --logappend --fork
    
    用 mongo 连接其中一台机器的 27018 端口的 mongod,初始化 Replica Sets “shard2” ,执行:
    [root@localhost bin]# ./mongo --port 27018
    MongoDB shell version: 1.8.1
    connecting to: 127.0.0.1:27018/test
    > config = {_id: 'shard2', members: [
    ... {_id: 0, host: '192.168.31.231:27018'},
    ... {_id: 1, host: '192.168.31.232:27018'},
    ... {_id: 2, host: '192.168.31.233:27018',arbiterOnly:true}]
    ... }
    {
            "_id" : "shard2",
            "members" : [
                    {
                            "_id" : 0,
                            "host" : "192.168.31.231:27018"
                    },
                    {
                            "_id" : 1,
                            "host" : "192.168.31.232:27018"
                    },
                    {
                            "_id" : 2,
                            "host" : "192.168.31.233:27018",
                            "arbiterOnly":true
                    }
            ]
    }
    > rs.initiate(config)
    { "ok" : 1 }
    
    1.4 配置 3  台 台 Config Server
    在 Server A 、B 、C  上 执行:
    /usr/local/mongo/bin/mongod  --configsvr  --dbpath  /data/config  --port  20000  --logpath /data/config/config.log --logappend --fork
    23.4  配置 3  台 台 Route Process
    在 在 Server A 、B 、C  上 执行:
    /usr/local/mongo/bin/mongos  --configdb 192.168.31.231:20000,192.168.31.232:20000,192.168.31.233:20000 --port 30000 --chunkSize 1 --logpath /data/mongos.log --logappend --fork
    
    [root@mongo1 bin]# netstat -tpln|grep mongo
    tcp        0      0 0.0.0.0:30000               0.0.0.0:*                   LISTEN      1573/mongos         
    tcp        0      0 0.0.0.0:20000               0.0.0.0:*                   LISTEN      1375/mongod         
    tcp        0      0 0.0.0.0:27017               0.0.0.0:*                   LISTEN      1061/mongod         
    tcp        0      0 0.0.0.0:27018               0.0.0.0:*                   LISTEN      1211/mongod         
    [root@mongo1 bin]# ps aux|grep mongo
    root      1061  2.0  4.1 3198136 41920 ?       Sl   21:02   0:35 /usr/local/mongo/bin/mongod --shardsvr --replSet shard1 --port 27017 --dbpath /data/shard1_1 --logpath /data/shard1_1/shard1_1.log --logappend --fork
    root      1211  2.5  7.6 3202220 77716 ?       Sl   21:09   0:33 /usr/local/mongo/bin/mongod --shardsvr --replSet shard2 --port 27018 --dbpath /data/shard2_1 --logpath /data/shard2_1/shard2_1.log --logappend --fork
    root      1375  1.3  3.7 487476 38484 ?        Sl   21:17   0:12 /usr/local/mongo/bin/mongod --configsvr --dbpath /data/config --port 20000 --logpath /data/config/config.log --logappend --fork
    root      1573  0.8  0.7 240152  7320 ?        Sl   21:28   0:02 /usr/local/mongo/bin/mongos --configdb 192.168.31.231:20000,192.168.31.232:20000,192.168.31.233:20000 --port 30000 --chunkSize 1 --logpath /data/mongos.log --logappend --fork
    root      1794  0.0  0.0 103244   848 pts/1    S+   21:31   0:00 grep mongo
    
    
    1.5配置 Shard Cluster
    连接到其中一台机器的端口 30000 的 mongos 进程,并切换到 admin 数据库做以下配置
    [root@localhost bin]# ./mongo --port 30000
    MongoDB shell version: 1.8.1
    connecting to: 127.0.0.1:30000/test
    > use admin
    switched to db admin
    >db.runCommand({addshard:"shard1/192.168.31.231:27017,192.168.31.232:27017,192.168.31.233:27017"});
    { "shardAdded" : "shard1", "ok" : 1 }
    >db.runCommand({addshard:"shard2/192.168.31.231:27018,192.168.31.232:27018,192.168.31.233:27018"});
    { "shardAdded" : "shard2", "ok" : 1 }
    >
    进入mongos进程config库可以看到目前分片的情况:
    mongos> use config
    switched to db config
    mongos> db.shards.find()
    { "_id" : "shard1", "host" : "shard1/192.168.31.231:27017,192.168.31.232:27017,192.168.31.233:27017" }
    { "_id" : "shard2", "host" : "shard2/192.168.31.231:27018,192.168.31.232:27018,192.168.31.233:27018" }
    mongos> 
    
    1.6激活数据库及集合的分片
    db.runCommand({ enablesharding:"test" })
    db.runCommand({ shardcollection: "test.users", key: { _id:1 }})
    
    
    1.7验证 Sharding  正常工作
    连接到其中一台机器的端口 30000 的 mongos 进程,并切换到 test 数据库,以便添加测试数
    据
    use test
    for(var i=1;i<=20000;i++) db.users.insert({id:i,addr_1:"Beijing",addr_2:"Shanghai"});
    db.users.stats()
    {
            "sharded" : true,
            "paddingFactorNote" : "paddingFactor is unused and unmaintained in 3.0. It remains hard coded to 1.0 for compatibility only.",
            "userFlags" : 1,
            "capped" : false,
            "ns" : "test.users",
            "count" : 20000,
            "numExtents" : 10,
            "size" : 2240000,
            "storageSize" : 5586944,
            "totalIndexSize" : 670432,
            "indexSizes" : {
                    "_id_" : 670432
            },
            "avgObjSize" : 112,
            "nindexes" : 1,
            "nchunks" : 5,
            "shards" : {
                    "shard1" : {
                            "ns" : "test.users",
                            "count" : 8697,
                            "size" : 974064,
                            "avgObjSize" : 112,
                            "numExtents" : 5,
                            "storageSize" : 2793472,
                            "lastExtentSize" : 2097152,
                            "paddingFactor" : 1,
                            "paddingFactorNote" : "paddingFactor is unused and unmaintained in 3.0. It remains hard coded to 1.0 for compatibility only.",
                            "userFlags" : 1,
                            "capped" : false,
                            "nindexes" : 1,
                            "totalIndexSize" : 294336,
                            "indexSizes" : {
                                    "_id_" : 294336
                            },
                            "ok" : 1,
                            "$gleStats" : {
                                    "lastOpTime" : Timestamp(0, 0),
                                    "electionId" : ObjectId("558aabf892f5b70479ade020")
                            }
                    },
                    "shard2" : {
                            "ns" : "test.users",
                            "count" : 11303,
                            "size" : 1265936,
                            "avgObjSize" : 112,
                            "numExtents" : 5,
                            "storageSize" : 2793472,
                            "lastExtentSize" : 2097152,
                            "paddingFactor" : 1,
                            "paddingFactorNote" : "paddingFactor is unused and unmaintained in 3.0. It remains hard coded to 1.0 for compatibility only.",
                            "userFlags" : 1,
                            "capped" : false,
                            "nindexes" : 1,
                            "totalIndexSize" : 376096,
                            "indexSizes" : {
                                    "_id_" : 376096
                            },
                            "ok" : 1,
                            "$gleStats" : {
                                    "lastOpTime" : Timestamp(0, 0),
                                    "electionId" : ObjectId("558aacdd8bf5d360e64002dd")
                            }
                    }
            },
            "ok" : 1
    }
    mongos> 
    ./mongo --port 27017
    shard2:PRIMARY> use test
    switched to db test
    shard2:PRIMARY> show collections;
    system.indexes
    users
    shard2:PRIMARY> db.users.find().count()
    11303  --和db.users.stats()查看的数据条数一样 20000=11303+8697
    shard2:PRIMARY> 
     
    ./mongo --port 27018
    shard2:PRIMARY> use test
    switched to db test
    shard2:PRIMARY> show collections;
    system.indexes
    users
    shard1:PRIMARY> db.users.find().count()
    8697   ----和db.users.stats()查看的数据条数一样  
    shard1:PRIMARY> 
    
    从SECONDARY读取数据:
    shard2:SECONDARY> show collections;
    2015-06-24T21:46:11.634+0800 E QUERY    Error: listCollections failed: { "note" : "from execCommand", "ok" : 0, "errmsg" : "not master" }
    查询时报错,说明SECONDARY节点不能执行查询操作。
    执行db.getMongo().setSlaveOk(),从库就可以进行读操作了,可以控制从哪些分片上面可以读。
    shard1:
    shard2:SECONDARY> db.getMongo().setSlaveOk()
    shard2:SECONDARY> show collections;
    system.indexes
    users
    
    shard2:
    shard1:SECONDARY> db.getMongo().setSlaveOk()
    shard1:SECONDARY> show collections;
    system.indexes
    users
    shard1:SECONDARY> db.users.find().count();
    8697
    shard1:SECONDARY> 
    
    模拟故障:
    [root@mongo1 bin]# ps aux|grep mongo
    root      1061  1.9  8.1 3200192 82956 ?       Sl   21:02   0:56 /usr/local/mongo/bin/mongod --shardsvr --replSet shard1 --port 27017 --dbpath /data/shard1_1 --logpath /data/shard1_1/shard1_1.log --logappend --fork
    root      1211  2.1  7.8 3205304 79928 ?       Sl   21:09   0:52 /usr/local/mongo/bin/mongod --shardsvr --replSet shard2 --port 27018 --dbpath /data/shard2_1 --logpath /data/shard2_1/shard2_1.log --logappend --fork
    root      1375  1.8  4.1 489528 42180 ?        Sl   21:17   0:37 /usr/local/mongo/bin/mongod --configsvr --dbpath /data/config --port 20000 --logpath /data/config/config.log --logappend --fork
    root      1573  0.9  0.7 240152  7340 ?        Sl   21:28   0:13 /usr/local/mongo/bin/mongos --configdb 192.168.31.231:20000,192.168.31.232:20000,192.168.31.233:20000 --port 30000 --chunkSize 1 --logpath /data/mongos.log --logappend --fork
    root      1901  0.0  1.8 721952 19132 pts/0    Sl+  21:38   0:00 ./mongo --port 27018
    root      2130  0.0  0.0 103244   848 pts/1    S+   21:51   0:00 grep mongo
    [root@mongo1 bin]# kill -2 1061  --这里不能使用kill -9 ,否则会导致丢失数据
    
    通过rs.status()可以查看复制集的状态:
    我们可以看到192.168.31.231:27017处于宕机状态,192.168.31.233:27017接管变成了PRIMARY
    shard1:PRIMARY>  rs.status()
    {
            "set" : "shard1",
            "date" : ISODate("2015-06-24T14:03:10.425Z"),
            "myState" : 1,
            "members" : [
                    {
                            "_id" : 0,
                            "name" : "192.168.31.231:27017",
                            "health" : 0,
                            "state" : 8,
                            "stateStr" : "(not reachable/healthy)",
                            "uptime" : 0,
                            "optime" : Timestamp(0, 0),
                            "optimeDate" : ISODate("1970-01-01T00:00:00Z"),
                            "lastHeartbeat" : ISODate("2015-06-24T14:03:09.083Z"),
                            "lastHeartbeatRecv" : ISODate("2015-06-24T13:51:38.834Z"),
                            "pingMs" : 0,
                            "lastHeartbeatMessage" : "Failed attempt to connect to 192.168.31.231:27017; couldn't connect to server 192.168.31.231:27017 (192.168.31.231), connection attempt failed",
                            "configVersion" : -1
                    },
                    {
                            "_id" : 1,
                            "name" : "192.168.31.232:27017",
                            "health" : 1,
                            "state" : 2,
                            "stateStr" : "SECONDARY",
                            "uptime" : 3242,
                            "optime" : Timestamp(1435152718, 13),
                            "optimeDate" : ISODate("2015-06-24T13:31:58Z"),
                            "lastHeartbeat" : ISODate("2015-06-24T14:03:09.814Z"),
                            "lastHeartbeatRecv" : ISODate("2015-06-24T14:03:09.813Z"),
                            "pingMs" : 1,
                            "configVersion" : 1
                    },
                    {
                            "_id" : 2,
                            "name" : "192.168.31.233:27017",
                            "health" : 1,
                            "state" : 1,
                            "stateStr" : "PRIMARY",
                            "uptime" : 3603,
                            "optime" : Timestamp(1435152718, 13),
                            "optimeDate" : ISODate("2015-06-24T13:31:58Z"),
                            "electionTime" : Timestamp(1435153900, 1),
                            "electionDate" : ISODate("2015-06-24T13:51:40Z"),
                            "configVersion" : 1,
                            "self" : true
                    }
            ],
            "ok" : 1,
            "$gleStats" : {
                    "lastOpTime" : Timestamp(0, 0),
                    "electionId" : ObjectId("558ab5ec28f234f7b5abda48")
            }
    }
    
    查看分片集合信息:
    mongos> db.collections.find()
    { "_id" : "test.users", "lastmod" : ISODate("2015-06-24T13:30:19.670Z"), "dropped" : false, "key" : { "_id" : 1 }, "unique" : false, "lastmodEpoch" : ObjectId("558ab0eb6a881136d7047be4") }
  • 相关阅读:
    数据库分库分表之后,你是如何解决事务问题?
    数据库周刊31丨openGauss 正式开源;7月数据库排行榜发布;浙江移动国产数据库AntDB迁移;oracle ADG跨版本搭建;PG解决社保问题;mysqlbinlog解析……
    2020年7月国产数据库排行:华为、腾讯发新品,中兴、阿里结硕果
    47%的MongoDB数据库遭黑客比特币勒索,你中招了吗?中招怎么办?
    数据库周刊30丨数据安全法草案将亮相;2020数据库产业报告;云南电网上线达梦;达梦7误删Redo Log;Oracle存储过程性能瓶颈;易鲸捷实践案例……
    数据库周刊29│2020数据库研究报告;Oracle取消今年技术大会;腾讯云DBbridge发布支持一键迁库;饿了么迁至阿里云;PG数组查询;Oracle被比特币勒索;DM8 安全管理…
    数据库周刊28│开发者最喜爱的数据库是什么?阿里云脱口秀聊程序员转型;MySQL update误操作;PG流复制踩坑;PG异机归档;MySQL架构选型;Oracle技能表;Oracle文件损坏处理……
    怎么查看HBase表的创建时间
    Hadoop/HBase Kerberos认证失败:Clock skew too great
    java线程莫名异常退出时,如何捕获异常信息
  • 原文地址:https://www.cnblogs.com/cheerong/p/4629588.html
Copyright © 2011-2022 走看看