zoukankan      html  css  js  c++  java
  • 高可用Kubernetes集群-7. 部署kube-controller-manager

     九.部署kube-controller-manager

    kube-controller-manager是Kube-Master相关的3个服务之一,是有状态的服务,会修改集群的状态信息。

    如果多个master节点上的相关服务同时生效,则会有同步与一致性问题,所以多master节点中的kube-controller-manager服务只能是主备的关系,kukubernetes采用租赁锁(lease-lock)实现leader的选举,具体到kube-controller-manager,设置启动参数"--leader-elect=true"。

    1. 创建kube-controller-manager证书

    1)创建kube-conftroller-manager证书签名请求

    # kube-controller-mamager与kubei-apiserver通信采用双向TLS认证;
    # kube-apiserver提取CN作为客户端的用户名,即system:kube-controller-manager。 kube-apiserver预定义的 RBAC使用的ClusterRoleBindings system:kube-controller-manager将用户system:kube-controller-manager与ClusterRole system:kube-controller-manager绑定
    [root@kubenode1 ~]# mkdir -p /etc/kubernetes/controller-manager
    [root@kubenode1 ~]# cd /etc/kubernetes/controller-manager/
    [root@kubenode1 controller-manager]# touch controller-manager-csr.json
    [root@kubenode1 controller-manager]# vim controller-manager-csr.json
    {
        "CN": "system:kube-controller-manager",
        "hosts": [
          "172.30.200.21",
          "172.30.200.22",
          "172.30.200.23"
        ],
        "key": {
            "algo": "rsa",
            "size": 2048
        },
        "names": [
            {
                "C": "CN",
                "ST": "ChengDu",
                "L": "ChengDu",
                "O": "system:kube-controller-manager",
                "OU": "cloudteam"
            }
        ]
    }

    2)生成kube-controller-mamager证书与私钥

    [root@kubenode1 controller-manager]# cfssl gencert -ca=/etc/kubernetes/ssl/ca.pem 
    -ca-key=/etc/kubernetes/ssl/ca-key.pem 
    -config=/etc/kubernetes/ssl/ca-config.json 
    -profile=kubernetes controller-manager-csr.json | cfssljson -bare controller-manager

    # 分发controller-manager.pem,controller-manager-key.pem
    [root@kubenode1 controller-manager]# scp controller-manager*.pem root@172.30.200.22:/etc/kubernetes/controller-manager/
    [root@kubenode1 controller-manager]# scp controller-manager*.pem root@172.30.200.23:/etc/kubernetes/controller-manager/

    2. 创建kube-controller-manager kubeconfig文件

    kube-controller-manager kubeconfig文件中包含Master地址信息与必要的认证信息。 

    # 配置集群参数;
    # --server:指定api-server,采用ha之后的vip;
    # cluster名自定义,设定之后需保持一致;
    # --kubeconfig:指定kubeconfig文件路径与文件名;如果不设置,默认生成在~/.kube/config文件
    [root@kubenode1 controller-manager]# kubectl config set-cluster kubernetes 
    --certificate-authority=/etc/kubernetes/ssl/ca.pem 
    --embed-certs=true 
    --server=https://172.30.200.10:6443 
    --kubeconfig=controller-manager.conf
    
    # 配置客户端认证参数;
    # 认证用户为前文签名中的“system:kube-controller-manager”;
    # 指定对应的公钥证书/私钥等
    [root@kubenode1 controller-manager]# kubectl config set-credentials system:kube-controller-manager 
    --client-certificate=/etc/kubernetes/controller-manager/controller-manager.pem 
    --embed-certs=true 
    --client-key=/etc/kubernetes/controller-manager/controller-manager-key.pem 
    --kubeconfig=controller-manager.conf
    
    # 配置上下文参数
    [root@kubenode1 controller-manager]# kubectl config set-context system:kube-controller-manager@kubernetes 
    --cluster=kubernetes 
    --user=system:kube-controller-manager 
    --kubeconfig=controller-manager.conf
    
    # 配置默认上下文
    [root@kubenode1 controller-manager]# kubectl config use-context system:kube-controller-manager@kubernetes --kubeconfig=controller-manager.conf

    # 分发controller-manager.conf文件到所有master节点;
    [root@kubenode1 controller-manager]# scp controller-manager.conf root@172.30.200.22:/etc/kubernetes/controller-manager/
    [root@kubenode1 controller-manager]# scp controller-manager.conf root@172.30.200.23:/etc/kubernetes/controller-manager/

    3. 配置kube-controller-manager的systemd unit文件

    相关可执行文件在部署kubectl时已部署完成。

    # kube-controller-manager在kube-apiserver启动之后启动
    [root@kubenode1 ~]# touch /usr/lib/systemd/system/kube-controller-manager.service
    [root@kubenode1 ~]# vim /usr/lib/systemd/system/kube-controller-manager.service
    [Unit]
    Description=Kubernetes Controller Manager
    Documentation=https://github.com/GoogleCloudPlatform/kubernetes
    After=network.target
    After=kube-apiserver.service
    
    [Service]
    EnvironmentFile=/usr/local/kubernetes/kube-controller-manager.conf
    ExecStart=/usr/local/kubernetes/bin/kube-controller-manager $KUBE_CONTROLLER_MANAGER_ARGS
    Restart=on-failure
    RestartSec=5
    
    [Install]
    WantedBy=multi-user.target
    
    # 启动参数文件
    # --kubeconfig:kubeconfig配置文件路径,配置文件中包含master地址信息与必要的认证信息;
    # --allocate-node:设置为true时,使用云服务商为Pod分配的cidrs,一般仅用在公有云;
    # --cluster-name:集群名称,默认即kubernetes;
    # --cluster-signing-cert-file / --cluster-signing-key-file:用于集群范围的认证;
    # --service-account-private-key-file:用于service account token签名的私钥文件路径;
    # --root-ca-file:根ca证书路径,被用于service account 的token secret中
    # --insecure-experimental-approve-all-kubelet-csrs-for-group:controller-manager自动授权kubelet客户端证书csr组
    # --use-service-account-credentials:设置为true时,表示为每个controller分别设置service account;
    # --controllers:启动的contrller列表,默认为”*”,启用所有的controller,但不包含” bootstrapsigner”与”tokencleaner”;
    # --leader-elect:设置为true时进行leader选举,集群高可用部署时controller-manager必须选举leader,默认即true
    [root@kubenode1 ~]# touch /usr/local/kubernetes/kube-controller-manager.conf
    [root@kubenode1 ~]# vim /usr/local/kubernetes/kube-controller-manager.conf
    KUBE_CONTROLLER_MANAGER_ARGS="--master=https://172.30.200.10:6443 
      --kubeconfig=/etc/kubernetes/controller-manager/controller-manager.conf 
      --allocate-node-cidrs=true 
      --service-cluster-ip-range=169.169.0.0/16 
      --cluster-cidr=10.254.0.0/16 
      --cluster-name=kubernetes 
      --cluster-signing-cert-file=/etc/kubernetes/ssl/ca.pem 
      --cluster-signing-key-file=/etc/kubernetes/ssl/ca-key.pem 
      --service-account-private-key-file=/etc/kubernetes/ssl/ca-key.pem 
      --root-ca-file=/etc/kubernetes/ssl/ca.pem 
      --insecure-experimental-approve-all-kubelet-csrs-for-group=system:bootstrappers 
      --use-service-account-credentials=true 
      --controllers=*,bootstrapsigner,tokencleaner 
      --leader-elect=true 
      --logtostderr=false 
      --log-dir=/var/log/kubernetes/controller-manager 
      --v=2  1>>/var/log/kubernetes/kube-controller-manager.log 2>&1"
    
    # 创建日志目录
    [root@kubenode1 ~]# mkdir -p /var/log/kubernetes/controller-manager 

    4. 启动并验证

    1)kube-conftroller-manager状态验证

    [root@kubenode1 ~]# systemctl daemon-reload
    [root@kubenode1 ~]# systemctl enable kube-controller-manager
    [root@kubenode1 ~]# systemctl start kube-controller-manager
    [root@kubenode1 ~]# systemctl status kube-controller-manager

    2)kube-conftroller-manager选举查看

    # 因kubenode1是第一个启动kube-controller-manager服务的节点,尝试获取leader权限,成功
    [root@kubenode1 ~]# cat /var/log/kubernetes/controller-manager/kube-controller-manager.INFO | grep "leaderelection"

    # 在kubenode2上观察,kubenode2在尝试获取leader权限,但未成功,后续操作挂起
    [root@kubenode2 ~]# tailf /var/log/kubernetes/controller-manager/kube-controller-manager.INFO

  • 相关阅读:
    mysql 主从复制
    confluence wiki搭建使用
    python 解析json loads dumps
    在linux下修改mysql的root密码
    [转]量子恒道统计
    php include
    php solr 查询
    php solr 扩展
    IKAnalyzer 和 solr4.3 冲突
    solr 4.3.0 配置
  • 原文地址:https://www.cnblogs.com/netonline/p/8747005.html
Copyright © 2011-2022 走看看