这一节我们在三个主节点部署高可用的etcd集群,官方建议5节点或者7节点,我们三节点也是够用的。我们开始吧!
  • 一、环境准备

      10.13.33.29  etcd-0
      10.13.33.40  etcd-1
      10.13.33.38  etcd-2

    etcd 数据目录:ETCD_DATA_DIR="/data/k8s/etcd/data"

    etcd WAL 目录:ETCD_WAL_DIR="/data/k8s/etcd/wal"
    (建议是 SSD 磁盘分区,或者和 ETCD_DATA_DIR 不同的磁盘分区)

  • 二、部署安装ETCD

1、创建etcd相关证书

cd /opt/k8s/work
cat > etcd-csr.json <<EOF
{
  "CN": "etcd",
  "hosts": [
    "127.0.0.1",
    "10.13.33.38",
    "10.13.33.29",
    "10.13.33.40"
  ],
  "key": {
    "algo": "rsa",
    "size": 2048
  },
  "names": [
    {
      "C": "CN",
      "ST": "BeiJing",
      "L": "BeiJing",
      "O": "k8s",
      "OU": "4Paradigm"
    }
  ]
}
EOF
###host字段指定授权使用该证书的etcd节点IP或子网列表,需要将etcd集群的3个节点都添加其中

###生成证书和私钥

cfssl gencert -ca=/opt/k8s/work/ca.pem \
    -ca-key=/opt/k8s/work/ca-key.pem \
    -config=/opt/k8s/work/ca-config.json \
    -profile=kubernetes etcd-csr.json | cfssljson -bare etcd
ls etcd*pem

###分发到etcd节点

mkdir -p /etc/etcd/cert
cp etcd*.pem /etc/etcd/cert/

2、安装etcd

wget https://github.com/etcd-io/etcd/releases/download/v3.3.13/etcd-v3.3.13-linux-amd64.tar.gz         ##下载etcd组件
tar -xvf etcd-v3.3.13-linux-amd64.tar.gz                                                                                                  ##解压 
cp etcd-v3.3.13-linux-amd64/etcd* /opt/k8s/bin/                                                                                     ##复制etcd命令到指定目录
chmod +x /opt/k8s/bin/*                                                                                                                          ##赋权
mkdir -p /data/k8s/etcd/data /data/k8s/etcd/wal                                                                                     ##创建etcd数据目录

3、创建etcd.service

cat > etcd.service.template <<EOF
[Unit]
Description=Etcd Server
After=network.target
After=network-online.target
Wants=network-online.target
Documentation=https://github.com/coreos
[Service]
Type=notify
WorkingDirectory=/data/k8s/etcd/data
ExecStart=/opt/k8s/bin/etcd \
  --data-dir '/data/k8s/etcd/data' \
  --wal-dir '/data/k8s/etcd/wal' \
  --name 'master-01' \                                                                                           ###根据etcd节点命名
  --cert-file '/etc/etcd/cert/etcd.pem' \
  --key-file '/etc/etcd/cert/etcd-key.pem' \
  --trusted-ca-file '/etc/kubernetes/cert/ca.pem' \
  --peer-cert-file '/etc/etcd/cert/etcd.pem' \
  --peer-key-file '/etc/etcd/cert/etcd-key.pem' \
  --peer-trusted-ca-file '/etc/kubernetes/cert/ca.pem' \
  --peer-client-cert-auth  \
  --client-cert-auth  \
  --listen-peer-urls 'https://10.13.33.38:2380' \
  --initial-advertise-peer-urls 'https://10.13.33.38:2380' \
  --listen-client-urls 'https://10.13.33.38:2379,http://127.0.0.1:2379' \
  --advertise-client-urls 'https://10.13.33.38:2379' \
  --initial-cluster-token 'etcd-cluster-0' \
  --initial-cluster 'master-03=https://10.13.33.40:2380,master-01=https://10.13.33.38:2380,master-02=https://10.13.33.29:2380' \                         ###各个etcd节点hostname及地址
  --initial-cluster-state 'new' \
  --auto-compaction-mode 'periodic' \
  --auto-compaction-retention '1' \
  --max-request-bytes '33554432' \
  --quota-backend-bytes '6442450944' \
  --heartbeat-interval '250' \
  --election-timeout '2000'
Restart=on-failure
RestartSec=5
LimitNOFILE=65536
[Install]
WantedBy=multi-user.target
EOF

3、启动etcd

systemctl daemon-reload && systemctl enable etcd && systemctl restart etcd && systemctl status etcd|grep Active
curl -L 127.0.0.1:2379/health          ###检查启动状态

4、确认集群健康

[root@master-01 work]# curl -L 127.0.0.1:2379/health   
{"health":"true"}
[root@master-01 work]# ETCDCTL_API=3 /opt/k8s/bin/etcdctl   -w table --cacert=/etc/kubernetes/cert/ca.pem   --cert=/etc/etcd/cert/etcd.pem   --key=/etc/etcd/cert/etcd-key.pem   --endpoints=https://10.13.33.29:2379,https://10.13.33.38:2379,https://10.13.33.40:2379  endpoint status
+--------------------------+------------------+---------+---------+-----------+-----------+------------+
|         ENDPOINT         |        ID        | VERSION | DB SIZE | IS LEADER | RAFT TERM | RAFT INDEX |
+--------------------------+------------------+---------+---------+-----------+-----------+------------+
| https://10.13.33.29:2379 | 3d57cc849d2a16** |  3.3.13 |   33 MB |      true |     36971 |    6201481 |
| https://10.13.33.38:2379 | bff9394ca77d77** |  3.3.13 |   33 MB |     false |     36971 |    6201481 |
| https://10.13.33.40:2379 | 2ce7325ad513f9** |  3.3.13 |   33 MB |     false |     36971 |    6201481 |
+--------------------------+------------------+---------+---------+-----------+-----------+------------+
[root@master-01 work]# ETCDCTL_API=3 /opt/k8s/bin/etcdctl   -w table --cacert=/etc/kubernetes/cert/ca.pem   --cert=/etc/etcd/cert/etcd.pem   --key=/etc/etcd/cert/etcd-key.pem   --endpoints=https://10.13.33.29:2379,https://10.13.33.38:2379,https://10.13.33.40:2379  member list
+------------------+---------+-----------+--------------------------+--------------------------+
|        ID        | STATUS  |   NAME    |        PEER ADDRS        |       CLIENT ADDRS       |
+------------------+---------+-----------+--------------------------+--------------------------+
| 2ce7325ad513f9** | started | master-03 | https://10.13.33.40:2380 | https://10.13.33.40:2379 |
| 3d57cc849d2a16** | started | master-02 | https://10.13.33.29:2380 | https://10.13.33.29:2379 |
| bff9394ca77d77** | started | master-01 | https://10.13.33.38:2380 | https://10.13.33.38:2379 |
+------------------+---------+-----------+--------------------------+--------------------------+
{"health":"true"}[root@ETCDCTL_API=3 /opt/k8s/bin/etcdctl   -w table --cacert=/etc/kubernetes/cert/ca.pem   --cert=/etc/etcd/cert/etcd.pem   --key=/etc/etcd/cert/etcd-key.pem   --endpoints=https://10.13.33.29:2379,https://10.13.33.38:2379,https://10.13.33.40:2379  endpoint health
https://10.13.33.38:2379 is healthy: successfully committed proposal: took = 2.174634ms
https://10.13.33.29:2379 is healthy: successfully committed proposal: took = 1.465878ms
https://10.13.33.40:2379 is healthy: successfully committed proposal: took = 2.36525ms
安装kubectl后可以通过kubectl查看
[root@master-01 ~]# kubectl get cs
NAME                 STATUS    MESSAGE             ERROR
controller-manager   Healthy   ok                  
scheduler            Healthy   ok                  
etcd-0               Healthy   {"health":"true"}   
etcd-1               Healthy   {"health":"true"}   
etcd-2               Healthy   {"health":"true"}   

5、向etcd写入POD网段信息

etcdctl \
  --endpoints=https://10.13.33.29:2379,https://10.13.33.38:2379,https://10.13.33.40:2379 \
  --ca-file=/opt/k8s/work/ca.pem \
  --cert-file=/opt/k8s/work/flanneld.pem \
  --key-file=/opt/k8s/work/flanneld-key.pem \
  mk /kubernetes/network/config '{"Network":"'172.30.0.0/16'", "SubnetLen": 21, "Backend": {"Type": "vxlan"}}'

6、重启flannel、docker

systemctl restart flanneld docker
systemctl status flanneld docker | grep Active
ip addr show| grep flannel.1

7、检查flannel、docker是否获取到正确的子网

[root@master-01 work]# ip addr show|grep flannel.1
5: flannel.1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue state UNKNOWN group default 
    inet 172.30.48.0/32 scope global flannel.1
[root@master-01 work]# ip addr show|grep docker 
6: docker0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1450 qdisc noqueue state UP group default 
    inet 172.30.48.1/21 brd 172.30.55.255 scope global docker0