安装apiserver
安装部署主控节点服务 -- apiserver
部署kube-apiserver集群
集群规划
主机名 角色 ip
rstx-203.rongbiz.cn kube-apiserver 192.168.1.203
rstx-204.rongbiz.cn kube-apiserver 192.168.1.204
rstx-201.rongbiz.cn 4层负载均衡 192.168.1.201
rstx-202.rongbiz.cn 4层负载均衡 192.168.1.202
注意:这里192.168.1.201和192.168.1.202使用nginx做4层负载均衡器,用keepalive跑一个vip:192.168.1.110,代理两个kube-apiserver,实现高可用
这里部署文档以HDSS7-14.host.com主机为例,另外一台运算节点部署方法类似
下载软件,解压,做软连接,安装1.15.2
[root@rstx-203 ~]# cd /opt/src/
[root@rstx-203 src]# ls
kubernetes-server-linux-amd64-v1.15.2.tar.gz
查看文件大小
[root@rstx-203 src]# du -sh kubernetes-server-linux-amd64-v1.15.2.tar.gz
[root@rstx-203 src]# tar xf kubernetes-server-linux-amd64-v1.15.4.tar.gz -C /opt
[root@rstx-203 src]# cd ..
[root@rstx-203 opt]# mv kubernetes/ kubernetes-v1.15.2
做软连接,方便以后更新
[root@rstx-203 opt]# ln -s /opt/kubernetes-v1.15.2/ /opt/kubernetes
[root@rstx-203 opt]# ll
总用量 0
drwx--x--x 4 root root 28 12月 10 14:28 containerd
lrwxrwxrwx 1 root root 17 12月 10 16:45 etcd -> /opt/etcd-v3.1.20
drwxr-xr-x 4 etcd etcd 166 12月 10 17:43 etcd-v3.1.20
lrwxrwxrwx 1 root root 24 12月 10 18:33 kubernetes -> /opt/kubernetes-v1.15.2/
drwxr-xr-x 4 root root 79 8月 5 18:01 kubernetes-v1.15.2
drwxr-xr-x 2 root root 97 12月 10 18:29 src
[root@rstx-203 opt]# cd kubernetes
[root@rstx-203 kubernetes]# ls
addons kubernetes-src.tar.gz LICENSES server
删除源码包
[root@rstx-203 kubernetes]# rm -rf kubernetes-src.tar.gz
[root@rstx-203 kubernetes]# cd server/bin/
删除没用的文件docker镜像等
[root@rstx-203 bin]# rm -rf *.tar
[root@rstx-203 bin]# rm -rf *_tag
剩余一些可执行文件
[root@rstx-203 bin]# ll
总用量 884636
-rwxr-xr-x 1 root root 43534816 8月 5 18:01 apiextensions-apiserver
-rwxr-xr-x 1 root root 100548640 8月 5 18:01 cloud-controller-manager
-rwxr-xr-x 1 root root 200648416 8月 5 18:01 hyperkube
-rwxr-xr-x 1 root root 40182208 8月 5 18:01 kubeadm
-rwxr-xr-x 1 root root 164501920 8月 5 18:01 kube-apiserver
-rwxr-xr-x 1 root root 116397088 8月 5 18:01 kube-controller-manager
-rwxr-xr-x 1 root root 42985504 8月 5 18:01 kubectl
-rwxr-xr-x 1 root root 119616640 8月 5 18:01 kubelet
-rwxr-xr-x 1 root root 36987488 8月 5 18:01 kube-proxy
-rwxr-xr-x 1 root root 38786144 8月 5 18:01 kube-scheduler
-rwxr-xr-x 1 root root 1648224 8月 5 18:01 mounter
签发apiserver-client证书:apiserver与etc通信用的证书。apiserver是客户端,etcd是服务端
运维主机HDSS-248.host.com上
创建生成证书签名请求(csr)的JSON配置文件 -- 此目录下有,直接上传修改,不要粘贴复制
[root@rstx-53 ~]# cd /opt/certs/
[root@rstx-53 ~]# vi /opt/certs/client-csr.json
{
"CN": "k8s-node",
"hosts": [
],
"key": {
"algo": "rsa",
"size": 2048
},
"names": [
{
"C": "CN",
"ST": "beijing",
"L": "beijing",
"O": "od",
"OU": "ops"
}
]
}
[root@rstx-53 certs]# cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=client client-csr.json |cfssl-json -bare client
创建签名请求(csr)的JSON配置文件,apiserver,server端证书 -- 直接上传
[root@rstx-53 certs]# vi apiserver-csr.json
{
"CN": "k8s-apiserver",
"hosts": [
"127.0.0.1",
"10.254.0.1",
"kubernetes.default",
"kubernetes.default.svc",
"kubernetes.default.svc.cluster",
"kubernetes.default.svc.cluster.local",
"192.168.1.200",
"192.168.1.203",
"192.168.1.204",
"192.168.1.205"
],
"key": {
"algo": "rsa",
"size": 2048
},
"names": [
{
"C": "CN",
"ST": "beijing",
"L": "beijing",
"O": "od",
"OU": "ops"
}
]
}
[root@rstx-53 certs]# cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=server apiserver-csr.json |cfssl-json -bare apiserver
[root@rstx-53 certs]# ll
总用量 80
total 24
-rw------- 1 root root 1679 Nov 10 17:42 apiserver-key.pem
-rw-r--r-- 1 root root 1598 Nov 10 17:42 apiserver.pem
-rw------- 1 root root 1679 Nov 10 17:41 ca-key.pem
-rw-r--r-- 1 root root 1346 Nov 10 17:41 ca.pem
-rw------- 1 root root 1675 Nov 10 17:41 client-key.pem
-rw-r--r-- 1 root root 1363 Nov 10 17:41 client.pem
拷贝证书
[root@rstx-203 ~]# cd /opt/kubernetes/server/bin/
[root@rstx-203 bin]# mkdir certs
[root@rstx-203 bin]# cd certs/
最后有个点,表示拷贝到当前目录
scp -rp root@192.168.1.53:/opt/certs/{ca.pem,ca-key.pem,client.pem,client-key.pem,apiserver.pem,apiserver-key.pem} .
[root@rstx-203 certs]# scp rstx-53:/opt/certs/ca.pem .
[root@rstx-203 certs]# scp rstx-53:/opt/certs/ca-key.pem .
[root@rstx-203 certs]# scp rstx-53:/opt/certs/client.pem .
[root@rstx-203 certs]# scp rstx-53:/opt/certs/client-key.pem .
[root@rstx-203 certs]# scp rstx-53:/opt/certs/apiserver.pem .
[root@rstx-203 certs]# scp rstx-53:/opt/certs/apiserver-key.pem .
创建启动配置脚本 -- 直接上传
[root@rstx-203 certs]# cd /opt/kubernetes/server/bin
[root@rstx-203 bin]# mkdir conf
[root@rstx-203 bin]# cd conf/
[root@rstx-203 conf]#
[root@rstx-203 conf]# vi audit.yaml
apiVersion: audit.k8s.io/v1beta1 # This is required.
kind: Policy
# Don't generate audit events for all requests in RequestReceived stage.
omitStages:
- "RequestReceived"
rules:
# Log pod changes at RequestResponse level
- level: RequestResponse
resources:
- group: ""
# Resource "pods" doesn't match requests to any subresource of pods,
# which is consistent with the RBAC policy.
resources: ["pods"]
# Log "pods/log", "pods/status" at Metadata level
- level: Metadata
resources:
- group: ""
resources: ["pods/log", "pods/status"]
# Don't log requests to a configmap called "controller-leader"
- level: None
resources:
- group: ""
resources: ["configmaps"]
resourceNames: ["controller-leader"]
# Don't log watch requests by the "system:kube-proxy" on endpoints or services
- level: None
users: ["system:kube-proxy"]
verbs: ["watch"]
resources:
- group: "" # core API group
resources: ["endpoints", "services"]
# Don't log authenticated requests to certsain non-resource URL paths.
- level: None
userGroups: ["system:authenticated"]
nonResourceURLs:
- "/api*" # Wildcard matching.
- "/version"
# Log the request body of configmap changes in kube-system.
- level: Request
resources:
- group: "" # core API group
resources: ["configmaps"]
# This rule only applies to resources in the "kube-system" namespace.
# The empty string "" can be used to select non-namespaced resources.
namespaces: ["kube-system"]
# Log configmap and secret changes in all other namespaces at the Metadata level.
- level: Metadata
resources:
- group: "" # core API group
resources: ["secrets", "configmaps"]
# Log all other resources in core and extensions at the Request level.
- level: Request
resources:
- group: "" # core API group
- group: "extensions" # Version of group should NOT be included.
# A catch-all rule to log all other requests at the Metadata level.
- level: Metadata
# Long-running requests like watches that fall under this rule will not
# generate an audit event in RequestReceived.
omitStages:
- "RequestReceived"
编写启动脚本 -- 直接上传
[root@rstx-203 config]# vi /opt/kubernetes/server/bin/kube-apiserver.sh
#!/bin/bash
./kube-apiserver \
--apiserver-count 2 \
--audit-log-path /data/logs/kubernetes/kube-apiserver/audit-log \
--audit-policy-file ./conf/audit.yaml \
--authorization-mode RBAC \
--client-ca-file ./certs/ca.pem \
--requestheader-client-ca-file ./certs/ca.pem \
--enable-admission-plugins NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota \
--etcd-cafile ./certs/ca.pem \
--etcd-certfile ./certs/client.pem \
--etcd-keyfile ./certs/client-key.pem \
--etcd-servers https://192.168.1.202:2379,https://192.168.1.203:2379,https://192.168.1.204:2379 \
--service-account-key-file ./certs/ca-key.pem \
--service-cluster-ip-range 10.254.0.0/16 \
--service-node-port-range 3000-29999 \
--target-ram-mb=1024 \
--kubelet-client-certificate ./certs/client.pem \
--kubelet-client-key ./certs/client-key.pem \
--log-dir /data/logs/kubernetes/kube-apiserver \
--tls-cert-file ./certs/apiserver.pem \
--tls-private-key-file ./certs/apiserver-key.pem \
--v 2
查看帮助命令,查看每行的意思
[root@rstx-203 bin]# ./kube-apiserver --help|grep -A 5 target-ram-mb
添加执行权限
[root@rstx-203 bin]# chmod +x kube-apiserver.sh
[root@rstx-203 bin]# mkdir /data/logs/kubernetes/kube-apiserver -p
创建后台启动
[root@rstx-203 bin]# vi /etc/supervisord.d/kube-apiserver.ini
[program:kube-apiserver-203] # 21根据实际IP地址更改
command=/opt/kubernetes/server/bin/kube-apiserver.sh ; the program (relative uses PATH, can take args)
numprocs=1 ; number of processes copies to start (def 1)
directory=/opt/kubernetes/server/bin ; directory to cwd to before exec (def no cwd)
autostart=true ; start at supervisord start (default: true)
autorestart=true ; retstart at unexpected quit (default: true)
startsecs=30 ; number of secs prog must stay running (def. 1)
startretries=3 ; max # of serial start failures (default 3)
exitcodes=0,2 ; 'expected' exit codes for process (default 0,2)
stopsignal=QUIT ; signal used to kill process (default TERM)
stopwaitsecs=10 ; max num secs to wait b4 SIGKILL (default 10)
user=root ; setuid to this UNIX account to run the program
redirect_stderr=true ; redirect proc stderr to stdout (default false)
stdout_logfile=/data/logs/kubernetes/kube-apiserver/apiserver.stdout.log ; stderr log path, NONE for none; default AUTO
stdout_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB)
stdout_logfile_backups=4 ; # of stdout logfile backups (default 10)
stdout_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0)
stdout_events_enabled=false ; emit events on stdout writes (default false)
killasgroup=true
stopasgroup=true
[root@rstx-203 bin]# mkdir -p /data/logs/kubernetes/kube-apiserver
[root@rstx-203 bin]# supervisorctl update
[root@rstx-203 bin]# netstat -luntp | grep kube-api
tcp 0 0 127.0.0.1:8080 0.0.0.0:* LISTEN 27303/./kube-apiser
tcp6 0 0 :::6443 :::* LISTEN 27303/./kube-apiser
4 设置四层反向代理及安装VIP
安装部署主控节点4层反向代理服务
部署在rstx-201 rstx-2012机器上,用VIP 192.168.1.110的7443端口,反代rstx-203、rstx-204的apiserver6443端口
rstx-201和rstx-2012上同时操作
nginx配置
[root@rstx-201 ~]# yum install yum-utils
sudo yum-config-manager --add-repo https://openresty.org/package/centos/openresty.repo
yum install openresty
ln -s /usr/local/openresty/nginx/ /etc/nginx
yum install nginx -y
[root@rstx-201 ~]# vi /etc/nginx/nginx.conf -- 直接上传
stream {
upstream kube-apiserver {
server 192.168.1.203:6443 max_fails=3 fail_timeout=30s;
server 192.168.1.204:6443 max_fails=3 fail_timeout=30s;
}
server {
listen 7443;
proxy_connect_timeout 2s;
proxy_timeout 900s;
proxy_pass kube-apiserver;
}
}
检查配置文件
[root@rstx-201 ~]# nginx -t
nginx: the configuration file /etc/nginx/nginx.conf syntax is ok
nginx: configuration file /etc/nginx/nginx.conf test is successful
[root@rstx-201 ~]# systemctl start nginx
[root@rstx-201 ~]# systemctl enable nginx
keepalived安装配置
[root@rstx-201 ~]# yum install keepalived -y
编写监听脚本 -- 直接上传
[root@rstx-201 ~]# vi /etc/keepalived/check_port.sh
#!/bin/bash
#keepalived 监控端口脚本
#使用方法:
#在keepalived的配置文件中
#vrrp_script check_port {#创建一个vrrp_script脚本,检查配置
# script "/etc/keepalived/check_port.sh 6379" #配置监听的端口
# interval 2 #检查脚本的频率,单位(秒)
#}
CHK_PORT=$1
if [ -n "$CHK_PORT" ];then
PORT_PROCESS=`ss -lnt|grep $CHK_PORT|wc -l`
if [ $PORT_PROCESS -eq 0 ];then
echo "Port $CHK_PORT Is Not Used,End."
exit 1
fi
else
echo "Check Port Cant Be Empty!"
fi
[root@rstx-201 ~]# # chmod +x /etc/keepalived/check_port.sh
配置keepalived -- 删除原文件,直接上传修改
keepalived 主:
[root@rstx-201 ~]# vi /etc/keepalived/keepalived.conf
! Configuration File for keepalived
global_defs {
router_id 192.168.1.201
}
vrrp_script chk_nginx {
script "/etc/keepalived/check_port.sh 7443"
interval 2
weight -20
}
vrrp_instance VI_1 {
state MASTER
interface eth0 # 根据实际网卡更改
virtual_router_id 251
priority 100
advert_int 1
mcast_src_ip 192.168.1.201
nopreempt
authentication {
auth_type PASS
auth_pass 11111111
}
track_script {
chk_nginx
}
virtual_ipaddress {
192.168.1.200
}
}
keepalived从: -- 直接上传
[root@rstx-202 ~]# vi /etc/keepalived/keepalived.conf
! Configuration File for keepalived
global_defs {
router_id 192.168.1.202
script_user root
enable_script_security
}
vrrp_script chk_nginx {
script "/etc/keepalived/check_port.sh 7443"
interval 2
weight -20
}
vrrp_instance VI_1 {
state BACKUP
interface eth0 # 根据实际网卡更改
virtual_router_id 251
mcast_src_ip 192.168.1.202
priority 90
advert_int 1
authentication {
auth_type PASS
auth_pass 11111111
}
track_script {
chk_nginx
}
virtual_ipaddress {
192.168.1.200
}
}
[root@rstx-201 keepalived]# ss -lnt|grep 7443|wc -l
1
[root@rstx-201 ~]# systemctl start keepalived
[root@rstx-201 ~]# systemctl enable keepalived
如果vip出现变动,主keepalived恢复后,一定要确认主keepalived端口起来, 服务搞好,重启keepalived,是vip变回主keepalived
[root@rstx-201 ~]# netstat -luntp | grep 7443
tcp 0 0 0.0.0.0:7443 0.0.0.0:* LISTEN 22071/nginx: master
apiserver api调用
[root@master ~]# kubectl proxy
Starting to serve on 127.0.0.1:8001
[root@master ~]# curl 127.0.0.1:8001/api/v1/namespaces/default/pods
{
"kind": "PodList",
"apiVersion": "v1",
"metadata": {
"resourceVersion": "152151"
},
"items": [
{
"metadata": {
"name": "nginx-6799fc88d8-d6ktm",
"generateName": "nginx-6799fc88d8-",
"namespace": "default",
"uid": "79b492da-40f7-4529-96ea-e7c72d341ec6",
"resourceVersion": "120158",
"creationTimestamp": "2021-07-30T01:24:06Z",
"labels": {
"app": "nginx",
"pod-template-hash": "6799fc88d8"
},
"ownerReferences": [
{
"apiVersion": "apps/v1",
"kind": "ReplicaSet",
"name": "nginx-6799fc88d8",
"uid": "1464e711-21a2-427e-93f9-2e9edabe5087",
"controller": true,
"blockOwnerDeletion": true
}
],
"managedFields": [
{
"manager": "kube-controller-manager",
"operation": "Update",
"apiVersion": "v1",
"time": "2021-07-30T01:24:06Z",
"fieldsType": "FieldsV1",
"fieldsV1": {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:app":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1464e711-21a2-427e-93f9-2e9edabe5087\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"nginx\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}
},
{
"manager": "kubelet",
"operation": "Update",
"apiVersion": "v1",
"time": "2021-07-30T01:24:06Z",
"fieldsType": "FieldsV1",
"fieldsV1": {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}
}
]
},