基础架构
主机名 |
角色 |
ip |
HDSS7-11.host.com |
k8s代理节点1 |
10.4.7.11 |
HDSS7-12.host.com |
k8s代理节点2 |
10.4.7.12 |
HDSS7-21.host.com |
k8s运算节点1 |
10.4.7.21 |
HDSS7-22.host.com |
k8s运算节点2 |
10.4.7.22 |
HDSS7-200.host.com |
k8s运维节点(docker仓库) |
10.4.7.200 |
硬件环境
软件环境
前置准备工作
DNS服务安装部署
- 创建主机域host.com
- 创建业务域od.com
- 主辅同步(10.4.7.11主、10.4.7.12辅)
- 客户端配置指向自建DNS
略
准备签发证书环境
运维主机HDSS7-200.host.com
上:
安装CFSSL
复制
1 2 3 4
|
[root@hdss7-200 ~]# curl -s -L -o /usr/bin/cfssl https://pkg.cfssl.org/R1.2/cfssl_linux-amd64 [root@hdss7-200 ~]# curl -s -L -o /usr/bin/cfssljson https://pkg.cfssl.org/R1.2/cfssljson_linux-amd64 [root@hdss7-200 ~]# curl -s -L -o /usr/bin/cfssl-certinfo https://pkg.cfssl.org/R1.2/cfssl-certinfo_linux-amd64 [root@hdss7-200 ~]# chmod +x /usr/bin/cfssl*
|
创建生成CA证书的JSON配置文件
复制/opt/certs/ca-config.json
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34
|
{ "signing": { "default": { "expiry": "175200h" }, "profiles": { "server": { "expiry": "175200h", "usages": [ "signing", "key encipherment", "server auth" ] }, "client": { "expiry": "175200h", "usages": [ "signing", "key encipherment", "client auth" ] }, "peer": { "expiry": "175200h", "usages": [ "signing", "key encipherment", "server auth", "client auth" ] } } } }
|
证书类型
client certificate: 客户端使用,用于服务端认证客户端,例如etcdctl、etcd proxy、fleetctl、docker客户端
server certificate: 服务端使用,客户端以此验证服务端身份,例如docker服务端、kube-apiserver
peer certificate: 双向证书,用于etcd集群成员间通信
创建生成CA证书签名请求(csr)的JSON配置文件
复制/opt/certs/ca-csr.json
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21
|
{ "CN": "kubernetes-ca", "hosts": [ ], "key": { "algo": "rsa", "size": 2048 }, "names": [ { "C": "CN", "ST": "beijing", "L": "beijing", "O": "od", "OU": "ops" } ], "ca": { "expiry": "175200h" } }
|
CN: Common Name,浏览器使用该字段验证网站是否合法,一般写的是域名。非常重要。浏览器使用该字段验证网站是否合法
C: Country, 国家
ST: State,州,省
L: Locality,地区,城市
O: Organization Name,组织名称,公司名称
OU: Organization Unit Name,组织单位名称,公司部门
生成CA证书和私钥
复制/opt/certs
1 2 3 4 5 6 7
|
[root@hdss7-200 certs]# cfssl gencert -initca ca-csr.json | cfssljson -bare ca - 2019/01/18 09:31:19 [INFO] generating a new CA key and certificate from CSR 2019/01/18 09:31:19 [INFO] generate received request 2019/01/18 09:31:19 [INFO] received CSR 2019/01/18 09:31:19 [INFO] generating key: rsa-2048 2019/01/18 09:31:19 [INFO] encoded CSR 2019/01/18 09:31:19 [INFO] signed certificate with serial number 345276964513449660162382535043012874724976422200
|
生成ca.pem、ca.csr、ca-key.pem(CA私钥,需妥善保管)
复制/opt/certs
1 2 3 4 5 6
|
[root@hdss7-200 certs]# ls -l -rw-r--r-- 1 root root 836 Jan 16 11:04 ca-config.json -rw-r--r-- 1 root root 332 Jan 16 11:10 ca-csr.json -rw------- 1 root root 1675 Jan 16 11:17 ca-key.pem -rw-r--r-- 1 root root 1001 Jan 16 11:17 ca.csr -rw-r--r-- 1 root root 1354 Jan 16 11:17 ca.pem
|
部署docker环境
HDSS7-200.host.com
,HDSS7-21.host.com
,HDSS7-22.host.com
上:
安装
复制
1 2 3 4
|
# ls -l|grep docker-engine -rw-r--r-- 1 root root 20013304 Jan 16 18:16 docker-engine-1.12.6-1.el7.centos.x86_64.rpm -rw-r--r-- 1 root root 29112 Jan 16 18:15 docker-engine-selinux-1.12.6-1.el7.centos.noarch.rpm # yum localinstall *.rpm
|
配置
复制/etc/docker/daemon.json
1 2 3 4 5 6 7 8 9
|
# vi /etc/docker/daemon.json { "graph": "/data/docker", "storage-driver": "overlay", "insecure-registries": ["registry.access.redhat.com","quay.io","harbor.od.com"], "bip": "172.7.21.1/24", "exec-opts": ["native.cgroupdriver=systemd"], "live-restore": true }
|
注意:这里bip要根据宿主机ip变化
启动脚本
复制/usr/lib/systemd/system/docker.service
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28
|
[Unit] Description=Docker Application Container Engine Documentation=https://docs.docker.com After=network.target
[Service] Type=notify # the default is not to use systemd for cgroups because the delegate issues still # exists and systemd currently does not support the cgroup feature set required # for containers run by docker ExecStart=/usr/bin/dockerd ExecReload=/bin/kill -s HUP $MAINPID # Having non-zero Limit*s causes performance problems due to accounting overhead # in the kernel. We recommend using cgroups to do container-local accounting. LimitNOFILE=infinity LimitNPROC=infinity LimitCORE=infinity # Uncomment TasksMax if your systemd version supports it. # Only systemd 226 and above support this version. #TasksMax=infinity TimeoutStartSec=0 # set delegate yes so that systemd does not reset the cgroups of docker containers Delegate=yes # kill only the docker process, not all processes in the cgroup KillMode=process
[Install] WantedBy=multi-user.target
|
启动
复制
1 2
|
# systemctl enable docker.service # systemctl start docker.service
|
部署docker镜像私有仓库harbor
HDSS7-200.host.com
上:
下载软件二进制包并解压
harbor下载地址
复制/opt/harbor
1 2 3 4 5 6
|
[root@hdss7-200 harbor]# tar xf harbor-offline-installer-v1.7.1.tgz -C /opt
[root@hdss7-200 harbor]# ll total 583848 drwxr-xr-x 3 root root 242 Jan 23 15:28 harbor -rw-r--r-- 1 root root 597857483 Jan 17 14:58 harbor-offline-installer-v1.7.1.tgz
|
配置
复制/opt/harbor/harbor.cfg
1
|
hostname = harbor.od.com
|
复制/opt/harbor/docker-compose.yml
1 2 3 4
|
ports: - 180:80 - 1443:443 - 4443:4443
|
安装docker-compose
复制
1 2 3
|
[root@hdss7-200 harbor]# yum install docker-compose -y [root@hdss7-200 harbor]# rpm -qa docker-compose docker-compose-1.18.0-2.el7.noarch
|
安装harbor
复制/opt/harbor
1
|
[root@hdss7-200 harbor]# ./install.sh
|
检查harbor启动情况
复制
1 2 3 4 5 6 7 8 9 10 11 12 13
|
[root@hdss7-200 harbor]# docker-compose ps Name Command State Ports -------------------------------------------------------------------------------------------------------------------------------- harbor-adminserver /harbor/start.sh Up harbor-core /harbor/start.sh Up harbor-db /entrypoint.sh postgres Up 5432/tcp harbor-jobservice /harbor/start.sh Up harbor-log /bin/sh -c /usr/local/bin/ ... Up 127.0.0.1:1514->10514/tcp harbor-portal nginx -g daemon off; Up 80/tcp nginx nginx -g daemon off; Up 0.0.0.0:1443->443/tcp, 0.0.0.0:4443->4443/tcp, 0.0.0.0:180->80/tcp redis docker-entrypoint.sh redis ... Up 6379/tcp registry /entrypoint.sh /etc/regist ... Up 5000/tcp registryctl /harbor/start.sh Up
|
配置harbor的dns内网解析
复制/var/named/od.com.zone
1
|
harbor 60 IN A 10.4.7.200
|
检查
复制
1 2
|
[root@hdss7-200 harbor]# dig -t A harbor.od.com @10.4.7.11 +short 10.4.7.200
|
安装nginx并配置
安装
复制
1 2 3
|
[root@hdss7-200 harbor]# yum install nginx -y [root@hdss7-200 harbor]# rpm -qa nginx nginx-1.12.2-2.el7.x86_64
|
配置
复制/etc/nginx/conf.d/harbor.od.com.conf
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26
|
server { listen 80; server_name harbor.od.com;
client_max_body_size 1000m;
location / { proxy_pass http://127.0.0.1:180; } } server { listen 443 ssl; server_name harbor.od.com;
ssl_certificate "certs/harbor.od.com.pem"; ssl_certificate_key "certs/harbor.od.com-key.pem"; ssl_session_cache shared:SSL:1m; ssl_session_timeout 10m; ssl_ciphers HIGH:!aNULL:!MD5; ssl_prefer_server_ciphers on; client_max_body_size 1000m;
location / { proxy_pass http://127.0.0.1:180; } }
|
注意:这里需要自签ssl证书,自签过程略
(umask 077; openssl genrsa -out od.key 2048)
openssl req -new -key od.key -out od.csr -subj “/CN=*.od.com/ST=Beijing/L=beijing/O=od/OU=ops”
openssl x509 -req -in od.csr -CA ca.crt -CAkey ca.key -CAcreateserial -out od.crt -days 365
启动
复制
1 2 3 4 5
|
[root@hdss7-200 harbor]# nginx
[root@hdss7-200 harbor]# netstat -luntp|grep nginx tcp 0 0 0.0.0.0:80 0.0.0.0:* LISTEN 6590/nginx: master tcp 0 0 0.0.0.0:443 0.0.0.0:* LISTEN 6590/nginx: master
|
- 用户名:admin
- 密码: Harbor12345
部署Master节点服务
部署etcd集群
集群规划
主机名 |
角色 |
ip |
HDSS7-12.host.com |
etcd lead |
10.4.7.12 |
HDSS7-21.host.com |
etcd follow |
10.4.7.21 |
HDSS7-22.host.com |
etcd follow |
10.4.7.22 |
注意:这里部署文档以HDSS7-12.host.com
主机为例,另外两台主机安装部署方法类似
创建生成证书签名请求(csr)的JSON配置文件
运维主机HDSS7-200.host.com
上:
复制/opt/certs/etcd-peer-csr.json
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22
|
{ "CN": "etcd-peer", "hosts": [ "10.4.7.11", "10.4.7.12", "10.4.7.21", "10.4.7.22" ], "key": { "algo": "rsa", "size": 2048 }, "names": [ { "C": "CN", "ST": "beijing", "L": "beijing", "O": "od", "OU": "ops" } ] }
|
生成etcd证书和私钥
复制/opt/certs
1 2 3 4 5 6 7 8 9 10 11
|
[root@hdss7-200 certs]# cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=peer etcd-peer-csr.json | cfssljson -bare etcd-peer 2019/01/18 09:35:09 [INFO] generate received request 2019/01/18 09:35:09 [INFO] received CSR 2019/01/18 09:35:09 [INFO] generating key: rsa-2048
2019/01/18 09:35:09 [INFO] encoded CSR 2019/01/18 09:35:10 [INFO] signed certificate with serial number 324191491384928915605254764031096067872154649010 2019/01/18 09:35:10 [WARNING] This certificate lacks a "hosts" field. This makes it unsuitable for websites. For more information see the Baseline Requirements for the Issuance and Management of Publicly-Trusted Certificates, v.1.1.6, from the CA/Browser Forum (https://cabforum.org); specifically, section 10.2.3 ("Information Requirements").
|
检查生成的证书、私钥
复制/opt/certs
1 2 3 4 5
|
[root@hdss7-200 certs]# ls -l|grep etcd -rw-r--r-- 1 root root 387 Jan 18 12:32 etcd-peer-csr.json -rw------- 1 root root 1679 Jan 18 12:32 etcd-peer-key.pem -rw-r--r-- 1 root root 1074 Jan 18 12:32 etcd-peer.csr -rw-r--r-- 1 root root 1432 Jan 18 12:32 etcd-peer.pem
|
创建etcd用户
HDSS7-12.host.com
上:
复制
1
|
[root@hdss7-12 ~]# useradd -s /sbin/nologin -M etcd
|
下载软件,解压,做软连接
etcd下载地址
HDSS7-12.host.com
上:
复制/opt/src
1 2 3 4 5 6 7 8 9 10
|
[root@hdss7-12 src]# ls -l total 9604 -rw-r--r-- 1 root root 9831476 Jan 18 10:45 etcd-v3.1.18-linux-amd64.tar.gz [root@hdss7-12 src]# tar xf etcd-v3.1.18-linux-amd64.tar.gz -C /opt [root@hdss7-12 src]# ln -s /opt/etcd-v3.1.18-linux-amd64 /opt/etcd [root@hdss7-12 src]# ls -l /opt total 0 lrwxrwxrwx 1 root root 24 Jan 18 14:21 etcd -> etcd-v3.1.18-linux-amd64 drwxr-xr-x 4 478493 89939 166 Jun 16 2018 etcd-v3.1.18-linux-amd64 drwxr-xr-x 2 root root 45 Jan 18 14:21 src
|
创建目录,拷贝证书、私钥
HDSS7-12.host.com
上:
复制
1 2 3
|
[root@hdss7-12 src]# mkdir -p /data/etcd /data/logs/etcd-server [root@hdss7-12 src]# chown -R etcd.etcd /data/etcd /data/logs/etcd-server/ [root@hdss7-12 src]# mkdir -p /opt/etcd/certs
|
将运维主机上生成的ca.pem
、etcd-peer-key.pem
、etcd-peer.pem
拷贝到/opt/etcd/certs
目录中,注意私钥文件权限600
复制/opt/etcd/certs
1 2 3 4 5 6 7
|
[root@hdss7-12 certs]# chmod 600 etcd-peer-key.pem [root@hdss7-12 certs]# chown -R etcd.etcd /opt/etcd/certs/ [root@hdss7-12 certs]# ls -l total 12 -rw-r--r-- 1 etcd etcd 1354 Jan 18 14:45 ca.pem -rw------- 1 etcd etcd 1679 Jan 18 17:00 etcd-peer-key.pem -rw-r--r-- 1 etcd etcd 1444 Jan 18 17:02 etcd-peer.pem
|
创建etcd服务启动脚本
HDSS7-12.host.com
上:
复制/opt/etcd/etcd-server-startup.sh
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
|
#!/bin/sh ./etcd --name etcd-server-7-12 \ --data-dir /data/etcd/etcd-server \ --listen-peer-urls https://10.4.7.12:2380 \ --listen-client-urls https://10.4.7.12:2379,http://127.0.0.1:2379 \ --quota-backend-bytes 8000000000 \ --initial-advertise-peer-urls https://10.4.7.12:2380 \ --advertise-client-urls https://10.4.7.12:2379,http://127.0.0.1:2379 \ --initial-cluster etcd-server-7-12=https://10.4.7.12:2380,etcd-server-7-21=https://10.4.7.21:2380,etcd-server-7-22=https://10.4.7.22:2380 \ --ca-file ./certs/ca.pem \ --cert-file ./certs/etcd-peer.pem \ --key-file ./certs/etcd-peer-key.pem \ --client-cert-auth \ --trusted-ca-file ./certs/ca.pem \ --peer-ca-file ./certs/ca.pem \ --peer-cert-file ./certs/etcd-peer.pem \ --peer-key-file ./certs/etcd-peer-key.pem \ --peer-client-cert-auth \ --peer-trusted-ca-file ./certs/ca.pem \ --log-output stdout
|
注意:etcd集群各主机的启动脚本略有不同,部署其他节点时注意修改。
调整权限和目录
HDSS7-12.host.com
上:
复制
1 2
|
[root@hdss7-12 certs]# chmod +x /opt/etcd/etcd-server-startup.sh [root@hdss7-12 certs]# mkdir -p /data/logs/etcd-server
|
安装supervisor软件
HDSS7-12.host.com
上:
复制
1 2 3
|
[root@hdss7-12 certs]# yum install supervisor -y [root@hdss7-12 certs]# systemctl start supervisord [root@hdss7-12 certs]# systemctl enable supervisord
|
创建etcd-server的启动配置
HDSS7-12.host.com
上:
复制/etc/supervisord.d/etcd-server.ini
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
[program:etcd-server-7-12] command=/opt/etcd/etcd-server-startup.sh ; the program (relative uses PATH, can take args) numprocs=1 ; number of processes copies to start (def 1) directory=/opt/etcd ; directory to cwd to before exec (def no cwd) autostart=true ; start at supervisord start (default: true) autorestart=true ; retstart at unexpected quit (default: true) startsecs=22 ; number of secs prog must stay running (def. 1) startretries=3 ; max # of serial start failures (default 3) exitcodes=0,2 ; 'expected' exit codes for process (default 0,2) stopsignal=QUIT ; signal used to kill process (default TERM) stopwaitsecs=10 ; max num secs to wait b4 SIGKILL (default 10) user=etcd ; setuid to this UNIX account to run the program redirect_stderr=false ; redirect proc stderr to stdout (default false) stdout_logfile=/data/logs/etcd-server/etcd.stdout.log ; stdout log path, NONE for none; default AUTO stdout_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stdout_logfile_backups=4 ; # of stdout logfile backups (default 10) stdout_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stdout_events_enabled=false ; emit events on stdout writes (default false) stderr_logfile=/data/logs/etcd-server/etcd.stderr.log ; stderr log path, NONE for none; default AUTO stderr_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stderr_logfile_backups=4 ; # of stderr logfile backups (default 10) stderr_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stderr_events_enabled=false ; emit events on stderr writes (default false)
|
注意:etcd集群各主机启动配置略有不同,配置其他节点时注意修改。
启动etcd服务并检查
HDSS7-12.host.com
上:
复制
1 2 3 4
|
[root@hdss7-12 certs]# supervisorctl start all etcd-server-7-12: started [root@hdss7-12 certs]# supervisorctl status etcd-server-7-12 RUNNING pid 6692, uptime 0:00:05
|
安装部署启动检查所有集群规划主机上的etcd服务
略
检查集群状态
3台均启动后,检查集群状态
复制
1 2 3 4 5 6 7 8 9 10
|
[root@hdss7-12 ~]# /opt/etcd/etcdctl cluster-health member 988139385f78284 is healthy: got healthy result from http://127.0.0.1:2379 member 5a0ef2a004fc4349 is healthy: got healthy result from http://127.0.0.1:2379 member f4a0cb0a765574a8 is healthy: got healthy result from http://127.0.0.1:2379 cluster is healthy
[root@hdss7-12 ~]# /opt/etcd/etcdctl member list 988139385f78284: name=etcd-server-7-22 peerURLs=https://10.4.7.22:2380 clientURLs=http://127.0.0.1:2379,https://10.4.7.22:2379 isLeader=false 5a0ef2a004fc4349: name=etcd-server-7-21 peerURLs=https://10.4.7.21:2380 clientURLs=http://127.0.0.1:2379,https://10.4.7.21:2379 isLeader=false f4a0cb0a765574a8: name=etcd-server-7-12 peerURLs=https://10.4.7.12:2380 clientURLs=http://127.0.0.1:2379,https://10.4.7.12:2379 isLeader=true
|
部署kube-apiserver集群
集群规划
主机名 |
角色 |
ip |
HDSS7-21.host.com |
kube-apiserver |
10.4.7.21 |
HDSS7-22.host.com |
kube-apiserver |
10.4.7.22 |
HDSS7-11.host.com |
4层负载均衡 |
10.4.7.11 |
HDSS7-12.host.com |
4层负载均衡 |
10.4.7.12 |
注意:这里10.4.7.11
和10.4.7.12
使用nginx做4层负载均衡器,用keepalived跑一个vip:10.4.7.10,代理两个kube-apiserver,实现高可用
这里部署文档以HDSS7-21.host.com
主机为例,另外一台运算节点安装部署方法类似
下载软件,解压,做软连接
HDSS7-21.host.com
上:
kubernetes下载地址
复制/opt/src
1 2 3 4 5 6 7 8 9
|
[root@hdss7-21 src]# ls -l|grep kubernetes -rw-r--r-- 1 root root 417761204 Jan 17 16:46 kubernetes-server-linux-amd64.tar.gz [root@hdss7-21 src]# tar xf kubernetes-server-linux-amd64.tar.gz -C /opt [root@hdss7-21 src]# mv /opt/kubernetes /opt/kubernetes-v1.13.2-linux-amd64 [root@hdss7-21 src]# ln -s /opt/kubernetes-v1.13.2-linux-amd64 /opt/kubernetes [root@hdss7-21 src]# mkdir /opt/kubernetes/server/bin/{cert,conf} [root@hdss7-21 src]# ls -l /opt|grep kubernetes lrwxrwxrwx 1 root root 31 Jan 18 10:49 kubernetes -> kubernetes-v1.13.2-linux-amd64/ drwxr-xr-x 4 root root 50 Jan 17 17:40 kubernetes-v1.13.2-linux-amd64
|
签发client证书
运维主机HDSS7-200.host.com
上:
创建生成证书签名请求(csr)的JSON配置文件
复制/opt/certs/client-csr.json
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
|
{ "CN": "k8s-node", "hosts": [ ], "key": { "algo": "rsa", "size": 2048 }, "names": [ { "C": "CN", "ST": "beijing", "L": "beijing", "O": "od", "OU": "ops" } ] }
|
生成client证书和私钥
复制
1 2 3 4 5 6 7 8 9 10
|
[root@hdss7-200 certs]# cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=client client-csr.json | cfssljson -bare client 2019/01/18 14:02:50 [INFO] generate received request 2019/01/18 14:02:50 [INFO] received CSR 2019/01/18 14:02:50 [INFO] generating key: rsa-2048 2019/01/18 14:02:51 [INFO] encoded CSR 2019/01/18 14:02:51 [INFO] signed certificate with serial number 423108651040279300242366884100637974155370861448 2019/01/18 14:02:51 [WARNING] This certificate lacks a "hosts" field. This makes it unsuitable for websites. For more information see the Baseline Requirements for the Issuance and Management of Publicly-Trusted Certificates, v.1.1.6, from the CA/Browser Forum (https://cabforum.org); specifically, section 10.2.3 ("Information Requirements").
|
检查生成的证书、私钥
复制
1 2 3 4
|
[root@hdss7-200 certs]# ls -l|grep client -rw------- 1 root root 1679 Jan 21 11:13 client-key.pem -rw-r--r-- 1 root root 989 Jan 21 11:13 client.csr -rw-r--r-- 1 root root 1367 Jan 21 11:13 client.pem
|
签发kube-apiserver证书
运维主机HDSS7-200.host.com
上:
创建生成证书签名请求(csr)的JSON配置文件
复制/opt/certs/apiserver-csr.json
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28
|
{ "CN": "apiserver", "hosts": [ "127.0.0.1", "192.168.0.1", "kubernetes.default", "kubernetes.default.svc", "kubernetes.default.svc.cluster", "kubernetes.default.svc.cluster.local", "10.4.7.10", "10.4.7.21", "10.4.7.22", "10.4.7.23" ], "key": { "algo": "rsa", "size": 2048 }, "names": [ { "C": "CN", "ST": "beijing", "L": "beijing", "O": "od", "OU": "ops" } ] }
|
生成kube-apiserver证书和私钥
复制
1 2 3 4 5 6 7 8 9 10
|
[root@hdss7-200 certs]# cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=server apiserver-csr.json | cfssljson -bare apiserver 2019/01/18 14:05:44 [INFO] generate received request 2019/01/18 14:05:44 [INFO] received CSR 2019/01/18 14:05:44 [INFO] generating key: rsa-2048 2019/01/18 14:05:46 [INFO] encoded CSR 2019/01/18 14:05:46 [INFO] signed certificate with serial number 633406650960616624590510576685608580490218676227 2019/01/18 14:05:46 [WARNING] This certificate lacks a "hosts" field. This makes it unsuitable for websites. For more information see the Baseline Requirements for the Issuance and Management of Publicly-Trusted Certificates, v.1.1.6, from the CA/Browser Forum (https://cabforum.org); specifically, section 10.2.3 ("Information Requirements").
|
检查生成的证书、私钥
复制
1 2 3 4 5 6
|
[root@hdss7-200 certs]# ls -l|grep apiserver total 72 -rw-r--r-- 1 root root 406 Jan 21 14:10 apiserver-csr.json -rw------- 1 root root 1675 Jan 21 14:11 apiserver-key.pem -rw-r--r-- 1 root root 1082 Jan 21 14:11 apiserver.csr -rw-r--r-- 1 root root 1599 Jan 21 14:11 apiserver.pem
|
拷贝证书至各运算节点,并创建配置
HDSS7-21.host.com
上:
拷贝证书、私钥,注意私钥文件属性600
复制/opt/kubernetes/server/bin/cert
1 2 3 4 5 6 7 8
|
[root@hdss7-21 cert]# ls -l /opt/kubernetes/server/bin/cert total 40 -rw------- 1 root root 1676 Jan 21 16:39 apiserver-key.pem -rw-r--r-- 1 root root 1599 Jan 21 16:36 apiserver.pem -rw------- 1 root root 1675 Jan 21 13:55 ca-key.pem -rw-r--r-- 1 root root 1354 Jan 21 13:50 ca.pem -rw------- 1 root root 1679 Jan 21 13:53 client-key.pem -rw-r--r-- 1 root root 1368 Jan 21 13:53 client.pem
|
创建配置
复制/opt/kubernetes/server/bin/conf/audit.yaml
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68
|
apiVersion: audit.k8s.io/v1beta1 # This is required. kind: Policy # Don't generate audit events for all requests in RequestReceived stage. omitStages: - "RequestReceived" rules: # Log pod changes at RequestResponse level - level: RequestResponse resources: - group: "" # Resource "pods" doesn't match requests to any subresource of pods, # which is consistent with the RBAC policy. resources: ["pods"] # Log "pods/log", "pods/status" at Metadata level - level: Metadata resources: - group: "" resources: ["pods/log", "pods/status"]
# Don't log requests to a configmap called "controller-leader" - level: None resources: - group: "" resources: ["configmaps"] resourceNames: ["controller-leader"]
# Don't log watch requests by the "system:kube-proxy" on endpoints or services - level: None users: ["system:kube-proxy"] verbs: ["watch"] resources: - group: "" # core API group resources: ["endpoints", "services"]
# Don't log authenticated requests to certain non-resource URL paths. - level: None userGroups: ["system:authenticated"] nonResourceURLs: - "/api*" # Wildcard matching. - "/version"
# Log the request body of configmap changes in kube-system. - level: Request resources: - group: "" # core API group resources: ["configmaps"] # This rule only applies to resources in the "kube-system" namespace. # The empty string "" can be used to select non-namespaced resources. namespaces: ["kube-system"]
# Log configmap and secret changes in all other namespaces at the Metadata level. - level: Metadata resources: - group: "" # core API group resources: ["secrets", "configmaps"]
# Log all other resources in core and extensions at the Request level. - level: Request resources: - group: "" # core API group - group: "extensions" # Version of group should NOT be included.
# A catch-all rule to log all other requests at the Metadata level. - level: Metadata # Long-running requests like watches that fall under this rule will not # generate an audit event in RequestReceived. omitStages: - "RequestReceived"
|
创建启动脚本
HDSS7-21.host.com
上:
复制/opt/kubernetes/server/bin/kube-apiserver.sh
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
#!/bin/bash ./kube-apiserver \ --apiserver-count 2 \ --audit-log-path /data/logs/kubernetes/kube-apiserver/audit-log \ --audit-policy-file ./conf/audit.yaml \ --authorization-mode RBAC \ --client-ca-file ./cert/ca.pem \ --requestheader-client-ca-file ./cert/ca.pem \ --enable-admission-plugins NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota \ --etcd-cafile ./cert/ca.pem \ --etcd-certfile ./cert/client.pem \ --etcd-keyfile ./cert/client-key.pem \ --etcd-servers https://10.4.7.12:2379,https://10.4.7.21:2379,https://10.4.7.22:2379 \ --service-account-key-file ./cert/ca-key.pem \ --service-cluster-ip-range 192.168.0.0/16 \ --service-node-port-range 3000-29999 \ --target-ram-mb=1024 \ --kubelet-client-certificate ./cert/client.pem \ --kubelet-client-key ./cert/client-key.pem \ --log-dir /data/logs/kubernetes/kube-apiserver \ --tls-cert-file ./cert/apiserver.pem \ --tls-private-key-file ./cert/apiserver-key.pem \ --v 2
|
调整权限和目录
HDSS7-21.host.com
上:
复制/opt/kubernetes/server/bin
1 2
|
[root@hdss7-21 bin]# chmod +x /opt/kubernetes/server/bin/kube-apiserver.sh [root@hdss7-21 bin]# mkdir -p /data/logs/kubernetes/kube-apiserver
|
创建supervisor配置
HDSS7-21.host.com
上:
复制/etc/supervisord.d/kube-apiserver.ini
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
[program:kube-apiserver] command=/opt/kubernetes/server/bin/kube-apiserver.sh ; the program (relative uses PATH, can take args) numprocs=1 ; number of processes copies to start (def 1) directory=/opt/kubernetes/server/bin ; directory to cwd to before exec (def no cwd) autostart=true ; start at supervisord start (default: true) autorestart=true ; retstart at unexpected quit (default: true) startsecs=22 ; number of secs prog must stay running (def. 1) startretries=3 ; max # of serial start failures (default 3) exitcodes=0,2 ; 'expected' exit codes for process (default 0,2) stopsignal=QUIT ; signal used to kill process (default TERM) stopwaitsecs=10 ; max num secs to wait b4 SIGKILL (default 10) user=root ; setuid to this UNIX account to run the program redirect_stderr=false ; redirect proc stderr to stdout (default false) stdout_logfile=/data/logs/kubernetes/kube-apiserver/apiserver.stdout.log ; stdout log path, NONE for none; default AUTO stdout_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stdout_logfile_backups=4 ; # of stdout logfile backups (default 10) stdout_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stdout_events_enabled=false ; emit events on stdout writes (default false) stderr_logfile=/data/logs/kubernetes/kube-apiserver/apiserver.stderr.log ; stderr log path, NONE for none; default AUTO stderr_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stderr_logfile_backups=4 ; # of stderr logfile backups (default 10) stderr_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stderr_events_enabled=false ; emit events on stderr writes (default false)
|
启动服务并检查
HDSS7-21.host.com
上:
复制
1 2 3 4 5
|
[root@hdss7-21 bin]# supervisorctl update kube-apiserverr: added process group [root@hdss7-21 bin]# supervisorctl status etcd-server-7-21 RUNNING pid 6661, uptime 1 day, 8:41:13 kube-apiserver RUNNING pid 43765, uptime 2:09:41
|
安装部署启动检查所有集群规划主机上的kube-apiserver
略
配4层反向代理
HDSS7-11.host.com
,HDSS7-12.host.com
上:
nginx配置
复制/etc/nginx/nginx.conf
1 2 3 4 5 6 7 8 9 10 11 12
|
stream { upstream kube-apiserver { server 10.4.7.21:6443 max_fails=3 fail_timeout=30s; server 10.4.7.22:6443 max_fails=3 fail_timeout=30s; } server { listen 7443; proxy_connect_timeout 2s; proxy_timeout 900s; proxy_pass kube-apiserver; } }
|
keepalived配置
check_port.sh
复制/etc/keepalived/check_port.sh
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
|
#!/bin/bash #keepalived 监控端口脚本 #使用方法: #在keepalived的配置文件中 #vrrp_script check_port {#创建一个vrrp_script脚本,检查配置 # script "/etc/keepalived/check_port.sh 6379" #配置监听的端口 # interval 2 #检查脚本的频率,单位(秒) #} CHK_PORT=$1 if [ -n "$CHK_PORT" ];then PORT_PROCESS=`ss -lt|grep $CHK_PORT|wc -l` if [ $PORT_PROCESS -eq 0 ];then echo "Port $CHK_PORT Is Not Used,End." exit 1 fi else echo "Check Port Cant Be Empty!" fi
|
keepalived主
HDSS7-11.host.com
上
复制
1 2
|
[root@hdss7-11 ~]# rpm -qa keepalived keepalived-1.3.5-6.el7.x86_64
|
复制/etc/keepalived/keepalived.conf
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33
|
! Configuration File for keepalived
global_defs { router_id 10.4.7.11
}
vrrp_script chk_nginx { script "/etc/keepalived/check_port.sh 7443" interval 2 weight -20 }
vrrp_instance VI_1 { state MASTER interface eth0 virtual_router_id 251 priority 100 advert_int 1 mcast_src_ip 10.4.7.11 nopreempt
authentication { auth_type PASS auth_pass 11111111 } track_script { chk_nginx } virtual_ipaddress { 10.4.7.10 } }
|
keepalived备
HDSS7-12.host.com
上
复制
1 2
|
[root@hdss7-12 ~]# rpm -qa keepalived keepalived-1.3.5-6.el7.x86_64
|
复制/etc/keepalived/keepalived.conf
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
|
! Configuration File for keepalived global_defs { router_id 10.4.7.12 } vrrp_script chk_nginx { script "/etc/keepalived/check_port.sh 7443" interval 2 weight -20 } vrrp_instance VI_1 { state BACKUP interface eth0 virtual_router_id 251 mcast_src_ip 10.4.7.12 priority 90 advert_int 1 authentication { auth_type PASS auth_pass 11111111 } track_script { chk_nginx } virtual_ipaddress { 10.4.7.10 } }
|
启动代理并检查
HDSS7-11.host.com
,HDSS7-12.host.com
上:
-
启动
复制
1 2 3 4 5 6 7
|
[root@hdss7-11 ~]# systemctl start keepalived [root@hdss7-11 ~]# systemctl enable keepalived [root@hdss7-11 ~]# nginx -s reload
[root@hdss7-12 ~]# systemctl start keepalived [root@hdss7-12 ~]# systemctl enable keepalived [root@hdss7-12 ~]# nginx -s reload
|
-
检查
复制
1 2 3 4 5 6 7 8
|
[root@hdss7-11 ~]## netstat -luntp|grep 7443 tcp 0 0 0.0.0.0:7443 0.0.0.0:* LISTEN 17970/nginx: master [root@hdss7-12 ~]## netstat -luntp|grep 7443 tcp 0 0 0.0.0.0:7443 0.0.0.0:* LISTEN 17970/nginx: master [root@hdss7-11 ~]# ip add|grep 10.4.9.10 inet 10.9.7.10/32 scope global vir0 [root@hdss7-11 ~]# ip add|grep 10.4.9.10 (空)
|
部署controller-manager
集群规划
主机名 |
角色 |
ip |
HDSS7-21.host.com |
controller-manager |
10.4.7.21 |
HDSS7-22.host.com |
controller-manager |
10.4.7.22 |
注意:这里部署文档以HDSS7-21.host.com
主机为例,另外一台运算节点安装部署方法类似
创建启动脚本
HDSS7-21.host.com
上:
复制/opt/kubernetes/server/bin/kube-controller-manager.sh
1 2 3 4 5 6 7 8 9 10
|
#!/bin/sh ./kube-controller-manager \ --cluster-cidr 172.7.0.0/16 \ --leader-elect true \ --log-dir /data/logs/kubernetes/kube-controller-manager \ --master http://127.0.0.1:8080 \ --service-account-private-key-file ./cert/ca-key.pem \ --service-cluster-ip-range 192.168.0.0/16 \ --root-ca-file ./cert/ca.pem \ --v 2
|
调整文件权限,创建目录
HDSS7-21.host.com
上:
复制/opt/kubernetes/server/bin
1 2
|
[root@hdss7-21 bin]# chmod +x /opt/kubernetes/server/bin/kube-controller-manager.sh [root@hdss7-21 bin]# mkdir -p /data/logs/kubernetes/kube-controller-manager
|
创建supervisor配置
HDSS7-21.host.com
上:
复制/etc/supervisord.d/kube-conntroller-manager.ini
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
[program:kube-controller-manager] command=/opt/kubernetes/server/bin/kube-controller-manager.sh ; the program (relative uses PATH, can take args) numprocs=1 ; number of processes copies to start (def 1) directory=/opt/kubernetes/server/bin ; directory to cwd to before exec (def no cwd) autostart=true ; start at supervisord start (default: true) autorestart=true ; retstart at unexpected quit (default: true) startsecs=22 ; number of secs prog must stay running (def. 1) startretries=3 ; max # of serial start failures (default 3) exitcodes=0,2 ; 'expected' exit codes for process (default 0,2) stopsignal=QUIT ; signal used to kill process (default TERM) stopwaitsecs=10 ; max num secs to wait b4 SIGKILL (default 10) user=root ; setuid to this UNIX account to run the program redirect_stderr=false ; redirect proc stderr to stdout (default false) stdout_logfile=/data/logs/kubernetes/kube-controller-manager/controll.stdout.log ; stdout log path, NONE for none; default AUTO stdout_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stdout_logfile_backups=4 ; # of stdout logfile backups (default 10) stdout_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stdout_events_enabled=false ; emit events on stdout writes (default false) stderr_logfile=/data/logs/kubernetes/kube-controller-manager/controll.stderr.log ; stderr log path, NONE for none; default AUTO stderr_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stderr_logfile_backups=4 ; # of stderr logfile backups (default 10) stderr_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stderr_events_enabled=false ; emit events on stderr writes (default false)
|
启动服务并检查
HDSS7-21.host.com
上:
复制
1 2 3 4 5 6
|
[root@hdss7-21 bin]# supervisorctl update kube-controller-manager: added process group [root@hdss7-21 bin]# supervisorctl status etcd-server-7-21 RUNNING pid 6661, uptime 1 day, 8:41:13 kube-apiserver RUNNING pid 43765, uptime 2:09:41 kube-controller-manager RUNNING pid 44230, uptime 2:05:01
|
安装部署启动检查所有集群规划主机上的kube-controller-manager服务
略
部署kube-scheduler
集群规划
主机名 |
角色 |
ip |
HDSS7-21.host.com |
kube-scheduler |
10.4.7.21 |
HDSS7-22.host.com |
kube-scheduler |
10.4.7.22 |
注意:这里部署文档以HDSS7-21.host.com
主机为例,另外一台运算节点安装部署方法类似
创建启动脚本
HDSS7-21.host.com
上:
复制/opt/kubernetes/server/bin/kube-scheduler.sh
1 2 3 4 5 6
|
#!/bin/sh ./kube-scheduler \ --leader-elect \ --log-dir /data/logs/kubernetes/kube-scheduler \ --master http://127.0.0.1:8080 \ --v 2
|
调整文件权限,创建目录
HDSS7-21.host.com
上:
复制/opt/kubernetes/server/bin
1 2
|
[root@hdss7-21 bin]# chmod +x /opt/kubernetes/server/bin/kube-scheduler.sh [root@hdss7-21 bin]# mkdir -p /data/logs/kubernetes/kube-scheduler
|
创建supervisor配置
HDSS7-21.host.com
上:
复制/etc/supervisord.d/kube-scheduler.ini
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
[program:kube-scheduler] command=/opt/kubernetes/server/bin/kube-scheduler.sh ; the program (relative uses PATH, can take args) numprocs=1 ; number of processes copies to start (def 1) directory=/opt/kubernetes/server/bin ; directory to cwd to before exec (def no cwd) autostart=true ; start at supervisord start (default: true) autorestart=true ; retstart at unexpected quit (default: true) startsecs=22 ; number of secs prog must stay running (def. 1) startretries=3 ; max # of serial start failures (default 3) exitcodes=0,2 ; 'expected' exit codes for process (default 0,2) stopsignal=QUIT ; signal used to kill process (default TERM) stopwaitsecs=10 ; max num secs to wait b4 SIGKILL (default 10) user=root ; setuid to this UNIX account to run the program redirect_stderr=false ; redirect proc stderr to stdout (default false) stdout_logfile=/data/logs/kubernetes/kube-scheduler/scheduler.stdout.log ; stdout log path, NONE for none; default AUTO stdout_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stdout_logfile_backups=4 ; # of stdout logfile backups (default 10) stdout_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stdout_events_enabled=false ; emit events on stdout writes (default false) stderr_logfile=/data/logs/kubernetes/kube-scheduler/scheduler.stderr.log ; stderr log path, NONE for none; default AUTO stderr_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stderr_logfile_backups=4 ; # of stderr logfile backups (default 10) stderr_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stderr_events_enabled=false ; emit events on stderr writes (default false)
|
启动服务并检查
HDSS7-21.host.com
上:
复制
1 2 3 4 5 6 7
|
[root@hdss7-21 bin]# supervisorctl update kube-scheduler: added process group [root@hdss7-21 bin]# supervisorctl status etcd-server-7-21 RUNNING pid 6661, uptime 1 day, 8:41:13 kube-apiserver RUNNING pid 43765, uptime 2:09:41 kube-controller-manager RUNNING pid 44230, uptime 2:05:01 kube-scheduler RUNNING pid 44779, uptime 2:02:27
|
安装部署启动检查所有集群规划主机上的kube-scheduler服务
略
部署Node节点服务
部署kubelet
集群规划
主机名 |
角色 |
ip |
HDSS7-21.host.com |
kubelet |
10.4.7.21 |
HDSS7-22.host.com |
kubelet |
10.4.7.22 |
注意:这里部署文档以HDSS7-21.host.com
主机为例,另外一台运算节点安装部署方法类似
签发kubelet证书
运维主机HDSS7-200.host.com
上:
创建生成证书签名请求(csr)的JSON配置文件
复制kubelet-csr.json
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28
|
{ "CN": "kubelet-node", "hosts": [ "127.0.0.1", "10.4.7.10", "10.4.7.21", "10.4.7.22", "10.4.7.23", "10.4.7.24", "10.4.7.25", "10.4.7.26", "10.4.7.27", "10.4.7.28" ], "key": { "algo": "rsa", "size": 2048 }, "names": [ { "C": "CN", "ST": "beijing", "L": "beijing", "O": "od", "OU": "ops" } ] }
|
生成kubelet证书和私钥
复制/opt/certs
1 2 3 4 5 6 7 8 9 10
|
[root@hdss7-200 certs]# cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=server kubelet-csr.json | cfssljson -bare kubelet 2019/01/18 17:51:16 [INFO] generate received request 2019/01/18 17:51:16 [INFO] received CSR 2019/01/18 17:51:16 [INFO] generating key: rsa-2048 2019/01/18 17:51:17 [INFO] encoded CSR 2019/01/18 17:51:17 [INFO] signed certificate with serial number 48870268157415133698067712395152321546974943470 2019/01/18 17:51:17 [WARNING] This certificate lacks a "hosts" field. This makes it unsuitable for websites. For more information see the Baseline Requirements for the Issuance and Management of Publicly-Trusted Certificates, v.1.1.6, from the CA/Browser Forum (https://cabforum.org); specifically, section 10.2.3 ("Information Requirements").
|
检查生成的证书、私钥
复制/opt/certs
1 2 3 4 5 6
|
[root@hdss7-200 certs]# ls -l|grep kubelet total 88 -rw-r--r-- 1 root root 415 Jan 22 16:58 kubelet-csr.json -rw------- 1 root root 1679 Jan 22 17:00 kubelet-key.pem -rw-r--r-- 1 root root 1086 Jan 22 17:00 kubelet.csr -rw-r--r-- 1 root root 1456 Jan 22 17:00 kubelet.pem
|
拷贝证书至各运算节点,并创建配置
HDSS7-21.host.com
上:
拷贝证书、私钥,注意私钥文件属性600
复制/opt/kubernetes/server/bin/cert
1 2 3 4 5 6 7 8 9 10
|
[root@hdss7-21 cert]# ls -l /opt/kubernetes/server/bin/cert total 40 -rw------- 1 root root 1676 Jan 21 16:39 apiserver-key.pem -rw-r--r-- 1 root root 1599 Jan 21 16:36 apiserver.pem -rw------- 1 root root 1675 Jan 21 13:55 ca-key.pem -rw-r--r-- 1 root root 1354 Jan 21 13:50 ca.pem -rw------- 1 root root 1679 Jan 21 13:53 client-key.pem -rw-r--r-- 1 root root 1368 Jan 21 13:53 client.pem -rw------- 1 root root 1679 Jan 22 17:00 kubelet-key.pem -rw-r--r-- 1 root root 1456 Jan 22 17:00 kubelet.pem
|
创建配置
HDSS7-21.host.com
上:
给kubectl创建软连接
复制/opt/kubernetes/server/bin
1 2 3
|
[root@hdss7-21 bin]# ln -s /opt/kubernetes/server/bin/kubectl /usr/bin/kubectl [root@hdss7-21 bin]# which kubectl /usr/bin/kubectl
|
set-cluster
注意:在conf目录下
复制/opt/kubernetes/server/conf
1 2 3 4 5 6 7
|
[root@hdss7-21 conf]# kubectl config set-cluster myk8s \ --certificate-authority=/opt/kubernetes/server/bin/cert/ca.pem \ --embed-certs=true \ --server=https://10.4.7.10:7443 \ --kubeconfig=kubelet.kubeconfig
Cluster "myk8s" set.
|
set-credentials
注意:在conf目录下
复制/opt/kubernetes/server/conf
1 2 3
|
[root@hdss7-21 conf]# kubectl config set-credentials k8s-node --client-certificate=/opt/kubernetes/server/bin/cert/client.pem --client-key=/opt/kubernetes/server/bin/cert/client-key.pem --embed-certs=true --kubeconfig=kubelet.kubeconfig
User "k8s-node" set.
|
set-context
注意:在conf目录下
复制/opt/kubernetes/server/conf
1 2 3 4 5 6
|
[root@hdss7-21 conf]# kubectl config set-context myk8s-context \ --cluster=myk8s \ --user=k8s-node \ --kubeconfig=kubelet.kubeconfig
Context "myk8s-context" created.
|
use-context
注意:在conf目录下
复制/opt/kubernetes/server/conf
1 2 3
|
[root@hdss7-21 conf]# kubectl config use-context myk8s-context --kubeconfig=kubelet.kubeconfig
Switched to context "myk8s-context".
|
k8s-node.yaml
复制/opt/kubernetes/server/bin/conf/k8s-node.yaml
1 2 3 4 5 6 7 8 9 10 11 12
|
apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRoleBinding metadata: name: k8s-node roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: system:node subjects: - apiGroup: rbac.authorization.k8s.io kind: User name: k8s-node
|
复制/opt/kubernetes/server/conf
1 2 3
|
[root@hdss7-21 conf]# kubectl create -f k8s-node.yaml
clusterrolebinding.rbac.authorization.k8s.io/k8s-node created
|
复制/opt/kubernetes/server/conf
1 2 3
|
[root@hdss7-21 conf]# kubectl get clusterrolebinding k8s-node NAME AGE k8s-node 3m
|
准备infra_pod基础镜像
运维主机HDSS7-200.host.com
上:
下载
复制
1 2 3 4 5 6 7 8
|
[root@hdss7-200 ~]# docker pull xplenty/rhel7-pod-infrastructure:v3.4 Trying to pull repository docker.io/xplenty/rhel7-pod-infrastructure ... sha256:9314554780673b821cb7113d8c048a90d15077c6e7bfeebddb92a054a1f84843: Pulling from docker.io/xplenty/rhel7-pod-infrastructure 615bc035f9f8: Pull complete 1c5fd9dfeaa8: Pull complete 7653a8c7f937: Pull complete Digest: sha256:9314554780673b821cb7113d8c048a90d15077c6e7bfeebddb92a054a1f84843 Status: Downloaded newer image for docker.io/xplenty/rhel7-pod-infrastructure:v3.4
|
提交至私有仓库(harbor)中
复制/root/.docker/config.json
1 2 3 4 5 6 7
|
{ "auths": { "harbor.od.com": { "auth": "YWRtaW46SGFyYm9yMTIzNDU=" } } }
|
这里代表:用户名admin,密码Harbor12345
[root@hdss7-200 ~]# echo YWRtaW46SGFyYm9yMTIzNDU=|base64 -d
admin:Harbor12345
注意:也可以在各运算节点使用docker login harbor.od.com,输入用户名,密码
复制
1 2 3
|
[root@hdss7-200 ~]# docker images|grep v3.4 xplenty/rhel7-pod-infrastructure v3.4 34d3450d733b 2 years ago 205 MB [root@hdss7-200 ~]# docker tag 34d3450d733b harbor.od.com/k8s/pod:v3.4
|
复制
1 2 3 4 5 6
|
[root@hdss7-200 ~]# docker push harbor.od.com/k8s/pod:v3.4 The push refers to a repository [harbor.od.com/k8s/pod] ba3d4cbbb261: Pushed 0a081b45cb84: Pushed df9d2808b9a9: Pushed v3.4: digest: sha256:73cc48728e707b74f99d17b4e802d836e22d373aee901fdcaa781b056cdabf5c size: 948
|
创建kubelet启动脚本
HDSS7-21.host.com
上:
复制/opt/kubernetes/server/bin/kubelet-721.sh
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
|
#!/bin/sh ./kubelet \ --anonymous-auth=false \ --cgroup-driver systemd \ --cluster-dns 192.168.0.2 \ --cluster-domain cluster.local \ --runtime-cgroups=/systemd/system.slice --kubelet-cgroups=/systemd/system.slice \ --fail-swap-on="false" \ --client-ca-file ./cert/ca.pem \ --tls-cert-file ./cert/kubelet.pem \ --tls-private-key-file ./cert/kubelet-key.pem \ --hostname-override 10.4.7.21 \ --image-gc-high-threshold 20 \ --image-gc-low-threshold 10 \ --kubeconfig ./conf/kubelet.kubeconfig \ --log-dir /data/logs/kubernetes/kube-kubelet \ --pod-infra-container-image harbor.od.com/k8s/pod:v3.4 \ --root-dir /data/kubelet
|
注意:kubelet集群各主机的启动脚本略有不同,部署其他节点时注意修改。
检查配置,权限,创建日志目录
HDSS7-21.host.com
上:
复制/opt/kubernetes/server/conf
1 2 3 4 5
|
[root@hdss7-21 conf]# ls -l|grep kubelet.kubeconfig -rw------- 1 root root 6471 Jan 22 17:33 kubelet.kubeconfig
[root@hdss7-21 conf]# chmod +x /opt/kubernetes/server/bin/kubelet-721.sh [root@hdss7-21 conf]# mkdir -p /data/logs/kubernetes/kube-kubelet /data/kubelet
|
创建supervisor配置
HDSS7-21.host.com
上:
复制/etc/supervisord.d/kube-kubelet.ini
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
[program:kube-kubelet] command=/opt/kubernetes/server/bin/kubelet-721.sh ; the program (relative uses PATH, can take args) numprocs=1 ; number of processes copies to start (def 1) directory=/opt/kubernetes/server/bin ; directory to cwd to before exec (def no cwd) autostart=true ; start at supervisord start (default: true) autorestart=true ; retstart at unexpected quit (default: true) startsecs=22 ; number of secs prog must stay running (def. 1) startretries=3 ; max # of serial start failures (default 3) exitcodes=0,2 ; 'expected' exit codes for process (default 0,2) stopsignal=QUIT ; signal used to kill process (default TERM) stopwaitsecs=10 ; max num secs to wait b4 SIGKILL (default 10) user=root ; setuid to this UNIX account to run the program redirect_stderr=false ; redirect proc stderr to stdout (default false) stdout_logfile=/data/logs/kubernetes/kube-kubelet/kubelet.stdout.log ; stdout log path, NONE for none; default AUTO stdout_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stdout_logfile_backups=4 ; # of stdout logfile backups (default 10) stdout_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stdout_events_enabled=false ; emit events on stdout writes (default false) stderr_logfile=/data/logs/kubernetes/kube-kubelet/kubelet.stderr.log ; stderr log path, NONE for none; default AUTO stderr_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stderr_logfile_backups=4 ; # of stderr logfile backups (default 10) stderr_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stderr_events_enabled=false ; emit events on stderr writes (default false)
|
启动服务并检查
HDSS7-21.host.com
上:
复制
1 2 3 4 5 6 7 8
|
[root@hdss7-21 bin]# supervisorctl update kube-kubelet: added process group [root@hdss7-21 bin]# supervisorctl status etcd-server-7-21 RUNNING pid 9507, uptime 22:44:48 kube-apiserver RUNNING pid 9770, uptime 21:10:49 kube-controller-manager RUNNING pid 10048, uptime 18:22:10 kube-kubelet STARTING kube-scheduler RUNNING pid 10041, uptime 18:22:13
|
检查运算节点
HDSS7-21.host.com
上:
复制
1 2 3
|
[root@hdss7-21 bin]# kubectl get node NAME STATUS ROLES AGE VERSION 10.4.7.21 Ready <none> 3m v1.13.2
|
非常重要!
安装部署启动检查所有集群规划主机上的kubelet服务
略
部署kube-proxy
集群规划
主机名 |
角色 |
ip |
HDSS7-21.host.com |
kube-proxy |
10.4.7.21 |
HDSS7-22.host.com |
kube-proxy |
10.4.7.22 |
注意:这里部署文档以HDSS7-21.host.com
主机为例,另外一台运算节点安装部署方法类似
签发kube-proxy证书
运维主机HDSS7-200.host.com
上:
创建生成证书签名请求(csr)的JSON配置文件
复制/opt/certs/kube-proxy-csr.json
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
|
{ "CN": "system:kube-proxy", "key": { "algo": "rsa", "size": 2048 }, "names": [ { "C": "CN", "ST": "beijing", "L": "beijing", "O": "od", "OU": "ops" } ] }
|
生成kube-proxy证书和私钥
复制/opt/certs
1 2 3 4 5 6 7 8 9 10
|
[root@hdss7-200 certs]# cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=client kube-proxy-csr.json | cfssljson -bare kube-proxy-client 2019/01/18 18:14:23 [INFO] generate received request 2019/01/18 18:14:23 [INFO] received CSR 2019/01/18 18:14:23 [INFO] generating key: rsa-2048 2019/01/18 18:14:23 [INFO] encoded CSR 2019/01/18 18:14:23 [INFO] signed certificate with serial number 375797145588654714099258750873820528127028390681 2019/01/18 18:14:23 [WARNING] This certificate lacks a "hosts" field. This makes it unsuitable for websites. For more information see the Baseline Requirements for the Issuance and Management of Publicly-Trusted Certificates, v.1.1.6, from the CA/Browser Forum (https://cabforum.org); specifically, section 10.2.3 ("Information Requirements").
|
检查生成的证书、私钥
复制/opt/certs
1 2 3 4 5
|
[root@hdss7-200 certs]# ls -l|grep kube-proxy -rw------- 1 root root 1679 Jan 22 17:31 kube-proxy-client-key.pem -rw-r--r-- 1 root root 1005 Jan 22 17:31 kube-proxy-client.csr -rw-r--r-- 1 root root 1383 Jan 22 17:31 kube-proxy-client.pem -rw-r--r-- 1 root root 268 Jan 22 17:23 kube-proxy-csr.json
|
拷贝证书至各运算节点,并创建配置
HDSS7-21.host.com
上:
拷贝证书、私钥,注意私钥文件属性600
复制/opt/kubernetes/server/bin/cert
1 2 3 4 5 6 7 8 9 10 11 12
|
[root@hdss7-21 cert]# ls -l /opt/kubernetes/server/bin/cert total 40 -rw------- 1 root root 1676 Jan 21 16:39 apiserver-key.pem -rw-r--r-- 1 root root 1599 Jan 21 16:36 apiserver.pem -rw------- 1 root root 1675 Jan 21 13:55 ca-key.pem -rw-r--r-- 1 root root 1354 Jan 21 13:50 ca.pem -rw------- 1 root root 1679 Jan 21 13:53 client-key.pem -rw-r--r-- 1 root root 1368 Jan 21 13:53 client.pem -rw------- 1 root root 1679 Jan 22 17:00 kubelet-key.pem -rw-r--r-- 1 root root 1456 Jan 22 17:00 kubelet.pem -rw------- 1 root root 1679 Jan 22 17:31 kube-proxy-client-key.pem -rw-r--r-- 1 root root 1383 Jan 22 17:31 kube-proxy-client.pem
|
创建配置
set-cluster
注意:在conf目录下
复制/opt/kubernetes/server/bin/conf
1 2 3 4 5 6 7
|
[root@hdss7-21 conf]# kubectl config set-cluster myk8s \ --certificate-authority=/opt/kubernetes/server/bin/cert/ca.pem \ --embed-certs=true \ --server=https://10.4.7.10:7443 \ --kubeconfig=kube-proxy.kubeconfig
Cluster "myk8s" set.
|
set-credentials
注意:在conf目录下
复制/opt/kubernetes/server/bin/conf
1 2 3 4 5 6 7
|
[root@hdss7-21 conf]# kubectl config set-credentials kube-proxy \ --client-certificate=/opt/kubernetes/server/bin/cert/kube-proxy-client.pem \ --client-key=/opt/kubernetes/server/bin/cert/kube-proxy-client-key.pem \ --embed-certs=true \ --kubeconfig=kube-proxy.kubeconfig
User "kube-proxy" set.
|
set-context
注意:在conf目录下
复制/opt/kubernetes/server/bin/conf
1 2 3 4 5 6
|
[root@hdss7-21 conf]# kubectl config set-context myk8s-context \ --cluster=myk8s \ --user=kube-proxy \ --kubeconfig=kube-proxy.kubeconfig
Context "myk8s-context" created.
|
use-context
注意:在conf目录下
复制/opt/kubernetes/server/bin/conf
1 2 3
|
[root@hdss7-21 conf]# kubectl config use-context myk8s-context --kubeconfig=kube-proxy.kubeconfig
Switched to context "myk8s-context".
|
创建kube-proxy启动脚本
HDSS7-21.host.com
上:
复制/opt/kubernetes/server/bin/kube-proxy-721.sh
1 2 3 4 5
|
#!/bin/sh ./kube-proxy \ --cluster-cidr 172.7.0.0/16 \ --hostname-override 10.4.7.21 \ --kubeconfig ./conf/kube-proxy.kubeconfig
|
注意:kube-proxy集群各主机的启动脚本略有不同,部署其他节点时注意修改。
检查配置,权限,创建日志目录
HDSS7-21.host.com
上:
复制/opt/kubernetes/server/conf
1 2 3 4 5
|
[root@hdss7-21 conf]# ls -l|grep kube-proxy.kubeconfig -rw------- 1 root root 6471 Jan 22 17:33 kube-proxy.kubeconfig
[root@hdss7-21 conf]# chmod +x /opt/kubernetes/server/bin/kube-proxy-721.sh [root@hdss7-21 conf]# mkdir -p /data/logs/kubernetes/kube-proxy
|
创建supervisor配置
HDSS7-21.host.com
上:
复制/etc/supervisord.d/kube-proxy.ini
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
[program:kube-proxy] command=/opt/kubernetes/server/bin/kube-proxy-721.sh ; the program (relative uses PATH, can take args) numprocs=1 ; number of processes copies to start (def 1) directory=/opt/kubernetes/server/bin ; directory to cwd to before exec (def no cwd) autostart=true ; start at supervisord start (default: true) autorestart=true ; retstart at unexpected quit (default: true) startsecs=22 ; number of secs prog must stay running (def. 1) startretries=3 ; max # of serial start failures (default 3) exitcodes=0,2 ; 'expected' exit codes for process (default 0,2) stopsignal=QUIT ; signal used to kill process (default TERM) stopwaitsecs=10 ; max num secs to wait b4 SIGKILL (default 10) user=root ; setuid to this UNIX account to run the program redirect_stderr=false ; redirect proc stderr to stdout (default false) stdout_logfile=/data/logs/kubernetes/kube-proxy/proxy.stdout.log ; stdout log path, NONE for none; default AUTO stdout_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stdout_logfile_backups=4 ; # of stdout logfile backups (default 10) stdout_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stdout_events_enabled=false ; emit events on stdout writes (default false) stderr_logfile=/data/logs/kubernetes/kube-proxy/proxy.stderr.log ; stderr log path, NONE for none; default AUTO stderr_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stderr_logfile_backups=4 ; # of stderr logfile backups (default 10) stderr_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stderr_events_enabled=false ; emit events on stderr writes (default false)
|
启动服务并检查
HDSS7-21.host.com
上:
复制
1 2 3 4 5 6 7 8 9
|
[root@hdss7-21 bin]# supervisorctl update kube-proxy: added process group [root@hdss7-21 bin]# supervisorctl status etcd-server-7-21 RUNNING pid 9507, uptime 22:44:48 kube-apiserver RUNNING pid 9770, uptime 21:10:49 kube-controller-manager RUNNING pid 10048, uptime 18:22:10 kube-kubelet RUNNING pid 14597, uptime 0:32:59 kube-proxy STARTING kube-scheduler RUNNING pid 10041, uptime 18:22:13
|
安装部署启动检查所有集群规划主机上的kube-proxy服务
略
部署addons插件
验证kubernetes集群
在任意一个运算节点,创建一个资源配置清单
这里我们选择HDSS7-21.host.com
主机
复制/root/nginx-ds.yaml
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34
|
apiVersion: v1 kind: Service metadata: name: nginx-ds labels: app: nginx-ds spec: type: NodePort selector: app: nginx-ds ports: - name: http port: 80 targetPort: 80
---
apiVersion: extensions/v1beta1 kind: DaemonSet metadata: name: nginx-ds labels: addonmanager.kubernetes.io/mode: Reconcile spec: template: metadata: labels: app: nginx-ds spec: containers: - name: my-nginx image: nginx:1.7.9 ports: - containerPort: 80
|
应用资源配置,并检查
复制/root
1 2 3 4 5
|
[root@hdss7-21 ~]# kubectl create -f nginx-ds.yaml [root@hdss7-21 ~]# kubectl get pods NAME READY STATUS RESTARTS AGE nginx-ds-6hnc7 1/1 Running 0 99m nginx-ds-m5q6j 1/1 Running 0 18h
|
验证
补
部署flannel
集群规划
主机名 |
角色 |
ip |
HDSS7-21.host.com |
flannel |
10.4.7.21 |
HDSS7-22.host.com |
flannel |
10.4.7.22 |
注意:这里部署文档以HDSS7-21.host.com
主机为例,另外一台运算节点安装部署方法类似
在各运算节点上增加iptables规则
注意:iptables规则各主机的略有不同,其他运算节点上执行时注意修改。
- 优化SNAT规则,各运算节点之间的各POD之间的网络通信不再出网
复制
1 2
|
# iptables -t nat -D POSTROUTING -s 172.7.21.0/24 ! -o docker0 -j MASQUERADE # iptables -t nat -I POSTROUTING -s 172.7.21.0/24 ! -d 172.7.0.0/16 ! -o docker0 -j MASQUERADE
|
10.4.7.21主机上的,来源是172.7.21.0/24段的docker的ip,目标ip不是172.7.0.0/16段,网络发包不从docker0桥设备出站的,才进行SNAT转换
各运算节点保存iptables规则
复制
1
|
[root@hdss7-21 ~]# iptables-save > /etc/sysconfig/iptables
|
下载软件,解压,做软连接
HDSS7-21.host.com
上:
复制/opt/src
1 2 3 4 5 6 7 8
|
[root@hdss7-21 src]# ls -l|grep flannel -rw-r--r-- 1 root root 417761204 Jan 17 18:46 flannel-v0.10.0-linux-amd64.tar.gz [root@hdss7-21 src]# mkdir -p /opt/flannel-v0.10.0-linux-amd64/cert [root@hdss7-21 src]# tar xf flannel-v0.10.0-linux-amd64.tar.gz -C /opt/flannel-v0.10.0-linux-amd64 [root@hdss7-21 src]# ln -s /opt/flannel-v0.10.0-linux-amd64 /opt/flannel [root@hdss7-21 src]# ls -l /opt|grep flannel lrwxrwxrwx 1 root root 31 Jan 17 18:49 flannel -> flannel-v0.10.0-linux-amd64/ drwxr-xr-x 4 root root 50 Jan 17 18:47 flannel-v0.10.0-linux-amd64
|
最终目录结构
复制/opt
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
|
[root@hdss7-21 opt]# tree -L 2 . |-- etcd -> etcd-v3.1.18-linux-amd64 |-- etcd-v3.1.18-linux-amd64 | |-- Documentation | |-- README-etcdctl.md | |-- README.md | |-- READMEv2-etcdctl.md | |-- certs | |-- etcd | |-- etcd-server-startup.sh | `-- etcdctl |-- flannel -> flannel-v0.10.0/ |-- flannel-v0.10.0 | |-- README.md | |-- cert | |-- flanneld | `-- mk-docker-opts.sh |-- kubernetes -> kubernetes-v1.13.2-linux-amd64/ |-- kubernetes-v1.13.2-linux-amd64 | |-- LICENSES | |-- addons | `-- server `-- src |-- etcd-v3.1.18-linux-amd64.tar.gz |-- flannel-v0.10.0-linux-amd64.tar.gz `-- kubernetes-server-linux-amd64.tar.gz
|
操作etcd,增加host-gw
HDSS7-21.host.com
上:
复制/opt/etcd
1 2
|
[root@hdss7-21 etcd]# ./etcdctl set /coreos.com/network/config '{"Network": "172.7.0.0/16", "Backend": {"Type": "host-gw"}}' {"Network": "172.7.0.0/16", "Backend": {"Type": "host-gw"}}
|
创建配置
HDSS7-21.host.com
上:
复制/opt/flannel/subnet.env
1 2 3 4
|
FLANNEL_NETWORK=172.7.0.0/16 FLANNEL_SUBNET=172.7.21.1/24 FLANNEL_MTU=1500 FLANNEL_IPMASQ=false
|
注意:flannel集群各主机的配置略有不同,部署其他节点时注意修改。
创建启动脚本
HDSS7-21.host.com
上:
复制/opt/flannel/flanneld.sh
1 2 3 4 5 6 7 8 9 10
|
#!/bin/sh ./flanneld \ --public-ip=10.4.7.21 \ --etcd-endpoints=https://10.4.7.12:2379,https://10.4.7.21:2379,https://10.4.7.22:2379 \ --etcd-keyfile=./cert/client-key.pem \ --etcd-certfile=./cert/client.pem \ --etcd-cafile=./cert/ca.pem \ --iface=eth0 \ --subnet-file=./subnet.env \ --healthz-port=2401
|
注意:flannel集群各主机的启动脚本略有不同,部署其他节点时注意修改。
检查配置,权限,创建日志目录
HDSS7-21.host.com
上:
复制/opt/flannel
1 2 3
|
[root@hdss7-21 flannel]# chmod +x /opt/flannel/flanneld.sh
[root@hdss7-21 flannel]# mkdir -p /data/logs/flanneld
|
创建supervisor配置
HDSS7-21.host.com
上:
复制/etc/supervisord.d/flanneld.ini
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
[program:flanneld] command=/opt/flannel/flanneld.sh ; the program (relative uses PATH, can take args) numprocs=1 ; number of processes copies to start (def 1) directory=/opt/flannel ; directory to cwd to before exec (def no cwd) autostart=true ; start at supervisord start (default: true) autorestart=true ; retstart at unexpected quit (default: true) startsecs=22 ; number of secs prog must stay running (def. 1) startretries=3 ; max # of serial start failures (default 3) exitcodes=0,2 ; 'expected' exit codes for process (default 0,2) stopsignal=QUIT ; signal used to kill process (default TERM) stopwaitsecs=10 ; max num secs to wait b4 SIGKILL (default 10) user=root ; setuid to this UNIX account to run the program redirect_stderr=false ; redirect proc stderr to stdout (default false) stdout_logfile=/data/logs/flanneld/flanneld.stdout.log ; stdout log path, NONE for none; default AUTO stdout_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stdout_logfile_backups=4 ; # of stdout logfile backups (default 10) stdout_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stdout_events_enabled=false ; emit events on stdout writes (default false) stderr_logfile=/data/logs/flanneld/flanneld.stderr.log ; stderr log path, NONE for none; default AUTO stderr_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stderr_logfile_backups=4 ; # of stderr logfile backups (default 10) stderr_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stderr_events_enabled=false ; emit events on stderr writes (default false)
|
启动服务并检查
HDSS7-21.host.com
上:
复制
1 2 3 4 5 6 7 8 9 10
|
[root@hdss7-21 flanneld]# supervisorctl update flanneld: added process group [root@hdss7-21 flanneld]# supervisorctl status etcd-server-7-21 RUNNING pid 9507, uptime 1 day, 20:35:42 flanneld STARTING kube-apiserver RUNNING pid 9770, uptime 1 day, 19:01:43 kube-controller-manager RUNNING pid 37646, uptime 0:58:48 kube-kubelet RUNNING pid 32640, uptime 17:16:36 kube-proxy RUNNING pid 15097, uptime 17:55:36 kube-scheduler RUNNING pid 37803, uptime 0:55:47
|
安装部署启动检查所有集群规划主机上的flannel服务
略
再次验证集群
部署k8s资源配置清单的内网http服务
在运维主机HDSS7-200.host.com
上,配置一个nginx虚拟主机,用以提供k8s统一的资源配置清单访问入口
复制/etc/nginx/conf.d/k8s-yaml.od.com.conf
1 2 3 4 5 6 7 8 9 10
|
server { listen 80; server_name k8s-yaml.od.com;
location / { autoindex on; default_type text/plain; root /data/k8s-yaml; } }
|
配置内网DNS解析
HDSS7-11.host.com
上
复制/var/named/od.com.zone
1
|
k8s-yaml 60 IN A 10.4.7.200
|
以后所有的资源配置清单统一放置在运维主机的/data/k8s-yaml
目录下即可
复制
1
|
[root@hdss7-200 ~]# nginx -s reload
|
部署kube-dns(coredns)
准备coredns-v1.3.1镜像
运维主机HDSS7-200.host.com
上:
复制
1 2 3 4 5 6 7 8 9 10 11 12 13
|
[root@hdss7-200 ~]# docker pull coredns/coredns:1.3.1 1.3.1: Pulling from coredns/coredns e0daa8927b68: Pull complete 3928e47de029: Pull complete Digest: sha256:02382353821b12c21b062c59184e227e001079bb13ebd01f9d3270ba0fcbf1e4 Status: Downloaded newer image for coredns/coredns:1.3.1 [root@hdss7-200 ~]# docker tag eb516548c180 harbor.od.com/k8s/coredns:v1.3.1 [root@hdss7-200 ~]# docker push harbor.od.com/k8s/coredns:v1.3.1 docker push harbor.od.com/k8s/coredns:v1.3.1 The push refers to a repository [harbor.od.com/k8s/coredns] c6a5fc8a3f01: Pushed fb61a074724d: Pushed v1.3.1: digest: sha256:e077b9680c32be06fc9652d57f64aa54770dd6554eb87e7a00b97cf8e9431fda size: 739
|
任意一台运算节点上:
复制
1
|
[root@hdss7-21 ~]# kubectl create secret docker-registry harbor --docker-server=harbor.od.com --docker-username=admin --docker-password=Harbor12345 --docker-email=stanley.wang.m@qq.com -n kube-system
|
准备资源配置清单
运维主机HDSS7-200.host.com
上:
复制
1
|
[root@hdss7-200 ~]# mkdir -p /data/k8s-yaml/coredns && cd /data/k8s-yaml/coredns
|
vi /data/k8s-yaml/coredns/rbac.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45
|
apiVersion: v1 kind: ServiceAccount metadata: name: coredns namespace: kube-system labels: kubernetes.io/cluster-service: "true" addonmanager.kubernetes.io/mode: Reconcile --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: labels: kubernetes.io/bootstrapping: rbac-defaults addonmanager.kubernetes.io/mode: Reconcile name: system:coredns rules: - apiGroups: - "" resources: - endpoints - services - pods - namespaces verbs: - list - watch --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRoleBinding metadata: annotations: rbac.authorization.kubernetes.io/autoupdate: "true" labels: kubernetes.io/bootstrapping: rbac-defaults addonmanager.kubernetes.io/mode: EnsureExists name: system:coredns roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: system:coredns subjects: - kind: ServiceAccount name: coredns namespace: kube-system
|
依次执行创建
浏览器打开:http://k8s-yaml.od.com/coredns 检查资源配置清单文件是否正确创建
在任意运算节点上应用资源配置清单
复制
1 2 3 4 5 6 7 8 9 10
|
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/coredns/rbac.yaml serviceaccount/coredns created clusterrole.rbac.authorization.k8s.io/system:coredns created clusterrolebinding.rbac.authorization.k8s.io/system:coredns created [root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/coredns/configmap.yaml configmap/coredns created [root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/coredns/deployment.yaml deployment.extensions/coredns created [root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/coredns/svc.yaml service/coredns created
|
检查
复制
1 2 3 4 5 6 7 8 9 10
|
[root@hdss7-21 ~]# kubectl get pods -n kube-system -o wide NAME READY STATUS RESTARTS AGE coredns-7ccccdf57c-5b9ch 1/1 Running 0 3m4s
[root@hdss7-21 coredns]# kubectl get svc -n kube-system NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE coredns ClusterIP 192.168.0.2 <none> 53/UDP,53/TCP 29s
[root@hdss7-21 ~]# dig -t A nginx-ds.default.svc.cluster.local. @192.168.0.2 +short 192.168.0.3
|
部署traefik(ingress)
准备traefik镜像
运维主机HDSS7-200.host.com
上:
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
|
[root@hdss7-200 ~]# docker pull traefik:v1.7-alpine v1.7-alpine: Pulling from library/traefik bdf0201b3a05: Pull complete 9dfd896cc066: Pull complete de06b5685128: Pull complete c4d82a21fa27: Pull complete Digest: sha256:0531581bde9da0670fc2c7a4e419e1cc38abff74e7ba06410bf2b1b55c70ef15 Status: Downloaded newer image for traefik:v1.7-alpine [root@hdss7-200 ~]# docker tag 1930b7508541 harbor.od.com/k8s/traefik:v1.7 [root@hdss7-200 ~]# docker push harbor.od.com/k8s/traefik:v1.7 The push refers to a repository [harbor.od.com/k8s/traefik] a3e3d574f6ae: Pushed a7c355c1a104: Pushed e89059911fc9: Pushed a464c54f93a9: Mounted from infra/apollo-portal v1.7: digest: sha256:8f92899f5feb08db600c89d3016145e838fa7ff0d316ee21ecd63d9623643410 size: 1157
|
准备资源配置清单
运维主机HDSS7-200.host.com
上:
复制
1
|
[root@hdss7-200 ~]# mkdir -p /data/k8s-yaml/traefik && cd /data/k8s-yaml/traefik
|
vi /data/k8s-yaml/traefik/rbac.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42
|
apiVersion: v1 kind: ServiceAccount metadata: name: traefik-ingress-controller namespace: kube-system --- apiVersion: rbac.authorization.k8s.io/v1beta1 kind: ClusterRole metadata: name: traefik-ingress-controller rules: - apiGroups: - "" resources: - services - endpoints - secrets verbs: - get - list - watch - apiGroups: - extensions resources: - ingresses verbs: - get - list - watch --- kind: ClusterRoleBinding apiVersion: rbac.authorization.k8s.io/v1beta1 metadata: name: traefik-ingress-controller roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: traefik-ingress-controller subjects: - kind: ServiceAccount name: traefik-ingress-controller namespace: kube-system
|
vi /data/k8s-yaml/coredns/configmap.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
apiVersion: v1 kind: ConfigMap metadata: name: coredns namespace: kube-system data: Corefile: | .:53 { errors log health kubernetes cluster.local 192.168.0.0/16 proxy . /etc/resolv.conf cache 30 }
|
vi /data/k8s-yaml/coredns/deployment.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55
|
apiVersion: extensions/v1beta1 kind: Deployment metadata: name: coredns namespace: kube-system labels: k8s-app: coredns kubernetes.io/cluster-service: "true" kubernetes.io/name: "CoreDNS" spec: replicas: 1 selector: matchLabels: k8s-app: coredns template: metadata: labels: k8s-app: coredns spec: serviceAccountName: coredns containers: - name: coredns image: harbor.od.com/k8s/coredns:v1.3.1 args: - -conf - /etc/coredns/Corefile volumeMounts: - name: config-volume mountPath: /etc/coredns ports: - containerPort: 53 name: dns protocol: UDP - containerPort: 53 name: dns-tcp protocol: TCP livenessProbe: httpGet: path: /health port: 8080 scheme: HTTP initialDelaySeconds: 60 timeoutSeconds: 5 successThreshold: 1 failureThreshold: 5 dnsPolicy: Default imagePullSecrets: - name: harbor volumes: - name: config-volume configMap: name: coredns items: - key: Corefile path: Corefile
|
vi /data/k8s-yaml/coredns/svc.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
|
apiVersion: v1 kind: Service metadata: name: coredns namespace: kube-system labels: k8s-app: coredns kubernetes.io/cluster-service: "true" kubernetes.io/name: "CoreDNS" spec: selector: k8s-app: coredns clusterIP: 192.168.0.2 ports: - name: dns port: 53 protocol: UDP - name: dns-tcp port: 53
|
解析域名
HDSS7-11.host.com
上
复制/var/named/od.com.zone
1
|
traefik 60 IN A 10.4.7.10
|
依次执行创建
浏览器打开:http://k8s-yaml.od.com/traefik 检查资源配置清单文件是否正确创建
在任意运算节点应用资源配置清单
复制
1 2 3 4 5 6 7 8 9 10 11 12 13
|
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/traefik/rbac.yaml serviceaccount/traefik-ingress-controller created clusterrole.rbac.authorization.k8s.io/traefik-ingress-controller created clusterrolebinding.rbac.authorization.k8s.io/traefik-ingress-controller created
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/traefik/daemonset.yaml daemonset.extensions/traefik-ingress-controller created
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/traefik/svc.yaml service/traefik-ingress-service created
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/traefik/ingress.yaml ingress.extensions/traefik-web-ui created
|
配置反代
HDSS7-11.host.com
和HDSS7-12.host.com
两台主机上的nginx均需要配置,这里可以考虑使用saltstack或者ansible进行统一配置管理
复制/etc/nginx/conf.d/od.com.conf
1 2 3 4 5 6 7 8 9 10 11 12 13
|
upstream default_backend_traefik { server 10.4.7.21:81 max_fails=3 fail_timeout=10s; server 10.4.7.22:81 max_fails=3 fail_timeout=10s; } server { server_name *.od.com;
location / { proxy_pass http://default_backend_traefik; proxy_set_header Host $http_host; proxy_set_header x-forwarded-for $proxy_add_x_forwarded_for; } }
|
浏览器访问
http://traefik.od.com
部署dashboard
准备dashboard镜像
运维主机HDSS7-200.host.com
上:
复制
1 2 3 4 5 6 7 8 9 10 11
|
[root@hdss7-200 ~]# docker pull k8scn/kubernetes-dashboard-amd64:v1.8.3 v1.8.3: Pulling from k8scn/kubernetes-dashboard-amd64 a4026007c47e: Pull complete Digest: sha256:ebc993303f8a42c301592639770bd1944d80c88be8036e2d4d0aa116148264ff Status: Downloaded newer image for k8scn/kubernetes-dashboard-amd64:v1.8.3 [root@hdss7-200 ~]# docker tag 0c60bcf89900 harbor.od.com/k8s/dashboard:v1.8.3 [root@hdss7-200 ~]# docker push harbor.od.com/k8s/dashboard:v1.8.3 docker push harbor.od.com/k8s/dashboard:v1.8.3 The push refers to a repository [harbor.od.com/k8s/dashboard] 23ddb8cbb75a: Pushed v1.8.3: digest: sha256:e76c5fe6886c99873898e4c8c0945261709024c4bea773fc477629455631e472 size: 529
|
准备资源配置清单
运维主机HDSS7-200.host.com
上:
复制
1
|
[root@hdss7-200 ~]# mkdir -p /data/k8s-yaml/dashboard && cd /data/k8s-yaml/dashboard
|
vi /data/k8s-yaml/dashboard/rbac.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25
|
apiVersion: v1 kind: ServiceAccount metadata: labels: k8s-app: kubernetes-dashboard addonmanager.kubernetes.io/mode: Reconcile name: kubernetes-dashboard-admin namespace: kube-system --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRoleBinding metadata: name: kubernetes-dashboard-admin namespace: kube-system labels: k8s-app: kubernetes-dashboard addonmanager.kubernetes.io/mode: Reconcile roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: cluster-admin subjects: - kind: ServiceAccount name: kubernetes-dashboard-admin namespace: kube-system
|
vi /data/k8s-yaml/dashboard/secret.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21
|
apiVersion: v1 kind: Secret metadata: labels: k8s-app: kubernetes-dashboard # Allows editing resource and makes sure it is created first. addonmanager.kubernetes.io/mode: EnsureExists name: kubernetes-dashboard-certs namespace: kube-system type: Opaque --- apiVersion: v1 kind: Secret metadata: labels: k8s-app: kubernetes-dashboard # Allows editing resource and makes sure it is created first. addonmanager.kubernetes.io/mode: EnsureExists name: kubernetes-dashboard-key-holder namespace: kube-system type: Opaque
|
vi /data/k8s-yaml/dashboard/configmap.yaml
复制
1 2 3 4 5 6 7 8 9
|
apiVersion: v1 kind: ConfigMap metadata: labels: k8s-app: kubernetes-dashboard # Allows editing resource and makes sure it is created first. addonmanager.kubernetes.io/mode: EnsureExists name: kubernetes-dashboard-settings namespace: kube-system
|
vi /data/k8s-yaml/dashboard/svc.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
apiVersion: v1 kind: Service metadata: name: kubernetes-dashboard namespace: kube-system labels: k8s-app: kubernetes-dashboard kubernetes.io/cluster-service: "true" addonmanager.kubernetes.io/mode: Reconcile spec: selector: k8s-app: kubernetes-dashboard ports: - port: 443 targetPort: 8443
|
vi /data/k8s-yaml/dashboard/ingress.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
apiVersion: extensions/v1beta1 kind: Ingress metadata: name: kubernetes-dashboard namespace: kube-system annotations: kubernetes.io/ingress.class: traefik spec: rules: - host: dashboard.od.com http: paths: - backend: serviceName: kubernetes-dashboard servicePort: 443
|
vi /data/k8s-yaml/dashboard/deployment.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61
|
apiVersion: apps/v1 kind: Deployment metadata: name: kubernetes-dashboard namespace: kube-system labels: k8s-app: kubernetes-dashboard kubernetes.io/cluster-service: "true" addonmanager.kubernetes.io/mode: Reconcile spec: selector: matchLabels: k8s-app: kubernetes-dashboard template: metadata: labels: k8s-app: kubernetes-dashboard annotations: scheduler.alpha.kubernetes.io/critical-pod: '' spec: priorityClassName: system-cluster-critical containers: - name: kubernetes-dashboard image: harbor.od.com/k8s/dashboard:v1.8.3 resources: limits: cpu: 100m memory: 300Mi requests: cpu: 50m memory: 100Mi ports: - containerPort: 8443 protocol: TCP args: # PLATFORM-SPECIFIC ARGS HERE - --auto-generate-certificates volumeMounts: - name: kubernetes-dashboard-certs mountPath: /certs - name: tmp-volume mountPath: /tmp livenessProbe: httpGet: scheme: HTTPS path: / port: 8443 initialDelaySeconds: 30 timeoutSeconds: 30 volumes: - name: kubernetes-dashboard-certs secret: secretName: kubernetes-dashboard-certs - name: tmp-volume emptyDir: {} serviceAccountName: kubernetes-dashboard-admin tolerations: - key: "CriticalAddonsOnly" operator: "Exists" imagePullSecrets: - name: harbor
|
解析域名
HDSS7-11.host.com
上
复制/var/named/od.com.zone
1
|
dashboard 60 IN A 10.4.7.10
|
依次执行创建
浏览器打开:http://k8s-yaml.od.com/dashboard 检查资源配置清单文件是否正确创建
在任意运算节点应用资源配置清单
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
|
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/dashboard/rbac.yaml serviceaccount/kubernetes-dashboard created role.rbac.authorization.k8s.io/kubernetes-dashboard-admin created rolebinding.rbac.authorization.k8s.io/kubernetes-dashboard-admin created
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/dashboard/secret.yaml secret/kubernetes-dashboard-certs created secret/kubernetes-dashboard-key-holder created
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/dashboard/configmap.yaml configmap/kubernetes-dashboard-settings created
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/dashboard/svc.yaml service/kubernetes-dashboard created
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/dashboard/ingress.yaml ingress.extensions/kubernetes-dashboard created
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/dashboard/deployment.yaml deployment.apps/kubernetes-dashboard created
|
浏览器访问
http://dashboard.od.com
配置认证
复制
1 2 3
|
[root@hdss7-200 ~]# docker pull hexun/kubernetes-dashboard-amd64:v1.10.1 [root@hdss7-200 ~]# docker tag f9aed6605b81 harbor.od.com/k8s/dashboard:v1.10.1 [root@hdss7-200 ~]# docker push harbor.od.com/k8s/dashboard:v1.10.1
|
应用新版dashboard
修改nginx配置,走https
复制/etc/nginx/conf.d/dashboard.od.com.conf
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
server { listen 80; server_name dashboard.od.com;
rewrite ^(.*)$ https://${server_name}$1 permanent; } server { listen 443 ssl; server_name dashboard.od.com;
ssl_certificate "certs/dashboard.od.com.crt"; ssl_certificate_key "certs/dashboard.od.com.key"; ssl_session_cache shared:SSL:1m; ssl_session_timeout 10m; ssl_ciphers HIGH:!aNULL:!MD5; ssl_prefer_server_ciphers on;
location / { proxy_pass http://default_backend_traefik; proxy_set_header Host $http_host; proxy_set_header x-forwarded-for $proxy_add_x_forwarded_for; } }
|
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14
|
[root@hdsss7-21 ~]# kubectl describe secret kubernetes-dashboard-admin-token-rhr62 -n kube-system Name: kubernetes-dashboard-admin-token-rhr62 Namespace: kube-system Labels: <none> Annotations: kubernetes.io/service-account.name: kubernetes-dashboard-admin kubernetes.io/service-account.uid: cdd3c552-856d-11e9-ae34-782bcb321c07
Type: kubernetes.io/service-account-token
Data ==== ca.crt: 1354 bytes namespace: 11 bytes token: eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJrdWJlcm5ldGVzLWRhc2hib2FyZC1hZG1pbi10b2tlbi1yaHI2MiIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJrdWJlcm5ldGVzLWRhc2hib2FyZC1hZG1pbiIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImNkZDNjNTUyLTg1NmQtMTFlOS1hZTM0LTc4MmJjYjMyMWMwNyIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTprdWJlcm5ldGVzLWRhc2hib2FyZC1hZG1pbiJ9.72OcJZCm_3I-7QZcEJTRPyIJSxQwSwZfVsB6Bx_RAZRJLOv3-BXy88PclYgxRy2dDqeX6cpjvFPBrmNOGQoxT9oD8_H49pvBnqdCdNuoJbXK7aBIZdkZxATzXd-63zmhHhUBsM3Ybgwy5XxD3vj8VUYfux5c5Mr4TzU_rnGLCj1H5mq_JJ3hNabv0rwil-ZAV-3HLikOMiIRhEK7RdMs1bfXF2yvse4VOabe9xv47TvbEYns97S4OlZvsurmOk0B8dD85OSaREEtqa8n_ND9GrHeeL4CcALqWYJHLrr7vLfndXi1QHDVrUzFKvgkAeYpDVAzGwIWL7rgHwp3sQguGA
|
部署heapster
heapster官方github地址
准备heapster镜像
运维主机HDSS7-200.host.com
上
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
[root@hdss7-200 ~]# docker pull quay.io/bitnami/heapster:1.5.4 1.5.4: Pulling from bitnami/heapster 4018396ca1ba: Pull complete 0e4723f815c4: Pull complete d8569f30adeb: Pull complete Digest: sha256:6d891479611ca06a5502bc36e280802cbf9e0426ce4c008dd2919c2294ce0324 Status: Downloaded newer image for quay.io/bitnami/heapster:1.5.4 [root@hdss7--200 ~]# docker tag c359b95ad38b harbor.od.com/k8s/heapster:v1.5.4 [root@hdss7--200 ~]# docker push !$ docker push harbor.od.com/k8s/heapster:v1.5.4 The push refers to a repository [harbor.od.com/k8s/heapster] 20d37d828804: Pushed b9b192015e25: Pushed b76dba5a0109: Pushed v1.5.4: digest: sha256:1203b49f2b2b07e02e77263bce8bb30563a91e1d7ee7c6742e9d125abcb3abe6 size: 952
|
准备资源配置清单
vi /data/k8s-yaml/dashboard/heapster/rbac.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
|
apiVersion: v1 kind: ServiceAccount metadata: name: heapster namespace: kube-system --- kind: ClusterRoleBinding apiVersion: rbac.authorization.k8s.io/v1beta1 metadata: name: heapster roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: system:heapster subjects: - kind: ServiceAccount name: heapster namespace: kube-system
|
vi /data/k8s-yaml/dashboard/heapster/deployment.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21
|
apiVersion: extensions/v1beta1 kind: Deployment metadata: name: heapster namespace: kube-system spec: replicas: 1 template: metadata: labels: task: monitoring k8s-app: heapster spec: serviceAccountName: heapster containers: - name: heapster image: harbor.od.com/k8s/heapster:v1.5.4 imagePullPolicy: IfNotPresent command: - /opt/bitnami/heapster/bin/heapster - --source=kubernetes:https://kubernetes.default
|
vi /data/k8s-yaml/dashboard/heapster/svc.yaml
复制
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17
|
apiVersion: v1 kind: Service metadata: labels: task: monitoring # For use as a Cluster add-on (https://github.com/kubernetes/kubernetes/tree/master/cluster/addons) # If you are NOT using this as an addon, you should comment out this line. kubernetes.io/cluster-service: 'true' kubernetes.io/name: Heapster name: heapster namespace: kube-system spec: ports: - port: 80 targetPort: 8082 selector: k8s-app: heapster
|
应用资源配置清单
任意运算节点上:
复制
1 2 3 4 5 6 7
|
[root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/dashboard/heapster/rbac.yaml serviceaccount/heapster created clusterrolebinding.rbac.authorization.k8s.io/heapster created [root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/dashboard/heapster/deployment.yaml deployment.extensions/heapster created [root@hdss7-21 ~]# kubectl apply -f http://k8s-yaml.od.com/dashboard/heapster/svc.yaml service/heapster created
|
重启dashboard
浏览器访问:http://dashboard.od.com