Centos7.8安装K8s 1.21.x 最新高可用集群

Centos7.8安装K8s 1.21.x高可用集群

目录

1. 安装说明

2. 节点规划

3. 基本配置

4. 内核配置

5. 基本组件安装

6. 高可用组件安装

7. 集群初始化

8. 高可用Master

9. 添加Node节点

10. Calico安装

11. Metrics Server部署

12. Dashboard部署

 

1.安装说明

本次安装采用的是Kubeadm安装工具,安装版本是K8s 1.20+,采用的系统为CentOS 7.8,其中Master节点3台,Node节点2台,高可用工具采用HAProxy + KeepAlived

 

2.节点规划

主机名

IP地址

角色

配置

K8s-master01~03

192.168.196.3-5

Master/worker节点

2C2G 40G

K8s-node01~02

192.168.196.6~7

worker节点

2C2G 40G

K8s-master-lb

192.168.196.2

VIP

VIP不占用机器

 

 

 

 

 

 

 

信息

备注

系统版本

Centos7.8

Docker版本

19.03.x

K8s版本

1.20.x

pod网段

172.168.0.0/16

serverice网段

10.96.0.0/12

 

 

 

 

 

 

 

 

 

 

 

 

 

 

3.基本配置

[root@k8s-master01 ~]# cat /etc/hosts

 

192.168.196.3 k8s-master01

192.168.196.4 k8s-master02

192.168.196.5 k8s-master03

192.168.196.2 k8s-master-lb     #如果不是高可用集群,该ip为master01的ip

192.168.196.6 k8s-node01

192.168.196.7 k8s-node02

 

 

Yum 源配置

 

curl -o /etc/yum.repos.d/CentOS-Base.repo

https://mirrors.aliyun.com/repo/Centos-7.repo

 

yum install -y yum-utils device-mapper-persistent-data lvm2

 

yum-config-manager --add-repo https://mirrors.aliyun.com/docker-ce/linux/centos/docker-ce.repo

 

cat <<EOF > /etc/yum.repos.d/kubernetes.repo [kubernetes] name=Kubernetes baseurl=https://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64/ enabled=1 gpgcheck=1 repo_gpgcheck=1 gpgkey=https://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpg https://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpg

EOF

 

sed -i -e '/mirrors.cloud.aliyuncs.com/d' -e '/mirrors.aliyuncs.com/d' /etc/yum.repos.d/CentOS-Base.reposetenforce 0 sed -i 's#SELINUX=enforcing#SELINUX=disabled#g' /etc/sysconfig/selinux sed -i 's#SELINU

 

必备工具安装

yum install wget jq psmisc vim net-tools telnet yum-utils device-mapper-persistent-data lvm2 git -y

 

所有节点关闭防火墙,selinux,dnsmasq,swap.服务配置如下:

systemctl disable --now firewalld  

systemctl disable --now dnsmasq

systemctl disable --now NetworkMa

 

setenforce 0

sed -i 's#SELINUX=enforcing#SELINUX=disabled#g' /etc/sysconfig/selinux

sed -i 's#SELINUX=enforcing#SELINUX=disabled#g' /etc/selinux/config

 

 

关闭swap分区

swapoff -a && sysctl -w vm.swappiness=0

sed -ri '/^[^#]*swap/s@^@#@' /etc/fstab

 

 

安装ntpdate

rpm -ivh http://mirrors.wlnmp.com/centos/wlnmp-release-centos.noarch.rpm

yum install ntpdate -y

 

所有节点同步时间,时间同步配置如下:

ln -sf /usr/share/zoneinfo/Asia/Shanghai /etc/localtime

echo 'Asia/Shanghai' >/etc/timezone

ntpdate time2.aliyun.com

 

加到定时任务crontab

crontab -e #编辑

*/5 * * * * ntpdate time2.aliyun.com

crontab -l  #查看

 

所有节点配置limit:

ulimit -SHn 65535

 

 cat /etc/security/limits.conf |grep -v "^#"

 

#末尾添加如下内容

 

* soft nofile 655360

* hard nofile 131072

* soft nproc 655350

* hard nproc 655350

* soft memlock unlimited

* hard memlock unlimited

 

Master01节点免密钥登陆其他节点:

ssh-keygen -t rsa

for i in k8s-master01 k8s-master02 k8s-master03 k8s-node01 k8s-node02;do ssh-copy-id -i .ssh/id_rsa.pub $i;done

 

下载安装所有的源码文件

cd /root

git clone https://github.com/dotbalo/k8s-ha-install.git

 

 

所有节点升级重启

yum update -y  && reboot

 

4.内核配置

所有节点安装ipvsadm

yum install ipvsadm ipset sysstat conntrack libseccomp -y

 

所有节点配置ipvs模块

 

vim /etc/modules-load.d/ipvs.conf

#加入以下内容

ip_vs_lc

ip_vs_wlc

ip_vs_rr

ip_vs_wrr

ip_vs_lblc

ip_vs_lblcr

ip_vs_dh

ip_vs_sh

ip_vs_fo

ip_vs_nq

ip_vs_sed

ip_vs_ftp

ip_vs_sh

nf_conntrack

ip_tables

ip_set

xt_set

ipt_set

ipt_rpfilter

ipt_REJECT

Ipip

 

 

加载内核配置

systemctl enable --now systemd-modules-load.service

 

开启一些k8s集群中必须的内核参数,所有节点配置k8s内核

 

cat <<EOF > /etc/sysctl.d/k8s.conf

net.ipv4.ip_forward = 1

net.bridge.bridge-nf-call-iptables = 1

net.bridge.bridge-nf-call-ip6tables = 1

fs.may_detach_mounts = 1

vm.overcommit_memory=1

vm.panic_on_oom=0

fs.inotify.max_user_watches=89100

fs.file-max=52706963

fs.nr_open=52706963

net.netfilter.nf_conntrack_max=2310720

 

net.ipv4.tcp_keepalive_time = 600

net.ipv4.tcp_keepalive_probes = 3

net.ipv4.tcp_keepalive_intvl =15

net.ipv4.tcp_max_tw_buckets = 36000

net.ipv4.tcp_tw_reuse = 1

net.ipv4.tcp_max_orphans = 327680

net.ipv4.tcp_orphan_retries = 3

net.ipv4.tcp_syncookies = 1

net.ipv4.tcp_max_syn_backlog = 16384

net.ipv4.ip_conntrack_max = 65536

net.ipv4.tcp_max_syn_backlog = 16384

net.ipv4.tcp_timestamps = 0

net.core.somaxconn = 16384

EOF

 

#应用

sysctl --system

 

5.基本组件安装

所有节点安装docker-ce 19.03

 

 

 

所有节点设置开机自启动docker

systemctl daemon-reload && systemctl enable --now docker

 

安装k8s组件

yum list kubeadm.x86_64 --showduplicates | sort -r

所有节点安装最新版本kubeadm

yum install kubeadm -y

 

默认配置的pause镜像使用gcr.io仓库,国内可能无法访问,所以这里配置Kubelet使用阿里云的pause镜像:

cat >/etc/sysconfig/kubelet<<EOF

KUBELET_EXTRA_ARGS="--pod-infra-container-image=registry.cn-hangzhou.aliyuncs.com/google_containers/pause-amd64:3.2"

EOF

 

设置kubelet开机启动

systemctl daemon-reload

systemctl enable --now kubelet

 

6.高可用组件安装

注意:如果不是高可用集群或者在云上安装,haproxy和keepalived无需安装

所有Master节点通过yum安装HAProxy和KeepAlived:、

 

yum install keepalived haproxy -y

 

所有Master节点配置HAProxy(详细配置参考HAProxy文档,所有Master节点的HAProxy配置相同):

 

vim /etc/haproxy/haproxy.cfg

global

  maxconn  2000

  ulimit-n  16384

  log  127.0.0.1 local0 err

  stats timeout 30s

 

defaults

  log global

  mode  http

  option  httplog

  timeout connect 5000

  timeout client  50000

  timeout server  50000

  timeout http-request 15s

  timeout http-keep-alive 15s

 

frontend monitor-in

  bind *:33305

  mode http

  option httplog

  monitor-uri /monitor

 

frontend k8s-master

  bind 0.0.0.0:16443

  bind 127.0.0.1:16443

  mode tcp

  option tcplog

  tcp-request inspect-delay 5s

  default_backend k8s-master

 

backend k8s-master

  mode tcp

  option tcplog

  option tcp-check

  balance roundrobin

  default-server inter 10s downinter 5s rise 2 fall 2 slowstart 60s maxconn 250 maxqueue 256 weight 100

  server k8s-master01 192.168.196.3:6443  check

  server k8s-master02 192.168.196.4:6443  check

  server k8s-master03 192.168.196.5:6443  check

 

 

所有Master节点配置KeepAlived,配置不一样,注意区分

注意每个节点的IP和网卡(interface参数)

Master01节点的配置:

vim /etc/keepalived/keepalived.conf

! Configuration File for keepalived

global_defs {

    router_id LVS_DEVEL

script_user root

    enable_script_security

}

vrrp_script chk_apiserver {

    script "/etc/keepalived/check_apiserver.sh"

    interval 5

    weight -5

    fall 2  

rise 1

}

vrrp_instance VI_1 {

    state MASTER

    interface ens33

    mcast_src_ip 192.168.196.3

    virtual_router_id 51

    priority 101

    advert_int 2

    authentication {

        auth_type PASS

        auth_pass K8SHA_KA_AUTH

    }

    virtual_ipaddress {

        192.168.196.2

    }

#    track_script {

#       chk_apiserver

#    }

}

 

Master02配置

cat /etc/keepalived/keepalived.conf

! Configuration File for keepalived

global_defs {

    router_id LVS_DEVEL

script_user root

    enable_script_security

}

vrrp_script chk_apiserver {

    script "/etc/keepalived/check_apiserver.sh"

   interval 5

    weight -5

    fall 2  

rise 1

}

vrrp_instance VI_1 {

    state BACKUP

    interface ens33

    mcast_src_ip 192.168.196.4

    virtual_router_id 51

    priority 100

    advert_int 2

    authentication {

        auth_type PASS

        auth_pass K8SHA_KA_AUTH

    }

    virtual_ipaddress {

        192.168.196.2

    }

#    track_script {

#       chk_apiserver

#    }

}

 

Master03节点配置:

! Configuration File for keepalived

global_defs {

    router_id LVS_DEVEL

script_user root

    enable_script_security

}

vrrp_script chk_apiserver {

    script "/etc/keepalived/check_apiserver.sh"

 interval 5

    weight -5

    fall 2  

rise 1

}

vrrp_instance VI_1 {

    state BACKUP

    interface ens33

    mcast_src_ip 192.168.196.5

    virtual_router_id 51

    priority 100

    advert_int 2

    authentication {

        auth_type PASS

        auth_pass K8SHA_KA_AUTH

    }

    virtual_ipaddress {

        192.168.196.2

    }

#    track_script {

#       chk_apiserver

#    }

}

 

注意上述的健康检查是关闭的,集群建立完成后再开启:

#    track_script {

#       chk_apiserver

#    }

}

 

配置KeepAlived健康检查文件:

cat /etc/keepalived/check_apiserver.sh

#!/bin/bash

 

err=0

for k in $(seq 1 3)

do

    check_code=$(pgrep haproxy)

    if [[ $check_code == "" ]]; then

        err=$(expr $err + 1)

        sleep 1

        continue

    else

        err=0

        break

    fi

done

 

if [[ $err != "0" ]]; then

    echo "systemctl stop keepalived"

    /usr/bin/systemctl stop keepalived

    exit 1

else

    exit 0

Fi

 

chmod +x /etc/keepalived/check_apiserver.sh

 

启动haproxy和keepalived

 

systemctl enable --now haproxy

systemctl status haproxy.service

systemctl enable --now keepalived

 

查看状态,注意如果起不来,查看selinux是否关闭,然后setenforce 0,重新启动

 

service haproxy.service start

systemctl status haproxy.service

systemctl status keepalived.service

 

测试

ping 192.168.196.2

PING 192.168.196.2 (192.168.196.2) 56(84) bytes of data.

64 bytes from 192.168.196.2: icmp_seq=1 ttl=64 time=0.691 ms

64 bytes from 192.168.196.2: icmp_seq=2 ttl=64 time=0.062 ms

64 bytes from 192.168.196.2: icmp_seq=3 ttl=64 time=0.045 ms

64 bytes from 192.168.196.2: icmp_seq=4 ttl=64 time=0.046 ms

^C

 

7.集群初始化

Master01节点创建new.yaml配置文件如下:

vim /root/new.yaml

apiVersion: kubeadm.k8s.io/v1beta2

bootstrapTokens:

- groups:

  - system:bootstrappers:kubeadm:default-node-token

  token: 7t2weq.bjbawausm0jaxury

  ttl: 24h0m0s

  usages:

  - signing

  - authentication

kind: InitConfiguration

localAPIEndpoint:

  advertiseAddress: 192.168.196.3

  bindPort: 6443

nodeRegistration:

  criSocket: /var/run/dockershim.sock

  name: k8s-master01

  taints:

  - effect: NoSchedule

    key: node-role.kubernetes.io/master

---

apiServer:

  certSANs:

  - 192.168.196.2

  timeoutForControlPlane: 4m0s

apiVersion: kubeadm.k8s.io/v1beta2

certificatesDir: /etc/kubernetes/pki

clusterName: kubernetes

controlPlaneEndpoint: 192.168.196.2:16443

controllerManager: {}

dns:

  type: CoreDNS

etcd:

  local:

    dataDir: /var/lib/etcd

imageRepository: registry.cn-hangzhou.aliyuncs.com/google_containers

kind: ClusterConfiguration

kubernetesVersion: v1.21.1

networking:

  dnsDomain: cluster.local

  podSubnet: 172.168.0.0/16

  serviceSubnet: 10.96.0.0/12

scheduler: {}

 

注意:如果不是高可用集群,192.168.0.236:16443改为master01的地址,16443改为apiserver的端口,默认是6443,注意更改v1.20.0为自己服务器kubeadm的版本:kubeadm version

将new.yaml文件复制到其他master节点,之后所有Master节点提前下载镜像,可以节省初始化时间:

scp new.yaml root@k8s-master02:/root

scp new.yaml root@k8s-master03:/root

 

kubeadm config images pull --config /root/new.yaml

报错

failed to pull image "registry.aliyuncs.com/google_containers/coredns/coredns:v1.8.0": output: Error response from daemon: pull access denied for registry.aliyuncs.com/google_containers/coredns/coredns, repository does not exist or may require 'docker login': denied: requested access to the resource is denied

 

 

解决方案

手动下载镜像

 

[root@k8s-master02 ~]# docker pull coredns/coredns

 

查看kubeadm需要镜像,并修改名称

[root@k8s-master02 ~]# kubeadm config images list --config new.yaml

 

查看镜像

[root@k8s-master02 ~]# docker images

 

 

打标签,修改名称

 

[root@k8s-master02 ~]# docker tag coredns/coredns:latest registry.cn-hangzhou.aliyuncs.com/google_containers/coredns/coredns:v1.8.0

 

删除多余镜像

[root@k8s-master02 ~]# docker rmi coredns/coredns:latest

 

所有节点设置开机自启动kubelet

systemctl restart kubelet(如果启动失败无需管理,初始化成功以后即可启动)

 

Master01节点初始化,初始化以后会在/etc/kubernetes目录下生成对应的证书和配置文件,之后其他Master节点加入Master01即可:

0

[root@k8s-master01 ~]# kubeadm init --config /root/new.yaml  --upload-certs

Your Kubernetes control-plane has initialized successfully!

 

To start using your cluster, you need to run the following as a regular user:

 

  mkdir -p $HOME/.kube

  sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config

  sudo chown $(id -u):$(id -g) $HOME/.kube/config

 

Alternatively, if you are the root user, you can run:

 

  export KUBECONFIG=/etc/kubernetes/admin.conf

 

You should now deploy a pod network to the cluster.

Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:

  https://kubernetes.io/docs/concepts/cluster-administration/addons/

 

You can now join any number of the control-plane node running the following command on each as root:

 

  kubeadm join 192.168.196.2:16443 --token 7t2weq.bjbawausm0jaxury \

--discovery-token-ca-cert-hash sha256:927471d6b447dd526eb15f1837e9cce8f5061b5d0ebc8a5ff9d17a2774bbbed6 \

--control-plane --certificate-key 51543575c9cde20808374c76c960e142f2f326b06c30267ce2ba470ce37c176c

 

Please note that the certificate-key gives access to cluster sensitive data, keep it secret!

As a safeguard, uploaded-certs will be deleted in two hours; If necessary, you can use

"kubeadm init phase upload-certs --upload-certs" to reload certs afterward.

 

Then you can join any number of worker nodes by running the following on each as root:

 

kubeadm join 192.168.196.2:16443 --token 7t2weq.bjbawausm0jaxury \

--discovery-token-ca-cert-hash sha256:927471d6b447dd526eb15f1837e9cce8f5061b5d0ebc8a5ff9d17a2774bbbed6  

如果出现报错,就是cpu核数没有2核

 

Master01节点配置环境变量,用于访问Kubernetes集群:

 

cat <<EOF >> /root/.bashrc

export KUBECONFIG=/etc/kubernetes/admin.conf

EOF

 

source /root/.bashrc

 

查看集群状态

root@k8s-master01 ~]# kubectl get nodes

NAME           STATUS     ROLES                  AGE     VERSION

k8s-master01   NotReady   control-plane,master   2m57s   v1.21.1

 

采用初始化安装方式,所有的系统组件均以容器的方式运行并且在kube-system命名空间内,此时可以查看Pod状态:

[root@k8s-master01 ~]# kubectl get pods -n kube-system -o wide

NAME                                   READY   STATUS    RESTARTS   AGE    IP              NODE           NOMINATED NODE   READINESS GATES

coredns-57d4cbf879-6t6c5               0/1     Pending   0          4m9s   <none>          <none>         <none>           <none>

coredns-57d4cbf879-p9px5               0/1     Pending   0          4m9s   <none>          <none>         <none>           <none>

etcd-k8s-master01                      1/1     Running   0          4m3s   192.168.196.3   k8s-master01   <none>           <none>

kube-apiserver-k8s-master01            1/1     Running   0          4m3s   192.168.196.3   k8s-master01   <none>           <none>

kube-controller-manager-k8s-master01   1/1     Running   0          4m3s   192.168.196.3   k8s-master01   <none>           <none>

kube-proxy-7gs9r                       1/1     Running   0          4m9s   192.168.196.3   k8s-master01   <none>           <none>

kube-scheduler-k8s-master01            1/1     Running   0          4m3s   192.168.196.3   k8s-master01   <none>           <none>

 

8. 高可用Master

注意:初始化其他master加入集群。注意一个是--control-plane --certificate-key

 

 

kubeadm join 192.168.196.2:16443 --token 7t2weq.bjbawausm0jaxury \

--discovery-token-ca-cert-hash sha256:927471d6b447dd526eb15f1837e9cce8f5061b5d0ebc8a5ff9d17a2774bbbed6 \

--control-plane --certificate-key 51543575c9cde20808374c76c960e142f2f326b06c30267ce2ba470ce37c176c

 

9.添加Node节点

kubeadm join 192.168.196.2:16443 --token 7t2weq.bjbawausm0jaxury \

--discovery-token-ca-cert-hash sha256:927471d6b447dd526eb15f1837e9cce8f5061b5d0ebc8a5ff9d17a2774bbbed6  

 

 

查看集群状态

 

[root@k8s-master01 ~]# kubectl get node

NAME           STATUS     ROLES                  AGE     VERSION

k8s-master01   NotReady   control-plane,master   21m     v1.21.1

k8s-master02   NotReady   control-plane,master   17m     v1.21.1

k8s-master03   NotReady   control-plane,master   9m29s   v1.21.1

k8s-node01     NotReady   <none>                 18s     v1.21.1

k8s-node02     NotReady   <none>                 10s     v1.21.1

 

注意:如果加入集群失败

执行kubeadm reset,重新添加密钥,如果没有,忽略

 

10.calico安装

以下步骤只在master01执行

cd /root/k8s-ha-install && git checkout manual-installation-v1.20.x && cd calico/

 

修改calico-etcd.yaml的以下位置

[root@k8s-master01 calico]#

sed -i 's#etcd_endpoints: "http://<ETCD_IP>:<ETCD_PORT>"#etcd_endpoints: "https://192.168.196.3:2379,https://192.168.196.4:2379,https://192.168.196.5:2379"#g' calico-etcd.yaml

 

ETCD_CA=`cat /etc/kubernetes/pki/etcd/ca.crt | base64 | tr -d '\n'`

ETCD_CERT=`cat /etc/kubernetes/pki/etcd/server.crt | base64 | tr -d '\n'`

ETCD_KEY=`cat /etc/kubernetes/pki/etcd/server.key | base64 | tr -d '\n'`

sed -i "s@# etcd-key: null@etcd-key: ${ETCD_KEY}@g; s@# etcd-cert: null@etcd-cert: ${ETCD_CERT}@g; s@# etcd-ca: null@etcd-ca: ${ETCD_CA}@g" calico-etcd.yaml

 

 

sed -i 's#etcd_ca: ""#etcd_ca: "/calico-secrets/etcd-ca"#g; s#etcd_cert: ""#etcd_cert: "/calico-secrets/etcd-cert"#g; s#etcd_key: "" #etcd_key: "/calico-secrets/etcd-key" #g' calico-etcd.yaml

 

POD_SUBNET=`cat /etc/kubernetes/manifests/kube-controller-manager.yaml | grep cluster-cidr= | awk -F= '{print $NF}'`

 

sed -i 's@# - name: CALICO_IPV4POOL_CIDR@- name: CALICO_IPV4POOL_CIDR@g; s@#   value: "192.168.0.0/16"@  value: '"${POD_SUBNET}"'@g' calico-etcd.yaml

 

创建calico

[root@k8s-master01 calico]# kubectl apply -f calico-etcd.yaml

 

 

11. Metrics Server部署

在新版的Kubernetes中系统资源的采集均使用Metrics-server,可以通过Metrics采集节点和Pod的内存、磁盘、CPU和网络的使用率。

将Master01节点的front-proxy-ca.crt复制到所有Node节点

 

[root@k8s-master01 calico]# scp /etc/kubernetes/pki/front-proxy-ca.crt k8s-node01:/etc/kubernetes/pki/front-proxy-ca.crt

 

[root@k8s-master01 calico]# scp /etc/kubernetes/pki/front-proxy-ca.crt k8s-node02:/etc/kubernetes/pki/front-proxy-ca.crt

 

安装metrics server

[root@k8s-master01 calico]# cd /root/k8s-ha-install/metrics-server-0.4.x-kubeadm/

[root@k8s-master01 metrics-server-0.4.x-kubeadm]#  kubectl  create -f comp.yaml

serviceaccount/metrics-server created

clusterrole.rbac.authorization.k8s.io/system:aggregated-metrics-reader created

clusterrole.rbac.authorization.k8s.io/system:metrics-server created

rolebinding.rbac.authorization.k8s.io/metrics-server-auth-reader created

clusterrolebinding.rbac.authorization.k8s.io/metrics-server:system:auth-delegator created

clusterrolebinding.rbac.authorization.k8s.io/system:metrics-server created

service/metrics-server created

deployment.apps/metrics-server created

apiservice.apiregistration.k8s.io/v1beta1.metrics.k8s.io created

 

等待kube-system命令空间下的Pod全部启动后,查看状态

[root@k8s-master01 metrics-server-0.4.x-kubeadm]# kubectl  top node

 

W0518 18:26:25.038045   31602 top_node.go:119] Using json format to get metrics. Next release will switch to protocol-buffers, switch early by passing --use-protocol-buffers flag

NAME           CPU(cores)   CPU%   MEMORY(bytes)   MEMORY%   

k8s-master01   235m         11%    1213Mi          65%       

k8s-master02   206m         10%    1203Mi          64%       

k8s-master03   232m         11%    1197Mi          64%       

k8s-node01     182m         18%    801Mi           43%       

k8s-node02     73m          7%     793Mi           42%    

 

 

12. Dashboard部署

[root@k8s-master01 metrics-server-0.4.x-kubeadm]# cd /root/k8s-ha-install/dashboard/

[root@k8s-master01 dashboard]# kubectl  create -f .

serviceaccount/admin-user created

clusterrolebinding.rbac.authorization.k8s.io/admin-user created

namespace/kubernetes-dashboard created

serviceaccount/kubernetes-dashboard created

service/kubernetes-dashboard created

secret/kubernetes-dashboard-certs created

secret/kubernetes-dashboard-csrf created

secret/kubernetes-dashboard-key-holder created

configmap/kubernetes-dashboard-settings created

role.rbac.authorization.k8s.io/kubernetes-dashboard created

clusterrole.rbac.authorization.k8s.io/kubernetes-dashboard created

rolebinding.rbac.authorization.k8s.io/kubernetes-dashboard created

clusterrolebinding.rbac.authorization.k8s.io/kubernetes-dashboard created

deployment.apps/kubernetes-dashboard created

service/dashboard-metrics-scraper created

deployment.apps/dashboard-metrics-scraper created

 

更改dashboard的svc为NodePort:

 

[root@k8s-master01 dashboard]# kubectl edit svc kubernetes-dashboard -n kubernetes-dashboard

 

 

selector:

    k8s-app: kubernetes-dashboard

  sessionAffinity: None

  type: NodePort

 

查看端口号:

[root@k8s-master01 dashboard]# kubectl get svc kubernetes-dashboard -n kubernetes-dashboard

NAME                   TYPE       CLUSTER-IP     EXTERNAL-IP   PORT(S)         AGE

kubernetes-dashboard   NodePort   10.97.249.90   <none>        443:30084/TCP   6m42s

 

 

根据自己的实例端口号,通过任意安装了kube-proxy的宿主机或者VIP的IP+端口即可访问到dashboard:

访问Dashboard:https://192.168.196.3:30084(请更改30084为自己的端口),选择登录方式为令牌(即token方式)

 

 

 

 

[root@k8s-master01 dashboard]# kubectl -n kube-system describe secret $(kubectl -n kube-system get secret | grep admin-user | awk '{print $1}')

 

 

 

将token值输入到令牌后,单击登录即可访问Dashboard


上一篇:Kubernetes容器集群管理环境 - Node节点的移除与加入


下一篇:Centos6.9下RocketMQ3.4.6高可用集群部署记录(双主双从+Nameserver+Console)