CentOS安装K8S图文教程(一次性成功)

2025-05-27 0 92

CentOS安装K8S图文教程(一次性成功)

一. 为什么是k8s v1.16.0?

最新版的v1.16.2试过了,一直无法安装完成,安装到kubeadm init那一步执行后,报了很多错,如:node xxx not found等。centos7都重装了几次,还是无法解决。用了一天都没安装完,差点放弃。后来在网上搜到的安装教程基本都是v1.16.0的,我不太相信是v1.16.2的坑所以先前没打算降级到v1.16.0。没办法了就试着安装v1.16.0版本,竟然成功了。记录在此,避免后来者踩坑。

本篇文章,安装大步骤如下:

  • 安装docker-ce 18.09.9(所有机器)
  • 设置k8s环境前置条件(所有机器)
  • 安装k8s v1.16.0 master管理节点
  • 安装k8s v1.16.0 node工作节点
  • 安装flannel(master)

这里有重要的一步,请记住自己master和node之间通信的ip,如我的master的ip为192.168.99.104,node的ip为:192.168.99.105. 请确保使用这两个ip在master和node上能互相ping通,这个master的ip 192.168.99.104接下来配置k8s的时候需要用到。

我的环境:

  • 操作系统:win10
  • 虚拟机:virtual box
  • linux发行版:CentOS7
  • linux内核(使用uname -r查看):3.10.0-957.el7.x86_64
  • master和node节点通信的ip(master):192.168.99.104

二. 安装docker-ce 18.09.9(所有机器)

所有安装k8s的机器都需要安装docker,命令如下:

#安装docker所需的工具

yuminstall-yyum-utilsdevice-mapper-persistent-datalvm2

#配置阿里云的docker源

yum-config-manager–add-repohttp://mirrors.aliyun.com/docker-ce/linux/centos/docker-ce.repo

#指定安装这个版本的docker-ce

yuminstall-ydocker-ce-18.09.9-3.el7

#启动docker

systemctlenabledocker&&systemctlstartdocker

三. 设置k8s环境准备条件(所有机器)

安装k8s的机器需要2个CPU和2g内存以上,这个简单,在虚拟机里面配置一下就可以了。然后执行以下脚本做一些准备操作。所有安装k8s的机器都需要这一步操作。

#关闭防火墙

systemctldisablefirewalld

systemctlstopfirewalld

#关闭selinux

#临时禁用selinux

setenforce0

#永久关闭修改/etc/sysconfig/selinux文件设置

sed-i's/SELINUX=permissive/SELINUX=disabled/'/etc/sysconfig/selinux

sed-i"s/SELINUX=enforcing/SELINUX=disabled/g"/etc/selinux/config

#禁用交换分区

swapoff-a

#永久禁用,打开/etc/fstab注释掉swap那一行。

sed-i's/.*swap.*/#&/'/etc/fstab

#修改内核参数

cat<<EOF>/etc/sysctl.d/k8s.conf

net.bridge.bridge-nf-call-ip6tables=1

net.bridge.bridge-nf-call-iptables=1

EOF

sysctl–system

四. 安装k8s v1.16.0 master管理节点

如果还没安装docker,请参照本文步骤二安装docker-ce 18.09.9(所有机器)安装。如果没设置k8s环境准备条件,请参照本文步骤三设置k8s环境准备条件(所有机器)执行。

以上两个步骤检查完毕之后,继续以下步骤。

1. 安装kubeadm、kubelet、kubectl

由于官方k8s源在google,国内无法访问,这里使用阿里云yum源

#执行配置k8s阿里云源

cat<<EOF>/etc/yum.repos.d/kubernetes.repo

[kubernetes]

name=Kubernetes

baseurl=https://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64/

enabled=1

gpgcheck=1

repo_gpgcheck=1

gpgkey=https://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpghttps://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpg

EOF

#安装kubeadm、kubectl、kubelet

yuminstall-ykubectl-1.16.0-0kubeadm-1.16.0-0kubelet-1.16.0-0

#启动kubelet服务

systemctlenablekubelet&&systemctlstartkubelet

2. 初始化k8s 以下这个命令开始安装k8s需要用到的docker镜像,因为无法访问到国外网站,所以这条命令使用的是国内的阿里云的源(registry.aliyuncs.com/google_containers)。另一个非常重要的是:这里的–apiserver-advertise-address使用的是master和node间能互相ping通的ip,我这里是192.168.99.104,刚开始在这里被坑了一个晚上,你请自己修改下ip执行。这条命令执行时会卡在[preflight] You can also perform this action in beforehand using ''kubeadm config images pull,大概需要2分钟,请耐心等待。

#下载管理节点中用到的6个docker镜像,你可以使用dockerimages查看到

#这里需要大概两分钟等待,会卡在[preflight]Youcanalsoperformthisactioninbeforehandusing''kubeadmconfigimagespull

kubeadminit–image-repositoryregistry.aliyuncs.com/google_containers–kubernetes-versionv1.16.0–apiserver-advertise-address192.168.99.104–pod-network-cidr=10.244.0.0/16–token-ttl0

上面安装完后,会提示你输入如下命令,复制粘贴过来,执行即可。

#上面安装完成后,k8s会提示你输入如下命令,执行

mkdir-p$HOME/.kube

sudocp-i/etc/kubernetes/admin.conf$HOME/.kube/config

sudochown$(id-u):$(id-g)$HOME/.kube/config

3. 记住node加入集群的命令 上面kubeadm init执行成功后会返回给你node节点加入集群的命令,等会要在node节点上执行,需要保存下来,如果忘记了,可以使用如下命令获取。

kubeadmtokencreate–print-join-command

以上,安装master节点完毕。可以使用kubectl get nodes查看一下,此时master处于NotReady状态,暂时不用管。

CentOS安装K8S图文教程(一次性成功)

五. 安装k8s v1.16.0 node工作节点

如果还没安装docker,请参照本文步骤二安装docker-ce 18.09.9(所有机器)安装。如果没设置k8s环境准备条件,请参照本文步骤三设置k8s环境准备条件(所有机器)执行。

以上两个步骤检查完毕之后,继续以下步骤。

1. 安装kubeadm、kubelet

#执行配置k8s阿里云源

cat<<EOF>/etc/yum.repos.d/kubernetes.repo

[kubernetes]

name=Kubernetes

baseurl=https://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64/

enabled=1

gpgcheck=1

repo_gpgcheck=1

gpgkey=https://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpghttps://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpg

EOF

#安装kubeadm、kubectl、kubelet

yuminstall-ykubeadm-1.16.0-0kubelet-1.16.0-0

#启动kubelet服务

systemctlenablekubelet&&systemctlstartkubelet

2. 加入集群 这里加入集群的命令每个人都不一样,可以登录master节点,使用kubeadm token create –print-join-command 来获取。获取后执行如下。

#加入集群,如果这里不知道加入集群的命令,可以登录master节点,使用kubeadmtokencreate–print-join-command来获取

kubeadmjoin192.168.99.104:6443–tokenncfrid.7ap0xiseuf97gikl\\

–discovery-token-ca-cert-hashsha256:47783e9851a1a517647f1986225f104e81dbfd8fb256ae55ef6d68ce9334c6a2

加入成功后,可以在master节点上使用kubectl get nodes命令查看到加入的节点。

六. 安装flannel(master机器)

以上步骤安装完后,机器搭建起来了,但状态还是NotReady状态,如下图,master机器需要安装flanneld。

CentOS安装K8S图文教程(一次性成功)

1. 下载官方fannel配置文件 使用wget命令,地址为:(https://raw.githubusercontent.com/coreos/flannel/master/Documentation/kube-flannel.yml),这个地址国内访问不了,所以我把内容复制下来,为了避免前面文章过长,我把它粘贴到文章末尾第八个步骤附录了。这个yml配置文件中配置了一个国内无法访问的地址(quay.io),我已经将其改为国内可以访问的地址(quay-mirror.qiniu.com)。我们新建一个kube-flannel.yml文件,复制粘贴该内容即可。

2. 安装fannel

kubectlapply-fkube-flannel.yml

七. 大功告成

至此,k8s集群搭建完成,如下图节点已为Ready状态,大功告成,完结撒花。

CentOS安装K8S图文教程(一次性成功)

八. 附录

这是kube-flannel.yml文件的内容,已经将无法访问的地址(quay.io)全部改为国内可以访问的地址(quay-mirror.qiniu.com)。我们新建一个kube-flannel.yml文件,复制粘贴该内容即可。

apiVersion:policy/v1beta1

kind:PodSecurityPolicy

metadata:

name:psp.flannel.unprivileged

annotations:

seccomp.security.alpha.kubernetes.io/allowedProfileNames:docker/default

seccomp.security.alpha.kubernetes.io/defaultProfileName:docker/default

apparmor.security.beta.kubernetes.io/allowedProfileNames:runtime/default

apparmor.security.beta.kubernetes.io/defaultProfileName:runtime/default

spec:

privileged:false

volumes:

-configMap

-secret

-emptyDir

-hostPath

allowedHostPaths:

-pathPrefix:"/etc/cni/net.d"

-pathPrefix:"/etc/kube-flannel"

-pathPrefix:"/run/flannel"

readOnlyRootFilesystem:false

#Usersandgroups

runAsUser:

rule:RunAsAny

supplementalGroups:

rule:RunAsAny

fsGroup:

rule:RunAsAny

#PrivilegeEscalation

allowPrivilegeEscalation:false

defaultAllowPrivilegeEscalation:false

#Capabilities

allowedCapabilities:['NET_ADMIN']

defaultAddCapabilities:[]

requiredDropCapabilities:[]

#Hostnamespaces

hostPID:false

hostIPC:false

hostNetwork:true

hostPorts:

-min:0

max:65535

#SELinux

seLinux:

#SELinuxisunusedinCaaSP

rule:'RunAsAny'

kind:ClusterRole

apiVersion:rbac.authorization.k8s.io/v1beta1

metadata:

name:flannel

rules:

-apiGroups:['extensions']

resources:['podsecuritypolicies']

verbs:['use']

resourceNames:['psp.flannel.unprivileged']

-apiGroups:

-""

resources:

-pods

verbs:

-get

-apiGroups:

-""

resources:

-nodes

verbs:

-list

-watch

-apiGroups:

-""

resources:

-nodes/status

verbs:

-patch

kind:ClusterRoleBinding

apiVersion:rbac.authorization.k8s.io/v1beta1

metadata:

name:flannel

roleRef:

apiGroup:rbac.authorization.k8s.io

kind:ClusterRole

name:flannel

subjects:

-kind:ServiceAccount

name:flannel

namespace:kube-system

apiVersion:v1

kind:ServiceAccount

metadata:

name:flannel

namespace:kube-system

kind:ConfigMap

apiVersion:v1

metadata:

name:kube-flannel-cfg

namespace:kube-system

labels:

tier:node

app:flannel

data:

cni-conf.json:|

{

"name":"cbr0",

"cniVersion":"0.3.1",

"plugins":[

{

"type":"flannel",

"delegate":{

"hairpinMode":true,

"isDefaultGateway":true

}

},

{

"type":"portmap",

"capabilities":{

"portMappings":true

}

}

]

}

net-conf.json:|

{

"Network":"10.244.0.0/16",

"Backend":{

"Type":"vxlan"

}

}

apiVersion:apps/v1

kind:DaemonSet

metadata:

name:kube-flannel-ds-amd64

namespace:kube-system

labels:

tier:node

app:flannel

spec:

selector:

matchLabels:

app:flannel

template:

metadata:

labels:

tier:node

app:flannel

spec:

affinity:

nodeAffinity:

requiredDuringSchedulingIgnoredDuringExecution:

nodeSelectorTerms:

-matchExpressions:

-key:beta.kubernetes.io/os

operator:In

values:

-linux

-key:beta.kubernetes.io/arch

operator:In

values:

-amd64

hostNetwork:true

tolerations:

-operator:Exists

effect:NoSchedule

serviceAccountName:flannel

initContainers:

-name:install-cni

image:quay-mirror.qiniu.com/coreos/flannel:v0.11.0-amd64

command:

-cp

args:

–f

-/etc/kube-flannel/cni-conf.json

-/etc/cni/net.d/10-flannel.conflist

volumeMounts:

-name:cni

mountPath:/etc/cni/net.d

-name:flannel-cfg

mountPath:/etc/kube-flannel/

containers:

-name:kube-flannel

image:quay-mirror.qiniu.com/coreos/flannel:v0.11.0-amd64

command:

-/opt/bin/flanneld

args:

—ip-masq

—kube-subnet-mgr

resources:

requests:

cpu:"100m"

memory:"50Mi"

limits:

cpu:"100m"

memory:"50Mi"

securityContext:

privileged:false

capabilities:

add:["NET_ADMIN"]

env:

-name:POD_NAME

valueFrom:

fieldRef:

fieldPath:metadata.name

-name:POD_NAMESPACE

valueFrom:

fieldRef:

fieldPath:metadata.namespace

volumeMounts:

-name:run

mountPath:/run/flannel

-name:flannel-cfg

mountPath:/etc/kube-flannel/

volumes:

-name:run

hostPath:

path:/run/flannel

-name:cni

hostPath:

path:/etc/cni/net.d

-name:flannel-cfg

configMap:

name:kube-flannel-cfg

apiVersion:apps/v1

kind:DaemonSet

metadata:

name:kube-flannel-ds-arm64

namespace:kube-system

labels:

tier:node

app:flannel

spec:

selector:

matchLabels:

app:flannel

template:

metadata:

labels:

tier:node

app:flannel

spec:

affinity:

nodeAffinity:

requiredDuringSchedulingIgnoredDuringExecution:

nodeSelectorTerms:

-matchExpressions:

-key:beta.kubernetes.io/os

operator:In

values:

-linux

-key:beta.kubernetes.io/arch

operator:In

values:

-arm64

hostNetwork:true

tolerations:

-operator:Exists

effect:NoSchedule

serviceAccountName:flannel

initContainers:

-name:install-cni

image:quay-mirror.qiniu.com/coreos/flannel:v0.11.0-arm64

command:

-cp

args:

–f

-/etc/kube-flannel/cni-conf.json

-/etc/cni/net.d/10-flannel.conflist

volumeMounts:

-name:cni

mountPath:/etc/cni/net.d

-name:flannel-cfg

mountPath:/etc/kube-flannel/

containers:

-name:kube-flannel

image:quay-mirror.qiniu.com/coreos/flannel:v0.11.0-arm64

command:

-/opt/bin/flanneld

args:

—ip-masq

—kube-subnet-mgr

resources:

requests:

cpu:"100m"

memory:"50Mi"

limits:

cpu:"100m"

memory:"50Mi"

securityContext:

privileged:false

capabilities:

add:["NET_ADMIN"]

env:

-name:POD_NAME

valueFrom:

fieldRef:

fieldPath:metadata.name

-name:POD_NAMESPACE

valueFrom:

fieldRef:

fieldPath:metadata.namespace

volumeMounts:

-name:run

mountPath:/run/flannel

-name:flannel-cfg

mountPath:/etc/kube-flannel/

volumes:

-name:run

hostPath:

path:/run/flannel

-name:cni

hostPath:

path:/etc/cni/net.d

-name:flannel-cfg

configMap:

name:kube-flannel-cfg

apiVersion:apps/v1

kind:DaemonSet

metadata:

name:kube-flannel-ds-arm

namespace:kube-system

labels:

tier:node

app:flannel

spec:

selector:

matchLabels:

app:flannel

template:

metadata:

labels:

tier:node

app:flannel

spec:

affinity:

nodeAffinity:

requiredDuringSchedulingIgnoredDuringExecution:

nodeSelectorTerms:

-matchExpressions:

-key:beta.kubernetes.io/os

operator:In

values:

-linux

-key:beta.kubernetes.io/arch

operator:In

values:

-arm

hostNetwork:true

tolerations:

-operator:Exists

effect:NoSchedule

serviceAccountName:flannel

initContainers:

-name:install-cni

image:quay-mirror.qiniu.com/coreos/flannel:v0.11.0-arm

command:

-cp

args:

–f

-/etc/kube-flannel/cni-conf.json

-/etc/cni/net.d/10-flannel.conflist

volumeMounts:

-name:cni

mountPath:/etc/cni/net.d

-name:flannel-cfg

mountPath:/etc/kube-flannel/

containers:

-name:kube-flannel

image:quay-mirror.qiniu.com/coreos/flannel:v0.11.0-arm

command:

-/opt/bin/flanneld

args:

—ip-masq

—kube-subnet-mgr

resources:

requests:

cpu:"100m"

memory:"50Mi"

limits:

cpu:"100m"

memory:"50Mi"

securityContext:

privileged:false

capabilities:

add:["NET_ADMIN"]

env:

-name:POD_NAME

valueFrom:

fieldRef:

fieldPath:metadata.name

-name:POD_NAMESPACE

valueFrom:

fieldRef:

fieldPath:metadata.namespace

volumeMounts:

-name:run

mountPath:/run/flannel

-name:flannel-cfg

mountPath:/etc/kube-flannel/

volumes:

-name:run

hostPath:

path:/run/flannel

-name:cni

hostPath:

path:/etc/cni/net.d

-name:flannel-cfg

configMap:

name:kube-flannel-cfg

apiVersion:apps/v1

kind:DaemonSet

metadata:

name:kube-flannel-ds-ppc64le

namespace:kube-system

labels:

tier:node

app:flannel

spec:

selector:

matchLabels:

app:flannel

template:

metadata:

labels:

tier:node

app:flannel

spec:

affinity:

nodeAffinity:

requiredDuringSchedulingIgnoredDuringExecution:

nodeSelectorTerms:

-matchExpressions:

-key:beta.kubernetes.io/os

operator:In

values:

-linux

-key:beta.kubernetes.io/arch

operator:In

values:

-ppc64le

hostNetwork:true

tolerations:

-operator:Exists

effect:NoSchedule

serviceAccountName:flannel

initContainers:

-name:install-cni

image:quay-mirror.qiniu.com/coreos/flannel:v0.11.0-ppc64le

command:

-cp

args:

–f

-/etc/kube-flannel/cni-conf.json

-/etc/cni/net.d/10-flannel.conflist

volumeMounts:

-name:cni

mountPath:/etc/cni/net.d

-name:flannel-cfg

mountPath:/etc/kube-flannel/

containers:

-name:kube-flannel

image:quay-mirror.qiniu.com/coreos/flannel:v0.11.0-ppc64le

command:

-/opt/bin/flanneld

args:

—ip-masq

—kube-subnet-mgr

resources:

requests:

cpu:"100m"

memory:"50Mi"

limits:

cpu:"100m"

memory:"50Mi"

securityContext:

privileged:false

capabilities:

add:["NET_ADMIN"]

env:

-name:POD_NAME

valueFrom:

fieldRef:

fieldPath:metadata.name

-name:POD_NAMESPACE

valueFrom:

fieldRef:

fieldPath:metadata.namespace

volumeMounts:

-name:run

mountPath:/run/flannel

-name:flannel-cfg

mountPath:/etc/kube-flannel/

volumes:

-name:run

hostPath:

path:/run/flannel

-name:cni

hostPath:

path:/etc/cni/net.d

-name:flannel-cfg

configMap:

name:kube-flannel-cfg

apiVersion:apps/v1

kind:DaemonSet

metadata:

name:kube-flannel-ds-s390x

namespace:kube-system

labels:

tier:node

app:flannel

spec:

selector:

matchLabels:

app:flannel

template:

metadata:

labels:

tier:node

app:flannel

spec:

affinity:

nodeAffinity:

requiredDuringSchedulingIgnoredDuringExecution:

nodeSelectorTerms:

-matchExpressions:

-key:beta.kubernetes.io/os

operator:In

values:

-linux

-key:beta.kubernetes.io/arch

operator:In

values:

-s390x

hostNetwork:true

tolerations:

-operator:Exists

effect:NoSchedule

serviceAccountName:flannel

initContainers:

-name:install-cni

image:quay-mirror.qiniu.com/coreos/flannel:v0.11.0-s390x

command:

-cp

args:

–f

-/etc/kube-flannel/cni-conf.json

-/etc/cni/net.d/10-flannel.conflist

volumeMounts:

-name:cni

mountPath:/etc/cni/net.d

-name:flannel-cfg

mountPath:/etc/kube-flannel/

containers:

-name:kube-flannel

image:quay-mirror.qiniu.com/coreos/flannel:v0.11.0-s390x

command:

-/opt/bin/flanneld

args:

—ip-masq

—kube-subnet-mgr

resources:

requests:

cpu:"100m"

memory:"50Mi"

limits:

cpu:"100m"

memory:"50Mi"

securityContext:

privileged:false

capabilities:

add:["NET_ADMIN"]

env:

-name:POD_NAME

valueFrom:

fieldRef:

fieldPath:metadata.name

-name:POD_NAMESPACE

valueFrom:

fieldRef:

fieldPath:metadata.namespace

volumeMounts:

-name:run

mountPath:/run/flannel

-name:flannel-cfg

mountPath:/etc/kube-flannel/

volumes:

-name:run

hostPath:

path:/run/flannel

-name:cni

hostPath:

path:/etc/cni/net.d

-name:flannel-cfg

configMap:

name:kube-flannel-cfg

原文地址:https://mp.weixin.qq.com/s/KGv6gbXRepG8Zu0-u8odIA

收藏 (0) 打赏

感谢您的支持,我会继续努力的!

打开微信/支付宝扫一扫,即可进行扫码打赏哦,分享从这里开始,精彩与您同在
点赞 (0)

声明:本站所有文章,如无特殊说明或标注,均为本站原创发布。任何个人或组织,在未征得本站同意时,禁止复制、盗用、采集、发布本站内容到任何网站、书籍等各类媒体平台。如若本站内容侵犯了原著者的合法权益,可联系我们进行处理。

快网idc优惠网 建站教程 CentOS安装K8S图文教程(一次性成功) https://www.kuaiidc.com/57053.html

相关文章

发表评论
暂无评论