企业营销型网站建设,wordpress 一键转载,有什么网站可以免费建站,网站功能分析目录 一、概述
二、探针类型
2.1、就绪探针#xff08;Readiness Probe#xff09;
2.2、存活探针#xff08;Liveness Probe#xff09;
三、探针探测方法
3.1、exec
3.2、httpGet
3.3、tcpSocket
四、探针配置项
五、探针使用
5.1、就绪探针#xff08;Readin…目录 一、概述
二、探针类型
2.1、就绪探针Readiness Probe
2.2、存活探针Liveness Probe
三、探针探测方法
3.1、exec
3.2、httpGet
3.3、tcpSocket
四、探针配置项
五、探针使用
5.1、就绪探针Readiness Probe
5.2、存活探针Liveness Probe
5.3、TCP就绪/存活探测
六、Liveness Probe存活探针 VS Readiness Probe就绪探针 一、概述
在k8s中我们不能仅仅通过查看应用的运行状态来判断应用是否正常因为在某些时候容器正常运行并不能代表应用健康因此k8s提供了探针Probe技术来帮助我们判断容器内运行的应用是否运行正常探针有点类似心跳检测。
二、探针类型
Kubernetes 的探针有三种类型
2.1、就绪探针Readiness Probe
判断容器是否启动完成即容器的 Ready 是否为 True可以接收请求如果ReadinessProbe 探测失败则容器的 Ready 将为 False控制器将此Pod 的Endpoint 从对应的Service的Endpoint 列表中移除从此不再将任何请求调度此Pod 上直到下次探测成功。通过使用 Readiness 探针Kubernetes 能够等待应用程序完全启动然后才允许服务将流量发送到新副本。
2.2、存活探针Liveness Probe
判断容器是否存活即 Pod 是否为 running 状态如果 LivenessProbe探测到容器不健康则 kubelet 将 kill 掉容器并根据容器的重启策略是否重启。如果一个容器不包含 LivenessProbe 探针则 Kubelet 认为容器的 LivenessProbe 探针的返回值永远成功。
有时应用程序可能因为某些原因后端服务故障等导致暂时无法对外提供服务但应用软件没有终止导致 k8s无法隔离有故障的 pod调用者可能会访问到有故障的pod导致业务不稳定。k8s提供 livenessProbe 来检测应用程序是否正常运行并且对相应状况进行相应的补救措施。
三、探针探测方法
每类探针都支持三种探测方法
3.1、exec
通过在容器内执行命令来检查服务是否正常针对复杂检测或无HTTP 接口的服务返回值为 0则表示容器健康。
3.2、httpGet
通过发送 http 请求检查服务是否正常返回 200-399 状态码则表明容器健康。
3.3、tcpSocket
通过容器的 IP 和 Port 执行 TCP 检查如果能够建立TCP 连接则表明容器健康。
四、探针配置项
探针(Probe)有许多可选字段可以用来更加精确的控制探针的行为。这些参数包括
initialDelaySeconds容器启动后第一次执行探测是需要等待多少秒periodSeconds执行探测的间隔时间默认是10秒timeoutSeconds超时时间当超过我们定义的时间后便会被视为失败默认 1 秒successThreshold探测失败后最少连续探测成功多少次才被认定为成功默认是1次。failureThreshold探测成功后最少连续探测失败多少次才被认定为失败默认是3次
五、探针使用
5.1、就绪探针Readiness Probe
创建Pod资源清单vim nginx-readiness-probe.yaml
apiVersion: v1
kind: Pod
metadata:name: nginx-readiness-probe
spec:containers:- name: nginx-readiness-probeimage: nginxreadinessProbe: # 就绪探针httpGet: # 对容器的IP地址、端口和URL路径来发送GET请求path: /healthzport: 80initialDelaySeconds: 10 # 等待10s后便开始就绪检查periodSeconds: 5 # 间隔5s检查一次successThreshold: 2 # 探测失败后最少连续探测成功多少次才被认定为成功
我们指定了探针检测方式为httpGet通过发送 http 请求检查服务是否正常返回 200-399 状态码则表明容器健康。
$ kubectl apply -f nginx-readiness-probe.yaml
pod/nginx-readiness-probe created$ kubectl get pod -o wide
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
nginx-readiness-probe 0/1 Running 0 32s 192.168.1.3 node01 none none$ kubectl describe pod nginx-readiness-probe
Name: nginx-readiness-probe
Namespace: default
Priority: 0
Service Account: default
Node: node01/172.30.2.2
Start Time: Mon, 16 Jan 2023 03:23:11 0000
Labels: none
Annotations: cni.projectcalico.org/containerID: 67b08cbc5b07020dcd7040cd47565c5405ee82641a9d3d68d9fd68b6b599c10f
cni.projectcalico.org/podIP: 192.168.1.3/32
cni.projectcalico.org/podIPs: 192.168.1.3/32
Status: Running
IP: 192.168.1.3
IPs:
IP: 192.168.1.3
Containers:
nginx-readiness-probe:
Container ID: containerd://23eca4eaeffce3e6801d3e7c26a60360d33b1fdb4046843ff9cf7c647adcf0a2
Image: nginx
Image ID: docker.io/library/nginxsha256:b8f2383a95879e1ae064940d9a200f67a6c79e710ed82ac42263397367e7cc4e
Port: none
Host Port: none
State: Running
Started: Mon, 16 Jan 2023 03:23:16 0000
Ready: False
Restart Count: 0
Readiness: http-get http://:80/healthz delay10s timeout1s period5s #success2 #failure3
Environment: none
Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-8xsm8 (ro)
Conditions:
Type Status
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-8xsm8:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: nil
DownwardAPI: true
QoS Class: BestEffort
Node-Selectors: none
Tolerations: node.kubernetes.io/not-ready:NoExecute opExists for 300s
node.kubernetes.io/unreachable:NoExecute opExists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 43s default-scheduler Successfully assigned default/nginx-readiness-probe to node01
Normal Pulling 43s kubelet Pulling image nginx
Normal Pulled 38s kubelet Successfully pulled image nginx in 4.968467021s (4.968471311s including waiting)
Normal Created 38s kubelet Created container nginx-readiness-probe
Normal Started 38s kubelet Started container nginx-readiness-probe
Warning Unhealthy 3s (x6 over 28s) kubelet Readiness probe failed: HTTP probe failed with statuscode: 404
通过describe查看Pod描述信息后可以看到Readiness probe就绪检测失败了失败原因就是我们的nginx容器中并不存在/healthz这个接口所以响应码是404并不在 200-399 状态码中所以我们看到的Pod的Ready一直都是未就绪状态。 5.2、存活探针Liveness Probe
创建Pod资源清单vim centos-liveness-probe.yaml
apiVersion: v1
kind: Pod
metadata:name: centos-liveness-probe
spec:containers:- name: centos-liveness-probeimage: centosargs: # 容器启动时执行如下的命令, 30s后删除/tmp/healthy文件- /bin/sh- -c- touch /tmp/healthy; sleep 30; rm -rf /tmp/healthy; sleep 600livenessProbe: # 存活探针exec: # 在容器内执行指定命令cat /tmp/healthycommand:- cat- /tmp/healthyinitialDelaySeconds: 5 # 等待5s后开始存活检查periodSeconds: 5 # 间隔5s检查一次
在这个配置文件中可以看到 Pod 中只有一个 Container。 periodSeconds 字段指定了 kubelet 应该每 5 秒执行一次存活探测。 initialDelaySeconds 字段告诉 kubelet 在执行第一次探测前应该等待 5 秒。 kubelet 在容器内执行命令 cat /tmp/healthy 来进行探测。 如果命令执行成功并且返回值为 0kubelet 就会认为这个容器是健康存活的。 如果这个命令返回非 0 值kubelet 会杀死这个容器并根据重启策略重新启动它。
当容器启动时执行如下的命令
/bin/sh -c touch /tmp/healthy; sleep 30; rm -f /tmp/healthy; sleep 600
这个容器生命的前 30 秒/tmp/healthy 文件是存在的。 所以在这最开始的 30 秒内执行命令 cat /tmp/healthy 会返回成功代码。 30 秒之后执行命令 cat /tmp/healthy 就会返回失败代码。
创建 Pod
$ kubectl apply -f centos-liveness-probe.yaml
pod/centos-liveness-probe created$ kubectl get pod -o wide
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
centos-liveness-probe 1/1 Running 0 9s 192.168.1.6 node01 none none# 在 30 秒内查看 Pod 的事件
$ kubectl describe pod centos-liveness-probe
Name: centos-liveness-probe
Namespace: default
Priority: 0
Service Account: default
Node: node01/172.30.2.2
Start Time: Mon, 16 Jan 2023 03:42:49 0000
Labels: none
Annotations: cni.projectcalico.org/containerID: 74ae52265e8236ec904a23c98f8eb6a929df6709c29643f8cf3a624274105ab6cni.projectcalico.org/podIP: 192.168.1.6/32cni.projectcalico.org/podIPs: 192.168.1.6/32
Status: Running
IP: 192.168.1.6
IPs:IP: 192.168.1.6
Containers:centos-liveness-probe:Container ID: containerd://272c3f8bf293271f3657d98e1e23922312de46afebc3ed76a65104bbe4209e39Image: centosImage ID: docker.io/library/centossha256:a27fd8080b517143cbbbab9dfb7c8571c40d67d534bbdee55bd6c473f432b177Port: noneHost Port: noneArgs:/bin/sh-ctouch /tmp/healthy; sleep 30; rm -rf /tmp/healthy; sleep 600State: RunningStarted: Mon, 16 Jan 2023 03:42:50 0000Ready: TrueRestart Count: 0Liveness: exec [cat /tmp/healthy] delay5s timeout1s period5s #success1 #failure3Environment: noneMounts:/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-d69p9 (ro)
Conditions:Type StatusInitialized True Ready True ContainersReady True PodScheduled True
Volumes:kube-api-access-d69p9:Type: Projected (a volume that contains injected data from multiple sources)TokenExpirationSeconds: 3607ConfigMapName: kube-root-ca.crtConfigMapOptional: nilDownwardAPI: true
QoS Class: BestEffort
Node-Selectors: none
Tolerations: node.kubernetes.io/not-ready:NoExecute opExists for 300snode.kubernetes.io/unreachable:NoExecute opExists for 300s
Events:Type Reason Age From Message---- ------ ---- ---- -------Normal Scheduled 17s default-scheduler Successfully assigned default/centos-liveness-probe to node01Normal Pulling 16s kubelet Pulling image centosNormal Pulled 16s kubelet Successfully pulled image centos in 458.064227ms (458.070248ms including waiting)Normal Created 16s kubelet Created container centos-liveness-probeNormal Started 16s kubelet Started container centos-liveness-probe
如上可以看到30s内输出结果表明还没有存活探针失败。 那么等待30s以后我们再次查看Pod详细信息
$ kubectl describe pod centos-liveness-probe
Name: centos-liveness-probe
Namespace: default
Priority: 0
Service Account: default
Node: node01/172.30.2.2
Start Time: Mon, 16 Jan 2023 03:42:49 0000
Labels: none
Annotations: cni.projectcalico.org/containerID: 74ae52265e8236ec904a23c98f8eb6a929df6709c29643f8cf3a624274105ab6cni.projectcalico.org/podIP: 192.168.1.6/32cni.projectcalico.org/podIPs: 192.168.1.6/32
Status: Running
IP: 192.168.1.6
IPs:IP: 192.168.1.6
Containers:centos-liveness-probe:Container ID: containerd://b03f2aaf2b854071223aae43cdfee4b9d1d4d3dd03f8ee7270b857817d362ca7Image: centosImage ID: docker.io/library/centossha256:a27fd8080b517143cbbbab9dfb7c8571c40d67d534bbdee55bd6c473f432b177Port: noneHost Port: noneArgs:/bin/sh-ctouch /tmp/healthy; sleep 30; rm -rf /tmp/healthy; sleep 600State: RunningStarted: Mon, 16 Jan 2023 03:44:05 0000Last State: TerminatedReason: ErrorExit Code: 137Started: Mon, 16 Jan 2023 03:42:50 0000Finished: Mon, 16 Jan 2023 03:44:04 0000Ready: TrueRestart Count: 1Liveness: exec [cat /tmp/healthy] delay5s timeout1s period5s #success1 #failure3Environment: noneMounts:/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-d69p9 (ro)
Conditions:Type StatusInitialized True Ready True ContainersReady True PodScheduled True
Volumes:kube-api-access-d69p9:Type: Projected (a volume that contains injected data from multiple sources)TokenExpirationSeconds: 3607ConfigMapName: kube-root-ca.crtConfigMapOptional: nilDownwardAPI: true
QoS Class: BestEffort
Node-Selectors: none
Tolerations: node.kubernetes.io/not-ready:NoExecute opExists for 300snode.kubernetes.io/unreachable:NoExecute opExists for 300s
Events:Type Reason Age From Message---- ------ ---- ---- -------Normal Scheduled 109s default-scheduler Successfully assigned default/centos-liveness-probe to node01Normal Pulled 108s kubelet Successfully pulled image centos in 458.064227ms (458.070248ms including waiting)Warning Unhealthy 64s (x3 over 74s) kubelet Liveness probe failed: cat: /tmp/healthy: No such file or directoryNormal Killing 64s kubelet Container centos-liveness-probe failed liveness probe, will be restartedNormal Pulling 33s (x2 over 108s) kubelet Pulling image centosNormal Created 33s (x2 over 108s) kubelet Created container centos-liveness-probeNormal Started 33s (x2 over 108s) kubelet Started container centos-liveness-probeNormal Pulled 33s kubelet Successfully pulled image centos in 405.865965ms (405.870664ms including waiting)
在输出结果的最下面有信息显示存活探针失败了Liveness probe failed: cat: /tmp/healthy: No such file or directory这个失败的容器被杀死并且被重建了。
再等 30 秒这个容器被重启了输出结果显示 RESTARTS 的值增加了 1。 请注意一旦失败的容器恢复为运行状态RESTARTS 计数器就会增加 1
controlplane $ kubectl get pod -o wide
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
centos-liveness-probe 1/1 Running 2 (69s ago) 3m39s 192.168.1.6 node01 none none
因为默认的重启策略restartPolicy是Always所以centos-liveness-probe将会一直重启。
5.3、TCP就绪/存活探测
前面两个示例分别演示了exec和httpGet探测方式这里演示一下基于tcpSocket的探测方式。
创建资源清单文件vim tcp-socket-probe.yaml
apiVersion: v1
kind: Pod
metadata:name: goproxylabels:app: goproxy
spec:containers:- name: goproxyimage: registry.k8s.io/goproxy:0.1ports:- containerPort: 8080readinessProbe: # 就绪探针tcpSocket:port: 8080initialDelaySeconds: 5periodSeconds: 10livenessProbe: # 存活探针tcpSocket:port: 8080initialDelaySeconds: 15periodSeconds: 20
如上配置kubelet 会在容器启动 5 秒后发送第一个就绪探针。 探针会尝试连接 goproxy 容器的 8080 端口。 如果探测成功这个 Pod 会被标记为就绪状态kubelet 将继续每隔 10 秒运行一次探测。
除了就绪探针这个配置包括了一个存活探针。 kubelet 会在容器启动 15 秒后进行第一次存活探测。 与就绪探针类似存活探针会尝试连接 goproxy 容器的 8080 端口。 如果存活探测失败容器会被重新启动。
$ kubectl apply -f tcp-socket-probe.yaml
pod/goproxy created$ kubectl get pod/goproxy -o wide
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
goproxy 1/1 Running 0 99s 192.168.1.3 node01 none none
可以看到goproxy容器的READY为1/1说明就绪探测成功了并且STATUS为Running运行状态说明goproxy容器当前是健康的。
$ kubectl describe pod/goproxy
Name: goproxy
Namespace: default
Priority: 0
Service Account: default
Node: node01/172.30.2.2
Start Time: Mon, 16 Jan 2023 05:15:18 0000
Labels: appgoproxy
Annotations: cni.projectcalico.org/containerID: 24cf48d7ee8e5ea9fe846afb16510c46cd63c9214c3a54aa2d548e647aa162fbcni.projectcalico.org/podIP: 192.168.1.3/32cni.projectcalico.org/podIPs: 192.168.1.3/32
Status: Running
IP: 192.168.1.3
IPs:IP: 192.168.1.3
Containers:goproxy:Container ID: containerd://c07e285dda94eb1ebc75c7aef01dc1816d4f027c007a7bb741b2a023ab4112d2Image: registry.k8s.io/goproxy:0.1Image ID: registry.k8s.io/goproxysha256:5334c7ad43048e3538775cb09aaf184f5e8acf4b0ea60e3bc8f1d93c209865a5Port: 8080/TCPHost Port: 0/TCPState: RunningStarted: Mon, 16 Jan 2023 05:15:21 0000Ready: TrueRestart Count: 0Liveness: tcp-socket :8080 delay15s timeout1s period20s #success1 #failure3Readiness: tcp-socket :8080 delay5s timeout1s period10s #success1 #failure3Environment: noneMounts:/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-4bcg8 (ro)
Conditions:Type StatusInitialized True Ready True ContainersReady True PodScheduled True
Volumes:kube-api-access-4bcg8:Type: Projected (a volume that contains injected data from multiple sources)TokenExpirationSeconds: 3607ConfigMapName: kube-root-ca.crtConfigMapOptional: nilDownwardAPI: true
QoS Class: BestEffort
Node-Selectors: none
Tolerations: node.kubernetes.io/not-ready:NoExecute opExists for 300snode.kubernetes.io/unreachable:NoExecute opExists for 300s
Events:Type Reason Age From Message---- ------ ---- ---- -------Normal Scheduled 3m27s default-scheduler Successfully assigned default/goproxy to node01Normal Pulling 3m26s kubelet Pulling image registry.k8s.io/goproxy:0.1Normal Pulled 3m24s kubelet Successfully pulled image registry.k8s.io/goproxy:0.1 in 2.531715761s (2.53172162s including waiting)Normal Created 3m24s kubelet Created container goproxyNormal Started 3m24s kubelet Started container goproxy
六、Liveness Probe存活探针 VS Readiness Probe就绪探针 liveness probe存活探针 readiness probe就绪探针 用途 判断容器是否存活 判断Pod是否就绪 检测期 Pod运行期 Pod启动期 失败处理 Kill容器 停止向Pod发送流量 探针类型 httpGet、exec、tcpSocket httpGet、exec、tcpSocket