728x90
반응형
Overview
오늘은 k8s 인증서 10년 만기 생성 방법에 대해서 알아보려고 한다.
나의 k8s 버전은 1.16.8이다.
참고자료
아래의 github의 소스코드를 가지고 진행했습니다.
https://github.com/yuyicai/update-kube-cert
해당 github를 보면, v1.9 이하인 경우와 v1.9이상인 경우에 대해서 전부 정리되어있다.
클러스터 버전이 v1.9 이하인 경우, etcd는 기본적으로 TLS 연결을 사용하지 않습니다.
따라서 마스터 노드의 인증서만 업데이트하면 됩니다.
1. 인증서 만료일 확인
$ sudo openssl x509 -in /etc/kubernetes/ssl/apiserver.crt -noout -text |grep Not
Not Before: Sep 16 01:33:38 2022 GMT
Not After : Sep 13 01:33:38 2023 GMT
2. 마스터 노드 인증서 갱신
- 각 마스터 노드 순차 작업 실행
1) 이전 .kube/config 백업
$ cp /home/clex/.kube/config /home/clex/.kube/config-old.$(date +%Y%m%d)
2) 새로운 인증서 발급 소스코드를 git에서 다운로드
$ git clone https://github.com/yuyicai/update-kube-cert.git
$ cd update-kube-cert
$ chmod 755 update-kubeadm-cert.sh
3) 인증서 생성 shell 내용
#!/usr/bin/env bash
set -o errexit
set -o pipefail
# set -o xtrace
# set output color
NC='\033[0m'
RED='\033[31m'
GREEN='\033[32m'
YELLOW='\033[33m'
BLUE='\033[34m'
log::err() {
printf "[$(date +'%Y-%m-%dT%H:%M:%S.%2N%z')][${RED}ERROR${NC}] %b\n" "$@"
}
log::info() {
printf "[$(date +'%Y-%m-%dT%H:%M:%S.%2N%z')][INFO] %b\n" "$@"
}
log::warning() {
printf "[$(date +'%Y-%m-%dT%H:%M:%S.%2N%z')][${YELLOW}WARNING${NC}] \033[0m%b\n" "$@"
}
check_file() {
if [[ ! -r ${1} ]]; then
log::err "can not find ${1}"
exit 1
fi
}
# get x509v3 subject alternative name from the old certificate
cert::get_subject_alt_name() {
local cert=${1}.crt
local alt_name
check_file "${cert}"
alt_name=$(openssl x509 -text -noout -in "${cert}" | grep -A1 'Alternative' | tail -n1 | sed 's/[[:space:]]*Address//g')
printf "%s\n" "${alt_name}"
}
# get subject from the old certificate
cert::get_subj() {
local cert=${1}.crt
local subj
check_file "${cert}"
subj=$(openssl x509 -text -noout -in "${cert}" | grep "Subject:" | sed 's/Subject:/\//g;s/\,/\//;s/[[:space:]]//g')
printf "%s\n" "${subj}"
}
cert::backup_file() {
local file=${1}
if [[ ! -e ${file}.old-$(date +%Y%m%d) ]]; then
cp -rp "${file}" "${file}.old-$(date +%Y%m%d)"
log::info "backup ${file} to ${file}.old-$(date +%Y%m%d)"
else
log::warning "does not backup, ${file}.old-$(date +%Y%m%d) already exists"
fi
}
# check certificate expiration
cert::check_cert_expiration() {
local cert=${1}.crt
local cert_expires
cert_expires=$(openssl x509 -text -noout -in "${cert}" | awk -F ": " '/Not After/{print$2}')
printf "%s\n" "${cert_expires}"
}
# check kubeconfig expiration
cert::check_kubeconfig_expiration() {
local config=${1}.conf
local cert
local cert_expires
cert=$(grep "client-certificate-data" "${config}" | awk '{print$2}' | base64 -d)
cert_expires=$(openssl x509 -text -noout -in <(printf "%s" "${cert}") | awk -F ": " '/Not After/{print$2}')
printf "%s\n" "${cert_expires}"
}
# check etcd certificates expiration
cert::check_etcd_certs_expiration() {
local cert
local certs
certs=(
"${ETCD_CERT_CA}"
"${ETCD_CERT_SERVER}"
"${ETCD_CERT_PEER}"
"${ETCD_CERT_HEALTHCHECK_CLIENT}"
"${ETCD_CERT_APISERVER_ETCD_CLIENT}"
)
for cert in "${certs[@]}"; do
if [[ ! -r ${cert} ]]; then
printf "%-50s%-30s\n" "${cert}.crt" "$(cert::check_cert_expiration "${cert}")"
fi
done
}
# check master certificates expiration
cert::check_master_certs_expiration() {
local certs
local kubeconfs
local cert
local conf
certs=(
"${CERT_CA}"
"${CERT_APISERVER}"
"${CERT_APISERVER_KUBELET_CLIENT}"
"${FRONT_PROXY_CA}"
"${FRONT_PROXY_CLIENT}"
)
kubeconfs=(
"${CONF_CONTROLLER_MANAGER}"
"${CONF_SCHEDULER}"
"${CONF_ADMIN}"
)
printf "%-50s%-30s\n" "CERTIFICATE" "EXPIRES"
for conf in "${kubeconfs[@]}"; do
if [[ ! -r ${conf} ]]; then
printf "%-50s%-30s\n" "${conf}.config" "$(cert::check_kubeconfig_expiration "${conf}")"
fi
done
for cert in "${certs[@]}"; do
if [[ ! -r ${cert} ]]; then
printf "%-50s%-30s\n" "${cert}.crt" "$(cert::check_cert_expiration "${cert}")"
fi
done
}
# check all certificates expiration
cert::check_all_expiration() {
cert::check_master_certs_expiration
cert::check_etcd_certs_expiration
}
# generate certificate whit client, server or peer
# Args:
# $1 (the name of certificate)
# $2 (the type of certificate, must be one of client, server, peer)
# $3 (the subject of certificates)
# $4 (the validity of certificates) (days)
# $5 (the name of ca)
# $6 (the x509v3 subject alternative name of certificate when the type of certificate is server or peer)
cert::gen_cert() {
local cert_name=${1}
local cert_type=${2}
local subj=${3}
local cert_days=${4}
local ca_name=${5}
local alt_name=${6}
local ca_cert=${ca_name}.crt
local ca_key=${ca_name}.key
local cert=${cert_name}.crt
local key=${cert_name}.key
local csr=${cert_name}.csr
local common_csr_conf='distinguished_name = dn\n[dn]\n[v3_ext]\nkeyUsage = critical, digitalSignature, keyEncipherment\n'
for file in "${ca_cert}" "${ca_key}" "${cert}" "${key}"; do
check_file "${file}"
done
case "${cert_type}" in
client)
csr_conf=$(printf "%bextendedKeyUsage = clientAuth\n" "${common_csr_conf}")
;;
server)
csr_conf=$(printf "%bextendedKeyUsage = serverAuth\nsubjectAltName = %b\n" "${common_csr_conf}" "${alt_name}")
;;
peer)
csr_conf=$(printf "%bextendedKeyUsage = serverAuth, clientAuth\nsubjectAltName = %b\n" "${common_csr_conf}" "${alt_name}")
;;
*)
log::err "unknow, unsupported certs type: ${YELLOW}${cert_type}${NC}, supported type: client, server, peer"
exit 1
;;
esac
# gen csr
openssl req -new -key "${key}" -subj "${subj}" -reqexts v3_ext \
-config <(printf "%b" "${csr_conf}") \
-out "${csr}" >/dev/null 2>&1
# gen cert
openssl x509 -in "${csr}" -req -CA "${ca_cert}" -CAkey "${ca_key}" -CAcreateserial -extensions v3_ext \
-extfile <(printf "%b" "${csr_conf}") \
-days "${cert_days}" -out "${cert}" >/dev/null 2>&1
rm -f "${csr}"
}
cert::update_kubeconf() {
local cert_name=${1}
local kubeconf_file=${cert_name}.conf
local cert=${cert_name}.crt
local key=${cert_name}.key
local subj
local cert_base64
check_file "${kubeconf_file}"
# get the key from the old kubeconf
grep "client-key-data" "${kubeconf_file}" | awk '{print$2}' | base64 -d >"${key}"
# get the old certificate from the old kubeconf
grep "client-certificate-data" "${kubeconf_file}" | awk '{print$2}' | base64 -d >"${cert}"
# get subject from the old certificate
subj=$(cert::get_subj "${cert_name}")
cert::gen_cert "${cert_name}" "client" "${subj}" "${CERT_DAYS}" "${CERT_CA}"
# get certificate base64 code
cert_base64=$(base64 -w 0 "${cert}")
# set certificate base64 code to kubeconf
sed -i 's/client-certificate-data:.*/client-certificate-data: '"${cert_base64}"'/g' "${kubeconf_file}"
rm -f "${cert}"
rm -f "${key}"
}
cert::update_etcd_cert() {
local subj
local subject_alt_name
local cert
# generate etcd server,peer certificate
# /etc/kubernetes/pki/etcd/server
# /etc/kubernetes/pki/etcd/peer
for cert in ${ETCD_CERT_SERVER} ${ETCD_CERT_PEER}; do
subj=$(cert::get_subj "${cert}")
subject_alt_name=$(cert::get_subject_alt_name "${cert}")
cert::gen_cert "${cert}" "peer" "${subj}" "${CERT_DAYS}" "${ETCD_CERT_CA}" "${subject_alt_name}"
log::info "${GREEN}updated ${BLUE}${cert}.conf${NC}"
done
# generate etcd healthcheck-client,apiserver-etcd-client certificate
# /etc/kubernetes/pki/etcd/healthcheck-client
# /etc/kubernetes/pki/apiserver-etcd-client
for cert in ${ETCD_CERT_HEALTHCHECK_CLIENT} ${ETCD_CERT_APISERVER_ETCD_CLIENT}; do
subj=$(cert::get_subj "${cert}")
cert::gen_cert "${cert}" "client" "${subj}" "${CERT_DAYS}" "${ETCD_CERT_CA}"
log::info "${GREEN}updated ${BLUE}${cert}.conf${NC}"
done
# restart etcd
docker ps | awk '/k8s_etcd/{print$1}' | xargs -r -I '{}' docker restart {} >/dev/null 2>&1 || true
log::info "restarted etcd"
}
cert::update_master_cert() {
local subj
local subject_alt_name
local conf
# generate apiserver server certificate
# /etc/kubernetes/pki/apiserver
subj=$(cert::get_subj "${CERT_APISERVER}")
subject_alt_name=$(cert::get_subject_alt_name "${CERT_APISERVER}")
cert::gen_cert "${CERT_APISERVER}" "server" "${subj}" "${CERT_DAYS}" "${CERT_CA}" "${subject_alt_name}"
log::info "${GREEN}updated ${BLUE}${CERT_APISERVER}.crt${NC}"
# generate apiserver-kubelet-client certificate
# /etc/kubernetes/pki/apiserver-kubelet-client
subj=$(cert::get_subj "${CERT_APISERVER_KUBELET_CLIENT}")
cert::gen_cert "${CERT_APISERVER_KUBELET_CLIENT}" "client" "${subj}" "${CERT_DAYS}" "${CERT_CA}"
log::info "${GREEN}updated ${BLUE}${CERT_APISERVER_KUBELET_CLIENT}.crt${NC}"
# generate kubeconf for controller-manager,scheduler and kubelet
# /etc/kubernetes/controller-manager,scheduler,admin,kubelet.conf
for conf in ${CONF_CONTROLLER_MANAGER} ${CONF_SCHEDULER} ${CONF_ADMIN} ${CONF_KUBELET}; do
if [[ ${conf##*/} == "kubelet" ]]; then
# https://github.com/kubernetes/kubeadm/issues/1753
set +e
grep kubelet-client-current.pem /etc/kubernetes/kubelet.conf >/dev/null 2>&1
kubelet_cert_auto_update=$?
set -e
if [[ "$kubelet_cert_auto_update" == "0" ]]; then
log::info "does not need to update kubelet.conf"
continue
fi
fi
# update kubeconf
cert::update_kubeconf "${conf}"
log::info "${GREEN}updated ${BLUE}${conf}.conf${NC}"
# copy admin.conf to ${HOME}/.kube/config
if [[ ${conf##*/} == "admin" ]]; then
mkdir -p "${HOME}/.kube"
local config=${HOME}/.kube/config
local config_backup
config_backup=${HOME}/.kube/config.old-$(date +%Y%m%d)
if [[ -f ${config} ]] && [[ ! -f ${config_backup} ]]; then
cp -fp "${config}" "${config_backup}"
log::info "backup ${config} to ${config_backup}"
fi
cp -fp "${conf}.conf" "${HOME}/.kube/config"
log::info "copy the admin.conf to ${HOME}/.kube/config"
fi
done
# generate front-proxy-client certificate
# /etc/kubernetes/pki/front-proxy-client
subj=$(cert::get_subj "${FRONT_PROXY_CLIENT}")
cert::gen_cert "${FRONT_PROXY_CLIENT}" "client" "${subj}" "${CERT_DAYS}" "${FRONT_PROXY_CA}"
log::info "${GREEN}updated ${BLUE}${FRONT_PROXY_CLIENT}.crt${NC}"
# restart apiserver, controller-manager, scheduler and kubelet
for item in "apiserver" "controller-manager" "scheduler"; do
docker ps | awk '/k8s_kube-'${item}'/{print$1}' | xargs -r -I '{}' docker restart {} >/dev/null 2>&1 || true
log::info "restarted ${item}"
done
systemctl restart kubelet || true
log::info "restarted kubelet"
}
main() {
local node_type=$1
CERT_DAYS=3650
KUBE_PATH=/etc/kubernetes
PKI_PATH=${KUBE_PATH}/pki
# master certificates path
# apiserver
CERT_CA=${PKI_PATH}/ca
CERT_APISERVER=${PKI_PATH}/apiserver
CERT_APISERVER_KUBELET_CLIENT=${PKI_PATH}/apiserver-kubelet-client
CONF_CONTROLLER_MANAGER=${KUBE_PATH}/controller-manager
CONF_SCHEDULER=${KUBE_PATH}/scheduler
CONF_ADMIN=${KUBE_PATH}/admin
CONF_KUBELET=${KUBE_PATH}/kubelet
# front-proxy
FRONT_PROXY_CA=${PKI_PATH}/front-proxy-ca
FRONT_PROXY_CLIENT=${PKI_PATH}/front-proxy-client
# etcd certificates path
ETCD_CERT_CA=${PKI_PATH}/etcd/ca
ETCD_CERT_SERVER=${PKI_PATH}/etcd/server
ETCD_CERT_PEER=${PKI_PATH}/etcd/peer
ETCD_CERT_HEALTHCHECK_CLIENT=${PKI_PATH}/etcd/healthcheck-client
ETCD_CERT_APISERVER_ETCD_CLIENT=${PKI_PATH}/apiserver-etcd-client
case ${node_type} in
# etcd)
# # update etcd certificates
# cert::update_etcd_cert
# ;;
master)
# check certificates expiration
cert::check_master_certs_expiration
# backup $KUBE_PATH to $KUBE_PATH.old-$(date +%Y%m%d)
cert::backup_file "${KUBE_PATH}"
# update master certificates and kubeconf
log::info "${GREEN}updating...${NC}"
cert::update_master_cert
log::info "${GREEN}done!!!${NC}"
# check certificates expiration after certificates updated
cert::check_master_certs_expiration
;;
all)
# check certificates expiration
cert::check_all_expiration
# backup $KUBE_PATH to $KUBE_PATH.old-$(date +%Y%m%d)
cert::backup_file "${KUBE_PATH}"
# update etcd certificates
log::info "${GREEN}updating...${NC}"
cert::update_etcd_cert
# update master certificates and kubeconf
cert::update_master_cert
log::info "${GREEN}done!!!${NC}"
# check certificates expiration after certificates updated
cert::check_all_expiration
;;
check)
# check certificates expiration
cert::check_all_expiration
;;
*)
log::err "unknown, unsupported cert type: ${node_type}, supported type: \"all\", \"master\""
printf "Documentation: https://github.com/yuyicai/update-kube-cert
example:
'\033[32m./update-kubeadm-cert.sh all\033[0m' update all etcd certificates, master certificates and kubeconf
/etc/kubernetes
├── admin.conf
├── controller-manager.conf
├── scheduler.conf
├── kubelet.conf
└── pki
├── apiserver.crt
├── apiserver-etcd-client.crt
├── apiserver-kubelet-client.crt
├── front-proxy-client.crt
└── etcd
├── healthcheck-client.crt
├── peer.crt
└── server.crt
'\033[32m./update-kubeadm-cert.sh master\033[0m' update only master certificates and kubeconf
/etc/kubernetes
├── admin.conf
├── controller-manager.conf
├── scheduler.conf
├── kubelet.conf
└── pki
├── apiserver.crt
├── apiserver-kubelet-client.crt
└── front-proxy-client.crt
"
exit 1
;;
esac
}
main "$@"
4) 인증서 생성 쉘 실행
권한 문제로 인하여, root로 쉘을 실행해줘야 합니다.
$ sudo su -
$ cd /home/clex/update-kube-cert
$ ./update-kubeadm-cert.sh master
CERTIFICATE EXPIRES
/etc/kubernetes/controller-manager.config Jun 22 15:49:01 2023 GMT
/etc/kubernetes/scheduler.config Jun 22 15:49:01 2023 GMT
/etc/kubernetes/admin.config Jun 22 15:49:01 2023 GMT
/etc/kubernetes/pki/ca.crt Jun 19 15:48:15 2032 GMT
/etc/kubernetes/pki/apiserver.crt Jun 22 15:48:15 2023 GMT
/etc/kubernetes/pki/apiserver-kubelet-client.crt Jun 22 15:48:15 2023 GMT
/etc/kubernetes/pki/front-proxy-ca.crt Jun 19 15:48:15 2032 GMT
/etc/kubernetes/pki/front-proxy-client.crt Jun 22 15:48:15 2023 GMT
[2022-09-16T10:33:38.04+0900][INFO] backup /etc/kubernetes to /etc/kubernetes.old-20220916
[2022-09-16T10:33:38.04+0900][INFO] updating...
[2022-09-16T10:33:38.12+0900][INFO] updated /etc/kubernetes/pki/apiserver.crt
[2022-09-16T10:33:38.17+0900][INFO] updated /etc/kubernetes/pki/apiserver-kubelet-client.crt
[2022-09-16T10:33:38.25+0900][INFO] updated /etc/kubernetes/controller-manager.conf
[2022-09-16T10:33:38.31+0900][INFO] updated /etc/kubernetes/scheduler.conf
[2022-09-16T10:33:38.37+0900][INFO] updated /etc/kubernetes/admin.conf
[2022-09-16T10:33:38.38+0900][INFO] backup /root/.kube/config to /root/.kube/config.old-20220916
[2022-09-16T10:33:38.38+0900][INFO] copy the admin.conf to /root/.kube/config
[2022-09-16T10:33:38.44+0900][INFO] updated /etc/kubernetes/kubelet.conf
[2022-09-16T10:33:38.49+0900][INFO] updated /etc/kubernetes/pki/front-proxy-client.crt
[2022-09-16T10:33:39.05+0900][INFO] restarted apiserver
[2022-09-16T10:33:39.59+0900][INFO] restarted controller-manager
[2022-09-16T10:33:40.30+0900][INFO] restarted scheduler
[2022-09-16T10:33:40.52+0900][INFO] restarted kubelet
[2022-09-16T10:33:40.52+0900][INFO] done!!!
CERTIFICATE EXPIRES
/etc/kubernetes/controller-manager.config Sep 13 01:33:38 2032 GMT
/etc/kubernetes/scheduler.config Sep 13 01:33:38 2032 GMT
/etc/kubernetes/admin.config Sep 13 01:33:38 2032 GMT
/etc/kubernetes/pki/ca.crt Jun 19 15:48:15 2032 GMT
/etc/kubernetes/pki/apiserver.crt Sep 13 01:33:38 2032 GMT
/etc/kubernetes/pki/apiserver-kubelet-client.crt Sep 13 01:33:38 2032 GMT
/etc/kubernetes/pki/front-proxy-ca.crt Jun 19 15:48:15 2032 GMT
/etc/kubernetes/pki/front-proxy-client.crt Sep 13 01:33:38 2032 GMT
5) admin.conf 복사 반영
따라서 위의 쉘 output 내용을 보면 root/.kube/config 파일만 복사해줍니다.
somaz로 돌아가 한번 더 복사를 해줘야 합니다.
$ su - somaz
$ rm -rf ~/.kube/config
$ sudo cp /etc/kubernetes/admin.conf ~/.kube/config
$ sudo chown somaz:somaz ~/.kube/config
$ export KUBECONFIG=.kube/config
6) 인증서 연장확인
$ ls -l /etc |grep kubernetes
drwxr-xr-x. 5 kube root 4096 Sep 16 10:33 kubernetes
drwxr-xr-x 5 kube root 4096 Jun 23 00:52 kubernetes.old-20220916
$ ls -al /home/clex/.kube |grep config
-rw-------. 1 clex clex 5461 Sep 16 11:31 config
-rw------- 1 root root 5461 Jun 23 00:52 config-old.20220916
$ sudo openssl x509 -in /etc/kubernetes/ssl/apiserver.crt -noout -text |grep Not
Not Before: Sep 16 01:33:38 2022 GMT
Not After : Sep 13 01:33:38 2032 GMT
3. 워커 노드 인증서 갱신
- 각 워커 노드 순차 작업 실행
1.) 마스터 노드에서 token 생성, hash 획득, 마스터 ip 획득
$ /usr/local/bin/kubeadm token create
ogpn0p.3q8s9dbvy6aptq3i
$ kubeadm token list
TOKEN TTL EXPIRES USAGES DESCRIPTION EXTRA GROUPS
ogpn0p.3q8s9dbvy6aptq3i 23h 2022-09-17T10:35:14+09:00 authentication,signing <none> system:bootstrappers:kubeadm:default-node-token
$ openssl x509 -pubkey -in /etc/kubernetes/pki/ca.crt | openssl rsa -pubin -outform der 2>/dev/null | openssl dgst -sha256 -hex | sed 's/^.* //'
20196eb13acb0f74bc3c41cd8b5a271f50d0822ef8372489cd106c5c9cb591ec
$ cat /etc/hosts|grep control01
127.0.1.1 control01 control01
192.168.20.11 control01
192.168.20.11 control01
192.168.20.11 control01.cluster.local control01
- token=ogpn0p.3q8s9dbvy6aptq3i
- hash=20196eb13acb0f74bc3c41cd8b5a271f50d0822ef8372489cd106c5c9cb591ec
- server=192.168.20.11:6443 아래 변수 지정시 server=192.168.162.11:6443 / controller kube-apiserver 지정)
2.) 워커 노드 접속 후 변수 지정
$ ssh compute01
$ token=ogpn0p.3q8s9dbvy6aptq3i
$ hash=20196eb13acb0f74bc3c41cd8b5a271f50d0822ef8372489cd106c5c9cb591ec
$ server=192.168.20.11:6443
3.) 새로운 토큰으로 부트스트랩
$ sudo ln -s /usr/local/bin/kubelet /usr/bin/ &&
> sudo systemctl stop kubelet &&
> sudo mv /etc/kubernetes/manifests/ ~/ &&
> sudo rm /etc/kubernetes/kubelet.conf &&
> sudo rm /etc/kubernetes/pki/ca.crt &&
> sudo /usr/local/bin/kubeadm join $server --discovery-token $token --discovery-token-ca-cert-hash sha256:$hash
[preflight] Running pre-flight checks
[preflight] Reading configuration from the cluster...
[preflight] FYI: You can look at this config file with 'kubectl -n kube-system get cm kubeadm-config -oyaml'
W0916 10:40:27.791411 1711 defaults.go:199] The recommended value for "clusterDNS" in "KubeletConfiguration" is: [10.233.0.10]; the provided value is: [169.254.25.10]
[kubelet-start] Downloading configuration for the kubelet from the "kubelet-config-1.16" ConfigMap in the kube-system namespace
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Activating the kubelet service
[kubelet-start] Waiting for the kubelet to perform the TLS Bootstrap...
This node has joined the cluster:
* Certificate signing request was sent to apiserver and a response was received.
* The Kubelet was informed of the new secure connection details.
Run 'kubectl get nodes' on the control-plane to see this node join the cluster.
$ sudo mv ~/manifests/ /etc/kubernetes/
- 모든 마스터, 워커 노드 인증서 갱신후 node 상태를 최소 10분여 동안 모니터링 필요.
- STATUS NotReady로 변경 되지 않는지 확인 필요
4.) Kubernetes Cluster 상태확인
$ kubectl cluster-info
Kubernetes master is running at https://192.168.20.11:6443
coredns is running at https://192.168.20.11:6443/api/v1/namespaces/kube-system/services/coredns:dns/proxy
To further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.
$ kubectl get nodes
NAME STATUS ROLES AGE VERSION
compute01 Ready <none> 85d v1.16.8
compute02 Ready <none> 85d v1.16.8
control01 Ready master 85d v1.16.8
5.) Jenkins config 파일 교체
jenkins랑 연동이 되어있다면 config 파일을 교체해줘야 한다.
$ ZONE=edu
$ cp .kube/config config-$ZONE
$ sudo scp -r config-$ZONE [user id]@[ip address]:~
$ sudo docker exec -ti 648d417c540e /bin/bash
bash-4.4# cd root/.kube/
bash-4.4# ZONE=edu
bash-4.4# cp config-$ZONE config-$ZONE.$(date +%Y%m%d)
$ ls
config-edu
$ sudo docker cp config-edu 648d417c540e:/root/.kube/
728x90
반응형
'Container Orchestration > Kubernetes' 카테고리의 다른 글
Kubernetes 플러그인 매니저 Krew란? (0) | 2023.04.30 |
---|---|
Kubernetes API Server, Group / RBAC란? (2) | 2023.04.19 |
kubectl 명령어 정리 (0) | 2022.08.10 |
Ingress(인그레스) (0) | 2022.08.08 |
Kubernetes Probe (Liveness, Readiness, Startup) (0) | 2022.07.15 |