文章是按https://blog.csdn.net/Excairun/article/details/88962769,来进行操作并记录相关结果

版本:k8s V14.0,docker-ce 18.09.03

1.环境准备

内核及版本

[root@k8s-node1 ~]# cat /etc/redhat-release
CentOS Linux release 7.6. (Core)
[root@k8s-node1 ~]# uname -r
3.10.-.el7.x86_64
[root@k8s-node1 ~]# uname -a
Linux k8s-node1 3.10.-.el7.x86_64 # SMP Thu Nov :: UTC x86_64 x86_64 x86_64 GNU/Linux
[root@k8s-node1 ~]#

修改主机名配置hosts

[root@localhost ~]# hostname
localhost.localdomain
[root@localhost ~]# hostname k8s-master ##临时生效
localhost.localdomain
[root@localhost ~]# hostnamectl set-hostname k8s-master ##重启后永久生效
[root@localhost ~]# reboot
[root@localhost ~]# cat /etc/hosts
127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4
:: localhost localhost.localdomain localhost6 localhost6.localdomain6
128.192.111.130 k8s-master
128.192.111.131 k8s-node1
[root@localhost ~]#

同步时间

yum install -y ntpdate
ntpdate -u ntp.api.bz

关闭防火墙,selinux,swap,桥接网络配置

# 所有主机:基本系统配置

# 关闭Selinux/firewalld
systemctl stop firewalld
systemctl disable firewalld
setenforce
sed -i "s/SELINUX=enforcing/SELINUX=disabled/g" /etc/selinux/config # 关闭交换分区
swapoff -a
yes | cp /etc/fstab /etc/fstab_bak
cat /etc/fstab_bak |grep -v swap > /etc/fstab # 设置网桥包经IPTables,core文件生成路径 modprobe br_netfilter
cat >/etc/sysctl.conf <<EOF
vm.swappiness =
net.bridge.bridge-nf-call-ip6tables =
net.bridge.bridge-nf-call-iptables =
EOF sysctl -p

2.docker安装

#安装yum操作的基本服务,如果已安装可跳过此步骤
yum install -y net-tools epel-release
yum install -y vim yum-utils device-mapper-persistent-data lvm2
#添加阿里云的docker-ce源
yum-config-manager --add-repo http://mirrors.aliyun.com/docker-ce/linux/centos/docker-ce.repo
#查看docker可安装版本
yum list docker-ce.x86_64 --showduplicates |sort -r
#安装docker如果不指定版本则会安装最新可安装版本
yum install docker-ce-18.09.3-3.el7
#设置开机启动
systemctl enable docker
#启动服务
systemctl start docker

3.kubernetes相关配置

配置k8s走阿里镜像

[root@localhost ~]# cat /etc/yum.repos.d/kubernetes.repo
[kubernetes]
name=Kubernetes
baseurl=https://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64/
enabled=
gpgcheck=
repo_gpgcheck=
gpgkey=https://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpg https://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpg

安装kubectl,kubelet,kubernetes-cni,kubeadm

yum install -y kubectl-1.14. kubelet-1.14. kubernetes-cni-1.14. kubeadm-1.14.0
systemctl enable kubelet #开机启动

设置docker Cgroup Driver为systemd,达到与kubelet相一致的要求

cat > /etc/docker/daemon.json <<EOF
{
"exec-opts": ["native.cgroupdriver=systemd"],
"log-driver": "json-file",
"log-opts": {
"max-size": "100m"
},
"storage-driver": "overlay2",
"storage-opts": [
"overlay2.override_kernel_check=true"
]
}
EOF

重启docker

[root@localhost ~]# mkdir -p /etc/systemd/system/docker.service.d
[root@localhost ~]#
[root@localhost ~]# systemctl daemon-reload
[root@localhost ~]# systemctl restart docker

查看kubeadm v1.14.0要求的版本

[root@localhost ~]# kubeadm config images list
I0402 ::29.358043 version.go:] could not fetch a Kubernetes version from the internet: unable to get URL "https://dl.k8s.io/release/stable-1.txt": Get https://dl.k8s.io/release/stable-1.txt: net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)
I0402 ::29.358209 version.go:] falling back to the local client version: v1.14.0
k8s.gcr.io/kube-apiserver:v1.14.0
k8s.gcr.io/kube-controller-manager:v1.14.0
k8s.gcr.io/kube-scheduler:v1.14.0
k8s.gcr.io/kube-proxy:v1.14.0
k8s.gcr.io/pause:3.1
k8s.gcr.io/etcd:3.3.
k8s.gcr.io/coredns:1.3.

拉取国内镜像

[root@localhost ~]# cat pullK8sImages.sh
#!/bin/bash
KUBE_VERSION=v1.14.0
KUBE_PAUSE_VERSION=3.1
ETCD_VERSION=3.3.
DNS_VERSION=1.3.
DASHBOARD_VERSION=v1.10.1
username=registry.cn-hangzhou.aliyuncs.com/google_containers
#多了一个dashboard组件是为后续安装做准备
images=(
kube-proxy-amd64:${KUBE_VERSION}
kube-scheduler-amd64:${KUBE_VERSION}
kube-controller-manager-amd64:${KUBE_VERSION}
kube-apiserver-amd64:${KUBE_VERSION}
pause:${KUBE_PAUSE_VERSION}
etcd-amd64:${ETCD_VERSION}
coredns:${DNS_VERSION}
kubernetes-dashboard-amd64:DASHBOARD_VERSION
)
for image in ${images[@]}
do
NEW_IMAGE=`echo ${image}|awk '{gsub(/-amd64/,"",$0);print}'`
echo ${NEW_IMAGE}
docker pull ${username}/${image}
docker tag ${username}/${image} k8s.gcr.io/${NEW_IMAGE}
docker rmi ${username}/${image}
done
[root@localhost ~]# sh pullK8sImages.sh

拉取结果

[root@localhost ~]# docker images
REPOSITORY TAG IMAGE ID CREATED SIZE
k8s.gcr.io/kube-proxy v1.14.0 5cd54e388aba days ago .1MB
k8s.gcr.io/kube-scheduler v1.14.0 00638a24688b days ago .6MB
k8s.gcr.io/kube-controller-manager v1.14.0 b95b1efa0436 days ago 158MB
k8s.gcr.io/kube-apiserver v1.14.0 ecf910f40d6e days ago 210MB
k8s.gcr.io/coredns 1.3. eb516548c180 months ago .3MB
k8s.gcr.io/etcd 3.3. 2c4adeb21b4f months ago 258MB
k8s.gcr.io/pause 3.1 da86e6ba6ca1 months ago 742kB
[root@localhost ~]#

4.kubeadm(仅master执行)安装

kubeadm init   --kubernetes-version=v1.14.0   --pod-network-cidr=10.244.0.0/  --apiserver-advertise-address=192.168.111.130  --token-ttl   --ignore-preflight-errors=Swap

安装结果

To start using your cluster, you need to run the following as a regular user:

  mkdir -p $HOME/.kube
sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
sudo chown $(id -u):$(id -g) $HOME/.kube/config You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
https://kubernetes.io/docs/concepts/cluster-administration/addons/ Then you can join any number of worker nodes by running the following on each as root: kubeadm join 192.168.111.130: --token ymtl8s.933t59qfezi9gjcq \
--discovery-token-ca-cert-hash sha256:7816d0b2572e6c569ed8e63ece15a7a08d06ed3fc89698245bf2aaa6acc345d7

出现最后一行,表示安装成功,若报错,可执行

kubeadm reset

rm -rf $HOME/.kube/config

调整后,再执行kubeadm init
为了在点上使用kubectl,需作以下设置

[root@localhost ~]# kubectl version
Client Version: version.Info{Major:"", Minor:"", GitVersion:"v1.14.0", GitCommit:"641856db18352033a0d96dbc99153fa3b27298e5", GitTreeState:"clean", BuildDate:"2019-03-25T15:53:57Z", GoVersion:"go1.12.1", Compiler:"gc", Platform:"linux/amd64"}
The connection to the server localhost: was refused - did you specify the right host or port?
[root@localhost ~]# hostname
k8s-master
[root@localhost ~]# mkdir -p $HOME/.kube
[root@localhost ~]# cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
[root@localhost ~]# chown $(id -u):$(id -g) $HOME/.kube/config
[root@localhost ~]# kubectl version
Client Version: version.Info{Major:"", Minor:"", GitVersion:"v1.14.0", GitCommit:"641856db18352033a0d96dbc99153fa3b27298e5", GitTreeState:"clean", BuildDate:"2019-03-25T15:53:57Z", GoVersion:"go1.12.1", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"", Minor:"", GitVersion:"v1.14.0", GitCommit:"641856db18352033a0d96dbc99153fa3b27298e5", GitTreeState:"clean", BuildDate:"2019-03-25T15:45:25Z", GoVersion:"go1.12.1", Compiler:"gc", Platform:"linux/amd64"}
[root@localhost ~]#

[root@localhost ~]# kubectl cluster-info Kubernetes master is running at https://192.168.111.130:6443 KubeDNS is running at https://192.168.111.130:6443/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy

To further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.

5.配置flannel网络(仅master节点执行)

不成功可多次尝试

[root@localhost ~]# yum -y install wget
[root@localhost ~]# wget https://raw.githubusercontent.com/coreos/flannel/master/Documentation/kube-flannel.yml
[root@localhost ~]# kubectl  apply -f kube-flannel.yml

6.添加子节点(节点机器上操作)

修改节点机器名称 linux centos7.5修改主机名和ip永久生效

##临时生效,重启失效
[root@localhost ~]# hostname
localhost.localdomain
[root@localhost ~]#hostname k8s-node1 ##永久有效,重启后生效
[root@localhost ~]#hostnamectl set-hostname k8s-node1
[root@localhost ~]# reboot
kubeadm join 192.168.111.130: --token ymtl8s.933t59qfezi9gjcq \
--discovery-token-ca-cert-hash sha256:7816d0b2572e6c569ed8e63ece15a7a08d06ed3fc89698245bf2aaa6acc345d7

执行结果

[root@localhost ~]# kubeadm join 192.168.111.130: --token ymtl8s.933t59qfezi9gjcq     --discovery-token-ca-cert-hash sha256:7816d0b2572e6c569ed8e63ece15a7a08d06ed3fc89698245bf2aaa6acc345d7
[preflight] Running pre-flight checks
[preflight] Reading configuration from the cluster...
[preflight] FYI: You can look at this config file with 'kubectl -n kube-system get cm kubeadm-config -oyaml'
[kubelet-start] Downloading configuration for the kubelet from the "kubelet-config-1.14" ConfigMap in the kube-system namespace
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Activating the kubelet service
[kubelet-start] Waiting for the kubelet to perform the TLS Bootstrap... This node has joined the cluster:
* Certificate signing request was sent to apiserver and a response was received.
* The Kubelet was informed of the new secure connection details. Run 'kubectl get nodes' on the control-plane to see this node join the cluster. [root@localhost ~]#

报错可执行对应操作,后再执行kubeadm

[root@localhost ~]# kubeadm join 192.168.111.130: --token ymtl8s.933t59qfezi9gjcq \
> --discovery-token-ca-cert-hash sha256:7816d0b2572e6c569ed8e63ece15a7a08d06ed3fc89698245bf2aaa6acc345d7
[preflight] Running pre-flight checks
error execution phase preflight: [preflight] Some fatal errors occurred:
[ERROR FileContent--proc-sys-net-ipv4-ip_forward]: /proc/sys/net/ipv4/ip_forward contents are not set to
[preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...`
[root@localhost ~]# echo > /proc/sys/net/ipv4/ip_forward
[root@localhost ~]# cat /proc/sys/net/ipv4/ip_forward

(master)查看要点信息

[root@localhost ~]# kubectl get nodes
NAME STATUS ROLES AGE VERSION
k8s-master Ready master 40m v1.14.0
k8s-node1 Ready <none> 4m31s v1.14.0

查看节点详情

[root@localhost ~]# kubectl describe nodes k8s-node1
Name: k8s-node1
Roles: <none>
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=k8s-node1
kubernetes.io/os=linux
Annotations: flannel.alpha.coreos.com/backend-data: {"VtepMAC":"ca:bc:84:14:09:94"}
flannel.alpha.coreos.com/backend-type: vxlan
flannel.alpha.coreos.com/kube-subnet-manager: true
flannel.alpha.coreos.com/public-ip: 192.168.111.131
kubeadm.alpha.kubernetes.io/cri-socket: /var/run/dockershim.sock
node.alpha.kubernetes.io/ttl:
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Tue, Apr :: +
Taints: <none>
Unschedulable: false
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Tue, Apr :: + Tue, Apr :: + KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Tue, Apr :: + Tue, Apr :: + KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Tue, Apr :: + Tue, Apr :: + KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Tue, Apr :: + Tue, Apr :: + KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.111.131
Hostname: k8s-node1
Capacity:
cpu:
ephemeral-storage: 17394Mi
hugepages-1Gi:
hugepages-2Mi:
memory: 995896Ki
pods:
Allocatable:
cpu:
ephemeral-storage:
hugepages-1Gi:
hugepages-2Mi:
memory: 893496Ki
pods:
System Info:
Machine ID: a5a43f5916c643bf83d6f99425a4b9d2
System UUID: FCCE4D56-202D-568C--7A69D9ADF401
Boot ID: 68e3c38d-d1d6--9af5-01a9699ce00c
Kernel Version: 3.10.-.el7.x86_64
OS Image: CentOS Linux (Core)
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://18.9.3
Kubelet Version: v1.14.0
Kube-Proxy Version: v1.14.0
PodCIDR: 10.244.1.0/
Non-terminated Pods: ( in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits AGE
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system kube-flannel-ds-amd64-bsrkx 100m (%) 100m (%) 50Mi (%) 50Mi (%) 7m19s
kube-system kube-proxy-2mj4q (%) (%) (%) (%) 7m19s
Allocated resources:
(Total limits may be over percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 100m (%) 100m (%)
memory 50Mi (%) 50Mi (%)
ephemeral-storage (%) (%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 7m20s kubelet, k8s-node1 Starting kubelet.
Normal NodeHasSufficientMemory 7m20s (x2 over 7m20s) kubelet, k8s-node1 Node k8s-node1 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 7m20s (x2 over 7m20s) kubelet, k8s-node1 Node k8s-node1 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 7m20s (x2 over 7m20s) kubelet, k8s-node1 Node k8s-node1 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 7m19s kubelet, k8s-node1 Updated Node Allocatable limit across pods
Normal Starting 7m16s kube-proxy, k8s-node1 Starting kube-proxy.
Normal NodeReady 6m29s kubelet, k8s-node1 Node k8s-node1 status is now: NodeReady
[root@localhost ~]#

kubectl查看系统pod信息

[root@k8s-master ~]# kubectl get namespaces ##get 可查看services,deployments,pods,replicasets
NAME STATUS AGE
default Active 3h17m
kube-node-lease Active 3h17m
kube-public Active 3h17m
kube-system Active 3h17m
root@k8s-master ~]# kubectl get po --all-namespaces -o wide ##-o查看pod运行在那个node上,--namespace=kube-system查看系统pod
NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
kube-system coredns-fb8b8dccf-ltlx4 / Running 3h12m 10.244.0.4 k8s-master <none> <none>
kube-system coredns-fb8b8dccf-q949f / Running 3h12m 10.244.0.5 k8s-master <none> <none>
kube-system etcd-k8s-master / Running 3h11m 192.168.111.130 k8s-master <none> <none>
kube-system kube-apiserver-k8s-master / Running 3h12m 192.168.111.130 k8s-master <none> <none>
kube-system kube-controller-manager-k8s-master / Running 3h11m 192.168.111.130 k8s-master <none> <none>
kube-system kube-flannel-ds-amd64-2gr2v / Running 177m 192.168.111.130 k8s-master <none> <none>
kube-system kube-flannel-ds-amd64-bsrkx / Running 157m 192.168.111.131 k8s-node1 <none> <none>
kube-system kube-flannel-ds-amd64-xdg5p / Running 15m 192.168.111.132 k8s-node2 <none> <none>
kube-system kube-proxy-2mj4q / Running 157m 192.168.111.131 k8s-node1 <none> <none>
kube-system kube-proxy-ffd8s / Running 15m 192.168.111.132 k8s-node2 <none> <none>
kube-system kube-proxy-qp5k7 / Running 3h12m 192.168.111.130 k8s-master <none> <none>
kube-system kube-scheduler-k8s-master / Running 3h12m 192.168.111.130 k8s-master <none> <none>
[root@k8s-master ~]#

docker ps -a查看container

[root@k8s-master docker]# docker ps -a
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
c8a00a7aab85 tomcat: "catalina.sh run" seconds ago Up seconds 0.0.0.0:->/tcp mytomcat_1
75bde1b990e7 b95b1efa0436 "kube-controller-man…" minutes ago Up minutes k8s_kube-controller-manager_kube-controller-manager-k8s-master_kube-system_0ff88c9b6e64cded3762e51ff18bce90_4
0073cac306f0 ff281650a721 "/opt/bin/flanneld -…" minutes ago Up minutes k8s_kube-flannel_kube-flannel-ds-amd64-2gr2v_kube-system_4cf85962--11e9-a96d-000c291ae345_1
f769c27066dc ff281650a721 "cp -f /etc/kube-fla…" minutes ago Exited () minutes ago k8s_install-cni_kube-flannel-ds-amd64-2gr2v_kube-system_4cf85962--11e9-a96d-000c291ae345_1
b54ff4716e39 k8s.gcr.io/pause:3.1 "/pause" minutes ago Up minutes k8s_POD_kube-flannel-ds-amd64-2gr2v_kube-system_4cf85962--11e9-a96d-000c291ae345_1
408f7e4b42e8 k8s.gcr.io/pause:3.1 "/pause" minutes ago Up minutes k8s_POD_kube-controller-manager-k8s-master_kube-system_0ff88c9b6e64cded3762e51ff18bce90_4
fecb42b177df eb516548c180 "/coredns -conf /etc…" minutes ago Up minutes k8s_coredns_coredns-fb8b8dccf-ltlx4_kube-system_2f8b4ffb-550e-11e9-a96d-000c291ae345_1
e5dbdf2d4af3 eb516548c180 "/coredns -conf /etc…" minutes ago Up minutes k8s_coredns_coredns-fb8b8dccf-q949f_kube-system_2f8c9997-550e-11e9-a96d-000c291ae345_1
de96a90cd8c8 ecf910f40d6e "kube-apiserver --ad…" minutes ago Up minutes k8s_kube-apiserver_kube-apiserver-k8s-master_kube-system_078b26a5af4c34641521cf85bb8b5ee7_3
2a5b2f1eff20 5cd54e388aba "/usr/local/bin/kube…" minutes ago Up minutes k8s_kube-proxy_kube-proxy-qp5k7_kube-system_2f1a1559-550e-11e9-a96d-000c291ae345_1
c9de0db0ce55 00638a24688b "kube-scheduler --bi…" minutes ago Up minutes k8s_kube-scheduler_kube-scheduler-k8s-master_kube-system_58272442e226c838b193bbba4c44091e_3
97364e091073 2c4adeb21b4f "etcd --advertise-cl…" minutes ago Up minutes k8s_etcd_etcd-k8s-master_kube-system_804ba6a1bef952d18f2040a1ff90dbc3_3
bd0c87c3d533 k8s.gcr.io/pause:3.1 "/pause" minutes ago Up minutes k8s_POD_coredns-fb8b8dccf-q949f_kube-system_2f8c9997-550e-11e9-a96d-000c291ae345_4
08f894c8c252 k8s.gcr.io/pause:3.1 "/pause" minutes ago Up minutes k8s_POD_coredns-fb8b8dccf-ltlx4_kube-system_2f8b4ffb-550e-11e9-a96d-000c291ae345_4
e0e7875598de k8s.gcr.io/pause:3.1 "/pause" minutes ago Up minutes k8s_POD_kube-scheduler-k8s-master_kube-system_58272442e226c838b193bbba4c44091e_3
461eaa7b2899 k8s.gcr.io/pause:3.1 "/pause" minutes ago Up minutes k8s_POD_kube-proxy-qp5k7_kube-system_2f1a1559-550e-11e9-a96d-000c291ae345_1
5233aca7eca5 k8s.gcr.io/pause:3.1 "/pause" minutes ago Up minutes k8s_POD_etcd-k8s-master_kube-system_804ba6a1bef952d18f2040a1ff90dbc3_3
841be3a3cc7d k8s.gcr.io/pause:3.1 "/pause" minutes ago Up minutes k8s_POD_kube-apiserver-k8s-master_kube-system_078b26a5af4c34641521cf85bb8b5ee7_4
4698599e362c b95b1efa0436 "kube-controller-man…" minutes ago Exited () minutes ago k8s_kube-controller-manager_kube-controller-manager-k8s-master_kube-system_0ff88c9b6e64cded3762e51ff18bce90_3
f0f87dd886f9 k8s.gcr.io/pause:3.1 "/pause" minutes ago Exited () minutes ago k8s_POD_kube-controller-manager-k8s-master_kube-system_0ff88c9b6e64cded3762e51ff18bce90_3
a0c24081b996 eb516548c180 "/coredns -conf /etc…" About an hour ago Exited () minutes ago k8s_coredns_coredns-fb8b8dccf-q949f_kube-system_2f8c9997-550e-11e9-a96d-000c291ae345_0
b8b4efead3b4 k8s.gcr.io/pause:3.1 "/pause" About an hour ago Exited () minutes ago k8s_POD_coredns-fb8b8dccf-q949f_kube-system_2f8c9997-550e-11e9-a96d-000c291ae345_3
96a92d82c0dc eb516548c180 "/coredns -conf /etc…" About an hour ago Exited () minutes ago k8s_coredns_coredns-fb8b8dccf-ltlx4_kube-system_2f8b4ffb-550e-11e9-a96d-000c291ae345_0
c2f167c6e9c8 k8s.gcr.io/pause:3.1 "/pause" About an hour ago Exited () minutes ago k8s_POD_coredns-fb8b8dccf-ltlx4_kube-system_2f8b4ffb-550e-11e9-a96d-000c291ae345_3
cb1e6e3458ee ff281650a721 "/opt/bin/flanneld -…" About an hour ago Exited () minutes ago k8s_kube-flannel_kube-flannel-ds-amd64-2gr2v_kube-system_4cf85962--11e9-a96d-000c291ae345_0
99acbfbb1e68 5cd54e388aba "/usr/local/bin/kube…" About an hour ago Exited () minutes ago k8s_kube-proxy_kube-proxy-qp5k7_kube-system_2f1a1559-550e-11e9-a96d-000c291ae345_0
09195555e12f k8s.gcr.io/pause:3.1 "/pause" About an hour ago Exited () minutes ago k8s_POD_kube-proxy-qp5k7_kube-system_2f1a1559-550e-11e9-a96d-000c291ae345_0
ed47844f108f k8s.gcr.io/pause:3.1 "/pause" About an hour ago Exited () minutes ago k8s_POD_kube-flannel-ds-amd64-2gr2v_kube-system_4cf85962--11e9-a96d-000c291ae345_0
f09401726136 00638a24688b "kube-scheduler --bi…" About an hour ago Exited () minutes ago k8s_kube-scheduler_kube-scheduler-k8s-master_kube-system_58272442e226c838b193bbba4c44091e_2
a03a28a6ef98 ecf910f40d6e "kube-apiserver --ad…" About an hour ago Exited () minutes ago k8s_kube-apiserver_kube-apiserver-k8s-master_kube-system_078b26a5af4c34641521cf85bb8b5ee7_2
4769961090db 2c4adeb21b4f "etcd --advertise-cl…" About an hour ago Exited () minutes ago k8s_etcd_etcd-k8s-master_kube-system_804ba6a1bef952d18f2040a1ff90dbc3_2
191090440ed3 k8s.gcr.io/pause:3.1 "/pause" About an hour ago Exited () minutes ago k8s_POD_etcd-k8s-master_kube-system_804ba6a1bef952d18f2040a1ff90dbc3_2
89001ab3e457 k8s.gcr.io/pause:3.1 "/pause" About an hour ago Exited () minutes ago k8s_POD_kube-apiserver-k8s-master_kube-system_078b26a5af4c34641521cf85bb8b5ee7_2
4e93b1057da0 k8s.gcr.io/pause:3.1 "/pause" About an hour ago Exited () minutes ago k8s_POD_kube-scheduler-k8s-master_kube-system_58272442e226c838b193bbba4c44091e_2

至此,kubernetes安装完成

centos7.5下kubeadm安装kubernetes集群安装的更多相关文章

  1. VirtualBox上使用kubeadm安装Kubernetes集群

    之前一直使用minikube练习,为了更贴近生产环境,使用VirtualBox搭建Kubernetes集群. 为了不是文章凌乱,把在搭建过程中遇到的问题及解决方法记在了另一篇文章:安装Kubernet ...

  2. Kubernetes(K8s) 安装(使用kubeadm安装Kubernetes集群)

    背景: 由于工作发生了一些变动,很长时间没有写博客了. 概述: 这篇文章是为了介绍使用kubeadm安装Kubernetes集群(可以用于生产级别).使用了Centos 7系统. 一.Centos7 ...

  3. kubeadm搭建kubernetes集群之二:创建master节点

    在上一章kubeadm搭建kubernetes集群之一:构建标准化镜像中我们用VMware安装了一个CentOS7虚拟机,并且打算用这个虚拟机的镜像文件作为后续整个kubernetes的标准化镜像,现 ...

  4. 从0到1使用Kubernetes系列(三):使用Ansible安装Kubernetes集群

    前两期的文章介绍了Kubernetes基本概念和架构,用Kubeadm+Ansible搭建Kubernetes集群所需要的工具及其作用.本篇介绍怎么使用Ansible安装Kubernetes集群. 启 ...

  5. kubeadm搭建kubernetes集群之三:加入node节点

    在上一章<kubeadm搭建kubernetes集群之二:创建master节点>的实战中,我们把kubernetes的master节点搭建好了,本章我们将加入node节点,使得整个环境可以 ...

  6. CentOS 7.5 使用 yum 安装 Kubernetes 集群(二)

    一.安装方式介绍 1.yum 安装 目前CentOS官方已经把Kubernetes源放入到自己的默认 extras 仓库里面,使用 yum 安装,好处是简单,坏处也很明显,需要官方更新 yum 源才能 ...

  7. 解决kubeadm部署kubernetes集群镜像问题

    kubeadm 是kubernetes 的集群安装工具,能够快速安装kubernetes 集群.kubeadm init 命令默认使用的docker镜像仓库为k8s.gcr.io,国内无法直接访问,需 ...

  8. 安装Kubernetes集群时遇到的问题及解决方法

    在搭建Kubernetes集群时遇到一些问题,记录在这里. 搭建过程在另一篇文章:VirtualBox上使用kubeadm安装Kubernetes集群 1. 虚拟机安装完CentOS7登录时遇到war ...

  9. kubeadm 搭建kubernetes集群环境

    需求 kubeadm 搭建kubernetes集群环境 准备条件 三台VPS(本文使用阿里云香港 - centos7.7) 一台能SSH连接到VPS的本地电脑 (推荐连接工具xshell) 安装步骤 ...

随机推荐

  1. DS二叉树--层次遍历

    题目描述 层次遍历二叉树,是从根结点开始遍历,按层次次序“自上而下,从左至右”访问树中的各结点. 建树方法采用“先序遍历+空树用0表示”的方法 要求:采用队列对象实现,函数框架如下: 输入 第一行输入 ...

  2. 【占位符替换】替换String中的占位符标志位{placeholder}

    概述 占位符替换, 占位符表示为:{placeholder}; 示例:替换如下{xxx}占位符中的内容 "名字:{name},年龄:{age},学校:{school}" 提供了两种 ...

  3. IE10 下系统出现Unable to get property 'PageRequestManager' of undefined or null reference错误

    在本地调试时没有任何问题,上传到测试服务器(win2003 framework 4.0)后打开网站出现Unable to get property 'PageRequestManager' of un ...

  4. 廖雪峰Java3异常处理-1错误处理-2捕获异常

    1捕获异常 1.1 finally语句保证有无错误都会执行 try{...}catch (){...}finally{...} 使用try...catch捕获异常 可能发生异常的语句放在try{... ...

  5. folly无锁队列,尝试添加新的函数

    1. folly是facebook开源的关于无锁队列的库,实现过程很精妙.folly向队列中添加节点过程,符合标准库中的队列的设计,而取出节点的过程,则会造成多个线程的分配不均.我曾经试着提供一次 取 ...

  6. Fibonacci数列的两种实现方式

    斐波那契数列的形式为:1,1,2,3,5,8,13,21......从第三项开始,后面的每一项都是前面两项的和. 实现的方式有一下 两种: 一:递归方式实现 def fib(n): if n < ...

  7. [UE4]圆形小地图

    一.创建一个名为M_RoundRetainer的材质 二.创建一个名为RoundMiniMap的UserWidget 三.TestMiniMap中将添加进来 四.运行游戏

  8. [UE4]世界坐标和相对坐标

    一.世界坐标:相对于整个世界的坐标 二.相对坐标是相对于组件父级的坐标.如下图: 1.Mesh组件和CameraPositionArrow组件的相对坐标是相对于Root组件的坐标 2.Cube组件的相 ...

  9. [UE4]Overlay容器:图片随着其他容器(比如Vertical Box)大小而同步改变

  10. [UE4]利用取模运算达到循环遍历数组的目的

    X mod Y: 1.X<Y: X mod Y = X.计算记过永远都是等于X 2.X=Y:X mod Y = 0.重新回到数组第一个索引位置