## 第三章——k8s集群 > 我们来回顾一下并学习一些必要知识 [k8s中文社区docs.kubernetes.org.cn](http://docs.kubernetes.org.cn/) ##### K8S核心资源管理方法 ~~~ # 任意机器(我是在21) # 查看名称空间 ~]# kubectl get namespace ~]# kubectl get ns ~~~ ![1579073760060](assets/1579073760060.png) ~~~ # 任意机器(我是在21) ~]# kubectl get all [-n default] ~]# kubectl create ns app # 增 ~]# kubectl create ns app # 删 ~]# kubectl delete namespace app # 查 ~]# kubectl get ns # 创建deployment资源 kubectl create deployment nginx-dp --image=harbor.od.com/public/nginx:v1.7.9 -n kube-public # 查指定空间 ~]# kubectl get deploy -n kube-public ~]# kubectl get pods -o wide -n kube-public ~]# kubectl describe deployment nginx-dp -n kube-public # 进入pod资源 21 ~]# kubectl get pods -n kube-public 21 ~]# kubectl exec -ti nginx-dp-5dfc689474-9zt9r /bin/bash -n kube-public ~~~ > **kubectl get deploy**:这里的deploy是容器类型,deploy也是deployment > > **kubectl exec**:进入容器 > > - -t:将标准输入控制台作为容器的控制台输入 > - -i:将控制台输入发送到容器 > - 一般是连起来用-it,后面带的是get出来的容器名 > - /bin/bash:终端模式 ![1579075596641](assets/1579075596641.png) ~~~ # 删除pod资源(重启),pod控制器预期你有一个pod,所以你删掉就会重启,后面force是强制删除 21 ~]# kubectl delete pod nginx-dp-5dfc689474-gtfvv -n kube-public [--force --grace-period=0] # 删掉deploy 21 ~]# kubectl delete deploy nginx-dp -n kube-public # 查看 21 ~]# kubectl get all -n kube-public ~~~ ![1579076172922](assets/1579076172922.png) ##### 管理service资源 ~~~ # 21机器 # 创建 ~]# kubectl create deployment nginx-dp --image=harbor.od.com/public/nginx:v1.7.9 -n kube-public ~]# kubectl get all -n kube-public # 暴露端口 ~]# kubectl expose deployment nginx-dp --port=80 -n kube-public ~]# kubectl get all -n kube-public -o wide ~~~ > **kubectl expose**:暴露端口,后面的--port=80 指的是暴露80端口 ![1579077073962](assets/1579077073962.png) ~~~ # 去22机器 ~]# curl 192.168.81.37 ~]# ipvsadm -Ln ~~~ ![1579077146418](assets/1579077146418.png) ![1579077275447](assets/1579077275447.png) ~~~ # 做成两份代理服务器,22机器 ~]# kubectl scale deployment nginx-dp --replicas=2 -n kube-public ~]# ipvsadm -Ln ~~~ > **kubectl scale:** 扩容或缩容 Deployment等中Pod数量 > > - --replicas=2:把Pod数量改为2,即如果之前是1则扩容变成2,如果之前是3则缩容变成2 可以看到下面的Pod已经变成了两个,而上图是只有一个的 ![1579077403598](assets/1579077403598.png) ~~~ # 获取资源配置清单,21机器 ~]# kubectl get pods -n kube-public ~]# kubectl get pods nginx-dp-5dfc689474-788xp -o yaml -n kube-public # 解释怎么用 ~]# kubectl explain service.metadata ~~~ > 资源清单的内容解释由于太多,这里就不做解析了,感兴趣的朋友可以网上搜下 ![1579079365291](assets/1579079365291.png) ~~~ # 声明式、21机器: ~]# vi nginx-ds-svc.yaml apiVersion: v1 kind: Service metadata: labels: app: nginx-ds name: nginx-ds namespace: default spec: ports: - port: 80 protocol: TCP targetPort: 80 selector: app: nginx-ds sessionAffinity: None type: ClusterIP ~]# kubectl create -f nginx-ds-svc.yaml # out: service/nginx-ds created ~]# kubectl get svc -n default ~]# kubectl get svc nginx-ds -o yaml ~~~ ![1579079818058](assets/1579079818058.png) ~~~ # 修改资源,在线方式: ~]# kubectl edit svc nginx-ds ~]# kubectl get svc # 离线:删了再打开,离线修改有记录 # 删除资源,实验,按照以下方法是无法删除的~去找一下吧 # 陈述式 ~]# kubectl delete -f nginx-ds # 声明式 ~]# kubectl delete -f nginx-dp-svc.yaml ~~~ > 当然删不了也无所谓 ![1579085057843](assets/1579085057843.png) 回顾完成 ### 安装部署flanneld > **WHAT**:通过给每台宿主机分配一个子网的方式为容器提供虚拟网络(覆盖网络),该网络中的结点可以看作通过虚拟或逻辑链路而连接起来的 > > **WHY**:我们生产上的集群宿主机/容器之间必须是互通的,因为只有互通才能形成集群,要是集群间的宿主机和容器都不互通,那就没有做集群的必要了 ~~~ # 你可以做如下尝试,21机器: ~]# kubectl get pods -o wide ~]# ping 172.7.21.2 ~]# ping 172.7.22.2 ~~~ ![1579141174992](assets/1579141174992.png) > 你可以发现,两个容器的宿主机之间是不互通的,更别说进入容器里面了。(当然ping10.4.7.22是没问题的) > > 这时候我们就需要CNI网络插件,CNI最主要的功能是实现POD资源能够跨宿主机进行通信,当然CNI网络插件有很多种,如Flannel、Calico等,而Flannel是目前市场上最为火热的 ~~~~ # 21/22机器: ~]# cd /opt/src/ src]# wget https://github.com/coreos/flannel/releases/download/v0.11.0/flannel-v0.11.0-linux-amd64.tar.gz src]# mkdir /opt/flannel-v0.11.0 src]# tar xf flannel-v0.11.0-linux-amd64.tar.gz -C /opt/flannel-v0.11.0/ src]# ln -s /opt/flannel-v0.11.0/ /opt/flannel src]# cd /opt/flannel flannel]# ll # out:总用量 34436 flannel]# mkdir cert flannel]# cd cert/ cert]# scp hdss7-200:/opt/certs/ca.pem . cert]# scp hdss7-200:/opt/certs/client.pem . cert]# scp hdss7-200:/opt/certs/client-key.pem . cert]# cd .. # 注意机器名,需要改一处:SUBNET=172.7.21.1/24,需要改成SUBNET=172.7.22.1/24 flannel]# vi subnet.env FLANNEL_NETWORK=172.7.0.0/16 FLANNEL_SUBNET=172.7.21.1/24 FLANNEL_MTU=1500 FLANNEL_IPMASQ=false ~~~~ ![1579144065680](assets/1579144065680.png) ~~~ # 21/22机器,注意,我的网络是eth0,新版的是ens33,如果是ens33,则需要改iface,其它需要改一处机器名:ip=10.4.7.21 flannel]# vi flanneld.sh #!/bin/sh ./flanneld \ --public-ip=10.4.7.21 \ --etcd-endpoints=https://10.4.7.12:2379,https://10.4.7.21:2379,https://10.4.7.22:2379 \ --etcd-keyfile=./cert/client-key.pem \ --etcd-certfile=./cert/client.pem \ --etcd-cafile=./cert/ca.pem \ --iface=eth0 \ --subnet-file=./subnet.env \ --healthz-port=2401 flannel]# chmod +x flanneld.sh flannel]# mkdir -p /data/logs/flanneld flannel]# cd /opt/etcd # 下面这一步在一部机器上执行即可,只需执行一次,我在21机器做的: etcd]# ./etcdctl set /coreos.com/network/config '{"Network": "172.7.0.0/16", "Backend": {"Type": "host-gw"}}' etcd]# ./etcdctl get /coreos.com/network/config # out:{"Network": "172.7.0.0/16", "Backend": {"Type": "host-gw"}} # 有一处要修改,21/22机器:flanneld-7-21] etcd]# vi /etc/supervisord.d/flannel.ini [program:flanneld-7-21] command=/opt/flannel/flanneld.sh ; the program (relative uses PATH, can take args) numprocs=1 ; number of processes copies to start (def 1) directory=/opt/flannel ; directory to cwd to before exec (def no cwd) autostart=true ; start at supervisord start (default: true) autorestart=true ; retstart at unexpected quit (default: true) startsecs=30 ; number of secs prog must stay running (def. 1) startretries=3 ; max # of serial start failures (default 3) exitcodes=0,2 ; 'expected' exit codes for process (default 0,2) stopsignal=QUIT ; signal used to kill process (default TERM) stopwaitsecs=10 ; max num secs to wait b4 SIGKILL (default 10) user=root ; setuid to this UNIX account to run the program redirect_stderr=true ; redirect proc stderr to stdout (default false) stdout_logfile=/data/logs/flanneld/flanneld.stdout.log ; stderr log path, NONE for none; default AUTO stdout_logfile_maxbytes=64MB ; max # logfile bytes b4 rotation (default 50MB) stdout_logfile_backups=4 ; # of stdout logfile backups (default 10) stdout_capture_maxbytes=1MB ; number of bytes in 'capturemode' (default 0) stdout_events_enabled=false ; emit events on stdout writes (default false) etcd]# supervisorctl update etcd]# supervisorctl status # 查看细节信息 etcd]# tail -fn 200 /data/logs/flanneld/flanneld.stdout.log # 两部机器完成后,在21和22机器ping对方,已经可以ping通 ~~~ ![1579147672612](assets/1579147672612.png) 完成 flannel原理:添加静态路由(前提条件,必须处在同一网关之下) 利用10.4.7.x本来互通的前提,172先去找10再转到其下面的172,形成互通 ![1582269815607](assets/1582269815607.png) > 再次复习一遍,10的21机器对应的172的21,这样方便知道那些pod在那些机器上 ### flannel之SNAT规则优化 > **WHAT**:使得容器之间的透明访问 > > **WHY**:解决两宿主机容器之间的透明访问,如不进行优化,容器之间的访问,日志记录为宿主机的IP地址 ~~~ # 把nginx:curl拉下来,21机器 ~]# docker login docker.io/909336740/nginx:curl ~]# docker pull 909336740/nginx:curl ~]# docker images|grep curl ~]# docker tag 34736e20b17b harbor.od.com/public/nginx:curl ~]# docker login harbor.od.com ~]# docker push harbor.od.com/public/nginx:curl ~~~ ![1579152363774](assets/1579152363774.png) ~~~ # 改以下内容,21机器: cd ~]# vi nginx-ds.yaml image: harbor.od.com/public/nginx:curl ~]# kubectl apply -f nginx-ds.yaml ~]# kubectl get pods # 删掉两个pod让它们自动重启以便应用新镜像 ~]# kubectl delete pod nginx-ds-5nhq6 # out:pod "nginx-ds-5nhq6" deleted ~]# kubectl delete pod nginx-ds-cfjvn #out:pod "nginx-ds-cfjvn" deleted ~]# kubectl get pods -o wide ~~~ ![1579152725660](assets/1579152725660.png) ~~~ # 21机器: ~]# kubectl exec -ti nginx-ds-6nmbr /bin/bash 6nmbr:/# curl 172.7.22.2 # 注意这个pod是起在了22网络了,如果网段没有在22上,就curl 172.7.22.2,只要有welcome的网页回应即可,而且log日志也有 # 22机器 etcd]# kubectl get pods -o wide etcd]# kubectl logs -f nginx-ds-drrkt ~~~ > **kubectl logs -f**:查看Pod日志 ![1579154925882](assets/1579154925882.png) ![1579155042838](assets/1579155042838.png) 确认启动正常 ~~~ # 21机器: ~]# iptables-save |grep -i postrouting ~~~ ![1582271259863](assets/1582271259863.png) > **iptables:** > > - `语法:iptables [-t 表名] 选项 [链名] [条件] [-j 控制类型]` > - **-A**:在规则链的末尾加入新规则 > - **-s**:匹配来源地址IP/MASK,加叹号"!"表示除这个IP外 > - **-o**:匹配从这块网卡流出的数据 > - **MASQUERADE**:动态伪装,能够自动的寻找外网地址并改为当前正确的外网IP地址 > - 上面红框内的可以理解为:如果是172.7.21.0/24段的docker的ip,网络发包不从docker0桥设备出战的,就进行SNAT转换,而我们需要的是如果出网的地址是172.7.21.0/24或者172.7.0.0/16网络(这是docker的大网络),就不要做源地址NAT转换,因为我们集群内部需要坦诚相见,自己人不需要伪装。 ~~~ # 21/22机器,我们开始改: ~]# yum install iptables-services -y ~]# systemctl start iptables ~]# systemctl enable iptables # 删掉对应的规则,以下需要对应机器,一处修改:-s 172.7.21 ~]# iptables -t nat -D POSTROUTING -s 172.7.21.0/24 ! -o docker0 -j MASQUERADE # 添加对应的规则,以下需要对应机器,一处修改:-s 172.7.21 ~]# iptables -t nat -I POSTROUTING -s 172.7.21.0/24 ! -d 172.7.0.0/16 ! -o docker0 -j MASQUERADE # 上面这条规则可以理解为:只有出网地址不是172.7.21.0/24或者172.7.0.0/16,网络发包不从docker0桥设备出战的,才做SNAT转换 ~]# iptables-save |grep -i postrouting ~]# iptables-save > /etc/sysconfig/iptables # 21机器curl22,22机器curl21 ~]# kubectl exec -ti nginx-ds-6nmbr /bin/bash 6nmbr:/# curl 172.7.22.2 ### 相关报错 # 如果报错:curl: (7) Failed to connect to 172.7.22.2 port 80: No route to host # 则执行以下操作,在删掉两台机器21/22的iptables的reject,两边同时执行 ~]# iptables-save|grep -i reject ~]# iptables -t filter -D [名字] ~]# iptables-save > /etc/sysconfig/iptables ### ~~~ > **iptables:** > > - `语法:iptables [-t 表名] 选项 [链名] [条件] [-j 控制类型]` > - -D:删除某一条规则 > - -I:在规则链的头部加入新规则 > - -s:匹配来源地址IP/MASK,加叹号"!"表示除这个IP外 > - -d:匹配目标地址 > - -o:匹配从这块网卡流出的数据 > - MASQUERADE:动态伪装,能够自动的寻找外网地址并改为当前正确的外网IP地址 ![1579157014119](assets/1579157014119.png) 成功图,在22上已经可以明确的看到对方是172.7.21.2了,在21上可以看到对方是172的22: ![1579157196199](assets/1579157196199.png) ![1579157441471](assets/1579157441471.png) 完成 ### 安装部署coredns(服务发现): > **WHAT**:服务(应用)之间相互定位的过程 > > **WHY:** > > - 服务发现对应的场景: > - 服务(应用)的动态性抢 > - 服务(应用)更新发布频繁 > - 服务(应用)支持自动伸缩 > > - kuberntes中的所有pod都是基于Service域名解析后,再负载均衡分发到service后端的各个pod服务中,POD的IP是不断变化的。如何解决: > - 抽象出Service资源,通过标签选择器,关联一组POD > - 抽象出集群网络,通过固定的“集群IP”,使服务接入点固定 > - 如何管理Service资源的“名称”和“集群网络IP” > - 我们前面做了传统的DNS模型:hdss7-21.host.com -> 10.4.7.21 > - 那么我们可以在K8S里做这样的模型:nginx-ds -> 192.168.0.1 ~~~ # 现在我们要开始用交付容器方式交付服务(非二进制),这也是以后最常用的方式 # 200机器 certs]# cd /etc/nginx/conf.d/ conf.d]# vi /etc/nginx/conf.d/k8s-yaml.od.com.conf server { listen 80; server_name k8s-yaml.od.com; location / { autoindex on; default_type text/plain; root /data/k8s-yaml; } } conf.d]# mkdir /data/k8s-yaml conf.d]# nginx -t conf.d]# nginx -s reload ~~~ ~~~ # 11机器,解析域名: ~]# vi /var/named/od.com.zone serial 前滚一位 # 最下面添加这个网段,以后也都是在最下面添加,后面我就加这个注释了 k8s-yaml A 10.4.7.200 ~]# systemctl restart named ~]# dig -t A k8s-yaml.od.com @10.4.7.11 +short # out:10.4.7.200 ~~~ > **dig -t A**:指的是找DNS里标记为A的相关记录,@用什么机器IP访问,+short是只返回IP ![1579158143760](assets/1579158143760.png) ~~~ # 200机器 conf.d]# cd /data/k8s-yaml/ k8s-yaml]# mkdir coredns ~~~ [k8s-yaml.od.com](k8s-yaml.od.com) ![1579158360896](assets/1579158360896.png) ~~~~ # 200机器,下载coredns镜像: cd /data/k8s-yaml/ k8s-yaml]# docker pull coredns/coredns:1.6.1 k8s-yaml]# docker images|grep coredns k8s-yaml]# docker tag c0f6e815079e harbor.od.com/public/coredns:v1.6.1 k8s-yaml]# docker push !$ ~~~~ > 这里我们需要注意的是,任何我用到的镜像都会推到我的本地私有仓库,原因前面也说了,1、是为了用的时候速度快保证不出现网络问题,2、保证版本是同样的版本,而不是突然被别人修改了 > > **docker push !$**:push上一个镜像的名字 ~~~ # 200机器,准备资源配置清单: cd /data/k8s-yaml/coredns coredns]# vi rbac.yaml apiVersion: v1 kind: ServiceAccount metadata: name: coredns namespace: kube-system labels: kubernetes.io/cluster-service: "true" addonmanager.kubernetes.io/mode: Reconcile --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: labels: kubernetes.io/bootstrapping: rbac-defaults addonmanager.kubernetes.io/mode: Reconcile name: system:coredns rules: - apiGroups: - "" resources: - endpoints - services - pods - namespaces verbs: - list - watch --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRoleBinding metadata: annotations: rbac.authorization.kubernetes.io/autoupdate: "true" labels: kubernetes.io/bootstrapping: rbac-defaults addonmanager.kubernetes.io/mode: EnsureExists name: system:coredns roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: system:coredns subjects: - kind: ServiceAccount name: coredns namespace: kube-system coredns]# vi cm.yaml apiVersion: v1 kind: ConfigMap metadata: name: coredns namespace: kube-system data: Corefile: | .:53 { errors log health ready kubernetes cluster.local 192.168.0.0/16 forward . 10.4.7.11 cache 30 loop reload loadbalance } coredns]# vi dp.yaml apiVersion: apps/v1 kind: Deployment metadata: name: coredns namespace: kube-system labels: k8s-app: coredns kubernetes.io/name: "CoreDNS" spec: replicas: 1 selector: matchLabels: k8s-app: coredns template: metadata: labels: k8s-app: coredns spec: priorityClassName: system-cluster-critical serviceAccountName: coredns containers: - name: coredns image: harbor.od.com/public/coredns:v1.6.1 args: - -conf - /etc/coredns/Corefile volumeMounts: - name: config-volume mountPath: /etc/coredns ports: - containerPort: 53 name: dns protocol: UDP - containerPort: 53 name: dns-tcp protocol: TCP - containerPort: 9153 name: metrics protocol: TCP livenessProbe: httpGet: path: /health port: 8080 scheme: HTTP initialDelaySeconds: 60 timeoutSeconds: 5 successThreshold: 1 failureThreshold: 5 dnsPolicy: Default volumes: - name: config-volume configMap: name: coredns items: - key: Corefile path: Corefile coredns]# vi svc.yaml apiVersion: v1 kind: Service metadata: name: coredns namespace: kube-system labels: k8s-app: coredns kubernetes.io/cluster-service: "true" kubernetes.io/name: "CoreDNS" spec: selector: k8s-app: coredns clusterIP: 192.168.0.2 ports: - name: dns port: 53 protocol: UDP - name: dns-tcp port: 53 - name: metrics port: 9153 protocol: TCP ~~~ ~~~~ # 21机器,应用资源配置清单(陈述式): ~]# kubectl apply -f http://k8s-yaml.od.com/coredns/rbac.yaml ~]# kubectl apply -f http://k8s-yaml.od.com/coredns/cm.yaml ~]# kubectl apply -f http://k8s-yaml.od.com/coredns/dp.yaml ~]# kubectl apply -f http://k8s-yaml.od.com/coredns/svc.yaml ~]# kubectl get all -n kube-system ~~~~ ![1579159195207](assets/1579159195207.png) > **CLUSTER-IP为什么是192.168.0.2**:因为我们之前已经写死了这是我们dns的统一接入点 > > ![1582278638665](assets/1582278638665.png) ~~~ # 21机器,测试(我的已经存在了,不过不影响): ~]# kubectl create deployment nginx-dp --image=harbor.od.com/public/nginx:v1.7.9 -n kube-public ~]# kubectl expose deployment nginx-dp --port=80 -n kube-public ~]# kubectl get svc -n kube-public ~]# dig -t A nginx-dp.kube-public.svc.cluster.local. @192.168.0.2 +short # out:192.168.81.37 ~~~ > **dig -t A**:指的是找DNS里标记为A的相关记录,@用什么机器IP访问,+short是只返回IP ![1579161063657](assets/1579161063657.png) 完成集群“内”被自动发现 ### K8S的服务暴露ingress > **WHAT**:K8S API的标准资源类型之一,也是核心资源,它是基于域名和URL路径,把用户的请求转发至指定Service资源的规则 > > - 将集群外部的请求流量,转发至集群内部,从而实现“服务暴露” > - nginx + go脚本 > > **WHY**:上面实现了服务在集群“内”被自动发现,那么需要使得服务在集群“外”被使用和访问,常规的两种方法: > > - 使用NodePort型的service > - 无法使用kube-proxy的ipvs模型,只能使用iptables模型 > - 使用ingress资源 > - 只能调度并暴露7蹭应用,特指http和https协议 ##### 以trafiker为例 > **WHAT**:为了让部署微服务更加便捷而诞生的现代HTTP反向代理、负载均衡工具。 > > **WHY**:可以监听你的服务发现/基础架构组件的管理API,并且每当你的微服务被添加、移除、杀死或更新都会被感知,并且可以自动生成它们的配置文件 ~~~ # 200机器,部署traefiker(ingress控制器) cd /data/k8s-yaml/ k8s-yaml]# mkdir traefik k8s-yaml]# cd traefik/ traefik]# docker pull traefik:v1.7.2-alpine traefik]# docker images|grep traefik traefik]# docker tag add5fac61ae5 harbor.od.com/public/traefik:v1.7.2 traefik]# docker push harbor.od.com/public/traefik:v1.7.2 ~~~ > 复习:mkdir 创建目录、cd 移动到其它目录、 > > docker pull 下载镜像、docker tag 打标签、docker push 上传到仓库 ~~~ # 200机器,准备资源配置清单(4个yaml): traefik]# vi rbac.yaml apiVersion: v1 kind: ServiceAccount metadata: name: traefik-ingress-controller namespace: kube-system --- apiVersion: rbac.authorization.k8s.io/v1beta1 kind: ClusterRole metadata: name: traefik-ingress-controller rules: - apiGroups: - "" resources: - services - endpoints - secrets verbs: - get - list - watch - apiGroups: - extensions resources: - ingresses verbs: - get - list - watch --- kind: ClusterRoleBinding apiVersion: rbac.authorization.k8s.io/v1beta1 metadata: name: traefik-ingress-controller roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: traefik-ingress-controller subjects: - kind: ServiceAccount name: traefik-ingress-controller namespace: kube-system traefik]# vi ds.yaml apiVersion: extensions/v1beta1 kind: DaemonSet metadata: name: traefik-ingress namespace: kube-system labels: k8s-app: traefik-ingress spec: template: metadata: labels: k8s-app: traefik-ingress name: traefik-ingress spec: serviceAccountName: traefik-ingress-controller terminationGracePeriodSeconds: 60 containers: - image: harbor.od.com/public/traefik:v1.7.2 name: traefik-ingress ports: - name: controller containerPort: 80 hostPort: 81 - name: admin-web containerPort: 8080 securityContext: capabilities: drop: - ALL add: - NET_BIND_SERVICE args: - --api - --kubernetes - --logLevel=INFO - --insecureskipverify=true - --kubernetes.endpoint=https://10.4.7.10:7443 - --accesslog - --accesslog.filepath=/var/log/traefik_access.log - --traefiklog - --traefiklog.filepath=/var/log/traefik.log - --metrics.prometheus traefik]# vi svc.yaml kind: Service apiVersion: v1 metadata: name: traefik-ingress-service namespace: kube-system spec: selector: k8s-app: traefik-ingress ports: - protocol: TCP port: 80 name: controller - protocol: TCP port: 8080 name: admin-web traefik]# vi ingress.yaml apiVersion: extensions/v1beta1 kind: Ingress metadata: name: traefik-web-ui namespace: kube-system annotations: kubernetes.io/ingress.class: traefik spec: rules: - host: traefik.od.com http: paths: - path: / backend: serviceName: traefik-ingress-service servicePort: 8080 ~~~ > 每次有ingress时,我们第一反应就是要去解析域名 > > 这里为什么我们都可以把什么都丢到80端口,是因为现在已经是Pod了,已经隔离了,无所谓你用什么端口 ~~~ # 21/22任意机器(我用的22),应用资源配置清单: ~]# kubectl apply -f http://k8s-yaml.od.com/traefik/rbac.yaml ~]# kubectl apply -f http://k8s-yaml.od.com/traefik/ds.yaml ~]# kubectl apply -f http://k8s-yaml.od.com/traefik/svc.yaml ~]# kubectl apply -f http://k8s-yaml.od.com/traefik/ingress.yaml # 下面重启docker服务要在21/22节点都执行,否则会有一个起不来 ~]# systemctl restart docker.service ~]# kubectl get pods -n kube-system ~]# netstat -luntp|grep 81 ~~~ ![1579165378812](assets/1579165378812.png) ~~~ # 11/12机器,做反代: ~]# vi /etc/nginx/conf.d/od.com.conf upstream default_backend_traefik { server 10.4.7.21:81 max_fails=3 fail_timeout=10s; server 10.4.7.22:81 max_fails=3 fail_timeout=10s; } server { server_name *.od.com; location / { proxy_pass http://default_backend_traefik; proxy_set_header Host $http_host; proxy_set_header x-forwarded-for $proxy_add_x_forwarded_for; } } ~]# nginx -t ~]# nginx -s reload # 11机器,解析域名: ~]# vi /var/named/od.com.zone 前滚serial traefik A 10.4.7.10 ~]# systemctl restart named ~~~ > **nginx -t**:检查nginx.conf文件有没有语法错误 > > **nginx -s reload**:不需要重启nginx的热配置 ![1579167500955](assets/1579167500955.png) [访问traefik.od.com](traefik.od.com) ![1579167546083](assets/1579167546083.png) 完成 ##### 用户访问流程: 当用户输入traefik.od.com时,被dns解析到10.4.7.10,而10则在11上,去找L7层服务,而反代配置的od.com.conf,则是将*.od.com无差别的抛给了ingress,ingress则通过noteselect找到pod ![1584961721898](assets/1584961721898.png) 再回顾上面的架构图,我们已经全部安装部署完。 接下来,我们就要开始安装部署K8S的周边生态,使其成为一个**真正的PaaS服务** kubernetes技能图谱