使用sealer-构建、交付、运行【kubernetes】-demo
Kubernetes 是一个开源的容器编排引擎,用来对容器化应用进行自动化部署、 扩缩和管理.
sealer
把整个集群看成一台服务器,把kubernetes看成云操作系统,吸取docker设计精髓实现分布式软件镜像化构建、交付、运行
- 构建: 使用Kubefile定义整个集群所有依赖,把kubernetes 中间件 数据库和SaaS软件所有依赖打包到集群镜像中
- 交付: 像交付Docker镜像一样交付整个集群和集群中的分布式软件
- 运行: 一条命令启动整个集群, 集群纬度保障一致性,兼容主流linux系统与多种架构
当前文档简单使用sealer,暂未应用,通过sealer看看Kubernetes。
参考
sealer
Kubernetes 文档
sealer是如何把sealos按在地上摩擦的
sealos-gitee
sealos-github
sealer-gitee
sealer-github
Containerd 的前世今生和保姆级入门教程
sealos爆火后依然要开发sealer,背后原因有哪些?
K8S高可用之Sealos
demo-操作
使用sealer创建一个kubernetes集群
主要参考:
下载
wget -c https://sealer.oss-cn-beijing.aliyuncs.com/sealers/sealer-v0.8.5-linux-amd64.tar.gz
解压并移动
tar -xvf sealer-v0.8.5-linux-amd64.tar.gz -C /usr/bin/
调整授权用户(如果是root请忽略)
压缩文件内
chown root:root /usr/bin/sealer
ll /usr/bin/ |grep seal
命令
版本
sealer version
{"gitVersion": "v0.8.5","gitCommit": "f9c3d99","buildDate": "2022-04-28 14:16:58","goVersion": "go1.16.15","compiler": "gc","platform": "linux/amd64"
}
当前版本命令
Available Commands:apply 应用kubernetes集群build 从Kubefile构建云imagecert 更新k8s API服务器证书check 检查群集的状态completion 为bash生成自动完成脚本debug 为POD和节点创建调试会话delete 删除集群exec 在所有节点上执行shell命令或脚本。gen 生成一个Clusterfile来接管未由sealer部署的正常集群gen-doc 使用降价格式为sealer CLI生成MarkDownhelp 关于任何命令的帮助images 列出所有集群imagesinspect 打印image信息或Clusterfilejoin 将节点加入集群load 从tar文件加载imagelogin 登录image存储库merge 将多个image合并为一个prune 削减sealer数据目录pull 将云image拉到本地push 将云image推送到注册表rmi 按名称删除本地imagerun 使用image和参数运行集群save 将image保存到tar文件search sealer搜索kubernetestag 将image标记为新imageupgrade 升级kubernetes集群version 显示sealer版本Flags:--config string 配置文件 (默认是 $HOME/.sealer.json)-d, --debug 启用调试模式-h, --help sealer帮助--hide-path 隐藏日志路径--hide-time 隐藏日志时间-t, --toggle 切换的帮助消息
一个命令部署整个 Kubernetes 集群
Flags:--cmd-args strings set args for image cmd instruction-e, --env strings set custom environment variables-h, --help help for run-m, --masters string set Count or IPList to masters-n, --nodes string set Count or IPList to nodes-p, --passwd string set cloud provider or baremetal server password--pk string set baremetal server private key (default "/root/.ssh/id_rsa")--pk-passwd string set baremetal server private key password--port string set the sshd service port number for the server (default port: 22) (default "22")--provider ALI_CLOUD set infra provider, example ALI_CLOUD, the local server need ignore this-u, --user string set baremetal server username (default "root")
Sealer运行命令行
# 默认root用户
sealer run kubernetes:v1.19.8 -m 192.168.xxx.xxx -p {root密码}
启动日志
2022-07-04 18:33:26 [INFO] [local.go:288] Start to create a new cluster: master [192.168.xxx.xxx], worker []
Start to Pull Image kubernetes:v1.19.8
70b6c7cb29b2: pull completed
6af04e85d342: pull completed
b66e77105fb0: pull completed
907a935ec2e7: pull completed
ebd0f721b47d: pull completed
1066d30573cf: pull completed
bdfff9613bac: pull completed
bdfff9613bac: already exists=========================> ] 207.8MB/346.4MB
ebd0f721b47d: already exists
1066d30573cf: already exists
6af04e85d342: already exists
869b45ef4055: already exists
907a935ec2e7: already exists
70b6c7cb29b2: already exists
b66e77105fb0: already exists
Success to Pull Image kubernetes:v1.19.8
2022-07-04 18:38:10 [INFO] [kube_certs.go:234] APIserver altNames : {map[apiserver.cluster.local:apiserver.cluster.local kubernetes:kubernetes kubernetes.default:kubernetes.default kubernetes.default.svc:kubernetes.default.svc kubernetes.default.svc.cluster.local:kubernetes.default.svc.cluster.local localhost:localhost localhost.localdomain:localhost.localdomain] map[10.103.xxx.xxx:10.103.xxx.xxx 10.96.xxx.xxx:10.96.xxx.xxx 127.0.0.1:127.0.0.1 172.16.xxx.xxx:172.16.xxx.xxx 192.168.xxx.xxx:192.168.xxx.xxx]}
2022-07-04 18:38:10 [INFO] [kube_certs.go:254] Etcd altnames : {map[localhost:localhost localhost.localdomain:localhost.localdomain] map[127.0.0.1:127.0.0.1 192.168.xxx.xxx:192.168.xxx.xxx ::1:::1]}, commonName : localhost.localdomain
2022-07-04 18:38:17 [INFO] [kubeconfig.go:267] [kubeconfig] Writing "admin.conf" kubeconfig file2022-07-04 18:38:19 [INFO] [kubeconfig.go:267] [kubeconfig] Writing "controller-manager.conf" kubeconfig file2022-07-04 18:38:19 [INFO] [kubeconfig.go:267] [kubeconfig] Writing "scheduler.conf" kubeconfig file2022-07-04 18:38:20 [INFO] [kubeconfig.go:267] [kubeconfig] Writing "kubelet.conf" kubeconfig file++ dirname init-registry.sh
+ cd .
+ REGISTRY_PORT=5000
+ VOLUME=/var/lib/sealer/data/my-cluster/rootfs/registry
+ REGISTRY_DOMAIN=sea.hub
+ container=sealer-registry
+++ pwd
++ dirname /var/lib/sealer/data/my-cluster/rootfs/scripts
+ rootfs=/var/lib/sealer/data/my-cluster/rootfs
+ config=/var/lib/sealer/data/my-cluster/rootfs/etc/registry_config.yml
+ htpasswd=/var/lib/sealer/data/my-cluster/rootfs/etc/registry_htpasswd
+ certs_dir=/var/lib/sealer/data/my-cluster/rootfs/certs
+ image_dir=/var/lib/sealer/data/my-cluster/rootfs/images
+ mkdir -p /var/lib/sealer/data/my-cluster/rootfs/registry
+ load_images
+ for image in '"$image_dir"/*'
+ '[' -f /var/lib/sealer/data/my-cluster/rootfs/images/registry.tar ']'
+ docker load -q -i /var/lib/sealer/data/my-cluster/rootfs/images/registry.tar
Loaded image: registry:2.7.1
++ docker ps -aq -f name=sealer-registry
+ '[' '' ']'
+ regArgs='-d --restart=always --net=host --name sealer-registry -v /var/lib/sealer/data/my-cluster/rootfs/certs:/certs -v /var/lib/sealer/data/my-cluster/rootfs/registry:/var/lib/registry -e REGISTRY_HTTP_TLS_CERTIFICATE=/certs/sea.hub.crt -e REGISTRY_HTTP_TLS_KEY=/certs/sea.hub.key'
+ '[' -f /var/lib/sealer/data/my-cluster/rootfs/etc/registry_config.yml ']'
+ sed -i s/5000/5000/g /var/lib/sealer/data/my-cluster/rootfs/etc/registry_config.yml
+ regArgs='-d --restart=always --net=host --name sealer-registry -v /var/lib/sealer/data/my-cluster/rootfs/certs:/certs -v /var/lib/sealer/data/my-cluster/rootfs/registry:/var/lib/registry -e REGISTRY_HTTP_TLS_CERTIFICATE=/certs/sea.hub.crt -e REGISTRY_HTTP_TLS_KEY=/certs/sea.hub.key -v /var/lib/sealer/data/my-cluster/rootfs/etc/registry_config.yml:/etc/docker/registry/config.yml'
+ '[' -f /var/lib/sealer/data/my-cluster/rootfs/etc/registry_htpasswd ']'
+ docker run -d --restart=always --net=host --name sealer-registry -v /var/lib/sealer/data/my-cluster/rootfs/certs:/certs -v /var/lib/sealer/data/my-cluster/rootfs/registry:/var/lib/registry -e REGISTRY_HTTP_TLS_CERTIFICATE=/certs/sea.hub.crt -e REGISTRY_HTTP_TLS_KEY=/certs/sea.hub.key -v /var/lib/sealer/data/my-cluster/rootfs/etc/registry_config.yml:/etc/docker/registry/config.yml registry:2.7.1
6d81axxx.xxx
+ check_registry
+ n=1
+ (( n <= 3 ))
++ docker inspect --format '{{json .State.Status}}' sealer-registry
+ registry_status='"running"'
+ [[ "running" == \"running\" ]]
+ break
2022-07-04 18:40:30 [INFO] [init.go:251] start to init master0...
2022-07-04 18:44:38 [INFO] [init.go:206] join command is: kubeadm join apiserver.cluster.local:6443 --token 2uywik.xxx.xxx \--discovery-token-ca-cert-hash sha256:f8efxxx.xxx \--control-plane --certificate-key 0830xxx.xxxcustomresourcedefinition.apiextensions.k8s.io/bgpconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/bgppeers.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/blockaffinities.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/caliconodestatuses.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/clusterinformations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/felixconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/globalnetworkpolicies.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/globalnetworksets.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/hostendpoints.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamblocks.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamconfigs.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamhandles.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ippools.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipreservations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/kubecontrollersconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/networkpolicies.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/networksets.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io created
customresourcedefinition.apiextensions.k8s.io/imagesets.operator.tigera.io created
customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io created
customresourcedefinition.apiextensions.k8s.io/tigerastatuses.operator.tigera.io created
namespace/tigera-operator created
podsecuritypolicy.policy/tigera-operator created
serviceaccount/tigera-operator created
clusterrole.rbac.authorization.k8s.io/tigera-operator created
clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created
deployment.apps/tigera-operator created
installation.operator.tigera.io/default created
apiserver.operator.tigera.io/default created
2022-07-04 18:44:48 [INFO] [local.go:298] Succeeded in creating a new cluster, enjoy it!
kubectl–help
kubectl controls the Kubernetes cluster manager.Find more information at: https://kubernetes.io/docs/reference/kubectl/overview/Basic Commands (Beginner):create Create a resource from a file or from stdin.expose 使用 replication controller, service, deployment 或者 pod 并暴露它作为一个 新的 Kubernetes
Servicerun 在集群中运行一个指定的镜像set 为 objects 设置一个指定的特征Basic Commands (Intermediate):explain 查看资源的文档get 显示一个或更多 resourcesedit 在服务器上编辑一个资源delete Delete resources by filenames, stdin, resources and names, or by resources and label selectorDeploy Commands:rollout Manage the rollout of a resourcescale Set a new size for a Deployment, ReplicaSet or Replication Controllerautoscale 自动调整一个 Deployment, ReplicaSet, 或者 ReplicationController 的副本数量Cluster Management Commands:certificate 修改 certificate 资源.cluster-info 显示集群信息top Display Resource (CPU/Memory/Storage) usage.cordon 标记 node 为 unschedulableuncordon 标记 node 为 schedulabledrain Drain node in preparation for maintenancetaint 更新一个或者多个 node 上的 taintsTroubleshooting and Debugging Commands:describe 显示一个指定 resource 或者 group 的 resources 详情logs 输出容器在 pod 中的日志attach Attach 到一个运行中的 containerexec 在一个 container 中执行一个命令port-forward Forward one or more local ports to a podproxy 运行一个 proxy 到 Kubernetes API servercp 复制 files 和 directories 到 containers 和从容器中复制 files 和 directories.auth Inspect authorizationAdvanced Commands:diff Diff live version against would-be applied versionapply 通过文件名或标准输入流(stdin)对资源进行配置patch 使用 strategic merge patch 更新一个资源的 field(s)replace 通过 filename 或者 stdin替换一个资源wait Experimental: Wait for a specific condition on one or many resources.convert 在不同的 API versions 转换配置文件kustomize Build a kustomization target from a directory or a remote url.Settings Commands:label 更新在这个资源上的 labelsannotate 更新一个资源的注解completion Output shell completion code for the specified shell (bash or zsh)Other Commands:alpha Commands for features in alphaapi-resources Print the supported API resources on the serverapi-versions Print the supported API versions on the server, in the form of "group/version"config 修改 kubeconfig 文件plugin Provides utilities for interacting with plugins.version 输出 client 和 server 的版本信息Usage:kubectl [flags] [options]Use "kubectl <command> --help" for more information about a given command.
Use "kubectl options" for a list of global command-line options (applies to all commands).
检查集群
# sealer正在启动的时候,是获取不到信息的
kubectl get node
#The connection to the server localhost:8080 was refused - did you specify the right host or port?
# sealer启动成功后就可以了
kubectl get node
#NAME STATUS ROLES AGE VERSION
#localhost.localdomain Ready master 5m30s v1.19.8
pod状态
Pod是Kubernetes创建或部署的最小/最简单的基本单位,一个Pod代表集群上正在运行的一个进程
kubectl get pod -A
监听模式
kubectl get pod -A -w
挂载情况
可以发现,挂载了很多容器
df -kh
sealer在启动的时候,容器数据是越来越多的,当前是启动成功后,路径做了压缩
文件系统 容量 已用 可用 已用% 挂载点
/dev/mapper/centos-root 50G 23G 28G 45% /
devtmpfs 7.8G 0 7.8G 0% /dev
tmpfs 7.8G 84K 7.8G 1% /dev/shm
tmpfs 7.8G 876M 7.0G 12% /run
tmpfs 7.8G 0 7.8G 0% /sys/fs/cgroup
/dev/sda1 1014M 180M 835M 18% /boot
/dev/mapper/centos-home 42G 3.4G 38G 9% /home
tmpfs 1.6G 0 1.6G 0% /run/user/0
tmpfs 1.6G 0 1.6G 0% /run/user/984
tmpfs 1.6G 0 1.6G 0% /run/user/1003
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/a5bc...e65e/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/5a3a...d865/merged
shm 64M 0 64M 0% /var/lib/docker/containers/d510...0a1e/mounts/shm
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/a21d...8e9b/merged
shm 64M 0 64M 0% /var/lib/docker/containers/e4b6...1173/mounts/shm
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/cc59...c72c/merged
shm 64M 0 64M 0% /var/lib/docker/containers/e567...6685/mounts/shm
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/5b6e...98d2/merged
shm 64M 0 64M 0% /var/lib/docker/containers/3f19...78ec/mounts/shm
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/446c...754e/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/98a0...3657/merged
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/ba6a...c6f/volumes/kubernetes.io~secret/kube-proxy-token-zd26q
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/84afc435a3ab38ab2421f6faf9cffdcb3acaeb7a18313ae1a34889357a46945c/merged
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/d039...b5c0/volumes/kubernetes.io~secret/tigera-operator-token-crqnv
shm 64M 0 64M 0% /var/lib/docker/containers/1667...028a/mounts/shm
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/496d...8cde/merged
shm 64M 0 64M 0% /var/lib/docker/containers/e64c...6364/mounts/shm
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/a36a...85cd/merged
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/d19d...3ba2/volumes/kubernetes.io~secret/calico-typha-token-k476g
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/d19d...3ba2/volumes/kubernetes.io~secret/typha-certs
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/36bd...a6e9/volumes/kubernetes.io~secret/felix-certs
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/36bd...a6e9/volumes/kubernetes.io~secret/calico-node-token-qh4s6
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/0529...5257/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/9f3d...4d93/merged
shm 64M 0 64M 0% /var/lib/docker/containers/f578...c67e/mounts/shm
shm 64M 0 64M 0% /var/lib/docker/containers/8db2...773c/mounts/shm
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/758f...afde/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/0173...b9af/merged
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/64f0...441e/volumes/kubernetes.io~secret/coredns-token-7sxtt
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/f660...eb6d/volumes/kubernetes.io~secret/coredns-token-7sxtt
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/918c...582c/volumes/kubernetes.io~secret/calico-kube-controllers-token-zx262
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/c5f2a...fb58/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/b808...e85a/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/2688...acfd/merged
shm 64M 0 64M 0% /var/lib/docker/containers/1b75...8e73/mounts/shm
shm 64M 0 64M 0% /var/lib/docker/containers/4c1b...c7e4/mounts/shm
shm 64M 0 64M 0% /var/lib/docker/containers/38e2...90b2/mounts/shm
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/1ed...cce6/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/56d6...14d4/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/4084...88ee/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/27b2...3bb2/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/3983...e1a5/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/1b6b...a007/merged
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/fbfc...882f/volumes/kubernetes.io~secret/calico-apiserver-token-h468b
tmpfs 7.8G 8.0K 7.8G 1% /var/lib/kubelet/pods/fbfc...882f/volumes/kubernetes.io~secret/calico-apiserver-certs
tmpfs 7.8G 8.0K 7.8G 1% /var/lib/kubelet/pods/6dab...d8f5/volumes/kubernetes.io~secret/calico-apiserver-certs
tmpfs 7.8G 12K 7.8G 1% /var/lib/kubelet/pods/6dab...d8f5/volumes/kubernetes.io~secret/calico-apiserver-token-h468b
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/4476...b167/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/a7c9...67b2/merged
shm 64M 0 64M 0% /var/lib/docker/containers/b0d7...3b1a/mounts/shm
shm 64M 0 64M 0% /var/lib/docker/containers/3770...d3f6/mounts/shm
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/4122...ca4b/merged
overlay 50G 23G 28G 45% /var/lib/docker/overlay2/ab2e...6cd8/merged
扩展和缩减(只有一个服务器-未测试)
略
清理集群
sealer delete --help
if provider is Bare Metal Server will delete kubernetes nodesUsage:sealer delete [flags]Examples:delete to default cluster: sealer delete --masters x.x.x.x --nodes x.x.x.xsealer delete --masters x.x.x.x-x.x.x.y --nodes x.x.x.x-x.x.x.y
delete all:sealer delete --all [--force]sealer delete -f /root/.sealer/mycluster/Clusterfile [--force]sealer delete -c my-cluster [--force]Flags:-f, --Clusterfile string delete a kubernetes cluster with Clusterfile Annotations-a, --all this flags is for delete nodes, if this is true, empty all node ip-c, --cluster string delete a kubernetes cluster with cluster name--force We also can input an --force flag to delete cluster by force-h, --help help for delete-m, --masters string reduce Count or IPList to masters-n, --nodes string reduce Count or IPList to nodesGlobal Flags:--config string config file (default is $HOME/.sealer.json)-d, --debug turn on debug mode--hide-path hide the log path--hide-time hide the log time
执行清理
sealer delete --all
2022-07-05 13:15:59 [INFO] [runtime.go:67] Start to delete cluster: master [192.168.xxx.xxx], node []
Are you sure to delete these nodes? Yes [y/yes], No [n/no] : y
/usr/bin/kubeadm
[reset] Reading configuration from the cluster...
[reset] FYI: You can look at this config file with 'kubectl -n kube-system get cm kubeadm-config -oyaml'
[preflight] Running pre-flight checks
[reset] Removing info for node "localhost.localdomain" from the ConfigMap "kubeadm-config" in the "kube-system" Namespace
[reset] Stopping the kubelet service
[reset] Unmounting mounted directories in "/var/lib/kubelet"
[reset] Deleting contents of config directories: [/etc/kubernetes/manifests /etc/kubernetes/pki]
[reset] Deleting files: [/etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/bootstrap-kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf]
[reset] Deleting contents of stateful directories: [/var/lib/etcd /var/lib/kubelet /var/lib/dockershim /var/run/kubernetes /var/lib/cni]The reset process does not clean CNI configuration. To do so, you must remove /etc/cni/net.dThe reset process does not reset or clean up iptables rules or IPVS tables.
If you wish to reset iptables, you must do so manually by using the "iptables" command.If your cluster was setup to utilize IPVS, run ipvsadm --clear (or similar)
to reset your system's IPVS tables.The reset process does not clean your kubeconfig files and you must remove them manually.
Please, check the contents of the $HOME/.kube/config file.
Module Size Used by
vxlan 49678 0
ip6_udp_tunnel 12755 1 vxlan
udp_tunnel 14423 1 vxlan
xt_multiport 12798 11
ipt_rpfilter 12606 1
iptable_raw 12678 1
ip_set_hash_ip 31658 1
ip_set_hash_net 36021 3
xt_set 18141 10
ip_set_hash_ipportnet 36255 1
ip_set_bitmap_port 17150 4
ip_set_hash_ipportip 31880 2
ip_set_hash_ipport 31806 8
dummy 12960 0
iptable_mangle 12695 1
xt_mark 12563 97
ip_vs_sh 12688 0
ip_vs_wrr 12697 0
ip_vs_rr 12600 7
ip_vs 145497 13 ip_vs_rr,ip_vs_sh,ip_vs_wrr
nf_conntrack_netlink 36354 0
xt_addrtype 12676 8
br_netfilter 22256 0
ip6table_nat 12864 0
nf_conntrack_ipv6 18935 1
nf_defrag_ipv6 35104 1 nf_conntrack_ipv6
nf_nat_ipv6 14131 1 ip6table_nat
ip6_tables 26912 1 ip6table_nat
xt_conntrack 12760 28
iptable_filter 12810 1
ipt_MASQUERADE 12678 5
nf_nat_masquerade_ipv4 13412 1 ipt_MASQUERADE
xt_comment 12504 315
iptable_nat 12875 1
nf_conntrack_ipv4 15053 29
nf_defrag_ipv4 12729 1 nf_conntrack_ipv4
nf_nat_ipv4 14115 1 iptable_nat
nf_nat 26787 3 nf_nat_ipv4,nf_nat_ipv6,nf_nat_masquerade_ipv4
nf_conntrack 133095 9 ip_vs,nf_nat,nf_nat_ipv4,nf_nat_ipv6,xt_conntrack,nf_nat_masquerade_ipv4,nf_conntrack_netlink,nf_conntrack_ipv4,nf_conntrack_ipv6
ip_tables 27126 4 iptable_filter,iptable_mangle,iptable_nat,iptable_raw
veth 13410 0
overlay 71964 2
tun 31740 1
devlink 48345 0
ip_set 45644 7 ip_set_hash_ipportnet,ip_set_hash_net,ip_set_bitmap_port,ip_set_hash_ipportip,ip_set_hash_ip,xt_set,ip_set_hash_ipport
nfnetlink 14490 3 ip_set,nf_conntrack_netlink
bridge 151336 1 br_netfilter
stp 12976 1 bridge
llc 14552 2 stp,bridge
vmw_vsock_vmci_transport 30577 1
vsock 36526 2 vmw_vsock_vmci_transport
sunrpc 353352 1
nfit 55016 0
libnvdimm 147731 1 nfit
iosf_mbi 15582 0
crc32_pclmul 13133 0
ghash_clmulni_intel 13273 0
aesni_intel 189414 1
lrw 13286 1 aesni_intel
gf128mul 15139 1 lrw
glue_helper 13990 1 aesni_intel
ablk_helper 13597 1 aesni_intel
ppdev 17671 0
cryptd 21190 3 ghash_clmulni_intel,aesni_intel,ablk_helper
vmw_balloon 18094 0
pcspkr 12718 0
sg 40721 0
joydev 17389 0
parport_pc 28205 0
i2c_piix4 22401 0
parport 46395 2 ppdev,parport_pc
vmw_vmci 67127 1 vmw_vsock_vmci_transport
sch_fq_codel 17571 6
xfs 997127 3
libcrc32c 12644 4 xfs,ip_vs,nf_nat,nf_conntrack
sd_mod 46281 3
sr_mod 22416 0
cdrom 42556 1 sr_mod
crc_t10dif 12912 1 sd_mod
crct10dif_generic 12647 0
crct10dif_pclmul 14307 1
crct10dif_common 12595 3 crct10dif_pclmul,crct10dif_generic,crc_t10dif
crc32c_intel 22094 1
serio_raw 13434 0
ata_generic 12923 0
pata_acpi 13053 0
vmwgfx 276430 2
floppy 69432 0
drm_kms_helper 179394 1 vmwgfx
syscopyarea 12529 1 drm_kms_helper
sysfillrect 12701 1 drm_kms_helper
sysimgblt 12640 1 drm_kms_helper
fb_sys_fops 12703 1 drm_kms_helper
ttm 114635 1 vmwgfx
ata_piix 35052 0
drm 429744 5 ttm,drm_kms_helper,vmwgfx
libata 243133 3 pata_acpi,ata_generic,ata_piix
vmxnet3 58059 0
mptspi 22628 2
scsi_transport_spi 30732 1 mptspi
mptscsih 40150 1 mptspi
mptbase 106036 2 mptspi,mptscsih
drm_panel_orientation_quirks 12957 1 drm
dm_mirror 22289 0
dm_region_hash 20813 1 dm_mirror
dm_log 18411 2 dm_region_hash,dm_mirror
dm_mod 124407 10 dm_log,dm_mirror
[{"Id": "6d81...9aff","Created": "2022-07-04T10:40:28.068520236Z","Path": "registry","Args": ["serve","/etc/docker/registry/config.yml"],"State": {"Status": "running","Running": true,"Paused": false,"Restarting": false,"OOMKilled": false,"Dead": false,"Pid": 767,"ExitCode": 0,"Error": "","StartedAt": "2022-07-04T10:40:30.304277835Z","FinishedAt": "0001-01-01T00:00:00Z"},"Image": "sha256:0d01...4fda","ResolvConfPath": "/var/lib/docker/containers/6d81...9aff/resolv.conf","HostnamePath": "/var/lib/docker/containers/6d81...9aff/hostname","HostsPath": "/var/lib/docker/containers/6d81...9aff/hosts","LogPath": "/var/lib/docker/containers/6d81...9aff/6d81...9aff-json.log","Name": "/sealer-registry","RestartCount": 0,"Driver": "overlay2","Platform": "linux","MountLabel": "","ProcessLabel": "","AppArmorProfile": "","ExecIDs": null,"HostConfig": {"Binds": ["/var/lib/sealer/data/my-cluster/rootfs/certs:/certs","/var/lib/sealer/data/my-cluster/rootfs/registry:/var/lib/registry","/var/lib/sealer/data/my-cluster/rootfs/etc/registry_config.yml:/etc/docker/registry/config.yml"],"ContainerIDFile": "","LogConfig": {"Type": "json-file","Config": {"max-file": "3","max-size": "10m"}},"NetworkMode": "host","PortBindings": {},"RestartPolicy": {"Name": "always","MaximumRetryCount": 0},"AutoRemove": false,"VolumeDriver": "","VolumesFrom": null,"CapAdd": null,"CapDrop": null,"Capabilities": null,"Dns": [],"DnsOptions": [],"DnsSearch": [],"ExtraHosts": null,"GroupAdd": null,"IpcMode": "private","Cgroup": "","Links": null,"OomScoreAdj": 0,"PidMode": "","Privileged": false,"PublishAllPorts": false,"ReadonlyRootfs": false,"SecurityOpt": null,"UTSMode": "","UsernsMode": "","ShmSize": 67108864,"Runtime": "runc","ConsoleSize": [0,0],"Isolation": "","CpuShares": 0,"Memory": 0,"NanoCpus": 0,"CgroupParent": "","BlkioWeight": 0,"BlkioWeightDevice": [],"BlkioDeviceReadBps": null,"BlkioDeviceWriteBps": null,"BlkioDeviceReadIOps": null,"BlkioDeviceWriteIOps": null,"CpuPeriod": 0,"CpuQuota": 0,"CpuRealtimePeriod": 0,"CpuRealtimeRuntime": 0,"CpusetCpus": "","CpusetMems": "","Devices": [],"DeviceCgroupRules": null,"DeviceRequests": null,"KernelMemory": 0,"KernelMemoryTCP": 0,"MemoryReservation": 0,"MemorySwap": 0,"MemorySwappiness": null,"OomKillDisable": false,"PidsLimit": null,"Ulimits": null,"CpuCount": 0,"CpuPercent": 0,"IOMaximumIOps": 0,"IOMaximumBandwidth": 0,"MaskedPaths": ["/proc/asound","/proc/acpi","/proc/kcore","/proc/keys","/proc/latency_stats","/proc/timer_list","/proc/timer_stats","/proc/sched_debug","/proc/scsi","/sys/firmware"],"ReadonlyPaths": ["/proc/bus","/proc/fs","/proc/irq","/proc/sys","/proc/sysrq-trigger"]},"GraphDriver": {"Data": {"LowerDir": "/var/lib/docker/overlay2/a5bc...e65e-init/diff:/var/lib/docker/overlay2/362e...7587f2d449685a3a6/diff:/var/lib/docker/overlay2/b907...c3f7/diff:/var/lib/docker/overlay2/8f53...ae07/diff","MergedDir": "/var/lib/docker/overlay2/a5bc12b6f24d74fb23c72e5b051cde8262853d0fdf612ccd060384663ec5e65e/merged","UpperDir": "/var/lib/docker/overlay2/a5bc...3ec5e65e/diff","WorkDir": "/var/lib/docker/overlay2/a5bc...e65e/work"},"Name": "overlay2"},"Mounts": [{"Type": "bind","Source": "/var/lib/sealer/data/my-cluster/rootfs/registry","Destination": "/var/lib/registry","Mode": "","RW": true,"Propagation": "rprivate"},{"Type": "bind","Source": "/var/lib/sealer/data/my-cluster/rootfs/etc/registry_config.yml","Destination": "/etc/docker/registry/config.yml","Mode": "","RW": true,"Propagation": "rprivate"},{"Type": "bind","Source": "/var/lib/sealer/data/my-cluster/rootfs/certs","Destination": "/certs","Mode": "","RW": true,"Propagation": "rprivate"}],"Config": {"Hostname": "localhost.localdomain","Domainname": "","User": "","AttachStdin": false,"AttachStdout": false,"AttachStderr": false,"ExposedPorts": {"5000/tcp": {}},"Tty": false,"OpenStdin": false,"StdinOnce": false,"Env": ["REGISTRY_HTTP_TLS_CERTIFICATE=/certs/sea.hub.crt","REGISTRY_HTTP_TLS_KEY=/certs/sea.hub.key","PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"],"Cmd": ["serve","/etc/docker/registry/config.yml"],"Image": "registry:2.7.1","Volumes": {"/var/lib/registry": {}},"WorkingDir": "","Entrypoint": ["registry"],"OnBuild": null,"Labels": {}},"NetworkSettings": {"Bridge": "","SandboxID": "ae1d...ac1b","HairpinMode": false,"LinkLocalIPv6Address": "","LinkLocalIPv6PrefixLen": 0,"Ports": {},"SandboxKey": "/var/run/docker/netns/default","SecondaryIPAddresses": null,"SecondaryIPv6Addresses": null,"EndpointID": "","Gateway": "","GlobalIPv6Address": "","GlobalIPv6PrefixLen": 0,"IPAddress": "","IPPrefixLen": 0,"IPv6Gateway": "","MacAddress": "","Networks": {"host": {"IPAMConfig": null,"Links": null,"Aliases": null,"NetworkID": "3ed7...98c3e","EndpointID": "ebaa...1ef4","Gateway": "","IPAddress": "","IPPrefixLen": 0,"IPv6Gateway": "","GlobalIPv6Address": "","GlobalIPv6PrefixLen": 0,"MacAddress": "","DriverOpts": null}}}}
]
sealer-registry
2022-07-05 13:16:28 [INFO] [local.go:316] Succeeded in deleting current cluster
删除镜像
sealer rmi kubernetes:v1.19.8
清理后挂载情况
df -kh
可以看到,归还了部分硬盘空间
查看sealer和docker的数据目录大小
du -h --max-depth=1 /var/lib/sealer/*
du -h --max-depth=1 /var/lib/docker/
sealer rmi
后,瞬间清净了。
find / -name docker
find / -name sealer
其他
sealer images
| IMAGE NAME | IMAGE ID | ARCH | VARIANT | CREATE | SIZE |
| kubernetes:v1.19.8 | 46f8...7618 | amd64 | | 2022-07-04 18:37:53 | 956.60MB |
# 清理docker,未找到运行docker运行,先这么处理了,暂不清楚后果如何
rm -rf /var/lib/docker
暂时没有测试其他的操作
END
使用sealer-构建、交付、运行【kubernetes】-demo相关推荐
- 通过Dapr实现一个简单的基于.net的微服务电商系统(十二)——istio+dapr构建多运行时服务网格...
多运行时是一个非常新的概念.在 2020 年,Bilgin Ibryam 提出了 Multi-Runtime(多运行时)的理念,对基于 Sidecar 模式的各种产品形态进行了实践总结和理论升华.那到 ...
- Rasa课程、Rasa培训、Rasa面试、Rasa实战系列之Understanding Rasa Deployments - Local Kubernetes Demo
Rasa课程.Rasa培训.Rasa面试.Rasa实战系列之Understanding Rasa Deployments - Local Kubernetes Demo 目录 Kind 简介 Gavi ...
- 编译+构建+链接+运行之间的关系分析
编译+构建+链接+运行之间的关系分析 源文件+编译+部署+测试+打包 Maven+default: 验证(validate)+编译(compile)+测试(test)+打包(package)+验证(v ...
- CoreOS 和 Kubernetes 1.5 自主运行 Kubernetes、Container Linux
CoreOS 的成立初衷是网络安全保护.显然,这是一个比较大的目标,但是毫无疑问对于隐私策略.商务和政府来说是十分必要的.而且这个目标也是可以完成的. 为了达到这个目标,我们在 Tectonic 大会 ...
- SwiftUI Xcode教程之在 iOS 中使用Schemes 和 .xcconfig 文件来组织构建交付
很多时候,我们需要组织我们的移动交付.这意味着,我们用于在 App Store 中发布我们的应用程序的版本,或者我们每天用于在本地构建我们的应用程序,或在我们公司内部分发的版本是不一样的. 在这里,我 ...
- kind:Kubernetes in Docker,单机运行 Kubernetes 群集的最佳方案?
作者:gc(at)sysin.org,主页:www.sysin.org 请访问原文发布链接:https://sysin.org/article/kind/,查看最新版. 是否没有足够的机器运行 Kub ...
- SpringBoot+Maven 多模块项目的构建、运行、打包实战
https://www.jb51.net/article/140772.htm?proxy=1 这篇文章主要介绍了SpringBoot+Maven 多模块项目的构建.运行.打包实战,小编觉得挺不错的, ...
- java web 服务 构建与运行_Java Web 服务:构建与运行(影印版)
基本详情信息 书名:Java Web 服务:构建与运行 ISBN:9787564119270 作者:卡林(Martin Kalin) 著 出版社:东南大学出版社 出版日期:2010-1-1 其他信息 ...
- 【社区分享 - 游戏篇】AWS云科技助力游戏内容开发 - 游戏项目构建篇 运行篇 增长篇
[社区分享 - 游戏篇]AWS云科技助力游戏内容开发 - 游戏项目构建篇 & 运行篇 & 增长篇 文章目录 [社区分享 - 游戏篇]AWS云科技助力游戏内容开发 - 游戏项目构建篇 & ...
- 小程序开发.mpvue.项目构建与运行
小程序开发.mpvue.项目构建与运行 本文介绍 mpvue 框架的特点以及相关生态,包括在 mpvue 中使用 typescript 和 css 预处理器的相关方法.最后介绍一个最基本 mpvue ...
最新文章
- 前端笔记-thymeleaf显示数据及隐藏数据
- 信息学奥赛一本通C++语言——1031:反向输出一个三位数
- excel显著性检验_怎样征服老板?教你用excel找到数据之间隐藏信息
- NBIOT(3)---NBIOT 移远BC28模块+stm32
- 使用MVC框架中要注意的问题(一):修改首页以支持主题
- 音频放大电路_集成电路技术汇总:检测技巧
- Atitit.http httpclient实践java c# .net php attilax总结
- datatables data bind json
- 多进程相关内容(IPC)
- 顺序栈和迷宫求解(C语言)
- html5扫雷代码,使用js开发网页版 扫雷(附代码详解)
- 协议知识点(2)- flash测试rtmp流是否正常
- 白纸不够白,签名照片还需要ps去除背景?Python大法半完美解决电子签名背景问题(20行代码去除签名背景)
- qt界面切换时出现的绿色背景修改
- 完整打造一个多功能音乐播放器项目(初步设想跟酷狗类似)
- 团队作业第五次—项目系统设计与数据库设计
- 锐龙R3 PRO 5450U性能怎么样?相当于什么水平级别档次?
- mysql高级学习笔记
- map中自定义比较函数
- 香港十大外汇交易商排名名单2020最新版
热门文章
- 【每日一练】后台管理系统小案例
- 服务器mysql 修改数据库密码怎么修改,服务器mysql修改数据库密码
- python能编plc吗_基于Python的丰炜系列PLC与PC串行通信的实现
- Linux下解压.tar.gz 和.gz文件解压的方式
- Mongoose aggregate
- 【新品发布】MP1013 MiniPCIe CANFD卡,应用于自动化数据采集和测试
- 14|跳数索引:后起新秀ClickHouse
- 专题-子项-1-mysqlf-Explain-执行计划-详解--索引--调优实践
- Pytorch model.train()
- Rust 中的 和 ref