各位大佬,我在安装集群的时候,启动TOS报错,之前已经重新安装好几遍了,都是无法启动TOS。
下面是错误日志:
02023-04-12T01:14:37.706 [Master] ========== Task 171 start to run. ==========
12023-04-12T01:14:37.708 [Master] Starting task local part ...
22023-04-12T01:14:37.710 [Master] Start handle role task...
32023-04-12T01:14:37.751 [Master] execute command: DirectiveDetail.SystemctlOp(action=EnableStart, service=kubelet, sleepSec=2)
42023-04-12T01:14:37.751 [Master] execute command: DirectiveDetail.SystemctlOp(action=EnableStart, service=haproxy, sleepSec=2)
52023-04-12T01:14:37.751 [Master] execute command: DirectiveDetail.RenderFileOp(templateType=FreeMarker, templatePath=tos-etcd.manifest, targetPath=/opt/kubernetes/manifests-multi/tos-etcd.manifest, mode=755, owner=null, group=null, opsTpl=false)
62023-04-12T01:14:37.751 [Master] rendering content of: /opt/kubernetes/manifests-multi/tos-etcd.manifest
72023-04-12T01:14:37.753 [Master] content of host node1 file /opt/kubernetes/manifests-multi/tos-etcd.manifest generated
82023-04-12T01:14:37.753 [Master] Start executing [chmod 755 /var/lib/transwarp-manager/master/content/resources/nodes/node1/@opt@kubernetes@manifests-multi@tos-etcd.manifest]
92023-04-12T01:14:37.759 [Master] Execute success.
102023-04-12T01:14:37.760 [Master] Start copy from [localhost] /var/lib/transwarp-manager/master/content/resources/nodes/node1/@opt@kubernetes@manifests-multi@tos-etcd.manifest to [node1] /var/lib/transwarp-manager/agent/resource-tmp/@opt@kubernetes@manifests-multi@tos-etcd.manifest
112023-04-12T01:14:37.765 [Master] Copy success.
122023-04-12T01:14:37.766 [Master] execute command: DirectiveDetail.RenderFileOp(templateType=FreeMarker, templatePath=tos-apiserver.manifest, targetPath=/opt/kubernetes/manifests-multi/tos-apiserver.manifest, mode=755, owner=null, group=null, opsTpl=false)
132023-04-12T01:14:37.766 [Master] rendering content of: /opt/kubernetes/manifests-multi/tos-apiserver.manifest
142023-04-12T01:14:37.772 [Master] content of host node1 file /opt/kubernetes/manifests-multi/tos-apiserver.manifest generated
152023-04-12T01:14:37.772 [Master] Start executing [chmod 755 /var/lib/transwarp-manager/master/content/resources/nodes/node1/@opt@kubernetes@manifests-multi@tos-apiserver.manifest]
162023-04-12T01:14:37.776 [Master] Execute success.
172023-04-12T01:14:37.776 [Master] Start copy from [localhost] /var/lib/transwarp-manager/master/content/resources/nodes/node1/@opt@kubernetes@manifests-multi@tos-apiserver.manifest to [node1] /var/lib/transwarp-manager/agent/resource-tmp/@opt@kubernetes@manifests-multi@tos-apiserver.manifest
182023-04-12T01:14:37.778 [Master] Copy success.
192023-04-12T01:14:37.778 [Master] execute command: DirectiveDetail.RenderFileOp(templateType=FreeMarker, templatePath=tos-controller.manifest, targetPath=/opt/kubernetes/manifests-multi/tos-controller.manifest, mode=755, owner=null, group=null, opsTpl=false)
202023-04-12T01:14:37.778 [Master] rendering content of: /opt/kubernetes/manifests-multi/tos-controller.manifest
212023-04-12T01:14:37.779 [Master] content of host node1 file /opt/kubernetes/manifests-multi/tos-controller.manifest generated
222023-04-12T01:14:37.779 [Master] Start executing [chmod 755 /var/lib/transwarp-manager/master/content/resources/nodes/node1/@opt@kubernetes@manifests-multi@tos-controller.manifest]
232023-04-12T01:14:37.783 [Master] Execute success.
242023-04-12T01:14:37.783 [Master] Start copy from [localhost] /var/lib/transwarp-manager/master/content/resources/nodes/node1/@opt@kubernetes@manifests-multi@tos-controller.manifest to [node1] /var/lib/transwarp-manager/agent/resource-tmp/@opt@kubernetes@manifests-multi@tos-controller.manifest
252023-04-12T01:14:37.785 [Master] Copy success.
262023-04-12T01:14:37.785 [Master] execute command: DirectiveDetail.RenderFileOp(templateType=FreeMarker, templatePath=tos-scheduler.manifest, targetPath=/opt/kubernetes/manifests-multi/tos-scheduler.manifest, mode=755, owner=null, group=null, opsTpl=false)
272023-04-12T01:14:37.785 [Master] rendering content of: /opt/kubernetes/manifests-multi/tos-scheduler.manifest
282023-04-12T01:14:37.787 [Master] content of host node1 file /opt/kubernetes/manifests-multi/tos-scheduler.manifest generated
292023-04-12T01:14:37.787 [Master] Start executing [chmod 755 /var/lib/transwarp-manager/master/content/resources/nodes/node1/@opt@kubernetes@manifests-multi@tos-scheduler.manifest]
302023-04-12T01:14:37.792 [Master] Execute success.
312023-04-12T01:14:37.792 [Master] Start copy from [localhost] /var/lib/transwarp-manager/master/content/resources/nodes/node1/@opt@kubernetes@manifests-multi@tos-scheduler.manifest to [node1] /var/lib/transwarp-manager/agent/resource-tmp/@opt@kubernetes@manifests-multi@tos-scheduler.manifest
322023-04-12T01:14:37.794 [Master] Copy success.
332023-04-12T01:14:37.794 [Master] Task local part ended.
342023-04-12T01:14:37.794 [Master] Starting task remote part ...
352023-04-12T01:14:37.797 [Agent] Execute command: systemctl daemon-reload
36systemctl enable kubelet
37systemctl restart kubelet
38sleep 2
39systemctl status kubelet
402023-04-12T01:14:40.032 [Agent] command output:
41● kubelet.service - Kubernetes Kubelet
42 Loaded: loaded (/usr/lib/systemd/system/kubelet.service; enabled; vendor preset: disabled)
43 Active: active (running) since Wed 2023-04-12 01:14:38 CST; 2s ago
44 Main PID: 124730 (kubelet)
45 Tasks: 27
46 CGroup: /system.slice/kubelet.service
47 └─124730 /opt/kubernetes/bin/kubelet --v=2 --hostname-override=node1 --log-dir=/var/log/kubernetes --node-labels=master=true,worker=true --node-ip=192.168.5.124 --pod-infra-container-image=transwarp/pause:tos-2.1 --network-plugin=cni --eviction-hard= --bootstrap-kubeconfig=/srv/kubernetes/bootstrap.kubeconfig --feature-gates=SupportPodPidsLimit=false,SupportNodePidsLimit=false --kubeconfig=/srv/kubernetes/kubeconfig --config=/opt/kubernetes/kubelet-config.yaml
48Apr 12 01:14:38 node1 kubelet[124730]: I0412 01:14:38.417799 124730 remote_image.go:50] scheme "" not registered, fallback to default scheme
49Apr 12 01:14:38 node1 kubelet[124730]: I0412 01:14:38.417808 124730 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{/var/run/dockershim.sock 0 <nil>}] <nil>}
50Apr 12 01:14:38 node1 kubelet[124730]: I0412 01:14:38.417814 124730 clientconn.go:577] ClientConn switching balancer to "pick_first"
51Apr 12 01:14:38 node1 kubelet[124730]: I0412 01:14:38.418010 124730 balancer_conn_wrappers.go:127] pickfirstBalancer: HandleSubConnStateChange: 0xc0006cb210, CONNECTING
52Apr 12 01:14:38 node1 kubelet[124730]: I0412 01:14:38.418060 124730 balancer_conn_wrappers.go:127] pickfirstBalancer: HandleSubConnStateChange: 0xc000f9c130, CONNECTING
53Apr 12 01:14:38 node1 kubelet[124730]: I0412 01:14:38.418579 124730 balancer_conn_wrappers.go:127] pickfirstBalancer: HandleSubConnStateChange: 0xc000f9c130, READY
54Apr 12 01:14:38 node1 kubelet[124730]: I0412 01:14:38.418605 124730 balancer_conn_wrappers.go:127] pickfirstBalancer: HandleSubConnStateChange: 0xc0006cb210, READY
55Apr 12 01:14:39 node1 kubelet[124730]: E0412 01:14:39.279058 124730 reflector.go:123] k8s.io/kubernetes/pkg/kubelet/kubelet.go:453: Failed to list *v1.Service: Get https://127.0.0.1:6443/api/v1/services?limit=500&resourceVersion=0: EOF
56Apr 12 01:14:39 node1 kubelet[124730]: E0412 01:14:39.280334 124730 reflector.go:123] k8s.io/kubernetes/pkg/kubelet/kubelet.go:462: Failed to list *v1.Node: Get https://127.0.0.1:6443/api/v1/nodes?fieldSelector=metadata.name%3Dnode1&limit=500&resourceVersion=0: EOF
57Apr 12 01:14:39 node1 kubelet[124730]: E0412 01:14:39.281175 124730 reflector.go:123] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:46: Failed to list *v1.Pod: Get https://127.0.0.1:6443/api/v1/pods?fieldSelector=spec.nodeName%3Dnode1&limit=500&resourceVersion=0: EOF
582023-04-12T01:14:40.033 [Agent] Execute command: systemctl daemon-reload
59systemctl enable haproxy
60systemctl restart haproxy
61sleep 2
62systemctl status haproxy
632023-04-12T01:14:42.250 [Agent] command output:
64● haproxy.service - HAProxy Load Balancer
65 Loaded: loaded (/usr/lib/systemd/system/haproxy.service; enabled; vendor preset: disabled)
66 Active: active (running) since Wed 2023-04-12 01:14:40 CST; 2s ago
67 Main PID: 124922 (haproxy-systemd)
68 Tasks: 3
69 CGroup: /system.slice/haproxy.service
70 ├─124922 /usr/sbin/haproxy-systemd-wrapper -f /etc/tos/conf/haproxy.cfg -p /run/haproxy.pid
71 ├─124924 /usr/sbin/haproxy -f /etc/tos/conf/haproxy.cfg -p /run/haproxy.pid -Ds
72 └─124925 /usr/sbin/haproxy -f /etc/tos/conf/haproxy.cfg -p /run/haproxy.pid -Ds
73Apr 12 01:14:40 node1 systemd[1]: Started HAProxy Load Balancer.
74Apr 12 01:14:40 node1 haproxy-systemd-wrapper[124922]: haproxy-systemd-wrapper: executing /usr/sbin/haproxy -f /etc/tos/conf/haproxy.cfg -p /run/haproxy.pid -Ds
752023-04-12T01:14:42.251 [Agent] Execute command: umask 0022 && mkdir -p $(dirname "/opt/kubernetes/manifests-multi/tos-etcd.manifest") && mv -f "/var/lib/transwarp-manager/agent/resource-tmp/@opt@kubernetes@manifests-multi@tos-etcd.manifest" "/opt/kubernetes/manifests-multi/tos-etcd.manifest" && chmod 755 "/opt/kubernetes/manifests-multi/tos-etcd.manifest"
762023-04-12T01:14:42.262 [Agent] Execute command: umask 0022 && mkdir -p $(dirname "/opt/kubernetes/manifests-multi/tos-apiserver.manifest") && mv -f "/var/lib/transwarp-manager/agent/resource-tmp/@opt@kubernetes@manifests-multi@tos-apiserver.manifest" "/opt/kubernetes/manifests-multi/tos-apiserver.manifest" && chmod 755 "/opt/kubernetes/manifests-multi/tos-apiserver.manifest"
772023-04-12T01:14:42.271 [Agent] Execute command: umask 0022 && mkdir -p $(dirname "/opt/kubernetes/manifests-multi/tos-controller.manifest") && mv -f "/var/lib/transwarp-manager/agent/resource-tmp/@opt@kubernetes@manifests-multi@tos-controller.manifest" "/opt/kubernetes/manifests-multi/tos-controller.manifest" && chmod 755 "/opt/kubernetes/manifests-multi/tos-controller.manifest"
782023-04-12T01:14:42.279 [Agent] Execute command: umask 0022 && mkdir -p $(dirname "/opt/kubernetes/manifests-multi/tos-scheduler.manifest") && mv -f "/var/lib/transwarp-manager/agent/resource-tmp/@opt@kubernetes@manifests-multi@tos-scheduler.manifest" "/opt/kubernetes/manifests-multi/tos-scheduler.manifest" && chmod 755 "/opt/kubernetes/manifests-multi/tos-scheduler.manifest"
792023-04-12T01:14:42.287 [Master] Waiting TOS Master (TOS,node1) to become Healthy within 600 s
802023-04-12T01:16:37.706 [Master] Task 171 timed out after 120000ms.
812023-04-12T01:16:37.708 [Master] The Task 171 run failed: java.util.concurrent.CancellationException
82 at java.util.concurrent.FutureTask.report(FutureTask.java:121)
83 at java.util.concurrent.FutureTask.get(FutureTask.java:192)
84 at org.springframework.util.concurrent.ListenableFutureTask.done(ListenableFutureTask.java:83)
85 at java.util.concurrent.FutureTask.finishCompletion(FutureTask.java:384)
86 at java.util.concurrent.FutureTask.cancel(FutureTask.java:180)
87 at io.transwarp.manager.master.operation.execution.TaskDriver.lambda$submitTask$0(TaskDriver.java:75)
88 at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
89 at java.util.concurrent.FutureTask.run(FutureTask.java:266)
90 at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
91 at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
92 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
93 at