leons9th

重启下ks-installer

kubectl rollout restart deploy -n kubesphere-system ks-installer

然后再关注下ks-installer的日志

    Cauchy
    重启后出现了新的错误,连接丢失,不确定具体是什么地址丢失了

    Start installing monitoring
    Start installing multicluster
    Start installing openpitrix
    Start installing network
    **************************************************
    Waiting for all tasks to be completed ...
    task multicluster status is successful  (1/4)
    task openpitrix status is successful  (2/4)
    task network status is successful  (3/4)
    
    error: http2: client connection lost

    重新查看日志,发现 monitoring 启动失败

    Start installing monitoring
    Start installing multicluster
    Start installing openpitrix
    Start installing network
    **************************************************
    Waiting for all tasks to be completed ...
    task multicluster status is successful  (1/4)
    task openpitrix status is successful  (2/4)
    task network status is successful  (3/4)
    task monitoring status is failed  (4/4)
    **************************************************
    Collecting installation results ...
    
    
    Task 'monitoring' failed:
    ******************************************************************************************************************************************************
    {
      "counter": 75,
      "created": "2021-05-10T09:29:39.402331",
      "end_line": 74,
      "event": "runner_on_failed",
      "event_data": {
        "duration": 117.614838,
        "end": "2021-05-10T09:29:39.402089",
        "event_loop": null,
        "host": "localhost",
        "ignore_errors": null,
        "play": "localhost",
        "play_pattern": "localhost",
        "play_uuid": "42453aca-4651-eb6e-4fd2-000000000005",
        "playbook": "/kubesphere/playbooks/monitoring.yaml",
        "playbook_uuid": "661aa835-1511-4178-ba20-17547f31be9b",
        "remote_addr": "127.0.0.1",
        "res": {
          "_ansible_no_log": false,
          "changed": true,
          "cmd": "/usr/local/bin/kubectl apply -f /kubesphere/kubesphere/prometheus/node-exporter --force",
          "delta": "0:01:57.271006",
          "end": "2021-05-10 05:29:39.379806",
          "invocation": {
            "module_args": {
              "_raw_params": "/usr/local/bin/kubectl apply -f /kubesphere/kubesphere/prometheus/node-exporter --force",
              "_uses_shell": true,
              "argv": null,
              "chdir": null,
              "creates": null,
              "executable": null,
              "removes": null,
              "stdin": null,
              "stdin_add_newline": true,
              "strip_empty_ends": true,
              "warn": true
            }
          },
          "msg": "non-zero return code",
          "rc": 1,
          "start": "2021-05-10 05:27:42.108800",
          "stderr": "error when retrieving current configuration of:\nResource: \"rbac.authorization.k8s.io/v1, Resource=clusterroles\", GroupVersionKind: \"rbac.authorization.k8s.io/v1, Kind=ClusterRole\"\nName: \"kubesphere-node-exporter\", Namespace: \"\"\nfrom server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-clusterRole.yaml\": Get https://10.233.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/kubesphere-node-exporter: http2: client connection lost\nerror when retrieving current configuration of:\nResource: \"rbac.authorization.k8s.io/v1, Resource=clusterrolebindings\", GroupVersionKind: \"rbac.authorization.k8s.io/v1, Kind=ClusterRoleBinding\"\nName: \"kubesphere-node-exporter\", Namespace: \"\"\nfrom server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-clusterRoleBinding.yaml\": Get https://10.233.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/kubesphere-node-exporter: net/http: TLS handshake timeout\nerror when retrieving current configuration of:\nResource: \"apps/v1, Resource=daemonsets\", GroupVersionKind: \"apps/v1, Kind=DaemonSet\"\nName: \"node-exporter\", Namespace: \"kubesphere-monitoring-system\"\nfrom server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-daemonset.yaml\": Get https://10.233.0.1:443/apis/apps/v1/namespaces/kubesphere-monitoring-system/daemonsets/node-exporter: net/http: TLS handshake timeout\nerror when retrieving current configuration of:\nResource: \"/v1, Resource=services\", GroupVersionKind: \"/v1, Kind=Service\"\nName: \"node-exporter\", Namespace: \"kubesphere-monitoring-system\"\nfrom server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-service.yaml\": Get https://10.233.0.1:443/api/v1/namespaces/kubesphere-monitoring-system/services/node-exporter: net/http: TLS handshake timeout\nerror when retrieving current configuration of:\nResource: \"/v1, Resource=serviceaccounts\", GroupVersionKind: \"/v1, Kind=ServiceAccount\"\nName: \"node-exporter\", Namespace: \"kubesphere-monitoring-system\"\nfrom server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-serviceAccount.yaml\": Get https://10.233.0.1:443/api/v1/namespaces/kubesphere-monitoring-system/serviceaccounts/node-exporter: net/http: TLS handshake timeout\nerror when retrieving current configuration of:\nResource: \"monitoring.coreos.com/v1, Resource=servicemonitors\", GroupVersionKind: \"monitoring.coreos.com/v1, Kind=ServiceMonitor\"\nName: \"node-exporter\", Namespace: \"kubesphere-monitoring-system\"\nfrom server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-serviceMonitor.yaml\": Get https://10.233.0.1:443/apis/monitoring.coreos.com/v1/namespaces/kubesphere-monitoring-system/servicemonitors/node-exporter: dial tcp 10.233.0.1:443: i/o timeout",
          "stderr_lines": [
            "error when retrieving current configuration of:",
            "Resource: \"rbac.authorization.k8s.io/v1, Resource=clusterroles\", GroupVersionKind: \"rbac.authorization.k8s.io/v1, Kind=ClusterRole\"",
            "Name: \"kubesphere-node-exporter\", Namespace: \"\"",
            "from server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-clusterRole.yaml\": Get https://10.233.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/kubesphere-node-exporter: http2: client connection lost",
            "error when retrieving current configuration of:",
            "Resource: \"rbac.authorization.k8s.io/v1, Resource=clusterrolebindings\", GroupVersionKind: \"rbac.authorization.k8s.io/v1, Kind=ClusterRoleBinding\"",
            "Name: \"kubesphere-node-exporter\", Namespace: \"\"",
            "from server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-clusterRoleBinding.yaml\": Get https://10.233.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/kubesphere-node-exporter: net/http: TLS handshake timeout",
            "error when retrieving current configuration of:",
            "Resource: \"apps/v1, Resource=daemonsets\", GroupVersionKind: \"apps/v1, Kind=DaemonSet\"",
            "Name: \"node-exporter\", Namespace: \"kubesphere-monitoring-system\"",
            "from server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-daemonset.yaml\": Get https://10.233.0.1:443/apis/apps/v1/namespaces/kubesphere-monitoring-system/daemonsets/node-exporter: net/http: TLS handshake timeout",
            "error when retrieving current configuration of:",
            "Resource: \"/v1, Resource=services\", GroupVersionKind: \"/v1, Kind=Service\"",
            "Name: \"node-exporter\", Namespace: \"kubesphere-monitoring-system\"",
            "from server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-service.yaml\": Get https://10.233.0.1:443/api/v1/namespaces/kubesphere-monitoring-system/services/node-exporter: net/http: TLS handshake timeout",
            "error when retrieving current configuration of:",
            "Resource: \"/v1, Resource=serviceaccounts\", GroupVersionKind: \"/v1, Kind=ServiceAccount\"",
            "Name: \"node-exporter\", Namespace: \"kubesphere-monitoring-system\"",
            "from server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-serviceAccount.yaml\": Get https://10.233.0.1:443/api/v1/namespaces/kubesphere-monitoring-system/serviceaccounts/node-exporter: net/http: TLS handshake timeout",
            "error when retrieving current configuration of:",
            "Resource: \"monitoring.coreos.com/v1, Resource=servicemonitors\", GroupVersionKind: \"monitoring.coreos.com/v1, Kind=ServiceMonitor\"",
            "Name: \"node-exporter\", Namespace: \"kubesphere-monitoring-system\"",
            "from server for: \"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-serviceMonitor.yaml\": Get https://10.233.0.1:443/apis/monitoring.coreos.com/v1/namespaces/kubesphere-monitoring-system/servicemonitors/node-exporter: dial tcp 10.233.0.1:443: i/o timeout"
          ],
          "stdout": "",
          "stdout_lines": []
        },
        "role": "ks-monitor",
        "start": "2021-05-10T09:27:41.787251",
        "task": "Monitoring | Installing node-exporter",
        "task_action": "command",
        "task_args": "",
        "task_path": "/kubesphere/installer/roles/ks-monitor/tasks/node-exporter.yaml:2",
        "task_uuid": "42453aca-4651-eb6e-4fd2-000000000033",
        "uuid": "7213fbbc-9a99-4995-8c76-c7d45a908a11"
      },
      "parent_uuid": "42453aca-4651-eb6e-4fd2-000000000033",
      "pid": 3959,
      "runner_ident": "monitoring",
      "start_line": 73,
      "stdout": "fatal: [localhost]: FAILED! => {\"changed\": true, \"cmd\": \"/usr/local/bin/kubectl apply -f /kubesphere/kubesphere/prometheus/node-exporter --force\", \"delta\": \"0:01:57.271006\", \"end\": \"2021-05-10 05:29:39.379806\", \"msg\": \"non-zero return code\", \"rc\": 1, \"start\": \"2021-05-10 05:27:42.108800\", \"stderr\": \"error when retrieving current configuration of:\\nResource: \\\"rbac.authorization.k8s.io/v1, Resource=clusterroles\\\", GroupVersionKind: \\\"rbac.authorization.k8s.io/v1, Kind=ClusterRole\\\"\\nName: \\\"kubesphere-node-exporter\\\", Namespace: \\\"\\\"\\nfrom server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-clusterRole.yaml\\\": Get https://10.233.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/kubesphere-node-exporter: http2: client connection lost\\nerror when retrieving current configuration of:\\nResource: \\\"rbac.authorization.k8s.io/v1, Resource=clusterrolebindings\\\", GroupVersionKind: \\\"rbac.authorization.k8s.io/v1, Kind=ClusterRoleBinding\\\"\\nName: \\\"kubesphere-node-exporter\\\", Namespace: \\\"\\\"\\nfrom server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-clusterRoleBinding.yaml\\\": Get https://10.233.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/kubesphere-node-exporter: net/http: TLS handshake timeout\\nerror when retrieving current configuration of:\\nResource: \\\"apps/v1, Resource=daemonsets\\\", GroupVersionKind: \\\"apps/v1, Kind=DaemonSet\\\"\\nName: \\\"node-exporter\\\", Namespace: \\\"kubesphere-monitoring-system\\\"\\nfrom server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-daemonset.yaml\\\": Get https://10.233.0.1:443/apis/apps/v1/namespaces/kubesphere-monitoring-system/daemonsets/node-exporter: net/http: TLS handshake timeout\\nerror when retrieving current configuration of:\\nResource: \\\"/v1, Resource=services\\\", GroupVersionKind: \\\"/v1, Kind=Service\\\"\\nName: \\\"node-exporter\\\", Namespace: \\\"kubesphere-monitoring-system\\\"\\nfrom server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-service.yaml\\\": Get https://10.233.0.1:443/api/v1/namespaces/kubesphere-monitoring-system/services/node-exporter: net/http: TLS handshake timeout\\nerror when retrieving current configuration of:\\nResource: \\\"/v1, Resource=serviceaccounts\\\", GroupVersionKind: \\\"/v1, Kind=ServiceAccount\\\"\\nName: \\\"node-exporter\\\", Namespace: \\\"kubesphere-monitoring-system\\\"\\nfrom server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-serviceAccount.yaml\\\": Get https://10.233.0.1:443/api/v1/namespaces/kubesphere-monitoring-system/serviceaccounts/node-exporter: net/http: TLS handshake timeout\\nerror when retrieving current configuration of:\\nResource: \\\"monitoring.coreos.com/v1, Resource=servicemonitors\\\", GroupVersionKind: \\\"monitoring.coreos.com/v1, Kind=ServiceMonitor\\\"\\nName: \\\"node-exporter\\\", Namespace: \\\"kubesphere-monitoring-system\\\"\\nfrom server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-serviceMonitor.yaml\\\": Get https://10.233.0.1:443/apis/monitoring.coreos.com/v1/namespaces/kubesphere-monitoring-system/servicemonitors/node-exporter: dial tcp 10.233.0.1:443: i/o timeout\", \"stderr_lines\": [\"error when retrieving current configuration of:\", \"Resource: \\\"rbac.authorization.k8s.io/v1, Resource=clusterroles\\\", GroupVersionKind: \\\"rbac.authorization.k8s.io/v1, Kind=ClusterRole\\\"\", \"Name: \\\"kubesphere-node-exporter\\\", Namespace: \\\"\\\"\", \"from server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-clusterRole.yaml\\\": Get https://10.233.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/kubesphere-node-exporter: http2: client connection lost\", \"error when retrieving current configuration of:\", \"Resource: \\\"rbac.authorization.k8s.io/v1, Resource=clusterrolebindings\\\", GroupVersionKind: \\\"rbac.authorization.k8s.io/v1, Kind=ClusterRoleBinding\\\"\", \"Name: \\\"kubesphere-node-exporter\\\", Namespace: \\\"\\\"\", \"from server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-clusterRoleBinding.yaml\\\": Get https://10.233.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/kubesphere-node-exporter: net/http: TLS handshake timeout\", \"error when retrieving current configuration of:\", \"Resource: \\\"apps/v1, Resource=daemonsets\\\", GroupVersionKind: \\\"apps/v1, Kind=DaemonSet\\\"\", \"Name: \\\"node-exporter\\\", Namespace: \\\"kubesphere-monitoring-system\\\"\", \"from server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-daemonset.yaml\\\": Get https://10.233.0.1:443/apis/apps/v1/namespaces/kubesphere-monitoring-system/daemonsets/node-exporter: net/http: TLS handshake timeout\", \"error when retrieving current configuration of:\", \"Resource: \\\"/v1, Resource=services\\\", GroupVersionKind: \\\"/v1, Kind=Service\\\"\", \"Name: \\\"node-exporter\\\", Namespace: \\\"kubesphere-monitoring-system\\\"\", \"from server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-service.yaml\\\": Get https://10.233.0.1:443/api/v1/namespaces/kubesphere-monitoring-system/services/node-exporter: net/http: TLS handshake timeout\", \"error when retrieving current configuration of:\", \"Resource: \\\"/v1, Resource=serviceaccounts\\\", GroupVersionKind: \\\"/v1, Kind=ServiceAccount\\\"\", \"Name: \\\"node-exporter\\\", Namespace: \\\"kubesphere-monitoring-system\\\"\", \"from server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-serviceAccount.yaml\\\": Get https://10.233.0.1:443/api/v1/namespaces/kubesphere-monitoring-system/serviceaccounts/node-exporter: net/http: TLS handshake timeout\", \"error when retrieving current configuration of:\", \"Resource: \\\"monitoring.coreos.com/v1, Resource=servicemonitors\\\", GroupVersionKind: \\\"monitoring.coreos.com/v1, Kind=ServiceMonitor\\\"\", \"Name: \\\"node-exporter\\\", Namespace: \\\"kubesphere-monitoring-system\\\"\", \"from server for: \\\"/kubesphere/kubesphere/prometheus/node-exporter/node-exporter-serviceMonitor.yaml\\\": Get https://10.233.0.1:443/apis/monitoring.coreos.com/v1/namespaces/kubesphere-monitoring-system/servicemonitors/node-exporter: dial tcp 10.233.0.1:443: i/o timeout\"], \"stdout\": \"\", \"stdout_lines\": []}",
      "uuid": "7213fbbc-9a99-4995-8c76-c7d45a908a11"
    }
    ******************************************************************************************************************************************************
    Failed to ansible-playbook ks-config.yaml
    E0510 05:36:26.988333       1 reflector.go:284] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to watch *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&resourceVersion=48786&timeoutSeconds=499&watch=true": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:27.990356       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:28.991146       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:29.991965       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:30.992636       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:31.993368       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:32.994235       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:33.998121       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:34.998777       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:35.999464       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:37.000144       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:38.000891       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:39.001655       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:40.002446       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:41.003137       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused
    E0510 05:36:42.005667       1 reflector.go:131] pkg/mod/k8s.io/client-go@v0.0.0-20190411052641-7a6b4715b709/tools/cache/reflector.go:99: Failed to list *unstructured.Unstructured: Get "https://10.233.0.1:443/apis/installer.kubesphere.io/v1alpha1/namespaces/kubesphere-system/clusterconfigurations?fieldSelector=metadata.name%3Dks-installer&limit=500&resourceVersion=0": dial tcp 10.233.0.1:443: connect: connection refused

      Cauchy
      现在已经安装成功,提示我访问 30880 的 web 端

      但是访问不了,连接超时是为什么呢

      用 telnet 能够连接的上,服务器资源也还够 CPU 使用了 6%,内存只用了 2G

      我服务器有 16G 内存,6核 CPU,120G 硬盘

      防火墙已经全部关闭 telnet 可以访问 curl 访问不了

      pod 信息

      NAME                                     READY   STATUS    RESTARTS   AGE
      ks-apiserver-7d4f466f75-5wc5v            1/1     Running   0          29m
      ks-console-5576fccbb8-4mqpt              1/1     Running   0          20h
      ks-controller-manager-7d9b98f467-r2km6   1/1     Running   0          29m
      ks-installer-8d7b5cfc4-lkffh             1/1     Running   0          32m
      NAMESPACE                      NAME                                               READY   STATUS      RESTARTS   AGE   IP                NODE      NOMINATED NODE   READINESS GATES
      kube-system                    calico-kube-controllers-8545b68dd4-8bzjr           1/1     Running     0          22h   10.233.90.1       node1     <none>           <none>
      kube-system                    calico-node-bhf2k                                  1/1     Running     0          22h   192.168.113.121   node1     <none>           <none>
      kube-system                    calico-node-k2r5k                                  1/1     Running     1          22h   192.168.113.120   master1   <none>           <none>
      kube-system                    calico-node-t7dbs                                  1/1     Running     0          22h   192.168.113.122   node2     <none>           <none>
      kube-system                    coredns-867b49865c-krk2l                           1/1     Running     1          22h   10.233.97.5       master1   <none>           <none>
      kube-system                    coredns-867b49865c-txt7n                           1/1     Running     1          22h   10.233.97.4       master1   <none>           <none>
      kube-system                    init-pvc-1dcd1a64-38c6-4250-8043-1f91ccc025a5      0/1     Completed   0          17h   10.233.96.4       node2     <none>           <none>
      kube-system                    init-pvc-c6e7bfc1-5485-4a59-89d8-691015562d39      0/1     Completed   0          17h   10.233.90.8       node1     <none>           <none>
      kube-system                    kube-apiserver-master1                             1/1     Running     2          22h   192.168.113.120   master1   <none>           <none>
      kube-system                    kube-controller-manager-master1                    1/1     Running     6          22h   192.168.113.120   master1   <none>           <none>
      kube-system                    kube-proxy-dj8nx                                   1/1     Running     1          22h   192.168.113.120   master1   <none>           <none>
      kube-system                    kube-proxy-mqmgw                                   1/1     Running     0          22h   192.168.113.121   node1     <none>           <none>
      kube-system                    kube-proxy-qhxwt                                   1/1     Running     0          22h   192.168.113.122   node2     <none>           <none>
      kube-system                    kube-scheduler-master1                             1/1     Running     5          22h   192.168.113.120   master1   <none>           <none>
      kube-system                    nodelocaldns-bkdlp                                 1/1     Running     0          22h   192.168.113.122   node2     <none>           <none>
      kube-system                    nodelocaldns-rsdn6                                 1/1     Running     0          22h   192.168.113.121   node1     <none>           <none>
      kube-system                    nodelocaldns-xkl8q                                 1/1     Running     1          22h   192.168.113.120   master1   <none>           <none>
      kube-system                    openebs-localpv-provisioner-c46f4fbd5-ntkw8        1/1     Running     57         20h   10.233.96.1       node2     <none>           <none>
      kube-system                    snapshot-controller-0                              1/1     Running     0          20h   10.233.96.2       node2     <none>           <none>
      kubesphere-controls-system     default-http-backend-76d9fb4bb7-pkkl7              1/1     Running     0          20h   10.233.90.4       node1     <none>           <none>
      kubesphere-controls-system     kubectl-admin-7b69cb97d5-6brvg                     1/1     Running     0          13h   10.233.90.12      node1     <none>           <none>
      kubesphere-monitoring-system   alertmanager-main-0                                2/2     Running     0          17h   10.233.90.9       node1     <none>           <none>
      kubesphere-monitoring-system   alertmanager-main-1                                2/2     Running     0          17h   10.233.96.5       node2     <none>           <none>
      kubesphere-monitoring-system   alertmanager-main-2                                2/2     Running     0          17h   10.233.96.6       node2     <none>           <none>
      kubesphere-monitoring-system   kube-state-metrics-687c7c4d86-nb4d4                3/3     Running     2          17h   10.233.90.7       node1     <none>           <none>
      kubesphere-monitoring-system   node-exporter-525tt                                2/2     Running     0          17h   192.168.113.121   node1     <none>           <none>
      kubesphere-monitoring-system   node-exporter-7dqgb                                2/2     Running     0          17h   192.168.113.122   node2     <none>           <none>
      kubesphere-monitoring-system   node-exporter-vgrhq                                2/2     Running     0          17h   192.168.113.120   master1   <none>           <none>
      kubesphere-monitoring-system   notification-manager-deployment-7bd887ffb4-pwjg8   1/1     Running     0          13h   10.233.90.13      node1     <none>           <none>
      kubesphere-monitoring-system   notification-manager-deployment-7bd887ffb4-vxhrp   1/1     Running     0          13h   10.233.96.8       node2     <none>           <none>
      kubesphere-monitoring-system   notification-manager-operator-78595d8666-d9vfn     2/2     Running     1          13h   10.233.90.11      node1     <none>           <none>
      kubesphere-monitoring-system   prometheus-k8s-0                                   0/3     Pending     0          17h   <none>            <none>    <none>           <none>
      kubesphere-monitoring-system   prometheus-k8s-1                                   0/3     Pending     0          17h   <none>            <none>    <none>           <none>
      kubesphere-monitoring-system   prometheus-operator-d7fdfccbf-7p2pv                2/2     Running     0          17h   10.233.96.3       node2     <none>           <none>
      kubesphere-system              ks-apiserver-7d4f466f75-5wc5v                      1/1     Running     0          29m   10.233.97.8       master1   <none>           <none>
      kubesphere-system              ks-console-5576fccbb8-4mqpt                        1/1     Running     0          20h   10.233.90.3       node1     <none>           <none>
      kubesphere-system              ks-controller-manager-7d9b98f467-r2km6             1/1     Running     0          29m   10.233.97.9       master1   <none>           <none>
      kubesphere-system              ks-installer-8d7b5cfc4-lkffh                       1/1     Running     0          33m   10.233.90.15      node1     <none>           <none>
      • warn 回复了此帖

        重装了一次,成功了但还是访问不了,都是 telnet 可以 curl 不可以
        Cauchy @Jeff

        #####################################################
        ###              Welcome to KubeSphere!           ###
        #####################################################
        
        Console: http://192.168.113.120:30880
        Account: admin
        Password: P@88w0rd
        
        NOTES:
          1. After you log into the console, please check the
             monitoring status of service components in
             "Cluster Management". If any service is not
             ready, please wait patiently until all components
             are up and running.
          2. Please change the default password after login.
        
        #####################################################
        https://kubesphere.io             2021-05-11 12:06:41
        #####################################################

        pod 信息

        [root@master1 kubesphere]# kubectl get pods -n kubesphere-system
        NAME                                    READY   STATUS      RESTARTS   AGE
        ks-apiserver-75958c7c85-wqdkr           1/1     Running     0          27m
        ks-console-5576fccbb8-ssgqs             1/1     Running     0          29m
        ks-controller-manager-fb858c67b-gsv29   1/1     Running     3          27m
        ks-installer-7fb74c656c-8887s           1/1     Running     0          34m
        minio-f69748945-j7qhm                   1/1     Running     0          31m
        openldap-0                              1/1     Running     1          31m
        openpitrix-import-job-kc22l             0/1     Completed   0          29m
        redis-658988fc5b-rfnnf                  1/1     Running     0          32m
          leons9th 更改标题为「安装 Kubesphere3.1.0 成功后无法访问 console 页面
          2 个月 后

          leons9th 现在已经安装成功

          怎么解决的?我的也是一直KubeSphere startup timeout

          • warn 回复了此帖

            warn

            `clusterconfiguration.installer.kubesphere.io/ks-installer unchanged
            WARN[17:35:42 HKT] Task failed …
            WARN[17:35:42 HKT] error: KubeSphere startup timeout.
            Error: Failed to deploy kubesphere: KubeSphere startup timeout.
            Usage:
            kk create cluster [flags]

            Flags:
            –download-cmd string The user defined command to download the necessary binary files. The first param ‘%s’ is output path, the second param ‘%s’, is the URL (default “curl -L -o %s %s”)
            -f, –filename string Path to a configuration file
            -h, –help help for cluster
            –skip-pull-images Skip pre pull images
            –with-kubernetes string Specify a supported version of kubernetes (default “v1.19.8”)
            –with-kubesphere Deploy a specific version of kubesphere (default v3.1.0)
            –with-local-storage Deploy a local PV provisioner
            -y, –yes Skip pre-check of the installation

            Global Flags:
            –debug Print detailed information (default true)
            –in-cluster Running inside the cluster

            Failed to deploy kubesphere: KubeSphere startup timeout.`

            9 天 后

            同样的问题 Failed to deploy kubesphere: KubeSphere startup timeout.

            是否被防火墙拦住了,看看防火墙关了没

            1 年 后
            1 个月 后
            5 个月 后

            我也是呀,这个问题很奇怪,但如果重装启动,就报其他错误