-
Notifications
You must be signed in to change notification settings - Fork 88
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issue in deployment #11
Comments
While enabling kube-proxy its throwing the error |
Out put for -----journalctl -u kubelet : Apr 20 01:10:09 k8s-kubespray-master-0 kubelet[16642]: E0420 01:10:09.535127 16642 kubelet.go:2236] node "k8s-kubespray-master-0" not found |
Hi, |
Only etcd is running CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES |
Was the kube-apiserver running at one point? (docker ps -a) |
Nope its not running at all and showing no entries. |
Any other steps to find the root cause for the issue |
Which Linux distribution are you using? On your k8s-kubespray-master-0: |
I think so its running properly. Below is the output : /usr/local/bin/etcdctl member list346d63ce2a2b4f8e, started, etcd2, https://172.16.10.244:2380, https://172.16.10.244:2379 |
Hey, I am having the exact same issue.
root@k8s-kubespray-master-0:~# docker ps -a CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 8d0cb68b5f7f quay.io/coreos/etcd:v3.2.24 "/usr/local/bin/etcd" About an hour ago Up About an hour etcd1
root@k8s-kubespray-master-0:~# /usr/local/bin/etcdctl member list 65b95689f280cf, started, etcd1, https://192.168.0.71:2380, https://192.168.0.71:2379 Unfortunately I had to perform some changes in diff --git a/vsphere-kubespray.tf b/vsphere-kubespray.tf index cbcaadd..c1ab2be 100644 --- a/vsphere-kubespray.tf +++ b/vsphere-kubespray.tf @@ -19,7 +19,7 @@ data "vsphere_datacenter" "dc" { name = "${var.vsphere_datacenter}" } -data "vsphere_compute_cluster" "cluster" { +data "vsphere_host" "cluster" { name = "${var.vsphere_drs_cluster}" datacenter_id = "${data.vsphere_datacenter.dc.id}" } @@ -367,7 +367,7 @@ resource "vsphere_folder" "folder" { # Create a resource pool for the Kubernetes VMs # resource "vsphere_resource_pool" "resource_pool" { name = "${var.vsphere_resource_pool}" - parent_resource_pool_id = "${data.vsphere_compute_cluster.cluster.resource_pool_id}" + parent_resource_pool_id = "${data.vsphere_host.cluster.resource_pool_id}" } # Create the Kubernetes master VMs # @@ -420,16 +420,6 @@ resource "vsphere_virtual_machine" "master" { depends_on = ["vsphere_virtual_machine.haproxy"] } -# Create anti affinity rule for the Kubernetes master VMs # -resource "vsphere_compute_cluster_vm_anti_affinity_rule" "master_anti_affinity_rule" { - count = "${var.vsphere_enable_anti_affinity == "true" ? 1 : 0}" - name = "${var.vm_name_prefix}-master-anti-affinity-rule" - compute_cluster_id = "${data.vsphere_compute_cluster.cluster.id}" - virtual_machine_ids = ["${vsphere_virtual_machine.master.*.id}"] - - depends_on = ["vsphere_virtual_machine.master"] -} - # Create the Kubernetes worker VMs # resource "vsphere_virtual_machine" "worker" { count = "${length(var.vm_worker_ips)}" |
Please find the log below :
TASK [kubernetes/master : kubeadm | Initialize first master] ************************************************************
Friday 19 April 2019 23:41:47 +0100 (0:00:03.591) 0:08:13.711 **********
TASK [kubernetes/master : kubeadm | Upgrade first master] ***************************************************************
Friday 19 April 2019 23:41:47 +0100 (0:00:00.229) 0:08:13.940 **********
TASK [kubernetes/master : kubeadm | Enable kube-proxy] ******************************************************************
Friday 19 April 2019 23:41:47 +0100 (0:00:00.229) 0:08:14.170 **********
FAILED - RETRYING: kubeadm | Enable kube-proxy (10 retries left).
FAILED - RETRYING: kubeadm | Enable kube-proxy (9 retries left).
FAILED - RETRYING: kubeadm | Enable kube-proxy (8 retries left).
FAILED - RETRYING: kubeadm | Enable kube-proxy (7 retries left).
FAILED - RETRYING: kubeadm | Enable kube-proxy (6 retries left).
FAILED - RETRYING: kubeadm | Enable kube-proxy (5 retries left).
FAILED - RETRYING: kubeadm | Enable kube-proxy (4 retries left).
FAILED - RETRYING: kubeadm | Enable kube-proxy (3 retries left).
FAILED - RETRYING: kubeadm | Enable kube-proxy (2 retries left).
FAILED - RETRYING: kubeadm | Enable kube-proxy (1 retries left).
fatal: [k8s-kubespray-master-0]: FAILED! => {"attempts": 10, "changed": false, "cmd": ["/usr/local/bin/kubeadm", "alpha", "phase", "addon", "kube-proxy", "--config=/etc/kubernetes/kubeadm-config.v1alpha3.yaml"], "delta": "0:00:00.053745", "end": "2019-04-19 23:42:40.261421", "msg": "non-zero return code", "rc": 1, "start": "2019-04-19 23:42:40.207676", "stderr": "error when creating kube-proxy service account: unable to create serviceaccount: Post https://172.16.10.249:6443/api/v1/namespaces/kube-system/serviceaccounts: EOF", "stderr_lines": ["error when creating kube-proxy service account: unable to create serviceaccount: Post https://172.16.10.249:6443/api/v1/namespaces/kube-system/serviceaccounts: EOF"], "stdout": "", "stdout_lines": []}
NO MORE HOSTS LEFT ******************************************************************************************************
to retry, use: --limit @/home/administrator/terraform-vsphere-kubespray/ansible/kubespray/cluster.retry
PLAY RECAP **************************************************************************************************************
k8s-kubespray-master-0 : ok=275 changed=11 unreachable=0 failed=1
k8s-kubespray-master-1 : ok=252 changed=9 unreachable=0 failed=0
k8s-kubespray-master-2 : ok=252 changed=9 unreachable=0 failed=0
k8s-kubespray-worker-0 : ok=204 changed=4 unreachable=0 failed=0
localhost : ok=1 changed=0 unreachable=0 failed=0
Friday 19 April 2019 23:42:40 +0100 (0:00:52.784) 0:09:06.955 **********
kubernetes/master : kubeadm | Enable kube-proxy ----------------------------------------------------------------- 52.78s
kubernetes/preinstall : Update package management cache (APT) --------------------------------------------------- 10.32s
download : Download items ---------------------------------------------------------------------------------------- 7.00s
gather facts from all instances ---------------------------------------------------------------------------------- 6.49s
download : Sync container ---------------------------------------------------------------------------------------- 5.85s
download : Download items ---------------------------------------------------------------------------------------- 4.81s
container-engine/docker : ensure docker packages are installed --------------------------------------------------- 4.75s
download : Sync container ---------------------------------------------------------------------------------------- 4.65s
download : Download items ---------------------------------------------------------------------------------------- 4.34s
download : Sync container ---------------------------------------------------------------------------------------- 4.34s
download : Sync container ---------------------------------------------------------------------------------------- 4.31s
download : Download items ---------------------------------------------------------------------------------------- 4.21s
kubernetes/node : Enable bridge-nf-call tables ------------------------------------------------------------------- 4.19s
download : Download items ---------------------------------------------------------------------------------------- 4.19s
download : Sync container ---------------------------------------------------------------------------------------- 4.16s
container-engine/docker : Ensure old versions of Docker are not installed. | Debian ------------------------------ 4.11s
kubernetes/preinstall : Hosts | populate inventory into hosts file ----------------------------------------------- 3.67s
kubernetes/master : kubeadm | Create kubeadm config -------------------------------------------------------------- 3.59s
download : Sync container ---------------------------------------------------------------------------------------- 3.48s
kubernetes/preinstall : Install packages requirements ------------------------------------------------------------ 3.42s
administrator@ubuntu:
/terraform-vsphere-kubespray/ansible/kubespray$ ^C/terraform-vsphere-kubespray/ansible/kubespray$ curl -k https://172.16.10.249:6443administrator@ubuntu:
curl: (35) gnutls_handshake() failed: The TLS connection was non-properly terminated.
administrator@ubuntu:
/terraform-vsphere-kubespray/ansible/kubespray$ ^C/terraform-vsphere-kubespray/ansible/kubespray$administrator@ubuntu:
The text was updated successfully, but these errors were encountered: