Jul 31 22:46:33 infra0 kubelet-wrapper[3175]: I0801 02:46:33.761626 3175 reconciler.go:253] MountVolume operation started for volume "kubernetes.io/host-path//usr/share
Jul 31 22:46:33 infra0 kubelet-wrapper[3175]: I0801 02:46:33.761754 3175 operation_executor.go:720] MountVolume.SetUp succeeded for volume "kubernetes.io/host-path//usr
Jul 31 22:46:33 infra0 kubelet-wrapper[3175]: I0801 02:46:33.761925 3175 operation_executor.go:720] MountVolume.SetUp succeeded for volume "kubernetes.io/host-path//etc
Jul 31 22:46:33 infra0 kubelet-wrapper[3175]: I0801 02:46:33.862395 3175 reconciler.go:253] MountVolume operation started for volume "kubernetes.io/host-path//usr/share
Jul 31 22:46:33 infra0 kubelet-wrapper[3175]: I0801 02:46:33.862631 3175 operation_executor.go:720] MountVolume.SetUp succeeded for volume "kubernetes.io/host-path//usr
Jul 31 22:46:34 infra0 kubelet-wrapper[3175]: E0801 02:46:34.032349 3175 docker_manager.go:345] NetworkPlugin cni failed on the status hook for pod 'kube-scheduler-10.1
Jul 31 22:46:34 infra0 kubelet-wrapper[3175]: with error: exit status 1
Jul 31 22:46:34 infra0 kubelet-wrapper[3175]: W0801 02:46:34.467274 3175 docker_manager.go:1380] No ref for pod '"bd8c9f70dad67a026bcf2a2c98a447730cdaf671daab2716e9947b
Jul 31 22:46:34 infra0 kubelet-wrapper[3175]: W0801 02:46:34.611620 3175 docker_manager.go:1380] No ref for pod '"b3c71a48a1cacd2d914bccfe40593666036b6ef41e678b44866182
Jul 31 22:46:35 infra0 kubelet-wrapper[3175]: E0801 02:46:35.095388 3175 docker_manager.go:345] NetworkPlugin cni failed on the status hook for pod 'kube-scheduler-10.1
Jul 31 22:46:35 infra0 kubelet-wrapper[3175]: with error: exit status 1
Jul 31 22:46:36 infra0 kubelet-wrapper[3175]: W0801 02:46:36.415003 3175 docker_manager.go:1380] No ref for pod '"bd8c9f70dad67a026bcf2a2c98a447730cdaf671daab2716e9947b
Jul 31 22:46:36 infra0 kubelet-wrapper[3175]: W0801 02:46:36.418328 3175 docker_manager.go:1380] No ref for pod '"b3c71a48a1cacd2d914bccfe40593666036b6ef41e678b44866182
Jul 31 23:33:08 infra0 kubelet-wrapper[3175]: E0801 03:33:08.855740 3175 kubelet.go:1899] Deleting mirror pod "kube-scheduler-10.1.10.110_kube-system(2058ac4a-5792-11e6
Jul 31 23:33:08 infra0 kubelet-wrapper[3175]: W0801 03:33:08.892731 3175 status_manager.go:447] Failed to update status for pod "_()": Operation cannot be fulfilled on
Jul 31 23:33:08 infra0 kubelet-wrapper[3175]: I0801 03:33:08.915472 3175 docker_manager.go:1834] pod "kube-scheduler-10.1.10.110_kube-system(0be27240583c9b189c7a18173fc
Jul 31 23:33:09 infra0 kubelet-wrapper[3175]: E0801 03:33:09.560590 3175 docker_manager.go:345] NetworkPlugin cni failed on the status hook for pod 'kube-scheduler-10.1
Jul 31 23:33:09 infra0 kubelet-wrapper[3175]: with error: exit status 1ROSELCDV0070481:~ cobbr$ kubectl get pods --namespace=kube-system
NAME READY STATUS RESTARTS AGE
kube-apiserver-10.1.10.110 1/1 Running 1 16h
kube-controller-manager-10.1.10.110 1/1 Running 1 16h
kube-dns-v11-71w7j 0/4 Pending 0 12h
kube-proxy-10.1.10.110 1/1 Running 1 16h
kube-scheduler-10.1.10.110 1/1 Running 1 9h
kubernetes-dashboard-3717423461-0pecu 0/1 Pending 0 12hROSELCDV0070481:~ cobbr$ kubectl get nodes
NAME STATUS AGE
10.1.10.110 Ready,SchedulingDisabled 16hI'm trying to get a Kubernetes cluster set up on 6 CoreOS nodes. Everything seems to work fine when set up using instructions from here: https://coreos.com/kubernetes/docs/latest/deploy-master.html.
However, I have made some progress... I found that I somehow missed the worker-kubeconfig.yaml step and it was getting created as a directory automatically on the worker kubelet launch, but now that I have the proper yaml file in place, the workers are registering with the master as minions, and do, now, show up in the list properly.
I still see the issues deploying the kube-dns and kubernetes-dashboard containers, but I think they may be related to something amiss in my configuration. Dashboard constantly fails with some sort of certificate issue, and the kube-dns container fails with a missing skydns or something setting. I'm still floundering trying to solve those 2 issues at this point.
I also think I have some networking issues to iron out, as well... I deployed the guestbook application after I got the minion issues settled, and I couldn't route to the guestbook app properly, but I can work on those once I get the kubernetes issues resolved.
--
You received this message because you are subscribed to the Google Groups "CoreOS User" group.
To unsubscribe from this group and stop receiving emails from it, send an email to coreos-user...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
On Wed, Aug 3, 2016 at 11:17 AM randal cobb <rco...@gmail.com> wrote:However, I have made some progress... I found that I somehow missed the worker-kubeconfig.yaml step and it was getting created as a directory automatically on the worker kubelet launch, but now that I have the proper yaml file in place, the workers are registering with the master as minions, and do, now, show up in the list properly.Great, good catch!I still see the issues deploying the kube-dns and kubernetes-dashboard containers, but I think they may be related to something amiss in my configuration. Dashboard constantly fails with some sort of certificate issue, and the kube-dns container fails with a missing skydns or something setting. I'm still floundering trying to solve those 2 issues at this point.Can you paste in the logs for dashboard and kube-dns logs so we can try and figure it out?
ROSELCDV0070481:.kubernetes cobbr$ kubectl get pods --namespace=kube-system
NAME READY STATUS RESTARTS AGE
kube-apiserver-10.1.10.110 1/1 Running 0 17h
kube-controller-manager-10.1.10.110 1/1 Running 9 1d
kube-dns-v11-6qoh5 3/4 Running 2 3m
kube-proxy-10.1.10.110 1/1 Running 7 1d
kube-proxy-10.1.10.111 1/1 Running 6 19h
kube-proxy-10.1.10.112 1/1 Running 26 1d
kube-proxy-10.1.10.113 1/1 Running 1 15h
kube-proxy-10.1.10.114 1/1 Running 289 1d
kube-proxy-10.1.10.115 1/1 Running 289 1d
kube-scheduler-10.1.10.110 1/1 Running 9 1d ROSELCDV0070481:.kubernetes cobbr$ kubectl logs kube-dns-v11-6qoh5 --namespace=kube-system
Error from server: a container name must be specified for pod kube-dns-v11-6qoh5, choose one of: [etcd kube2sky skydns healthz]
ROSELCDV0070481:.kubernetes cobbr$ kubectl create -f kubernetes-dashboard.yaml
deployment "kubernetes-dashboard" created
You have exposed your service on an external port on all nodes in your
cluster. If you want to expose this service to the external internet, you may
need to set up firewall rules for the service port(s) (tcp:32560) to serve traffic.
See http://releases.k8s.io/release-1.3/docs/user-guide/services-firewalls.md for more details.
service "kubernetes-dashboard" created
ROSELCDV0070481:.kubernetes cobbr$ kubectl get pods --namespace=kube-system
NAME READY STATUS RESTARTS AGE
kube-apiserver-10.1.10.110 1/1 Running 0 18h
kube-controller-manager-10.1.10.110 1/1 Running 9 1d
kube-dns-v11-6qoh5 3/4 CrashLoopBackOff 7 15m
kube-proxy-10.1.10.110 1/1 Running 7 1d
kube-proxy-10.1.10.111 1/1 Running 6 19h
kube-proxy-10.1.10.112 1/1 Running 26 1d
kube-proxy-10.1.10.113 1/1 Running 1 16h
kube-proxy-10.1.10.114 1/1 Running 289 1d
kube-proxy-10.1.10.115 1/1 Running 289 1d
kube-scheduler-10.1.10.110 1/1 Running 9 1d
kubernetes-dashboard-3954469829-nyvie 1/1 Running 1 10s
ROSELCDV0070481:.kubernetes cobbr$ kubectl get pods --namespace=kube-system
NAME READY STATUS RESTARTS AGE
kube-apiserver-10.1.10.110 1/1 Running 0 18h
kube-controller-manager-10.1.10.110 1/1 Running 9 1d
kube-dns-v11-6qoh5 3/4 CrashLoopBackOff 7 15m
kube-proxy-10.1.10.110 1/1 Running 7 1d
kube-proxy-10.1.10.111 1/1 Running 6 19h
kube-proxy-10.1.10.112 1/1 Running 26 1d
kube-proxy-10.1.10.113 1/1 Running 1 16h
kube-proxy-10.1.10.114 1/1 Running 289 1d
kube-proxy-10.1.10.115 1/1 Running 289 1d
kube-scheduler-10.1.10.110 1/1 Running 9 1d
kubernetes-dashboard-3954469829-nyvie 0/1 CrashLoopBackOff 1 20sROSELCDV0070481:.kubernetes cobbr$ kubectl logs kubernetes-dashboard-3954469829-nyvie --namespace=kube-system
Starting HTTP server on port 9090
Creating API server client for http://10.1.12.1:8080
Error while initializing connection to Kubernetes apiserver. This most likely means that the cluster is misconfigured (e.g., it has invalid apiserver certificates or service accounts configuration) or the --apiserver-host param points to a server that does not exist. Reason: Get http://10.1.12.1:8080/version: dial tcp 10.1.12.1:8080: getsockopt: no route to hostcore@kube-worke-5 ~ $ curl http://10.1.12.1:8080
curl: (7) Failed to connect to 10.1.12.1 port 8080: No route to host
core@kubernetes ~ $ curl http://127.0.0.1:8080
{
"paths": [
"/api",
"/api/v1",
"/apis",
"/apis/apps",
"/apis/apps/v1alpha1",
"/apis/autoscaling",
"/apis/autoscaling/v1",
"/apis/batch",
"/apis/batch/v1",
"/apis/batch/v2alpha1",
"/apis/extensions",
"/apis/extensions/v1beta1",
"/apis/policy",
"/apis/policy/v1alpha1",
"/apis/rbac.authorization.k8s.io",
"/apis/rbac.authorization.k8s.io/v1alpha1",
"/healthz",
"/healthz/ping",
"/logs/",
"/metrics",
"/swaggerapi/",
"/ui/",
"/version"
]
}core@kubernetes ~ $ curl http://10.1.10.110:8080
curl: (7) Failed to connect to 10.1.10.110 port 8080: Connection refused
core@kubernetes ~ $Brandon
<blockquote class="gm
kubernetes ~ # route -n
Kernel IP routing table
Destination Gateway Genmask Flags Metric Ref Use Iface
0.0.0.0 10.1.10.1 0.0.0.0 UG 0 0 0 ens18
10.0.0.0 0.0.0.0 255.252.0.0 U 0 0 0 ens18
10.2.0.0 0.0.0.0 255.255.0.0 U 0 0 0 flannel.1
10.2.58.0 0.0.0.0 255.255.255.0 U 0 0 0 docker0Those logs finally led me to the fact that the kube2sky container wasn't able to properly authenticate to the kube-apiserver. More Googling led me to an identical situation that was listed as a "bug" in an earlier version of Kubernetes (1.1.0, I believe) where the generated authtoken was being cached by etcd and the old cached token was being used instead of the current token. So, I basically shut down the entire cluster, purges all entries from etcd, re-created namespaces (because I stoopidly purged them, too), restarted the cluster, and re-deployed the dns-addon. Viola! It came up first try! So, I rolled the dice and redeployed the dashboard again, and BOOM! It came up, 1st try, too!
I think I now have a stable environment to play with!
I'll let you know how it goes.