/cc @kubernetes/sig-api-machinery-bugs
—
You are receiving this because you are on a team that was mentioned.
Reply to this email directly, view it on GitHub, or mute the thread.
I don't think this is necessary a bug - this may be working as intended.
If you have 500 pods, each of the generating 10qps of "get configmap), this gives us 5000qps. Depending on the size of the master machine/number of apiservers and/or etcd instances, this just may be too many requests to handle.
@wojtek-t Not actually the case, I think I need to emphasis it a little bit:
please look at and/or paste here the apiserver logs.
We updated the grpc package whose version is same as what kubernetes 1.7 is used, the hung has gone.
@zhouhaibing089 @keyingliu
grpc/grpc-java#2258
there is one dead lock bug in grpc 1.0.0.
/assign
Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale
.
Stale issues rot after an additional 30d of inactivity and eventually close.
Prevent issues from auto-closing with an /lifecycle frozen
comment.
If this issue is safe to close now please do so with /close
.
Send feedback to sig-testing, kubernetes/test-infra and/or @fejta
.
/lifecycle stale
/remove-lifecycle stale
Below is the goroutine profile based on k8s 1.6.3.
goroutine profile: total 49531
20276 @ 0x42ce8a 0x43c5d5 0x43b36c 0x294b926 0x293a799 0x25b3f05 0x25b4a12 0x25cd6e6 0x25b42fd 0x255c692 0x2233292 0x222c55f 0x222c3b7 0x2224388 0x221d67e 0x221cf5d 0x221ca62 0x220bf92 0xf7d136 0xf8ac74 0x225a357 0x2259f70 0x223abbc 0xfc4bf0 0xfd3048 0x603bc4 0x6050bf 0x102a205 0x603bc4 0x102c10d 0x603bc4 0x1029792
# 0x294b925 k8s.io/kubernetes/vendor/google.golang.org/grpc/transport.wait+0x445 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/google.golang.org/grpc/transport/transport.go:577
# 0x293a798 k8s.io/kubernetes/vendor/google.golang.org/grpc/transport.(*http2Client).NewStream+0x658 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/google.golang.org/grpc/transport/http2_client.go:319
# 0x25b3f04 k8s.io/kubernetes/vendor/google.golang.org/grpc.sendRequest+0x94 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/google.golang.org/grpc/call.go:80
# 0x25b4a11 k8s.io/kubernetes/vendor/google.golang.org/grpc.invoke+0x621 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/google.golang.org/grpc/call.go:191
# 0x25cd6e5 k8s.io/kubernetes/vendor/github.com/grpc-ecosystem/go-grpc-prometheus.(*ClientMetrics).UnaryClientInterceptor.func1+0x125 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/grpc-ecosystem/go-grpc-prometheus/client_metrics.go:84
# 0x25b42fc k8s.io/kubernetes/vendor/google.golang.org/grpc.Invoke+0xdc /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/google.golang.org/grpc/call.go:116
# 0x255c691 k8s.io/kubernetes/vendor/github.com/coreos/etcd/etcdserver/etcdserverpb.(*kVClient).Range+0xd1 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/etcdserver/etcdserverpb/rpc.pb.go:2203
# 0x2233291 k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3.(*retryWriteKVClient).Range+0x91 <autogenerated>:174
# 0x222c55e k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3.(*retryKVClient).Range.func1+0x8e /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3/retry.go:92
# 0x222c3b6 k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3.(*Client).newAuthRetryWrapper.func1+0x46 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3/retry.go:61
# 0x2224387 k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3.(*retryKVClient).Range+0x157 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3/retry.go:94
# 0x221d67d k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3.(*kv).do+0x4ed /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3/kv.go:145
# 0x221cf5c k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3.(*kv).Do+0x7c /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3/kv.go:124
# 0x221ca61 k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3.(*kv).Get+0xe1 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/clientv3/kv.go:98
# 0x220bf91 k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/storage/etcd3.(*store).Get+0x131 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/storage/etcd3/store.go:128
# 0xf7d135 k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/storage.(*Cacher).Get+0xc5 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/storage/cacher.go:360
# 0xf8ac73 k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/registry/generic/registry.(*Store).Get+0x183 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/registry/generic/registry/store.go:517
# 0x225a356 k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/handlers.GetResource.func1+0x1f6 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:162
# 0x2259f6f k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/handlers.getResourceHandler.func1+0x19f /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:123
# 0x223abbb k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/metrics.InstrumentRouteFunc.func1+0x1eb /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/metrics/metrics.go:104
# 0xfc4bef k8s.io/kubernetes/vendor/github.com/emicklei/go-restful.(*Container).dispatch+0xb9f /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/emicklei/go-restful/container.go:272
# 0xfd3047 k8s.io/kubernetes/vendor/github.com/emicklei/go-restful.(*Container).(k8s.io/kubernetes/vendor/github.com/emicklei/go-restful.dispatch)-fm+0x47 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/emicklei/go-restful/container.go:120
# 0x603bc3 net/http.HandlerFunc.ServeHTTP+0x43 /usr/local/go/src/net/http/server.go:1726
# 0x6050be net/http.(*ServeMux).ServeHTTP+0x7e /usr/local/go/src/net/http/server.go:2022
# 0x102a204 k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1+0x364 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters/authorization.go:50
# 0x603bc3 net/http.HandlerFunc.ServeHTTP+0x43 /usr/local/go/src/net/http/server.go:1726
# 0x102c10c k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1+0x1e4c /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters/impersonation.go:47
# 0x603bc3 net/http.HandlerFunc.ServeHTTP+0x43 /usr/local/go/src/net/http/server.go:1726
# 0x1029791 k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAudit.func1+0x911 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters/audit.go:137
Looks like exactly the same as #57061
Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale
.
Stale issues rot after an additional 30d of inactivity and eventually close.
If this issue is safe to close now please do so with /close
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/lifecycle stale
/remove-lifecycle stale
Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale
.
Stale issues rot after an additional 30d of inactivity and eventually close.
If this issue is safe to close now please do so with /close
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/lifecycle stale
/remove-lifecycle stale
Same issue
kubectl create -f kubia-liveness.yml
Error from server (ServerTimeout): error when creating "kubia-liveness.yml": the server cannot complete the requested operation at this time, try again later (post pods)
Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale
.
Stale issues rot after an additional 30d of inactivity and eventually close.
If this issue is safe to close now please do so with /close
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/lifecycle stale
—
Stale issues rot after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle rotten
.
Rotten issues close after an additional 30d of inactivity.
If this issue is safe to close now please do so with /close
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/lifecycle rotten
Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen
.
Mark the issue as fresh with /remove-lifecycle rotten
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/close
@fejta-bot: Closing this issue.
In response to this:
Rotten issues close after 30d of inactivity.
Reopen the issue with/reopen
.
Mark the issue as fresh with/remove-lifecycle rotten
.Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/close
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.
Closed #45811.
/reopen
—
You are receiving this because you are on a team that was mentioned.
Reply to this email directly, view it on GitHub, or unsubscribe.
@innovate-invent: You can't reopen an issue/PR unless you authored it or you are a collaborator.
In response to this:
/reopen
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.
—