Re: [kubernetes/kubernetes] Flaky 1.8 sig-cli conformance test (#64110)

5 views
Skip to first unread message

Benjamin Elder

unread,
May 21, 2018, 2:26:23 PM5/21/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

@kubernetes/sig-cli-bugs


You are receiving this because you are on a team that was mentioned.
Reply to this email directly, view it on GitHub, or mute the thread.

Benjamin Elder

unread,
May 21, 2018, 2:28:58 PM5/21/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

Actually looking back through the logs more, it's more broadly kubectl tests that flake:

  • [sig-cli] Kubectl client [k8s.io] Kubectl run rc should create an rc from an image [Conformance]
  • [sig-cli] Kubectl client [k8s.io] Kubectl run --rm job should create a job from an image, then delete the job [Conformance]

https://k8s-gubernator.appspot.com/builds/kubernetes-jenkins/logs/ci-kubernetes-gce-conformance-latest-1-8/

Kim Min

unread,
May 25, 2018, 4:00:41 AM5/25/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

@BenTheElder These flakes are all caused by network connection, and we could do nothing but retry when it happens.
Should these test explicitly handle the network problem caused by test environment? OTOH maybe we could try to fix the network.

Benjamin Elder

unread,
May 25, 2018, 4:10:17 AM5/25/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention
Are you saying that the inter cluster networking is flaking or do these
rely on the broader internet?

I'm not sure what the best change would be, but the rest of the conformance
suite is not flaky in the slightest, which is why I question the
reliability of these tests. It's also possible that GCE 1.8 clusters are
flaky but given that the rest of the tests pass flawlessly many many times
I'd guess we just need to improve these tests...

On Fri, May 25, 2018, 1:00 AM Kim Min <notifi...@github.com> wrote:

> @BenTheElder <https://github.com/BenTheElder> These flakes are all caused

> by network connection, and we could do nothing but retry when it happens.
> Should these test explicitly handle the network problem caused by test
> environment? OTOH maybe we could try to fix the network.
>
> —
> You are receiving this because you were mentioned.

> Reply to this email directly, view it on GitHub
> <https://github.com/kubernetes/kubernetes/issues/64110#issuecomment-391974006>,
> or mute the thread
> <https://github.com/notifications/unsubscribe-auth/AA4BqyYOrdQ5zrxiMmTJA5p78y0axeUrks5t17ongaJpZM4UHaca>

Francois Tur

unread,
May 25, 2018, 10:46:29 AM5/25/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

@BenTheElder, @yue9944882 : we can see in the test dashboard of "sig-sluster-lifecycle" or the test dashboard of sig-network-gce, that this serie of test "sig-cli" is falking when it is a deployment with kubeadm in gce.
But looks like it does not depend the type of network used (default, ipvs, calico, flannel - although obviously calico has more issues).

and when the test fail, the error is always the same:
could not convert scale update to external Scale: scheme.Scale is not suitable for converting to \"v1\"

These same tests are not flaking, when the deployment is usual kube-up with gce. see here : http://k8s-testgrid.appspot.com/sig-network-gce#gci-gce-coredns

Does it tilt a light for one of you ?

Benjamin Elder

unread,
May 25, 2018, 6:55:12 PM5/25/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

@fturib I think that warrants its own, higher priority issue. these flake with low regularity in the oldest supported release.

Also as a note, I am setting up GCE conformance tests for the ongoing release today (separate from master) and next week I plan to raise the question of blocking releases on a conformance suite.

Francois Tur

unread,
May 29, 2018, 11:14:52 AM5/29/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

I created an issue #64450 for these flaky sig-cli tests related to kubeadm/gce.

Benjamin Elder

unread,
May 29, 2018, 4:42:32 PM5/29/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention
thanks!

On Tue, May 29, 2018 at 8:14 AM Francois Tur <notifi...@github.com>
wrote:


> I created an issue #64450
> <https://github.com/kubernetes/kubernetes/issues/64450> for these flaky

> sig-cli tests related to kubeadm/gce.
>
> —
> You are receiving this because you were mentioned.

> Reply to this email directly, view it on GitHub
> <https://github.com/kubernetes/kubernetes/issues/64110#issuecomment-392814223>,
> or mute the thread
> <https://github.com/notifications/unsubscribe-auth/AA4Bq-mEbo-UEu4Eo8Xe3mmKPMSnLG19ks5t3WXrgaJpZM4UHaca>

fejta-bot

unread,
Aug 27, 2018, 4:56:57 PM8/27/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale.
Stale issues rot after an additional 30d of inactivity and eventually close.

If this issue is safe to close now please do so with /close.

Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/lifecycle stale

fejta-bot

unread,
Sep 26, 2018, 5:20:03 PM9/26/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

Stale issues rot after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle rotten.
Rotten issues close after an additional 30d of inactivity.

If this issue is safe to close now please do so with /close.

Send feedback to sig-testing, kubernetes/test-infra and/or fejta.

/lifecycle rotten

fejta-bot

unread,
Oct 26, 2018, 6:07:33 PM10/26/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen.


Mark the issue as fresh with /remove-lifecycle rotten.

Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/close

k8s-ci-robot

unread,
Oct 26, 2018, 6:07:33 PM10/26/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

Closed #64110.

k8s-ci-robot

unread,
Oct 26, 2018, 6:07:44 PM10/26/18
to kubernetes/kubernetes, k8s-mirror-cli-bugs, Team mention

@fejta-bot: Closing this issue.

In response to this:

Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen.
Mark the issue as fresh with /remove-lifecycle rotten.

Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/close

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

Reply all
Reply to author
Forward
0 new messages