--
You received this message because you are subscribed to the Google Groups "Kubernetes developer/contributor discussion" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-dev+unsubscribe@googlegroups.com.
To post to this group, send email to kubernetes-dev@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/67c2c101-78b1-4d44-bffc-05a4ee5126d6%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
To post to this group, send email to kuberne...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/67c2c101-78b1-4d44-bffc-05a4ee5126d6%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
--
You received this message because you are subscribed to the Google Groups "Kubernetes developer/contributor discussion" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-dev+unsubscribe@googlegroups.com.
To post to this group, send email to kubernetes-dev@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/CAH16ShKU%3DrDnYS5emxK%3DErixMShgfxP-zUoSDZNiqOC45OkmBA%40mail.gmail.com.
I tend to agree w/ Clayton. I think we may need one more release first, though. I think we are still shaking bugs out of etcd3.
Also we should have a discussion about how we adopt minor versions of etcd3. 3.2 is out, but we are still rolling out 3.0.
Should we validate and recommend specific etcd releases for specific Kubernetes versions? Should we decouple as much as possible and leave the choice of minor version up to the cluster operator?
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/CAH16ShKU%3DrDnYS5emxK%3DErixMShgfxP-zUoSDZNiqOC45OkmBA%40mail.gmail.com.
--
You received this message because you are subscribed to the Google Groups "Kubernetes developer/contributor discussion" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-dev+unsubscribe@googlegroups.com.
To post to this group, send email to kubernetes-dev@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/CAB_J3bbFLvqpk%2Bp8BbdX2NNA89hJg6JCtSffn5HJLUaaTHPu4A%40mail.gmail.com.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-de...@googlegroups.com.
To post to this group, send email to kuberne...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/67c2c101-78b1-4d44-bffc-05a4ee5126d6%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
--
You received this message because you are subscribed to the Google Groups "Kubernetes developer/contributor discussion" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-de...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/CAH16ShKU%3DrDnYS5emxK%3DErixMShgfxP-zUoSDZNiqOC45OkmBA%40mail.gmail.com.
--
You received this message because you are subscribed to the Google Groups "Kubernetes developer/contributor discussion" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-de...@googlegroups.com.
Gotcha - thanks Joe. There is this sentence which I think clarifies the policy: "This applies only to significant, user-visible behaviors which impact the correctness of applications running on Kubernetes or that impact the administration of Kubernetes clusters, and which are being removed entirely." But I also don't feel I'm in a place to make a determination - this may need to be the steering committee's first order of business!I'm signed up to keep etcd2 running :-) In practice that means getting our etcd2 e2e tests green, and ensuring that they are healthy enough to be blocking.I think we're looking for someone to volunteer to do the upgrade e2e/automation work so that we can deprecate etcd2, that is where our resources are lacking.
On Thu, Jul 27, 2017 at 12:11 PM Justin Santa Barbara <jus...@fathomdb.com> wrote:We discussed this in the contributor meeting - I'm sure Aaron can fill in more details:* I volunteered to get etcd2 tests that are broken working again, so that they can be considered for release signal* There are lots of users using etcd2, though it's unclear exactly how many - Azure & kops, likely self-install operators. Not OpenShift / GKE users.
* I believe in general everyone agrees that we want to get off etcd2* We don't seem to have an explicit policy for deprecating for something like etcd (*)* We have an issue containing a checklist for what should happen to deprecate etcd2 (proposed by me)* We didn't have any volunteers to do that work to deprecate etcd2
* sig-apimachinery was nominated to (initially?) own the process of deprecation
(*) I didn't think anyone read down this far in the meeting, but Rule #7 in our deprecation policy seems to say 1 year after announcement of deprecation. (Joe: I think you called attention to our deprecation policy, am I misinterpreting?)
I think we're looking for someone to volunteer to do the upgrade e2e/automation work so that we can deprecate etcd2, that is where our resources are lacking.I am confused about this--haven't we shipped a container that does this upgrade already? Distributions of Kubernetes that don't use our default container will have to write something special for their own use (but this is not the last upgrade we will ever do, so writing that mechanism is not wasted effort).
I tend to think an entire year is maybe an excessive maintenance cost. Especially since etcd2 doesn't scale and will never pass tests in big enough clusters.
I think we're looking for someone to volunteer to do the upgrade e2e/automation work so that we can deprecate etcd2, that is where our resources are lacking.I am confused about this--haven't we shipped a container that does this upgrade already? Distributions of Kubernetes that don't use our default container will have to write something special for their own use (but this is not the last upgrade we will ever do, so writing that mechanism is not wasted effort).The shipped container did not work for HA clusters (we discovered this very late in the 1.5 cycle I believe, but shipped etcd3 anyway, but this was why kops at least didn't adopt etcd3). wojtekt has a PR to fix it; I haven't had the chance to test it. I'd love to see that get under e2e testing though - this may turn out to actually not be that hard after all :-)
I tend to think an entire year is maybe an excessive maintenance cost. Especially since etcd2 doesn't scale and will never pass tests in big enough clusters.I don't disagree that a year is a long time, but I'd say it's pretty short for a database migration. Re-considering this time period may well be item #2 for the steering committee!
Re etcd2 scaling - I would like to see the metrics on the various options. My intuition is that having the Node healthcheck not write to etcd except on state transitions would have a bigger performance impact, but I hope someone has a doc that proves me wrong (and that would be an API change of course) :-)
Justin
--
You received this message because you are subscribed to the Google Groups "Kubernetes developer/contributor discussion" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-dev+unsubscribe@googlegroups.com.
To post to this group, send email to kubernetes-dev@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/f0a4cc78-b4ac-4cef-be4c-02dd6e51da98%40googlegroups.com.
On Thu, Jul 27, 2017 at 1:10 PM, Justin Santa Barbara <jus...@fathomdb.com> wrote:I think we're looking for someone to volunteer to do the upgrade e2e/automation work so that we can deprecate etcd2, that is where our resources are lacking.I am confused about this--haven't we shipped a container that does this upgrade already? Distributions of Kubernetes that don't use our default container will have to write something special for their own use (but this is not the last upgrade we will ever do, so writing that mechanism is not wasted effort).The shipped container did not work for HA clusters (we discovered this very late in the 1.5 cycle I believe, but shipped etcd3 anyway, but this was why kops at least didn't adopt etcd3). wojtekt has a PR to fix it; I haven't had the chance to test it. I'd love to see that get under e2e testing though - this may turn out to actually not be that hard after all :-)I haven't checked recently-- do we test anything about HA configurations? We should probably start if not...I tend to think an entire year is maybe an excessive maintenance cost. Especially since etcd2 doesn't scale and will never pass tests in big enough clusters.I don't disagree that a year is a long time, but I'd say it's pretty short for a database migration. Re-considering this time period may well be item #2 for the steering committee!etcd seems to pop out a new version ~quarterly if not faster. If our migrations take more than one release we will always be far behind. Database migrations without schema changes shouldn't be a big deal, right? ;) ;)Re etcd2 scaling - I would like to see the metrics on the various options. My intuition is that having the Node healthcheck not write to etcd except on state transitions would have a bigger performance impact, but I hope someone has a doc that proves me wrong (and that would be an API change of course) :-)That would be a major change, yes. Also it is really easy to break an etcd2 cluster (at least on a small machine) just by e.g. writing a bunch of big configmaps, or by accidentally making controllers fight.
My intuition is that having the Node healthcheck not write to etcd except on state transitions would have a bigger performance impact,
--
You received this message because you are subscribed to the Google Groups "Kubernetes developer/contributor discussion" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-dev+unsubscribe@googlegroups.com.
To post to this group, send email to kubernetes-dev@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/f0a4cc78-b4ac-4cef-be4c-02dd6e51da98%40googlegroups.com.
My intuition is that having the Node healthcheck not write to etcd except on state transitions would have a bigger performance impact,We've talked about decoupling node liveness heartbeat from NodeStatus update (there is probably a Github issue for it) but as things stand today every node needs to update NodeStatus at least as often as the node failure timeout since we use the lastUpdated field there to detect node failure.
-Eric
--
You received this message because you are subscribed to the Google Groups "Kubernetes developer/contributor discussion" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-dev+unsubscribe@googlegroups.com.
To post to this group, send email to kubernetes-dev@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/CAFoXKmp7h5Txteo6rvawzWuddfjwXR2N_rY%3D%3D%3Dvm0%3DAEwEfiyw%40mail.gmail.com.
To post to this group, send email to kuberne...@googlegroups.com.
https://github.com/kubernetes/kubernetes/issues/43600 has been closed for months, I thought that was the blocker for kops? *we* use the technique / container mentioned there. It works.I consider etcd2 deprecated, the clock should definitely start no later than 1.9... etcd3 has been the default for how long now? Almost an entire year, I think?I consider rolling out an etcd 2 -> 3 upgrade the responsibility of distributions, not api machinery or other parts of "core kubernetes" (whatever that means) for the simple reason that it is incredibly environment dependent. We provided a container which does the upgrade, but we can't e.g. exhaustively test it in every environment, we don't know where a random Kubernetes install wants to store its backups, etc.I am painfully aware of how much work it is to operate a Kubernetes distro, so please don't take this the wrong way; but doing updates is part of that. Rolling over GKE clusters was a 2 ish person 6 ish month process; every reusable output of that process is available already, so hopefully it won't be that hard for anyone else.
To post to this group, send email to kuberne...@googlegroups.com.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-de...@googlegroups.com.
To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-dev+unsubscribe@googlegroups.com.
To post to this group, send email to kubernetes-dev@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/ccdca081-0261-440a-9392-6e87b0daefdc%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/ccdca081-0261-440a-9392-6e87b0daefdc%40googlegroups.com.
--
You received this message because you are subscribed to a topic in the Google Groups "Kubernetes developer/contributor discussion" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/kubernetes-dev/DoOl77xjpDA/unsubscribe.
To unsubscribe from this group and all its topics, send an email to kubernetes-dev+unsubscribe@googlegroups.com.
To post to this group, send email to kubernetes-dev@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/CAB_J3bZSsknyE4qDS9Ezh%2BxOSSZ4hw6xEFpkcSLk1st%2BD%3Dyg2g%40mail.gmail.com.
I put an item on the agenda for sig-cluster-lifecycle for next Tuesday (Feb 6th), to demo/discuss https://github.com/kopeio/etcd-manager . It allows for etcd management and backups without relying on kubernetes, including upgrades. It avoids some of the potential circular dependencies of the operator approach, which have been a blocker for consensus up until now, so I'm hoping we can reach agreement on this approach - or at the very least the backup structure on S3/GCS etc.
Justin
Justin
To view this discussion on the web visit https://groups.google.com/d/msgid/kubernetes-dev/ccdca081-0261-440a-9392-6e87b0daefdc%40googlegroups.com.
--
You received this message because you are subscribed to a topic in the Google Groups "Kubernetes developer/contributor discussion" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/kubernetes-dev/DoOl77xjpDA/unsubscribe.
To unsubscribe from this group and all its topics, send an email to kubernetes-de...@googlegroups.com.
To post to this group, send email to kuberne...@googlegroups.com.