diff --git a/contributors/design-proposals/bootstrap-discovery.md b/contributors/design-proposals/bootstrap-discovery.md new file mode 100644 index 000000000..dbcf793ca --- /dev/null +++ b/contributors/design-proposals/bootstrap-discovery.md @@ -0,0 +1,178 @@ +# Super Simple Discovery API + +## Overview + +It is surprisingly hard to figure out how to talk to a Kubernetes cluster. Not only do clients need to know where to look on the network, they also need to identify the set of root certificates to trust when talking to that endpoint. + +This presents a set of problems: +* It should be super easy for users to configure client systems with a minimum of effort `kubectl` or `kubeadm init` (or other client systems). + * Establishing this should be doable even in the face of nodes and master components booting out of order. + * We should have mechanisms that don't require users to ever have to manually manage certificate files. +* Over the life of the cluster this information could change and client systems should be able to adapt. + +While this design is mainly being created to help `kubeadm` possible, these problems aren't isolated there and can be used outside of the kubeadm context. + +Mature organizations should be able to distribute and manage root certificates out of band of Kubernetes installations. In that case, clients will defer to corporation wide system installed root certificates or root certificates distributed through other means. However, for smaller and more casual users distributing or obtaining certificates represents a challenge. + +Similarly, mature organizations will be able to rely on a centrally managed DNS system to distribute the location of a set of API servers and keep those names up to date over time. Those DNS servers will be managed for high availability. + +With that in mind, the proposals here will devolve into simply using DNS names that are validated with system installed root certificates. + +## Cluster Location information + +First we define a set of information that identifies a cluster and how to talk to it. + +While we could define a new format for communicating the set of information needed here, we'll start by using the standard [`kubeconfig`](http://kubernetes.io/docs/user-guide/kubeconfig-file/) file format. + +It is expected that the `kubeconfig` file will have a single unnamed `Cluster` entry. Other information (especially authentication secrets) must be omitted. + +### Evolving kubeconfig + +In the future we look forward to enhancing `kubeconfig` to address some issues. These are out of scope for this design. Some of this is covered in [#30395](https://github.com/kubernetes/kubernetes/issues/30395). + +Additions include: + +* A cluster serial number/identifier. + * In an HA world, API servers may come and go and it is necessary to make sure we are talking to the same cluster as we thought we were talking to. +* A _set_ of addresses for finding the cluster. + * It is implied that all of these are equivalent and that a client can try multiple until an appropriate target is found. + * Initially I’m proposing a flat set here. In the future we can introduce more structure that hints to the user which addresses to try first. +* Better documentation and exposure of: + * The root certificates can be a bundle to enable rotation. + * If no root certificates are given (and the insecure bit isn't set) then the client trusts the system managed list of CAs. + +### Client caching and update + +**This is to be implemented in a later phase** + +Any client of the cluster will want to have this information. As the configuration of the cluster changes we need the client to keep this information up to date. It is assumed that the information here won’t drift so fast that clients won’t be able to find *some* way to connect. + +In exceptional circumstances it is possible that this information may be out of date and a client would be unable to connect to a cluster. Consider the case where a user has kubectl set up and working well and then doesn't run kubectl for quite a while. It is possible that over this time (a) the set of servers will have migrated so that all endpoints are now invalid or (b) the root certificates will have rotated so that the user can no longer trust any endpoint. + +## Methods + +Now that we know *what* we want to get to the client, the question is how. We want to do this in as secure a way possible (as there are cryptographic keys involved) without requiring a lot of overhead in terms of information that needs to be copied around. + +### Method: Out of Band + +The simplest way to do this would be to simply put this object in a file and copy it around. This is more overhead for the user, but it is easy to implement and lets users rely on existing systems to distribute configuration. + +For the `kubeadm` flow, the command line might look like: + +``` +kubeadm join --cluster-info-file=my-cluster.yaml +``` + +Note that TLS bootstrap (which establishes a way for a client to authenticate itself to the server) is a separate issue and has its own set of methods. This command line may have a TLS bootstrap token (or config file) on the command line also. + +### Method: HTTPS Endpoint + +If the ClusterInfo information is hosted in a trusted place via HTTPS you can just request it that way. This will use the root certificates that are installed on the system. It may or may not be appropriate based on the user's constraints. + +``` +kubeadm join --cluster-info-url="https://example/mycluster.yaml" +``` + +This is really a shorthand for someone doing something like (assuming we support stdin with `-`): + +``` +curl https://example.com/mycluster.json | kubeadm join --cluster-info-file=- +``` + +If the user requires some auth to the HTTPS server (to keep the ClusterInfo object private) that can be done in the curl command equivalent. Or we could eventually add it to `kubeadm` directly. + +### Method: Bootstrap Token + +There won’t always be a trusted external endpoint to talk to and transmitting +the locator file out of band is a pain. However, we want something more secure +than just hitting HTTP and trusting whatever we get back. In this case, we +assume we have the following: + + * An address for at least one of the API servers (which will implement this API). + * This address is technically an HTTPS URL base but is often expressed as a bare domain or IP. + * A shared secret token + +An interesting aspect here is that this information is often easily obtained before the API server is configured or started. This makes some cluster bring-up scenarios much easier. + +The user experience for joining a cluster would be something like: + +``` +kubeadm join --token=ae23dc.faddc87f5a5ab458
+``` + +**Note:** This is logically a different use of the token from TLS bootstrap. We harmonize these usages and allow the same token to play double duty. + +#### Implementation Flow + +`kubeadm` will implement the following flow: + +* `kubeadm` connects to the API server address specified over TLS. As we don't yet have a root certificate to trust, this is an insecure connection and the server certificate is not validated. `kubeadm` provides no authentication credentials at all. + * Implementation note: the API server doesn't have to expose a new and special insecure HTTP endpoint. +* `kubeadm` requests a ConfigMap containing the kubeconfig file defined above. + * This ConfigMap exists at a well known URL: `https://