kops/nodeup/pkg/model/kube_apiserver.go

872 lines
27 KiB
Go

/*
Copyright 2019 The Kubernetes Authors.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
package model
import (
"context"
"fmt"
"net"
"path/filepath"
"sort"
"strings"
"k8s.io/klog/v2"
"k8s.io/kops/pkg/apis/kops"
"k8s.io/kops/pkg/flagbuilder"
"k8s.io/kops/pkg/k8scodecs"
"k8s.io/kops/pkg/kubeconfig"
"k8s.io/kops/pkg/kubemanifest"
"k8s.io/kops/pkg/model/components"
"k8s.io/kops/pkg/tokens"
"k8s.io/kops/pkg/wellknownports"
"k8s.io/kops/pkg/wellknownusers"
"k8s.io/kops/upup/pkg/fi"
"k8s.io/kops/upup/pkg/fi/nodeup/nodetasks"
"k8s.io/kops/util/pkg/proxy"
v1 "k8s.io/api/core/v1"
"k8s.io/apimachinery/pkg/api/resource"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/apimachinery/pkg/util/intstr"
)
// PathAuthnConfig is the path to the custom webhook authentication config.
const PathAuthnConfig = "/etc/kubernetes/authn.config"
// KubeAPIServerBuilder installs kube-apiserver.
type KubeAPIServerBuilder struct {
*NodeupModelContext
}
var _ fi.NodeupModelBuilder = &KubeAPIServerBuilder{}
// Build is responsible for generating the configuration for the kube-apiserver.
func (b *KubeAPIServerBuilder) Build(c *fi.NodeupModelBuilderContext) error {
ctx := c.Context()
if !b.HasAPIServer {
return nil
}
pathSrvKAPI := filepath.Join(b.PathSrvKubernetes(), "kube-apiserver")
var kubeAPIServer kops.KubeAPIServerConfig
if b.NodeupConfig.APIServerConfig.KubeAPIServer != nil {
kubeAPIServer = *b.NodeupConfig.APIServerConfig.KubeAPIServer
}
if b.CloudProvider() == kops.CloudProviderHetzner {
localIP, err := b.GetMetadataLocalIP(c.Context())
if err != nil {
return err
}
if localIP != "" {
kubeAPIServer.AdvertiseAddress = localIP
}
}
if b.CloudProvider() == kops.CloudProviderMetal {
// Workaround for https://github.com/kubernetes/kubernetes/issues/111671
if b.IsIPv6Only() {
interfaces, err := net.Interfaces()
if err != nil {
return fmt.Errorf("getting local network interfaces: %w", err)
}
var ipv6s []net.IP
for _, intf := range interfaces {
addresses, err := intf.Addrs()
if err != nil {
return fmt.Errorf("getting addresses for network interface %q: %w", intf.Name, err)
}
for _, addr := range addresses {
ip, _, err := net.ParseCIDR(addr.String())
if ip == nil {
return fmt.Errorf("parsing ip address %q (bound to network %q): %w", addr.String(), intf.Name, err)
}
if ip.To4() != nil {
// We're only looking for ipv6
continue
}
if ip.IsLinkLocalUnicast() {
klog.V(4).Infof("ignoring link-local unicast addr %v", addr)
continue
}
if ip.IsLinkLocalMulticast() {
klog.V(4).Infof("ignoring link-local multicast addr %v", addr)
continue
}
if ip.IsLoopback() {
klog.V(4).Infof("ignoring loopback addr %v", addr)
continue
}
ipv6s = append(ipv6s, ip)
}
}
if len(ipv6s) > 1 {
klog.Warningf("found multiple ipv6s, choosing first: %v", ipv6s)
}
if len(ipv6s) == 0 {
klog.Warningf("did not find ipv6 address for kube-apiserver --advertise-address")
}
if len(ipv6s) > 0 {
kubeAPIServer.AdvertiseAddress = ipv6s[0].String()
}
}
}
b.configureOIDC(&kubeAPIServer)
if err := b.writeAuthenticationConfig(c, &kubeAPIServer); err != nil {
return err
}
if b.NodeupConfig.APIServerConfig.EncryptionConfigSecretHash != "" {
encryptionConfigPath := fi.PtrTo(filepath.Join(pathSrvKAPI, "encryptionconfig.yaml"))
kubeAPIServer.EncryptionProviderConfig = encryptionConfigPath
key := "encryptionconfig"
encryptioncfg, err := b.SecretStore.Secret(key)
if err == nil {
contents := string(encryptioncfg.Data)
t := &nodetasks.File{
Path: *encryptionConfigPath,
Contents: fi.NewStringResource(contents),
Mode: fi.PtrTo("600"),
Type: nodetasks.FileType_File,
}
c.AddTask(t)
} else {
return fmt.Errorf("encryptionConfig enabled, but could not load encryptionconfig secret: %v", err)
}
}
kubeAPIServer.ServiceAccountKeyFile = append(kubeAPIServer.ServiceAccountKeyFile, filepath.Join(pathSrvKAPI, "service-account.pub"))
c.AddTask(&nodetasks.File{
Path: filepath.Join(pathSrvKAPI, "service-account.pub"),
Contents: fi.NewStringResource(b.NodeupConfig.APIServerConfig.ServiceAccountPublicKeys),
Type: nodetasks.FileType_File,
Mode: s("0600"),
})
// Set the signing key if we're using Service Account Token VolumeProjection
if kubeAPIServer.ServiceAccountSigningKeyFile == nil {
s := filepath.Join(pathSrvKAPI, "service-account.key")
kubeAPIServer.ServiceAccountSigningKeyFile = &s
if err := b.BuildPrivateKeyTask(c, "service-account", pathSrvKAPI, "service-account", nil, nil); err != nil {
return err
}
}
{
c.AddTask(&nodetasks.File{
Path: filepath.Join(pathSrvKAPI, "etcd-ca.crt"),
Contents: fi.NewStringResource(b.NodeupConfig.CAs["etcd-clients-ca"]),
Type: nodetasks.FileType_File,
Mode: fi.PtrTo("0644"),
})
kubeAPIServer.EtcdCAFile = filepath.Join(pathSrvKAPI, "etcd-ca.crt")
issueCert := &nodetasks.IssueCert{
Name: "etcd-client",
Signer: "etcd-clients-ca",
KeypairID: b.NodeupConfig.KeypairIDs["etcd-clients-ca"],
Type: "client",
Subject: nodetasks.PKIXName{
CommonName: "kube-apiserver",
},
}
c.AddTask(issueCert)
if err := issueCert.AddFileTasks(c, pathSrvKAPI, issueCert.Name, "", nil); err != nil {
return err
}
}
kubeAPIServer.EtcdCertFile = filepath.Join(pathSrvKAPI, "etcd-client.crt")
kubeAPIServer.EtcdKeyFile = filepath.Join(pathSrvKAPI, "etcd-client.key")
{
c.AddTask(&nodetasks.File{
Path: filepath.Join(pathSrvKAPI, "apiserver-aggregator-ca.crt"),
Contents: fi.NewStringResource(b.NodeupConfig.CAs["apiserver-aggregator-ca"]),
Type: nodetasks.FileType_File,
Mode: fi.PtrTo("0644"),
})
kubeAPIServer.RequestheaderClientCAFile = filepath.Join(pathSrvKAPI, "apiserver-aggregator-ca.crt")
issueCert := &nodetasks.IssueCert{
Name: "apiserver-aggregator",
Signer: "apiserver-aggregator-ca",
KeypairID: b.NodeupConfig.KeypairIDs["apiserver-aggregator-ca"],
Type: "client",
// Must match RequestheaderAllowedNames
Subject: nodetasks.PKIXName{CommonName: "aggregator"},
}
c.AddTask(issueCert)
err := issueCert.AddFileTasks(c, pathSrvKAPI, "apiserver-aggregator", "", nil)
if err != nil {
return err
}
kubeAPIServer.ProxyClientCertFile = fi.PtrTo(filepath.Join(pathSrvKAPI, "apiserver-aggregator.crt"))
kubeAPIServer.ProxyClientKeyFile = fi.PtrTo(filepath.Join(pathSrvKAPI, "apiserver-aggregator.key"))
}
if err := b.writeServerCertificate(c, &kubeAPIServer); err != nil {
return err
}
if err := b.writeKubeletAPICertificate(c, &kubeAPIServer); err != nil {
return err
}
if err := b.writeStaticCredentials(c, &kubeAPIServer); err != nil {
return err
}
{
pod, err := b.buildPod(ctx, &kubeAPIServer)
if err != nil {
return fmt.Errorf("error building kube-apiserver manifest: %v", err)
}
manifest, err := k8scodecs.ToVersionedYaml(pod)
if err != nil {
return fmt.Errorf("error marshaling manifest to yaml: %v", err)
}
c.AddTask(&nodetasks.File{
Path: "/etc/kubernetes/manifests/kube-apiserver.manifest",
Contents: fi.NewBytesResource(manifest),
Type: nodetasks.FileType_File,
})
}
// If we're using kube-apiserver-healthcheck, we need to set up the client cert etc
if b.findHealthcheckManifest() != nil {
if err := b.addHealthcheckSidecarTasks(c); err != nil {
return err
}
}
c.AddTask(&nodetasks.File{
Path: "/var/log/kube-apiserver.log",
Contents: fi.NewStringResource(""),
Type: nodetasks.FileType_File,
Mode: s("0400"),
IfNotExists: true,
})
return nil
}
func (b *KubeAPIServerBuilder) configureOIDC(kubeAPIServer *kops.KubeAPIServerConfig) {
if b.NodeupConfig.APIServerConfig.Authentication == nil || b.NodeupConfig.APIServerConfig.Authentication.OIDC == nil {
return
}
oidc := b.NodeupConfig.APIServerConfig.Authentication.OIDC
kubeAPIServer.OIDCClientID = oidc.ClientID
if oidc.GroupsClaims != nil {
join := strings.Join(oidc.GroupsClaims, ",")
kubeAPIServer.OIDCGroupsClaim = &join
}
kubeAPIServer.OIDCGroupsPrefix = oidc.GroupsPrefix
kubeAPIServer.OIDCIssuerURL = oidc.IssuerURL
if oidc.RequiredClaims != nil {
kubeAPIServer.OIDCRequiredClaim = make([]string, 0, len(oidc.RequiredClaims))
for claim, value := range oidc.RequiredClaims {
kubeAPIServer.OIDCRequiredClaim = append(kubeAPIServer.OIDCRequiredClaim, claim+"="+value)
}
sort.Strings(kubeAPIServer.OIDCRequiredClaim)
}
kubeAPIServer.OIDCUsernameClaim = oidc.UsernameClaim
kubeAPIServer.OIDCUsernamePrefix = oidc.UsernamePrefix
}
func (b *KubeAPIServerBuilder) writeAuthenticationConfig(c *fi.NodeupModelBuilderContext, kubeAPIServer *kops.KubeAPIServerConfig) error {
if b.NodeupConfig.APIServerConfig.Authentication == nil {
return nil
}
if b.NodeupConfig.APIServerConfig.Authentication.AWS == nil && b.NodeupConfig.APIServerConfig.Authentication.Kopeio == nil {
return nil
}
if b.NodeupConfig.APIServerConfig.Authentication.Kopeio != nil {
cluster := kubeconfig.KubectlCluster{
Server: "http://127.0.0.1:9001/hooks/authn",
}
context := kubeconfig.KubectlContext{
Cluster: "webhook",
User: "kube-apiserver",
}
config := kubeconfig.KubectlConfig{
Kind: "Config",
ApiVersion: "v1",
}
config.Clusters = append(config.Clusters, &kubeconfig.KubectlClusterWithName{
Name: "webhook",
Cluster: cluster,
})
config.Users = append(config.Users, &kubeconfig.KubectlUserWithName{
Name: "kube-apiserver",
})
config.CurrentContext = "webhook"
config.Contexts = append(config.Contexts, &kubeconfig.KubectlContextWithName{
Name: "webhook",
Context: context,
})
manifest, err := kops.ToRawYaml(config)
if err != nil {
return fmt.Errorf("error marshaling authentication config to yaml: %v", err)
}
c.AddTask(&nodetasks.File{
Path: PathAuthnConfig,
Contents: fi.NewBytesResource(manifest),
Type: nodetasks.FileType_File,
})
return nil
}
if b.NodeupConfig.APIServerConfig.Authentication.AWS != nil {
id := "aws-iam-authenticator"
kubeAPIServer.AuthenticationTokenWebhookConfigFile = fi.PtrTo(PathAuthnConfig)
{
cluster := kubeconfig.KubectlCluster{
Server: "https://127.0.0.1:21362/authenticate",
CertificateAuthorityData: []byte(b.NodeupConfig.CAs[fi.CertificateIDCA]),
}
context := kubeconfig.KubectlContext{
Cluster: "aws-iam-authenticator",
User: "kube-apiserver",
}
config := kubeconfig.KubectlConfig{}
config.Clusters = append(config.Clusters, &kubeconfig.KubectlClusterWithName{
Name: "aws-iam-authenticator",
Cluster: cluster,
})
config.Users = append(config.Users, &kubeconfig.KubectlUserWithName{
Name: "kube-apiserver",
})
config.CurrentContext = "webhook"
config.Contexts = append(config.Contexts, &kubeconfig.KubectlContextWithName{
Name: "webhook",
Context: context,
})
manifest, err := kops.ToRawYaml(config)
if err != nil {
return fmt.Errorf("error marshaling authentication config to yaml: %v", err)
}
c.AddTask(&nodetasks.File{
Path: PathAuthnConfig,
Contents: fi.NewBytesResource(manifest),
Type: nodetasks.FileType_File,
Mode: fi.PtrTo("600"),
})
}
// We create user aws-iam-authenticator and hardcode its UID to 10000 as
// that is the ID used inside the aws-iam-authenticator container.
// The owner/group for the keypair to aws-iam-authenticator
{
c.AddTask(&nodetasks.UserTask{
Name: "aws-iam-authenticator",
UID: wellknownusers.AWSAuthenticator,
Shell: "/sbin/nologin",
Home: "/srv/kubernetes/aws-iam-authenticator",
})
}
{
issueCert := &nodetasks.IssueCert{
Name: id,
Signer: fi.CertificateIDCA,
KeypairID: b.NodeupConfig.KeypairIDs[fi.CertificateIDCA],
Type: "server",
Subject: nodetasks.PKIXName{CommonName: id},
AlternateNames: []string{
"localhost",
"127.0.0.1",
},
}
c.AddTask(issueCert)
certificate, privateKey, _ := issueCert.GetResources()
c.AddTask(&nodetasks.File{
Path: "/srv/kubernetes/aws-iam-authenticator/cert.pem",
Contents: certificate,
Type: nodetasks.FileType_File,
Mode: fi.PtrTo("600"),
Owner: fi.PtrTo("aws-iam-authenticator"),
Group: fi.PtrTo("aws-iam-authenticator"),
})
c.AddTask(&nodetasks.File{
Path: "/srv/kubernetes/aws-iam-authenticator/key.pem",
Contents: privateKey,
Type: nodetasks.FileType_File,
Mode: fi.PtrTo("600"),
Owner: fi.PtrTo("aws-iam-authenticator"),
Group: fi.PtrTo("aws-iam-authenticator"),
})
}
return nil
}
return fmt.Errorf("unrecognized authentication config %v", b.NodeupConfig.APIServerConfig.Authentication)
}
func (b *KubeAPIServerBuilder) writeServerCertificate(c *fi.NodeupModelBuilderContext, kubeAPIServer *kops.KubeAPIServerConfig) error {
pathSrvKAPI := filepath.Join(b.PathSrvKubernetes(), "kube-apiserver")
{
// A few names used from inside the cluster, which all resolve the same based on our default suffixes
alternateNames := []string{
"kubernetes",
"kubernetes.default",
"kubernetes.default.svc",
"kubernetes.default.svc." + b.NodeupConfig.APIServerConfig.ClusterDNSDomain,
}
// Names specified in the cluster spec
if b.NodeupConfig.APIServerConfig.API.PublicName != "" {
alternateNames = append(alternateNames, b.NodeupConfig.APIServerConfig.API.PublicName)
}
alternateNames = append(alternateNames, b.APIInternalName())
alternateNames = append(alternateNames, b.NodeupConfig.APIServerConfig.API.AdditionalSANs...)
// Load balancer IPs passed in through NodeupConfig
alternateNames = append(alternateNames, b.NodeupConfig.ApiserverAdditionalIPs...)
// Referencing it by internal IP should work also
{
ip, err := components.WellKnownServiceIP(&b.NodeupConfig.Networking, 1)
if err != nil {
return err
}
alternateNames = append(alternateNames, ip.String())
}
// We also want to be able to reference it locally via https://127.0.0.1
alternateNames = append(alternateNames, "127.0.0.1")
if b.CloudProvider() == kops.CloudProviderHetzner {
localIP, err := b.GetMetadataLocalIP(c.Context())
if err != nil {
return err
}
if localIP != "" {
alternateNames = append(alternateNames, localIP)
}
}
if b.CloudProvider() == kops.CloudProviderOpenstack {
instanceAddress, err := getInstanceAddress()
if err != nil {
return err
}
alternateNames = append(alternateNames, instanceAddress)
}
issueCert := &nodetasks.IssueCert{
Name: "master",
Signer: fi.CertificateIDCA,
KeypairID: b.NodeupConfig.KeypairIDs[fi.CertificateIDCA],
Type: "server",
Subject: nodetasks.PKIXName{CommonName: "kubernetes-master"},
AlternateNames: alternateNames,
}
// Including the CA certificate is more correct, and is needed for e.g. AWS WebIdentity federation
issueCert.IncludeRootCertificate = true
c.AddTask(issueCert)
err := issueCert.AddFileTasks(c, pathSrvKAPI, "server", "", nil)
if err != nil {
return err
}
}
// If clientCAFile is not specified, set it to the default value ${PathSrvKubernetes}/ca.crt
if kubeAPIServer.ClientCAFile == "" {
kubeAPIServer.ClientCAFile = filepath.Join(b.PathSrvKubernetes(), "ca.crt")
}
kubeAPIServer.TLSCertFile = filepath.Join(pathSrvKAPI, "server.crt")
kubeAPIServer.TLSPrivateKeyFile = filepath.Join(pathSrvKAPI, "server.key")
return nil
}
func (b *KubeAPIServerBuilder) writeKubeletAPICertificate(c *fi.NodeupModelBuilderContext, kubeAPIServer *kops.KubeAPIServerConfig) error {
pathSrvKAPI := filepath.Join(b.PathSrvKubernetes(), "kube-apiserver")
issueCert := &nodetasks.IssueCert{
Name: "kubelet-api",
Signer: fi.CertificateIDCA,
KeypairID: b.NodeupConfig.KeypairIDs[fi.CertificateIDCA],
Type: "client",
Subject: nodetasks.PKIXName{CommonName: "kubelet-api"},
}
c.AddTask(issueCert)
err := issueCert.AddFileTasks(c, pathSrvKAPI, "kubelet-api", "", nil)
if err != nil {
return err
}
// @note we are making assumption were using the ones created by the pki model, not custom defined ones
kubeAPIServer.KubeletClientCertificate = filepath.Join(pathSrvKAPI, "kubelet-api.crt")
kubeAPIServer.KubeletClientKey = filepath.Join(pathSrvKAPI, "kubelet-api.key")
return nil
}
func (b *KubeAPIServerBuilder) writeStaticCredentials(c *fi.NodeupModelBuilderContext, kubeAPIServer *kops.KubeAPIServerConfig) error {
pathSrvKAPI := filepath.Join(b.PathSrvKubernetes(), "kube-apiserver")
if b.SecretStore != nil {
allTokens, err := b.allAuthTokens()
if err != nil {
return err
}
var lines []string
for id, token := range allTokens {
if id == adminUser {
lines = append(lines, token+","+id+","+id+","+adminGroup)
} else {
lines = append(lines, token+","+id+","+id)
}
}
csv := strings.Join(lines, "\n")
c.AddTask(&nodetasks.File{
Path: filepath.Join(pathSrvKAPI, "known_tokens.csv"),
Contents: fi.NewStringResource(csv),
Type: nodetasks.FileType_File,
Mode: s("0600"),
})
}
return nil
}
// allAuthTokens returns a map of all auth tokens that are present
func (b *KubeAPIServerBuilder) allAuthTokens() (map[string]string, error) {
possibleTokens := tokens.GetKubernetesAuthTokens_Deprecated()
tokens := make(map[string]string)
for _, id := range possibleTokens {
token, err := b.SecretStore.FindSecret(id)
if err != nil {
return nil, err
}
if token != nil {
tokens[id] = string(token.Data)
}
}
return tokens, nil
}
// buildPod is responsible for generating the kube-apiserver pod and thus manifest file
func (b *KubeAPIServerBuilder) buildPod(ctx context.Context, kubeAPIServer *kops.KubeAPIServerConfig) (*v1.Pod, error) {
// we need to replace 127.0.0.1 for etcd urls with the dns names in case this apiserver is not
// running on master nodes
if !b.IsMaster {
clusterName := b.NodeupConfig.ClusterName
mainEtcdDNSName := "main.etcd.internal." + clusterName
eventsEtcdDNSName := "events.etcd.internal." + clusterName
for i := range kubeAPIServer.EtcdServers {
kubeAPIServer.EtcdServers[i] = strings.ReplaceAll(kubeAPIServer.EtcdServers[i], "127.0.0.1", mainEtcdDNSName)
}
for i := range kubeAPIServer.EtcdServersOverrides {
if strings.HasPrefix(kubeAPIServer.EtcdServersOverrides[i], "/events") {
kubeAPIServer.EtcdServersOverrides[i] = strings.ReplaceAll(kubeAPIServer.EtcdServersOverrides[i], "127.0.0.1", eventsEtcdDNSName)
}
}
}
// @fixup: the admission controller migrated from --admission-control to --enable-admission-plugins, but
// most people will still have c.Spec.KubeAPIServer.AdmissionControl references into their configuration we need
// to fix up. A PR https://github.com/kubernetes/kops/pull/5221/ introduced the issue and since the command line
// flags are mutually exclusive the API refuses to come up.
{
// @note: note sure if this is the best place to put it, I could place into the validation.go which has the benefit of
// fixing up the manifests itself, but that feels VERY hacky
// @note: it's fine to use AdmissionControl here and it's not populated by the model, thus the only data could have come from the cluster spec
if len(kubeAPIServer.AdmissionControl) > 0 {
kubeAPIServer.EnableAdmissionPlugins = append([]string(nil), kubeAPIServer.AdmissionControl...)
kubeAPIServer.AdmissionControl = []string{}
}
}
// build the kube-apiserver flags for the service
flags, err := flagbuilder.BuildFlagsList(kubeAPIServer)
if err != nil {
return nil, fmt.Errorf("error building kube-apiserver flags: %v", err)
}
if b.IsKubernetesLT("1.33") {
flags = append(flags, fmt.Sprintf("--cloud-config=%s", InTreeCloudConfigFilePath))
}
pod := &v1.Pod{
TypeMeta: metav1.TypeMeta{
APIVersion: "v1",
Kind: "Pod",
},
ObjectMeta: metav1.ObjectMeta{
Name: "kube-apiserver",
Namespace: "kube-system",
Annotations: b.buildAnnotations(),
Labels: map[string]string{
"k8s-app": "kube-apiserver",
},
},
Spec: v1.PodSpec{
HostNetwork: true,
},
}
useHealthcheckProxy := b.findHealthcheckManifest() != nil
livenessProbe := &v1.Probe{
ProbeHandler: v1.ProbeHandler{
HTTPGet: &v1.HTTPGetAction{
Host: "127.0.0.1",
Path: "/livez",
Port: intstr.FromInt(wellknownports.KubeAPIServerHealthCheck),
},
},
InitialDelaySeconds: 10,
TimeoutSeconds: 15,
FailureThreshold: 8,
PeriodSeconds: 10,
}
readinessProbe := &v1.Probe{
ProbeHandler: v1.ProbeHandler{
HTTPGet: &v1.HTTPGetAction{
Host: "127.0.0.1",
Path: "/healthz",
Port: intstr.FromInt(wellknownports.KubeAPIServerHealthCheck),
},
},
InitialDelaySeconds: 0,
TimeoutSeconds: 15,
FailureThreshold: 3,
PeriodSeconds: 1,
}
startupProbe := &v1.Probe{
ProbeHandler: v1.ProbeHandler{
HTTPGet: &v1.HTTPGetAction{
Host: "127.0.0.1",
Path: "/livez",
Port: intstr.FromInt(wellknownports.KubeAPIServerHealthCheck),
},
},
InitialDelaySeconds: 10,
TimeoutSeconds: 5 * 60,
FailureThreshold: 5 * 60 / 10,
PeriodSeconds: 10,
}
allProbes := []*v1.Probe{
startupProbe,
livenessProbe,
readinessProbe,
}
insecurePort := fi.ValueOf(kubeAPIServer.InsecurePort)
if useHealthcheckProxy {
// kube-apiserver-healthcheck sidecar container runs on port 3990
} else if insecurePort != 0 {
for _, probe := range allProbes {
probe.HTTPGet.Port = intstr.FromInt(int(insecurePort))
}
} else if kubeAPIServer.SecurePort != 0 {
for _, probe := range allProbes {
probe.HTTPGet.Port = intstr.FromInt(int(kubeAPIServer.SecurePort))
probe.HTTPGet.Scheme = v1.URISchemeHTTPS
}
}
if b.IsKubernetesLT("1.31") {
// Compatibility: Use the old healthz probe for older clusters
for _, probe := range allProbes {
probe.HTTPGet.Path = "/healthz"
}
// Compatibility: Don't use startup probe / readiness probe
startupProbe = nil
readinessProbe = nil
// Compatibility: use old livenessProbe values
livenessProbe.FailureThreshold = 0
livenessProbe.PeriodSeconds = 0
livenessProbe.InitialDelaySeconds = 45
}
resourceRequests := v1.ResourceList{}
resourceLimits := v1.ResourceList{}
cpuRequest := resource.MustParse("150m")
if kubeAPIServer.CPURequest != nil {
cpuRequest = *kubeAPIServer.CPURequest
}
resourceRequests["cpu"] = cpuRequest
if kubeAPIServer.CPULimit != nil {
resourceLimits["cpu"] = *kubeAPIServer.CPULimit
}
if kubeAPIServer.MemoryRequest != nil {
resourceRequests["memory"] = *kubeAPIServer.MemoryRequest
}
if kubeAPIServer.MemoryLimit != nil {
resourceLimits["memory"] = *kubeAPIServer.MemoryLimit
}
image := b.RemapImage(kubeAPIServer.Image)
container := &v1.Container{
Name: "kube-apiserver",
Image: image,
Env: append(kubeAPIServer.Env, proxy.GetProxyEnvVars(b.NodeupConfig.Networking.EgressProxy)...),
LivenessProbe: livenessProbe,
ReadinessProbe: readinessProbe,
StartupProbe: startupProbe,
Ports: []v1.ContainerPort{
{
Name: "https",
ContainerPort: kubeAPIServer.SecurePort,
HostPort: kubeAPIServer.SecurePort,
},
},
Resources: v1.ResourceRequirements{
Requests: resourceRequests,
Limits: resourceLimits,
},
}
if insecurePort != 0 {
container.Ports = append(container.Ports, v1.ContainerPort{
Name: "local",
ContainerPort: insecurePort,
HostPort: insecurePort,
})
}
// Log both to docker and to the logfile
kubemanifest.AddHostPathMapping(pod, container, "logfile", "/var/log/kube-apiserver.log", kubemanifest.WithReadWrite())
// We use lighter containers that don't include shells
// But they have richer logging support via klog
{
container.Command = []string{"/go-runner"}
container.Args = []string{
"--log-file=/var/log/kube-apiserver.log",
"--also-stdout",
"/usr/local/bin/kube-apiserver",
}
container.Args = append(container.Args, sortedStrings(flags)...)
for _, issuer := range kubeAPIServer.AdditionalServiceAccountIssuers {
container.Args = append(container.Args, "--service-account-issuer="+issuer)
}
}
for _, path := range b.SSLHostPaths() {
name := strings.ReplaceAll(path, "/", "")
kubemanifest.AddHostPathMapping(pod, container, name, path)
}
kubemanifest.AddHostPathMapping(pod, container, "cloudconfig", InTreeCloudConfigFilePath)
kubemanifest.AddHostPathMapping(pod, container, "kubernetesca", filepath.Join(b.PathSrvKubernetes(), "ca.crt"))
pathSrvKAPI := filepath.Join(b.PathSrvKubernetes(), "kube-apiserver")
kubemanifest.AddHostPathMapping(pod, container, "srvkapi", pathSrvKAPI)
pathSrvSshproxy := b.PathSrvSshproxy()
if pathSrvSshproxy != "" {
kubemanifest.AddHostPathMapping(pod, container, "srvsshproxy", pathSrvSshproxy)
}
auditLogPath := fi.ValueOf(kubeAPIServer.AuditLogPath)
// Don't mount a volume if the mount path is set to '-' for stdout logging
// See https://kubernetes.io/docs/tasks/debug-application-cluster/audit/#audit-backends
if auditLogPath != "" && auditLogPath != "-" {
// Mount the directory of the path instead, as kube-apiserver rotates the log by renaming the file.
// Renaming is not possible when the file is mounted as the host path, and will return a
// 'Device or resource busy' error
auditLogPathDir := filepath.Dir(auditLogPath)
kubemanifest.AddHostPathMapping(pod, container, "auditlogpathdir", auditLogPathDir, kubemanifest.WithReadWrite())
}
if kubeAPIServer.AuditPolicyFile != "" {
// The audit config dir will be used for both the audit policy and the audit webhook config
auditConfigDir := filepath.Dir(kubeAPIServer.AuditPolicyFile)
if pathSrvKAPI != auditConfigDir {
kubemanifest.AddHostPathMapping(pod, container, "auditconfigdir", auditConfigDir)
}
}
if b.NodeupConfig.APIServerConfig.Authentication != nil {
if b.NodeupConfig.APIServerConfig.Authentication.Kopeio != nil || b.NodeupConfig.APIServerConfig.Authentication.AWS != nil {
kubemanifest.AddHostPathMapping(pod, container, "authn-config", PathAuthnConfig)
}
}
pod.Spec.Containers = append(pod.Spec.Containers, *container)
kubemanifest.MarkPodAsCritical(pod)
kubemanifest.MarkPodAsClusterCritical(pod)
kubemanifest.AddHostPathSELinuxContext(pod, b.NodeupConfig)
if useHealthcheckProxy {
if err := b.addHealthcheckSidecar(ctx, pod); err != nil {
return nil, err
}
}
return pod, nil
}
func (b *KubeAPIServerBuilder) buildAnnotations() map[string]string {
annotations := make(map[string]string)
annotations["kubectl.kubernetes.io/default-container"] = "kube-apiserver"
if b.NodeupConfig.UsesNoneDNS {
return annotations
}
if b.NodeupConfig.APIServerConfig.API.LoadBalancer == nil || !b.NodeupConfig.APIServerConfig.API.LoadBalancer.UseForInternalAPI {
annotations["dns.alpha.kubernetes.io/internal"] = b.APIInternalName()
}
if b.NodeupConfig.APIServerConfig.API.DNS != nil && b.NodeupConfig.APIServerConfig.API.PublicName != "" {
annotations["dns.alpha.kubernetes.io/external"] = b.NodeupConfig.APIServerConfig.API.PublicName
}
return annotations
}