kops/pkg/commands/toolbox_enroll.go

979 lines
26 KiB
Go

/*
Copyright 2023 The Kubernetes Authors.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
package commands
import (
"bytes"
"context"
cryptorand "crypto/rand"
"encoding/hex"
"errors"
"fmt"
"io"
"io/fs"
"net"
"os"
"path"
"sort"
"strconv"
"strings"
"golang.org/x/crypto/ssh"
"golang.org/x/crypto/ssh/agent"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/apimachinery/pkg/runtime"
"k8s.io/client-go/rest"
"k8s.io/klog/v2"
"sigs.k8s.io/controller-runtime/pkg/client"
"sigs.k8s.io/yaml"
"k8s.io/kops/pkg/apis/kops"
"k8s.io/kops/pkg/apis/kops/v1alpha2"
"k8s.io/kops/pkg/apis/nodeup"
"k8s.io/kops/pkg/assets"
"k8s.io/kops/pkg/client/simple"
"k8s.io/kops/pkg/commands/commandutils"
"k8s.io/kops/pkg/featureflag"
"k8s.io/kops/pkg/kubeconfig"
"k8s.io/kops/pkg/model"
"k8s.io/kops/pkg/model/resources"
"k8s.io/kops/pkg/nodemodel"
"k8s.io/kops/pkg/wellknownservices"
"k8s.io/kops/upup/pkg/fi"
"k8s.io/kops/upup/pkg/fi/cloudup"
"k8s.io/kops/util/pkg/vfs"
)
type ToolboxEnrollOptions struct {
ClusterName string
InstanceGroup string
Host string
SSHUser string
SSHPort int
// BuildHost is a flag to only build the host resource, don't apply it or enroll the node
BuildHost bool
// PodCIDRs is the list of IP Address ranges to use for pods that run on this node
PodCIDRs []string
kubeconfig.CreateKubecfgOptions
}
func (o *ToolboxEnrollOptions) InitDefaults() {
o.SSHUser = "root"
o.SSHPort = 22
}
func RunToolboxEnroll(ctx context.Context, f commandutils.Factory, out io.Writer, options *ToolboxEnrollOptions) error {
if !featureflag.Metal.Enabled() {
return fmt.Errorf("bare-metal support requires the Metal feature flag to be enabled")
}
if options.ClusterName == "" {
return fmt.Errorf("cluster is required")
}
if options.InstanceGroup == "" {
return fmt.Errorf("instance-group is required")
}
if options.Host == "" {
// Technically we could build the host resource without the PKI, but this isn't the case we are targeting right now.
return fmt.Errorf("host is required")
}
clientset, err := f.KopsClient()
if err != nil {
return err
}
configBuilder := &ConfigBuilder{
Clientset: clientset,
ClusterName: options.ClusterName,
InstanceGroupName: options.InstanceGroup,
}
fullCluster, err := configBuilder.GetFullCluster(ctx)
if err != nil {
return err
}
// Enroll the node over SSH.
restConfig, err := f.RESTConfig(ctx, fullCluster, options.CreateKubecfgOptions)
if err != nil {
return err
}
sudo := true
if options.SSHUser == "root" {
sudo = false
}
sshTarget, err := NewSSHHost(ctx, options.Host, options.SSHPort, options.SSHUser, sudo)
if err != nil {
return err
}
defer sshTarget.Close()
hostData, err := buildHostData(ctx, sshTarget, options)
if err != nil {
return err
}
if options.BuildHost {
klog.Infof("building host data for %+v", hostData)
b, err := yaml.Marshal(hostData)
if err != nil {
return fmt.Errorf("error marshalling host data: %w", err)
}
fmt.Fprintf(out, "%s\n", string(b))
return nil
}
fullInstanceGroup, err := configBuilder.GetFullInstanceGroup(ctx)
if err != nil {
return err
}
bootstrapData, err := configBuilder.GetBootstrapData(ctx)
if err != nil {
return err
}
if err := enrollHost(ctx, fullInstanceGroup, bootstrapData, restConfig, hostData, sshTarget); err != nil {
return err
}
return nil
}
// buildHostData builds an instance of the Host CRD, based on information in the options and by SSHing to the target host.
func buildHostData(ctx context.Context, sshTarget *SSHHost, options *ToolboxEnrollOptions) (*v1alpha2.Host, error) {
publicKeyPath := "/etc/kubernetes/kops/pki/machine/public.pem"
publicKeyBytes, err := sshTarget.readFile(ctx, publicKeyPath)
if err != nil {
if errors.Is(err, fs.ErrNotExist) {
publicKeyBytes = nil
} else {
return nil, fmt.Errorf("error reading public key %q: %w", publicKeyPath, err)
}
}
// Create the key if it doesn't exist
publicKeyBytes = bytes.TrimSpace(publicKeyBytes)
if len(publicKeyBytes) == 0 {
if _, err := sshTarget.runScript(ctx, scriptCreateKey, ExecOptions{Echo: true}); err != nil {
return nil, err
}
b, err := sshTarget.readFile(ctx, publicKeyPath)
if err != nil {
return nil, fmt.Errorf("error reading public key %q (after creation): %w", publicKeyPath, err)
}
publicKeyBytes = b
}
klog.Infof("public key is %s", string(publicKeyBytes))
hostname, err := sshTarget.getHostname(ctx)
if err != nil {
return nil, err
}
host := &v1alpha2.Host{}
host.SetGroupVersionKind(v1alpha2.SchemeGroupVersion.WithKind("Host"))
host.Namespace = "kops-system"
host.Name = hostname
host.Spec.InstanceGroup = options.InstanceGroup
host.Spec.PublicKey = string(publicKeyBytes)
host.Spec.PodCIDRs = options.PodCIDRs
return host, nil
}
func enrollHost(ctx context.Context, ig *kops.InstanceGroup, bootstrapData *BootstrapData, restConfig *rest.Config, hostData *v1alpha2.Host, sshTarget *SSHHost) error {
scheme := runtime.NewScheme()
if err := v1alpha2.AddToScheme(scheme); err != nil {
return fmt.Errorf("building kubernetes scheme: %w", err)
}
kubeClient, err := client.New(restConfig, client.Options{
Scheme: scheme,
})
if err != nil {
return fmt.Errorf("building kubernetes client: %w", err)
}
// We can't create the host resource in the API server for control-plane nodes,
// because the API server (likely) isn't running yet.
if !ig.IsControlPlane() {
if err := kubeClient.Create(ctx, hostData); err != nil {
return fmt.Errorf("failed to create host %s/%s: %w", hostData.Namespace, hostData.Name, err)
}
}
for k, v := range bootstrapData.NodeupScriptAdditionalFiles {
if err := sshTarget.writeFile(ctx, k, bytes.NewReader(v)); err != nil {
return fmt.Errorf("writing file %q over SSH: %w", k, err)
}
}
if len(bootstrapData.NodeupScript) != 0 {
if _, err := sshTarget.runScript(ctx, string(bootstrapData.NodeupScript), ExecOptions{Echo: true}); err != nil {
return err
}
}
return nil
}
const scriptCreateKey = `
#!/bin/bash
set -o errexit
set -o nounset
set -o pipefail
set -x
DIR=/etc/kubernetes/kops/pki/machine/
mkdir -p ${DIR}
if [[ ! -f "${DIR}/private.pem" ]]; then
openssl ecparam -name prime256v1 -genkey -noout -out "${DIR}/private.pem"
fi
if [[ ! -f "${DIR}/public.pem" ]]; then
openssl ec -in "${DIR}/private.pem" -pubout -out "${DIR}/public.pem"
fi
`
// SSHHost is a wrapper around an SSH connection to a host machine.
type SSHHost struct {
hostname string
sshClient *ssh.Client
sudo bool
}
// Close closes the connection.
func (s *SSHHost) Close() error {
if s.sshClient != nil {
if err := s.sshClient.Close(); err != nil {
return err
}
s.sshClient = nil
}
return nil
}
// NewSSHHost creates a new SSHHost.
func NewSSHHost(ctx context.Context, host string, sshPort int, sshUser string, sudo bool) (*SSHHost, error) {
socket := os.Getenv("SSH_AUTH_SOCK")
if socket == "" {
return nil, fmt.Errorf("cannot connect to SSH agent; SSH_AUTH_SOCK env variable not set")
}
conn, err := net.Dial("unix", socket)
if err != nil {
return nil, fmt.Errorf("failed to connect to SSH agent with SSH_AUTH_SOCK %q: %w", socket, err)
}
agentClient := agent.NewClient(conn)
signers, err := agentClient.Signers()
if err != nil {
_ = conn.Close()
return nil, fmt.Errorf("failed to get signers: %w", err)
}
if len(signers) == 0 {
return nil, fmt.Errorf("SSH agent has no keys")
}
sshConfig := &ssh.ClientConfig{
HostKeyCallback: func(hostname string, remote net.Addr, key ssh.PublicKey) error {
klog.Warningf("accepting SSH key %v for %q", key, hostname)
return nil
},
Auth: []ssh.AuthMethod{
// Use a callback rather than PublicKeys so we only consult the
// agent once the remote server wants it.
ssh.PublicKeysCallback(agentClient.Signers),
},
User: sshUser,
}
sshClient, err := ssh.Dial("tcp", host+":"+strconv.Itoa(sshPort), sshConfig)
if err != nil {
return nil, fmt.Errorf("failed to SSH to %q (with user %q): %w", host, sshUser, err)
}
return &SSHHost{
hostname: host,
sshClient: sshClient,
sudo: sudo,
}, nil
}
func (s *SSHHost) readFile(ctx context.Context, path string) ([]byte, error) {
p := vfs.NewSSHPath(s.sshClient, s.hostname, path, s.sudo)
return p.ReadFile(ctx)
}
func (s *SSHHost) writeFile(ctx context.Context, path string, data io.ReadSeeker) error {
p := vfs.NewSSHPath(s.sshClient, s.hostname, path, s.sudo)
return p.WriteFile(ctx, data, nil)
}
func (s *SSHHost) runScript(ctx context.Context, script string, options ExecOptions) (*CommandOutput, error) {
var tempDir string
{
b := make([]byte, 32)
if _, err := cryptorand.Read(b); err != nil {
return nil, fmt.Errorf("error getting random data: %w", err)
}
tempDir = path.Join("/tmp", hex.EncodeToString(b))
}
scriptPath := path.Join(tempDir, "script.sh")
p := vfs.NewSSHPath(s.sshClient, s.hostname, scriptPath, s.sudo)
defer func() {
if _, err := s.runCommand(ctx, "rm -rf "+tempDir, ExecOptions{Echo: false}); err != nil {
klog.Warningf("error cleaning up temp directory %q: %v", tempDir, err)
}
}()
if err := p.WriteFile(ctx, bytes.NewReader([]byte(script)), nil); err != nil {
return nil, fmt.Errorf("error writing script to SSH target: %w", err)
}
scriptCommand := "/bin/bash " + scriptPath
return s.runCommand(ctx, scriptCommand, options)
}
// CommandOutput holds the results of running a command.
type CommandOutput struct {
Stdout bytes.Buffer
Stderr bytes.Buffer
}
// ExecOptions holds options for running a command remotely.
type ExecOptions struct {
Echo bool
}
func (s *SSHHost) runCommand(ctx context.Context, command string, options ExecOptions) (*CommandOutput, error) {
session, err := s.sshClient.NewSession()
if err != nil {
return nil, fmt.Errorf("failed to start SSH session: %w", err)
}
defer session.Close()
output := &CommandOutput{}
session.Stdout = &output.Stdout
session.Stderr = &output.Stderr
if options.Echo {
// We send both to stderr, so we don't "corrupt" stdout
session.Stdout = io.MultiWriter(os.Stderr, session.Stdout)
session.Stderr = io.MultiWriter(os.Stderr, session.Stderr)
}
if s.sudo {
command = "sudo " + command
}
if err := session.Run(command); err != nil {
return output, fmt.Errorf("error running command %q: %w", command, err)
}
return output, nil
}
// getHostname gets the hostname of the SSH target.
// This is used as the node name when registering the node.
func (s *SSHHost) getHostname(ctx context.Context) (string, error) {
output, err := s.runCommand(ctx, "hostname", ExecOptions{Echo: true})
if err != nil {
return "", fmt.Errorf("failed to get hostname: %w", err)
}
hostname := output.Stdout.String()
hostname = strings.TrimSpace(hostname)
if len(hostname) == 0 {
return "", fmt.Errorf("hostname was empty")
}
return hostname, nil
}
type BootstrapData struct {
// NodeupScript is a script that can be used to bootstrap the node.
NodeupScript []byte
// NodeupConfig is structured configuration, provided by kops-controller (for example).
NodeupConfig *nodeup.Config
// NodeupScriptAdditionalFiles are additional files that are needed by the nodeup script.
NodeupScriptAdditionalFiles map[string][]byte
}
// ConfigBuilder builds bootstrap configuration for a node.
type ConfigBuilder struct {
// ClusterName is the name of the cluster to build.
// Required.
ClusterName string
// InstanceGroupName is the name of the InstanceGroup we are building configuration for.
// Required.
InstanceGroupName string
// Clientset is the clientset to use to query for clusters / instancegroups etc
// Required.
Clientset simple.Clientset
// Cloud is the cloud implementation
// Use GetCloud to read and auto-populate.
Cloud fi.Cloud
// AssetBuilder holds the assets used by the cluster.
// Use GetAssetBuilder to read and auto-populate.
AssetBuilder *assets.AssetBuilder
// Cluster holds the (unexpanded) cluster configuration.
// Use GetCluster to read and auto-populate.
Cluster *kops.Cluster
// InstanceGroup holds the (unexpanded) instance group configuration.
// Use GetInstanceGroup to read and auto-populate.
InstanceGroup *kops.InstanceGroup
// instanceGroups holds the (unexpanded) instance group configurations
// Use GetInstanceGroups to read and auto-populate
instanceGroups *kops.InstanceGroupList
// wellKnownAddresses holds the known IP/host endpoints for the cluster.
// Use GetWellKnownAddresses to read and auto-populate.
wellKnownAddresses *model.WellKnownAddresses
// fullCluster holds the fully-expanded cluster configuration
// Use GetFullCluster to read and auto-populate.
fullCluster *kops.Cluster
// fullInstanceGroup holds the fully-expanded instance group configuration
// Use GetFullIntsanceGroup to read and auto-populate.
fullInstanceGroup *kops.InstanceGroup
// bootstrapData holds the final computed bootstrap configuration.
// Use GetBootstrapData to read and auto-populate.
bootstrapData *BootstrapData
}
func (b *ConfigBuilder) GetClientset(ctx context.Context) (simple.Clientset, error) {
if b.Clientset != nil {
return b.Clientset, nil
}
return nil, fmt.Errorf("clientset is required")
}
func (b *ConfigBuilder) GetFullCluster(ctx context.Context) (*kops.Cluster, error) {
if b.fullCluster != nil {
return b.fullCluster, nil
}
clientset, err := b.GetClientset(ctx)
if err != nil {
return nil, err
}
cluster, err := b.GetCluster(ctx)
if err != nil {
return nil, err
}
cloud, err := b.GetCloud(ctx)
if err != nil {
return nil, err
}
assetBuilder, err := b.GetAssetBuilder(ctx)
if err != nil {
return nil, err
}
instanceGroupList, err := b.GetInstanceGroups(ctx)
if err != nil {
return nil, err
}
var instanceGroups []*kops.InstanceGroup
for i := range instanceGroupList.Items {
instanceGroup := &instanceGroupList.Items[i]
instanceGroups = append(instanceGroups, instanceGroup)
}
fullCluster, err := cloudup.PopulateClusterSpec(ctx, clientset, cluster, instanceGroups, cloud, assetBuilder)
if err != nil {
return nil, fmt.Errorf("building full cluster spec: %w", err)
}
b.fullCluster = fullCluster
return fullCluster, nil
}
func (b *ConfigBuilder) GetInstanceGroup(ctx context.Context) (*kops.InstanceGroup, error) {
if b.InstanceGroup != nil {
return b.InstanceGroup, nil
}
instanceGroups, err := b.GetInstanceGroups(ctx)
if err != nil {
return nil, err
}
if b.InstanceGroupName == "" {
return nil, fmt.Errorf("InstanceGroup name is missing")
}
// Build full IG spec to ensure we end up with a valid IG
for i := range instanceGroups.Items {
ig := &instanceGroups.Items[i]
if ig.Name == b.InstanceGroupName {
b.InstanceGroup = ig
return ig, nil
}
}
return nil, fmt.Errorf("instance group %q not found", b.InstanceGroupName)
}
func (b *ConfigBuilder) GetFullInstanceGroup(ctx context.Context) (*kops.InstanceGroup, error) {
if b.fullInstanceGroup != nil {
return b.fullInstanceGroup, nil
}
clientset, err := b.GetClientset(ctx)
if err != nil {
return nil, err
}
fullCluster, err := b.GetFullCluster(ctx)
if err != nil {
return nil, err
}
cloud, err := b.GetCloud(ctx)
if err != nil {
return nil, err
}
ig, err := b.GetInstanceGroup(ctx)
if err != nil {
return nil, err
}
channel, err := cloudup.ChannelForCluster(clientset.VFSContext(), fullCluster)
if err != nil {
return nil, fmt.Errorf("getting channel for cluster %q: %w", fullCluster.Name, err)
}
// Build full IG spec to ensure we end up with a valid IG
fullInstanceGroup, err := cloudup.PopulateInstanceGroupSpec(fullCluster, ig, cloud, channel)
if err != nil {
return nil, err
}
b.fullInstanceGroup = fullInstanceGroup
return fullInstanceGroup, nil
}
func (b *ConfigBuilder) GetCloud(ctx context.Context) (fi.Cloud, error) {
if b.Cloud != nil {
return b.Cloud, nil
}
cluster, err := b.GetCluster(ctx)
if err != nil {
return nil, err
}
cloud, err := cloudup.BuildCloud(cluster)
if err != nil {
return nil, err
}
b.Cloud = cloud
return cloud, nil
}
func (b *ConfigBuilder) GetInstanceGroups(ctx context.Context) (*kops.InstanceGroupList, error) {
if b.instanceGroups != nil {
return b.instanceGroups, nil
}
cluster, err := b.GetCluster(ctx)
if err != nil {
return nil, err
}
clientset, err := b.GetClientset(ctx)
if err != nil {
return nil, err
}
instanceGroupList, err := clientset.InstanceGroupsFor(cluster).List(ctx, metav1.ListOptions{})
if err != nil {
return nil, fmt.Errorf("reading instance groups: %w", err)
}
b.instanceGroups = instanceGroupList
return instanceGroupList, nil
}
func (b *ConfigBuilder) GetCluster(ctx context.Context) (*kops.Cluster, error) {
if b.Cluster != nil {
return b.Cluster, nil
}
if b.ClusterName == "" {
return nil, fmt.Errorf("ClusterName is missing")
}
clientset, err := b.GetClientset(ctx)
if err != nil {
return nil, err
}
cluster, err := clientset.GetCluster(ctx, b.ClusterName)
if err != nil {
return nil, err
}
if cluster == nil {
return nil, fmt.Errorf("cluster %q not found", b.ClusterName)
}
b.Cluster = cluster
return cluster, nil
}
func (b *ConfigBuilder) GetAssetBuilder(ctx context.Context) (*assets.AssetBuilder, error) {
if b.AssetBuilder != nil {
return b.AssetBuilder, nil
}
clientset, err := b.GetClientset(ctx)
if err != nil {
return nil, err
}
cluster, err := b.GetCluster(ctx)
if err != nil {
return nil, err
}
cloud, err := b.GetCloud(ctx)
if err != nil {
return nil, err
}
// ApplyClusterCmd is get the assets.
// We use DryRun and GetAssets to do this without applying any changes.
apply := &cloudup.ApplyClusterCmd{
Cloud: cloud,
Cluster: cluster,
Clientset: clientset,
DryRun: true,
GetAssets: true,
TargetName: cloudup.TargetDryRun,
}
applyResults, err := apply.Run(ctx)
if err != nil {
return nil, fmt.Errorf("error during apply: %w", err)
}
b.AssetBuilder = applyResults.AssetBuilder
return b.AssetBuilder, nil
}
func (b *ConfigBuilder) GetWellKnownAddresses(ctx context.Context) (model.WellKnownAddresses, error) {
if b.wellKnownAddresses != nil {
return *b.wellKnownAddresses, nil
}
cloud, err := b.GetCloud(ctx)
if err != nil {
return nil, err
}
fullCluster, err := b.GetFullCluster(ctx)
if err != nil {
return nil, err
}
// Determine the well-known addresses for the cluster.
wellKnownAddresses := make(model.WellKnownAddresses)
{
ingresses, err := cloud.GetApiIngressStatus(fullCluster)
if err != nil {
return nil, fmt.Errorf("error getting ingress status: %v", err)
}
for _, ingress := range ingresses {
// TODO: Do we need to support hostnames?
// if ingress.Hostname != "" {
// apiserverAdditionalIPs = append(apiserverAdditionalIPs, ingress.Hostname)
// }
if ingress.IP != "" {
wellKnownAddresses[wellknownservices.KubeAPIServer] = append(wellKnownAddresses[wellknownservices.KubeAPIServer], ingress.IP)
}
}
}
if len(wellKnownAddresses[wellknownservices.KubeAPIServer]) == 0 {
// TODO: Should we support DNS?
return nil, fmt.Errorf("unable to determine IP address for kube-apiserver")
}
for k := range wellKnownAddresses {
sort.Strings(wellKnownAddresses[k])
}
b.wellKnownAddresses = &wellKnownAddresses
return wellKnownAddresses, nil
}
func (b *ConfigBuilder) GetBootstrapData(ctx context.Context) (*BootstrapData, error) {
if b.bootstrapData != nil {
return b.bootstrapData, nil
}
cluster, err := b.GetFullCluster(ctx)
if err != nil {
return nil, err
}
assetBuilder, err := b.GetAssetBuilder(ctx)
if err != nil {
return nil, err
}
clientset, err := b.GetClientset(ctx)
if err != nil {
return nil, err
}
wellKnownAddresses, err := b.GetWellKnownAddresses(ctx)
if err != nil {
return nil, err
}
ig, err := b.GetFullInstanceGroup(ctx)
if err != nil {
return nil, err
}
bootstrapData := &BootstrapData{}
bootstrapData.NodeupScriptAdditionalFiles = make(map[string][]byte)
encryptionConfigSecretHash := ""
// TODO: Support encryption config?
// if fi.ValueOf(c.Cluster.Spec.EncryptionConfig) {
// secret, err := secretStore.FindSecret("encryptionconfig")
// if err != nil {
// return fmt.Errorf("could not load encryptionconfig secret: %v", err)
// }
// if secret == nil {
// fmt.Println("")
// fmt.Println("You have encryptionConfig enabled, but no encryptionconfig secret has been set.")
// fmt.Println("See `kops create secret encryptionconfig -h` and https://kubernetes.io/docs/tasks/administer-cluster/encrypt-data/")
// return fmt.Errorf("could not find encryptionconfig secret")
// }
// hashBytes := sha256.Sum256(secret.Data)
// encryptionConfigSecretHash = base64.URLEncoding.EncodeToString(hashBytes[:])
nodeUpAssets, err := nodemodel.BuildNodeUpAssets(ctx, assetBuilder)
if err != nil {
return nil, err
}
configBuilder, err := nodemodel.NewNodeUpConfigBuilder(cluster, assetBuilder, encryptionConfigSecretHash)
if err != nil {
return nil, err
}
keysets := make(map[string]*fi.Keyset)
keystore, err := clientset.KeyStore(cluster)
if err != nil {
return nil, err
}
keyNames := model.KeypairNamesForInstanceGroup(cluster, ig)
for _, keyName := range keyNames {
keyset, err := keystore.FindKeyset(ctx, keyName)
if err != nil {
return nil, fmt.Errorf("getting keyset %q: %w", keyName, err)
}
if keyset == nil {
return nil, fmt.Errorf("did not find keyset %q", keyName)
}
keysets[keyName] = keyset
}
nodeupConfig, bootConfig, err := configBuilder.BuildConfig(ig, wellKnownAddresses, keysets)
if err != nil {
return nil, err
}
var nodeupScript resources.NodeUpScript
nodeupScript.NodeUpAssets = nodeUpAssets.NodeUpAssets
nodeupScript.BootConfig = bootConfig
nodeupScript.WithEnvironmentVariables(cluster, ig)
nodeupScript.WithProxyEnv(cluster)
nodeupScript.WithSysctls()
nodeupScript.CloudProvider = string(cluster.GetCloudProvider())
bootConfig.ConfigBase = fi.PtrTo("file:///etc/kubernetes/kops/config")
nodeupScriptResource, err := nodeupScript.Build()
if err != nil {
return nil, err
}
vfsContext := clientset.VFSContext()
// If this is the control plane, we want to copy the config from s3/gcs to the local file system on the target node,
// so that we don't need credentials to the state store.
if bootConfig.InstanceGroupRole == kops.InstanceGroupRoleControlPlane {
remapPrefix := "s3://" // TODO: Support GCS?
// targetDir is the location of the config on the target node.
targetDir := "/etc/kubernetes/kops/config"
// remapFile remaps a file from s3/gcs etc to the local file system on the target node.
remapFile := func(pSrc *string, destDir string) error {
src := *pSrc
if !strings.HasPrefix(src, remapPrefix) {
return nil
}
srcPath, err := vfsContext.BuildVfsPath(src)
if err != nil {
return fmt.Errorf("building vfs path: %w", err)
}
b, err := srcPath.ReadFile(ctx)
if err != nil {
return fmt.Errorf("reading file: %w", err)
}
dest := strings.TrimPrefix(src, remapPrefix)
dest = path.Join(destDir, dest)
bootstrapData.NodeupScriptAdditionalFiles[dest] = b
*pSrc = dest
return nil
}
// remapTree remaps a file tree from s3/gcs etc to the local file system on the target node.
remapTree := func(pSrc *string, dest string) error {
src := *pSrc
if !strings.HasPrefix(src, remapPrefix) {
return nil
}
srcPath, err := vfsContext.BuildVfsPath(src)
if err != nil {
return fmt.Errorf("building vfs path: %w", err)
}
srcFiles, err := srcPath.ReadTree(ctx)
if err != nil {
return fmt.Errorf("reading tree: %w", err)
}
basePath := srcPath.Path()
for _, srcFile := range srcFiles {
b, err := srcFile.ReadFile(ctx)
if err != nil {
return fmt.Errorf("reading file: %w", err)
}
if !strings.HasPrefix(srcFile.Path(), basePath) {
return fmt.Errorf("unexpected path: %q", srcFile.Path())
}
relativePath := strings.TrimPrefix(srcFile.Path(), basePath)
bootstrapData.NodeupScriptAdditionalFiles[path.Join(dest, relativePath)] = b
}
*pSrc = dest
return nil
}
for i := range nodeupConfig.EtcdManifests {
if err := remapFile(&nodeupConfig.EtcdManifests[i], path.Join(targetDir)); err != nil {
return nil, err
}
}
configBase, err := vfs.Context.BuildVfsPath(cluster.Spec.ConfigStore.Base)
if err != nil {
return nil, fmt.Errorf("parsing configStore.base %q: %w", cluster.Spec.ConfigStore.Base, err)
}
for i, channel := range nodeupConfig.Channels {
bootstrapChannelPath := configBase.Join("addons", "bootstrap-channel.yaml").Path()
if channel != bootstrapChannelPath {
klog.Infof("not remapping non-bootstrap channel %q", channel)
continue
}
parentPath := configBase.Join("addons").Path()
if err := remapTree(&parentPath, path.Join(targetDir, "addons")); err != nil {
return nil, err
}
nodeupConfig.Channels[i] = path.Join(parentPath, "bootstrap-channel.yaml")
// The channels tool requires a file:// prefix
if strings.HasPrefix(nodeupConfig.Channels[i], "/") {
nodeupConfig.Channels[i] = "file://" + nodeupConfig.Channels[i]
}
}
if nodeupConfig.ConfigStore != nil {
if err := remapTree(&nodeupConfig.ConfigStore.Keypairs, path.Join(targetDir, "pki/etcd")); err != nil {
return nil, err
}
if err := remapTree(&nodeupConfig.ConfigStore.Secrets, path.Join(targetDir, "pki")); err != nil {
return nil, err
}
}
nodeupConfigBytes, err := yaml.Marshal(nodeupConfig)
if err != nil {
return nil, fmt.Errorf("error converting nodeup config to yaml: %w", err)
}
// Not much reason to hash this, since we're reading it from the local file system
// sum256 := sha256.Sum256(nodeupConfigBytes)
// bootConfig.NodeupConfigHash = base64.StdEncoding.EncodeToString(sum256[:])
p := path.Join(targetDir, "igconfig", bootConfig.InstanceGroupRole.ToLowerString(), ig.Name, "nodeupconfig.yaml")
bootstrapData.NodeupScriptAdditionalFiles[p] = nodeupConfigBytes
// Copy any static manifests we need on the control plane
for _, staticManifest := range assetBuilder.StaticManifests {
if !staticManifest.AppliesToRole(bootConfig.InstanceGroupRole) {
continue
}
p := path.Join(targetDir, staticManifest.Path)
bootstrapData.NodeupScriptAdditionalFiles[p] = staticManifest.Contents
}
}
nodeupScriptBytes, err := fi.ResourceAsBytes(nodeupScriptResource)
if err != nil {
return nil, err
}
bootstrapData.NodeupScript = nodeupScriptBytes
bootstrapData.NodeupConfig = nodeupConfig
b.bootstrapData = bootstrapData
return bootstrapData, nil
}