cas: remove unused simulator removal funcs
This commit is contained in:
		
							parent
							
								
									a9292351c3
								
							
						
					
					
						commit
						d0686d9023
					
				|  | @ -115,29 +115,6 @@ func NewRemovalSimulator(listers kube_util.ListerRegistry, clusterSnapshot clust | |||
| 	} | ||||
| } | ||||
| 
 | ||||
| // FindNodesToRemove finds nodes that can be removed.
 | ||||
| func (r *RemovalSimulator) FindNodesToRemove( | ||||
| 	candidates []string, | ||||
| 	destinations []string, | ||||
| 	timestamp time.Time, | ||||
| 	remainingPdbTracker pdb.RemainingPdbTracker, | ||||
| ) (nodesToRemove []NodeToBeRemoved, unremovableNodes []*UnremovableNode) { | ||||
| 	destinationMap := make(map[string]bool, len(destinations)) | ||||
| 	for _, destination := range destinations { | ||||
| 		destinationMap[destination] = true | ||||
| 	} | ||||
| 
 | ||||
| 	for _, nodeName := range candidates { | ||||
| 		rn, urn := r.SimulateNodeRemoval(nodeName, destinationMap, timestamp, remainingPdbTracker) | ||||
| 		if rn != nil { | ||||
| 			nodesToRemove = append(nodesToRemove, *rn) | ||||
| 		} else if urn != nil { | ||||
| 			unremovableNodes = append(unremovableNodes, urn) | ||||
| 		} | ||||
| 	} | ||||
| 	return nodesToRemove, unremovableNodes | ||||
| } | ||||
| 
 | ||||
| // SimulateNodeRemoval simulates removing a node from the cluster to check
 | ||||
| // whether it is possible to move its pods. Depending on
 | ||||
| // the outcome, exactly one of (NodeToBeRemoved, UnremovableNode) will be
 | ||||
|  | @ -178,24 +155,6 @@ func (r *RemovalSimulator) SimulateNodeRemoval( | |||
| 	}, nil | ||||
| } | ||||
| 
 | ||||
| // FindEmptyNodesToRemove finds empty nodes that can be removed.
 | ||||
| func (r *RemovalSimulator) FindEmptyNodesToRemove(candidates []string, timestamp time.Time) []string { | ||||
| 	result := make([]string, 0) | ||||
| 	for _, node := range candidates { | ||||
| 		nodeInfo, err := r.clusterSnapshot.GetNodeInfo(node) | ||||
| 		if err != nil { | ||||
| 			klog.Errorf("Can't retrieve node %s from snapshot, err: %v", node, err) | ||||
| 			continue | ||||
| 		} | ||||
| 		// Should block on all pods
 | ||||
| 		podsToRemove, _, _, err := GetPodsToMove(nodeInfo, r.deleteOptions, r.drainabilityRules, nil, nil, timestamp) | ||||
| 		if err == nil && len(podsToRemove) == 0 { | ||||
| 			result = append(result, node) | ||||
| 		} | ||||
| 	} | ||||
| 	return result | ||||
| } | ||||
| 
 | ||||
| func (r *RemovalSimulator) withForkedSnapshot(f func() error) (err error) { | ||||
| 	r.clusterSnapshot.Fork() | ||||
| 	defer func() { | ||||
|  |  | |||
|  | @ -1,267 +0,0 @@ | |||
| /* | ||||
| Copyright 2016 The Kubernetes Authors. | ||||
| 
 | ||||
| Licensed under the Apache License, Version 2.0 (the "License"); | ||||
| you may not use this file except in compliance with the License. | ||||
| You may obtain a copy of the License at | ||||
| 
 | ||||
|     http://www.apache.org/licenses/LICENSE-2.0
 | ||||
| 
 | ||||
| Unless required by applicable law or agreed to in writing, software | ||||
| distributed under the License is distributed on an "AS IS" BASIS, | ||||
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||||
| See the License for the specific language governing permissions and | ||||
| limitations under the License. | ||||
| */ | ||||
| 
 | ||||
| package simulator | ||||
| 
 | ||||
| import ( | ||||
| 	"fmt" | ||||
| 	"testing" | ||||
| 	"time" | ||||
| 
 | ||||
| 	"github.com/stretchr/testify/assert" | ||||
| 
 | ||||
| 	appsv1 "k8s.io/api/apps/v1" | ||||
| 	apiv1 "k8s.io/api/core/v1" | ||||
| 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||
| 	"k8s.io/autoscaler/cluster-autoscaler/simulator/clustersnapshot" | ||||
| 	"k8s.io/autoscaler/cluster-autoscaler/simulator/clustersnapshot/testsnapshot" | ||||
| 	"k8s.io/autoscaler/cluster-autoscaler/simulator/framework" | ||||
| 	"k8s.io/autoscaler/cluster-autoscaler/simulator/options" | ||||
| 	"k8s.io/autoscaler/cluster-autoscaler/utils/drain" | ||||
| 	kube_util "k8s.io/autoscaler/cluster-autoscaler/utils/kubernetes" | ||||
| 	. "k8s.io/autoscaler/cluster-autoscaler/utils/test" | ||||
| 	"k8s.io/kubernetes/pkg/kubelet/types" | ||||
| ) | ||||
| 
 | ||||
| func TestFindEmptyNodes(t *testing.T) { | ||||
| 	nodes := []*apiv1.Node{} | ||||
| 	nodeNames := []string{} | ||||
| 	for i := 0; i < 4; i++ { | ||||
| 		nodeName := fmt.Sprintf("n%d", i) | ||||
| 		node := BuildTestNode(nodeName, 1000, 2000000) | ||||
| 		SetNodeReadyState(node, true, time.Time{}) | ||||
| 		nodes = append(nodes, node) | ||||
| 		nodeNames = append(nodeNames, nodeName) | ||||
| 	} | ||||
| 
 | ||||
| 	pod1 := BuildTestPod("p1", 300, 500000) | ||||
| 	pod1.Spec.NodeName = "n1" | ||||
| 
 | ||||
| 	pod2 := BuildTestPod("p2", 300, 500000) | ||||
| 	pod2.Spec.NodeName = "n2" | ||||
| 	pod2.Annotations = map[string]string{ | ||||
| 		types.ConfigMirrorAnnotationKey: "", | ||||
| 	} | ||||
| 
 | ||||
| 	clusterSnapshot := testsnapshot.NewTestSnapshotOrDie(t) | ||||
| 	clustersnapshot.InitializeClusterSnapshotOrDie(t, clusterSnapshot, []*apiv1.Node{nodes[0], nodes[1], nodes[2], nodes[3]}, []*apiv1.Pod{pod1, pod2}) | ||||
| 	testTime := time.Date(2020, time.December, 18, 17, 0, 0, 0, time.UTC) | ||||
| 	r := NewRemovalSimulator(nil, clusterSnapshot, testDeleteOptions(), nil, false) | ||||
| 	emptyNodes := r.FindEmptyNodesToRemove(nodeNames, testTime) | ||||
| 	assert.Equal(t, []string{nodeNames[0], nodeNames[2], nodeNames[3]}, emptyNodes) | ||||
| } | ||||
| 
 | ||||
| type findNodesToRemoveTestConfig struct { | ||||
| 	name        string | ||||
| 	pods        []*apiv1.Pod | ||||
| 	allNodes    []*apiv1.Node | ||||
| 	candidates  []string | ||||
| 	toRemove    []NodeToBeRemoved | ||||
| 	unremovable []*UnremovableNode | ||||
| } | ||||
| 
 | ||||
| func TestFindNodesToRemove(t *testing.T) { | ||||
| 	emptyNode := BuildTestNode("n1", 1000, 2000000) | ||||
| 
 | ||||
| 	// two small pods backed by ReplicaSet
 | ||||
| 	drainableNode := BuildTestNode("n2", 1000, 2000000) | ||||
| 	drainableNodeInfo := framework.NewTestNodeInfo(drainableNode) | ||||
| 
 | ||||
| 	// one small pod, not backed by anything
 | ||||
| 	nonDrainableNode := BuildTestNode("n3", 1000, 2000000) | ||||
| 	nonDrainableNodeInfo := framework.NewTestNodeInfo(nonDrainableNode) | ||||
| 
 | ||||
| 	// one very large pod
 | ||||
| 	fullNode := BuildTestNode("n4", 1000, 2000000) | ||||
| 	fullNodeInfo := framework.NewTestNodeInfo(fullNode) | ||||
| 
 | ||||
| 	SetNodeReadyState(emptyNode, true, time.Time{}) | ||||
| 	SetNodeReadyState(drainableNode, true, time.Time{}) | ||||
| 	SetNodeReadyState(nonDrainableNode, true, time.Time{}) | ||||
| 	SetNodeReadyState(fullNode, true, time.Time{}) | ||||
| 
 | ||||
| 	replicas := int32(5) | ||||
| 	replicaSets := []*appsv1.ReplicaSet{ | ||||
| 		{ | ||||
| 			ObjectMeta: metav1.ObjectMeta{ | ||||
| 				Name:      "rs", | ||||
| 				Namespace: "default", | ||||
| 				SelfLink:  "api/v1/namespaces/default/replicasets/rs", | ||||
| 			}, | ||||
| 			Spec: appsv1.ReplicaSetSpec{ | ||||
| 				Replicas: &replicas, | ||||
| 			}, | ||||
| 		}, | ||||
| 	} | ||||
| 	rsLister, err := kube_util.NewTestReplicaSetLister(replicaSets) | ||||
| 	assert.NoError(t, err) | ||||
| 	registry := kube_util.NewListerRegistry(nil, nil, nil, nil, nil, nil, nil, rsLister, nil) | ||||
| 
 | ||||
| 	ownerRefs := GenerateOwnerReferences("rs", "ReplicaSet", "extensions/v1beta1", "") | ||||
| 
 | ||||
| 	pod1 := BuildTestPod("p1", 100, 100000) | ||||
| 	pod1.OwnerReferences = ownerRefs | ||||
| 	pod1.Spec.NodeName = "n2" | ||||
| 	drainableNodeInfo.AddPod(&framework.PodInfo{Pod: pod1}) | ||||
| 
 | ||||
| 	pod2 := BuildTestPod("p2", 100, 100000) | ||||
| 	pod2.OwnerReferences = ownerRefs | ||||
| 	pod2.Spec.NodeName = "n2" | ||||
| 	drainableNodeInfo.AddPod(&framework.PodInfo{Pod: pod2}) | ||||
| 
 | ||||
| 	pod3 := BuildTestPod("p3", 100, 100000) | ||||
| 	pod3.Spec.NodeName = "n3" | ||||
| 	nonDrainableNodeInfo.AddPod(&framework.PodInfo{Pod: pod3}) | ||||
| 
 | ||||
| 	pod4 := BuildTestPod("p4", 1000, 100000) | ||||
| 	pod4.Spec.NodeName = "n4" | ||||
| 	fullNodeInfo.AddPod(&framework.PodInfo{Pod: pod4}) | ||||
| 
 | ||||
| 	emptyNodeToRemove := NodeToBeRemoved{ | ||||
| 		Node: emptyNode, | ||||
| 	} | ||||
| 	drainableNodeToRemove := NodeToBeRemoved{ | ||||
| 		Node:             drainableNode, | ||||
| 		PodsToReschedule: []*apiv1.Pod{pod1, pod2}, | ||||
| 	} | ||||
| 
 | ||||
| 	clusterSnapshot := testsnapshot.NewTestSnapshotOrDie(t) | ||||
| 
 | ||||
| 	topoNode1 := BuildTestNode("topo-n1", 1000, 2000000) | ||||
| 	topoNode2 := BuildTestNode("topo-n2", 1000, 2000000) | ||||
| 	topoNode3 := BuildTestNode("topo-n3", 1000, 2000000) | ||||
| 	topoNode1.Labels = map[string]string{"kubernetes.io/hostname": "topo-n1"} | ||||
| 	topoNode2.Labels = map[string]string{"kubernetes.io/hostname": "topo-n2"} | ||||
| 	topoNode3.Labels = map[string]string{"kubernetes.io/hostname": "topo-n3"} | ||||
| 
 | ||||
| 	SetNodeReadyState(topoNode1, true, time.Time{}) | ||||
| 	SetNodeReadyState(topoNode2, true, time.Time{}) | ||||
| 	SetNodeReadyState(topoNode3, true, time.Time{}) | ||||
| 
 | ||||
| 	minDomains := int32(2) | ||||
| 	maxSkew := int32(1) | ||||
| 	topoConstraint := apiv1.TopologySpreadConstraint{ | ||||
| 		MaxSkew:           maxSkew, | ||||
| 		TopologyKey:       "kubernetes.io/hostname", | ||||
| 		WhenUnsatisfiable: apiv1.DoNotSchedule, | ||||
| 		MinDomains:        &minDomains, | ||||
| 		LabelSelector: &metav1.LabelSelector{ | ||||
| 			MatchLabels: map[string]string{ | ||||
| 				"app": "topo-app", | ||||
| 			}, | ||||
| 		}, | ||||
| 	} | ||||
| 
 | ||||
| 	pod5 := BuildTestPod("p5", 100, 100000) | ||||
| 	pod5.Labels = map[string]string{"app": "topo-app"} | ||||
| 	pod5.OwnerReferences = ownerRefs | ||||
| 	pod5.Spec.NodeName = "topo-n1" | ||||
| 	pod5.Spec.TopologySpreadConstraints = []apiv1.TopologySpreadConstraint{topoConstraint} | ||||
| 
 | ||||
| 	pod6 := BuildTestPod("p6", 100, 100000) | ||||
| 	pod6.Labels = map[string]string{"app": "topo-app"} | ||||
| 	pod6.OwnerReferences = ownerRefs | ||||
| 	pod6.Spec.NodeName = "topo-n2" | ||||
| 	pod6.Spec.TopologySpreadConstraints = []apiv1.TopologySpreadConstraint{topoConstraint} | ||||
| 
 | ||||
| 	pod7 := BuildTestPod("p7", 100, 100000) | ||||
| 	pod7.Labels = map[string]string{"app": "topo-app"} | ||||
| 	pod7.OwnerReferences = ownerRefs | ||||
| 	pod7.Spec.NodeName = "topo-n3" | ||||
| 	pod7.Spec.TopologySpreadConstraints = []apiv1.TopologySpreadConstraint{topoConstraint} | ||||
| 
 | ||||
| 	blocker1 := BuildTestPod("blocker1", 100, 100000) | ||||
| 	blocker1.Spec.NodeName = "topo-n2" | ||||
| 	blocker2 := BuildTestPod("blocker2", 100, 100000) | ||||
| 	blocker2.Spec.NodeName = "topo-n3" | ||||
| 
 | ||||
| 	topoNodeToRemove := NodeToBeRemoved{ | ||||
| 		Node:             topoNode1, | ||||
| 		PodsToReschedule: []*apiv1.Pod{pod5}, | ||||
| 	} | ||||
| 
 | ||||
| 	tests := []findNodesToRemoveTestConfig{ | ||||
| 		{ | ||||
| 			name:       "just an empty node, should be removed", | ||||
| 			candidates: []string{emptyNode.Name}, | ||||
| 			allNodes:   []*apiv1.Node{emptyNode}, | ||||
| 			toRemove:   []NodeToBeRemoved{emptyNodeToRemove}, | ||||
| 		}, | ||||
| 		{ | ||||
| 			name:        "just a drainable node, but nowhere for pods to go to", | ||||
| 			pods:        []*apiv1.Pod{pod1, pod2}, | ||||
| 			candidates:  []string{drainableNode.Name}, | ||||
| 			allNodes:    []*apiv1.Node{drainableNode}, | ||||
| 			unremovable: []*UnremovableNode{{Node: drainableNode, Reason: NoPlaceToMovePods}}, | ||||
| 		}, | ||||
| 		{ | ||||
| 			name:        "drainable node, and a mostly empty node that can take its pods", | ||||
| 			pods:        []*apiv1.Pod{pod1, pod2, pod3}, | ||||
| 			candidates:  []string{drainableNode.Name, nonDrainableNode.Name}, | ||||
| 			allNodes:    []*apiv1.Node{drainableNode, nonDrainableNode}, | ||||
| 			toRemove:    []NodeToBeRemoved{drainableNodeToRemove}, | ||||
| 			unremovable: []*UnremovableNode{{Node: nonDrainableNode, Reason: BlockedByPod, BlockingPod: &drain.BlockingPod{Pod: pod3, Reason: drain.NotReplicated}}}, | ||||
| 		}, | ||||
| 		{ | ||||
| 			name:        "drainable node, and a full node that cannot fit anymore pods", | ||||
| 			pods:        []*apiv1.Pod{pod1, pod2, pod4}, | ||||
| 			candidates:  []string{drainableNode.Name}, | ||||
| 			allNodes:    []*apiv1.Node{drainableNode, fullNode}, | ||||
| 			unremovable: []*UnremovableNode{{Node: drainableNode, Reason: NoPlaceToMovePods}}, | ||||
| 		}, | ||||
| 		{ | ||||
| 			name:       "4 nodes, 1 empty, 1 drainable", | ||||
| 			pods:       []*apiv1.Pod{pod1, pod2, pod3, pod4}, | ||||
| 			candidates: []string{emptyNode.Name, drainableNode.Name}, | ||||
| 			allNodes:   []*apiv1.Node{emptyNode, drainableNode, fullNode, nonDrainableNode}, | ||||
| 			toRemove:   []NodeToBeRemoved{emptyNodeToRemove, drainableNodeToRemove}, | ||||
| 		}, | ||||
| 		{ | ||||
| 			name:       "topology spread constraint test - one node should be removable", | ||||
| 			pods:       []*apiv1.Pod{pod5, pod6, pod7, blocker1, blocker2}, | ||||
| 			allNodes:   []*apiv1.Node{topoNode1, topoNode2, topoNode3}, | ||||
| 			candidates: []string{topoNode1.Name, topoNode2.Name, topoNode3.Name}, | ||||
| 			toRemove:   []NodeToBeRemoved{topoNodeToRemove}, | ||||
| 			unremovable: []*UnremovableNode{ | ||||
| 				{Node: topoNode2, Reason: BlockedByPod, BlockingPod: &drain.BlockingPod{Pod: blocker1, Reason: drain.NotReplicated}}, | ||||
| 				{Node: topoNode3, Reason: BlockedByPod, BlockingPod: &drain.BlockingPod{Pod: blocker2, Reason: drain.NotReplicated}}, | ||||
| 			}, | ||||
| 		}, | ||||
| 	} | ||||
| 
 | ||||
| 	for _, test := range tests { | ||||
| 		t.Run(test.name, func(t *testing.T) { | ||||
| 			destinations := make([]string, 0, len(test.allNodes)) | ||||
| 			for _, node := range test.allNodes { | ||||
| 				destinations = append(destinations, node.Name) | ||||
| 			} | ||||
| 			clustersnapshot.InitializeClusterSnapshotOrDie(t, clusterSnapshot, test.allNodes, test.pods) | ||||
| 			r := NewRemovalSimulator(registry, clusterSnapshot, testDeleteOptions(), nil, false) | ||||
| 			toRemove, unremovable := r.FindNodesToRemove(test.candidates, destinations, time.Now(), nil) | ||||
| 			fmt.Printf("Test scenario: %s, found len(toRemove)=%v, expected len(test.toRemove)=%v\n", test.name, len(toRemove), len(test.toRemove)) | ||||
| 			assert.Equal(t, test.toRemove, toRemove) | ||||
| 			assert.Equal(t, test.unremovable, unremovable) | ||||
| 		}) | ||||
| 	} | ||||
| } | ||||
| 
 | ||||
| func testDeleteOptions() options.NodeDeleteOptions { | ||||
| 	return options.NodeDeleteOptions{ | ||||
| 		SkipNodesWithSystemPods:           true, | ||||
| 		SkipNodesWithLocalStorage:         true, | ||||
| 		SkipNodesWithCustomControllerPods: true, | ||||
| 	} | ||||
| } | ||||
		Loading…
	
		Reference in New Issue