cas: remove unused simulator removal funcs
This commit is contained in:
parent
a9292351c3
commit
d0686d9023
|
|
@ -115,29 +115,6 @@ func NewRemovalSimulator(listers kube_util.ListerRegistry, clusterSnapshot clust
|
|||
}
|
||||
}
|
||||
|
||||
// FindNodesToRemove finds nodes that can be removed.
|
||||
func (r *RemovalSimulator) FindNodesToRemove(
|
||||
candidates []string,
|
||||
destinations []string,
|
||||
timestamp time.Time,
|
||||
remainingPdbTracker pdb.RemainingPdbTracker,
|
||||
) (nodesToRemove []NodeToBeRemoved, unremovableNodes []*UnremovableNode) {
|
||||
destinationMap := make(map[string]bool, len(destinations))
|
||||
for _, destination := range destinations {
|
||||
destinationMap[destination] = true
|
||||
}
|
||||
|
||||
for _, nodeName := range candidates {
|
||||
rn, urn := r.SimulateNodeRemoval(nodeName, destinationMap, timestamp, remainingPdbTracker)
|
||||
if rn != nil {
|
||||
nodesToRemove = append(nodesToRemove, *rn)
|
||||
} else if urn != nil {
|
||||
unremovableNodes = append(unremovableNodes, urn)
|
||||
}
|
||||
}
|
||||
return nodesToRemove, unremovableNodes
|
||||
}
|
||||
|
||||
// SimulateNodeRemoval simulates removing a node from the cluster to check
|
||||
// whether it is possible to move its pods. Depending on
|
||||
// the outcome, exactly one of (NodeToBeRemoved, UnremovableNode) will be
|
||||
|
|
@ -178,24 +155,6 @@ func (r *RemovalSimulator) SimulateNodeRemoval(
|
|||
}, nil
|
||||
}
|
||||
|
||||
// FindEmptyNodesToRemove finds empty nodes that can be removed.
|
||||
func (r *RemovalSimulator) FindEmptyNodesToRemove(candidates []string, timestamp time.Time) []string {
|
||||
result := make([]string, 0)
|
||||
for _, node := range candidates {
|
||||
nodeInfo, err := r.clusterSnapshot.GetNodeInfo(node)
|
||||
if err != nil {
|
||||
klog.Errorf("Can't retrieve node %s from snapshot, err: %v", node, err)
|
||||
continue
|
||||
}
|
||||
// Should block on all pods
|
||||
podsToRemove, _, _, err := GetPodsToMove(nodeInfo, r.deleteOptions, r.drainabilityRules, nil, nil, timestamp)
|
||||
if err == nil && len(podsToRemove) == 0 {
|
||||
result = append(result, node)
|
||||
}
|
||||
}
|
||||
return result
|
||||
}
|
||||
|
||||
func (r *RemovalSimulator) withForkedSnapshot(f func() error) (err error) {
|
||||
r.clusterSnapshot.Fork()
|
||||
defer func() {
|
||||
|
|
|
|||
|
|
@ -1,267 +0,0 @@
|
|||
/*
|
||||
Copyright 2016 The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
package simulator
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/stretchr/testify/assert"
|
||||
|
||||
appsv1 "k8s.io/api/apps/v1"
|
||||
apiv1 "k8s.io/api/core/v1"
|
||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||
"k8s.io/autoscaler/cluster-autoscaler/simulator/clustersnapshot"
|
||||
"k8s.io/autoscaler/cluster-autoscaler/simulator/clustersnapshot/testsnapshot"
|
||||
"k8s.io/autoscaler/cluster-autoscaler/simulator/framework"
|
||||
"k8s.io/autoscaler/cluster-autoscaler/simulator/options"
|
||||
"k8s.io/autoscaler/cluster-autoscaler/utils/drain"
|
||||
kube_util "k8s.io/autoscaler/cluster-autoscaler/utils/kubernetes"
|
||||
. "k8s.io/autoscaler/cluster-autoscaler/utils/test"
|
||||
"k8s.io/kubernetes/pkg/kubelet/types"
|
||||
)
|
||||
|
||||
func TestFindEmptyNodes(t *testing.T) {
|
||||
nodes := []*apiv1.Node{}
|
||||
nodeNames := []string{}
|
||||
for i := 0; i < 4; i++ {
|
||||
nodeName := fmt.Sprintf("n%d", i)
|
||||
node := BuildTestNode(nodeName, 1000, 2000000)
|
||||
SetNodeReadyState(node, true, time.Time{})
|
||||
nodes = append(nodes, node)
|
||||
nodeNames = append(nodeNames, nodeName)
|
||||
}
|
||||
|
||||
pod1 := BuildTestPod("p1", 300, 500000)
|
||||
pod1.Spec.NodeName = "n1"
|
||||
|
||||
pod2 := BuildTestPod("p2", 300, 500000)
|
||||
pod2.Spec.NodeName = "n2"
|
||||
pod2.Annotations = map[string]string{
|
||||
types.ConfigMirrorAnnotationKey: "",
|
||||
}
|
||||
|
||||
clusterSnapshot := testsnapshot.NewTestSnapshotOrDie(t)
|
||||
clustersnapshot.InitializeClusterSnapshotOrDie(t, clusterSnapshot, []*apiv1.Node{nodes[0], nodes[1], nodes[2], nodes[3]}, []*apiv1.Pod{pod1, pod2})
|
||||
testTime := time.Date(2020, time.December, 18, 17, 0, 0, 0, time.UTC)
|
||||
r := NewRemovalSimulator(nil, clusterSnapshot, testDeleteOptions(), nil, false)
|
||||
emptyNodes := r.FindEmptyNodesToRemove(nodeNames, testTime)
|
||||
assert.Equal(t, []string{nodeNames[0], nodeNames[2], nodeNames[3]}, emptyNodes)
|
||||
}
|
||||
|
||||
type findNodesToRemoveTestConfig struct {
|
||||
name string
|
||||
pods []*apiv1.Pod
|
||||
allNodes []*apiv1.Node
|
||||
candidates []string
|
||||
toRemove []NodeToBeRemoved
|
||||
unremovable []*UnremovableNode
|
||||
}
|
||||
|
||||
func TestFindNodesToRemove(t *testing.T) {
|
||||
emptyNode := BuildTestNode("n1", 1000, 2000000)
|
||||
|
||||
// two small pods backed by ReplicaSet
|
||||
drainableNode := BuildTestNode("n2", 1000, 2000000)
|
||||
drainableNodeInfo := framework.NewTestNodeInfo(drainableNode)
|
||||
|
||||
// one small pod, not backed by anything
|
||||
nonDrainableNode := BuildTestNode("n3", 1000, 2000000)
|
||||
nonDrainableNodeInfo := framework.NewTestNodeInfo(nonDrainableNode)
|
||||
|
||||
// one very large pod
|
||||
fullNode := BuildTestNode("n4", 1000, 2000000)
|
||||
fullNodeInfo := framework.NewTestNodeInfo(fullNode)
|
||||
|
||||
SetNodeReadyState(emptyNode, true, time.Time{})
|
||||
SetNodeReadyState(drainableNode, true, time.Time{})
|
||||
SetNodeReadyState(nonDrainableNode, true, time.Time{})
|
||||
SetNodeReadyState(fullNode, true, time.Time{})
|
||||
|
||||
replicas := int32(5)
|
||||
replicaSets := []*appsv1.ReplicaSet{
|
||||
{
|
||||
ObjectMeta: metav1.ObjectMeta{
|
||||
Name: "rs",
|
||||
Namespace: "default",
|
||||
SelfLink: "api/v1/namespaces/default/replicasets/rs",
|
||||
},
|
||||
Spec: appsv1.ReplicaSetSpec{
|
||||
Replicas: &replicas,
|
||||
},
|
||||
},
|
||||
}
|
||||
rsLister, err := kube_util.NewTestReplicaSetLister(replicaSets)
|
||||
assert.NoError(t, err)
|
||||
registry := kube_util.NewListerRegistry(nil, nil, nil, nil, nil, nil, nil, rsLister, nil)
|
||||
|
||||
ownerRefs := GenerateOwnerReferences("rs", "ReplicaSet", "extensions/v1beta1", "")
|
||||
|
||||
pod1 := BuildTestPod("p1", 100, 100000)
|
||||
pod1.OwnerReferences = ownerRefs
|
||||
pod1.Spec.NodeName = "n2"
|
||||
drainableNodeInfo.AddPod(&framework.PodInfo{Pod: pod1})
|
||||
|
||||
pod2 := BuildTestPod("p2", 100, 100000)
|
||||
pod2.OwnerReferences = ownerRefs
|
||||
pod2.Spec.NodeName = "n2"
|
||||
drainableNodeInfo.AddPod(&framework.PodInfo{Pod: pod2})
|
||||
|
||||
pod3 := BuildTestPod("p3", 100, 100000)
|
||||
pod3.Spec.NodeName = "n3"
|
||||
nonDrainableNodeInfo.AddPod(&framework.PodInfo{Pod: pod3})
|
||||
|
||||
pod4 := BuildTestPod("p4", 1000, 100000)
|
||||
pod4.Spec.NodeName = "n4"
|
||||
fullNodeInfo.AddPod(&framework.PodInfo{Pod: pod4})
|
||||
|
||||
emptyNodeToRemove := NodeToBeRemoved{
|
||||
Node: emptyNode,
|
||||
}
|
||||
drainableNodeToRemove := NodeToBeRemoved{
|
||||
Node: drainableNode,
|
||||
PodsToReschedule: []*apiv1.Pod{pod1, pod2},
|
||||
}
|
||||
|
||||
clusterSnapshot := testsnapshot.NewTestSnapshotOrDie(t)
|
||||
|
||||
topoNode1 := BuildTestNode("topo-n1", 1000, 2000000)
|
||||
topoNode2 := BuildTestNode("topo-n2", 1000, 2000000)
|
||||
topoNode3 := BuildTestNode("topo-n3", 1000, 2000000)
|
||||
topoNode1.Labels = map[string]string{"kubernetes.io/hostname": "topo-n1"}
|
||||
topoNode2.Labels = map[string]string{"kubernetes.io/hostname": "topo-n2"}
|
||||
topoNode3.Labels = map[string]string{"kubernetes.io/hostname": "topo-n3"}
|
||||
|
||||
SetNodeReadyState(topoNode1, true, time.Time{})
|
||||
SetNodeReadyState(topoNode2, true, time.Time{})
|
||||
SetNodeReadyState(topoNode3, true, time.Time{})
|
||||
|
||||
minDomains := int32(2)
|
||||
maxSkew := int32(1)
|
||||
topoConstraint := apiv1.TopologySpreadConstraint{
|
||||
MaxSkew: maxSkew,
|
||||
TopologyKey: "kubernetes.io/hostname",
|
||||
WhenUnsatisfiable: apiv1.DoNotSchedule,
|
||||
MinDomains: &minDomains,
|
||||
LabelSelector: &metav1.LabelSelector{
|
||||
MatchLabels: map[string]string{
|
||||
"app": "topo-app",
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
pod5 := BuildTestPod("p5", 100, 100000)
|
||||
pod5.Labels = map[string]string{"app": "topo-app"}
|
||||
pod5.OwnerReferences = ownerRefs
|
||||
pod5.Spec.NodeName = "topo-n1"
|
||||
pod5.Spec.TopologySpreadConstraints = []apiv1.TopologySpreadConstraint{topoConstraint}
|
||||
|
||||
pod6 := BuildTestPod("p6", 100, 100000)
|
||||
pod6.Labels = map[string]string{"app": "topo-app"}
|
||||
pod6.OwnerReferences = ownerRefs
|
||||
pod6.Spec.NodeName = "topo-n2"
|
||||
pod6.Spec.TopologySpreadConstraints = []apiv1.TopologySpreadConstraint{topoConstraint}
|
||||
|
||||
pod7 := BuildTestPod("p7", 100, 100000)
|
||||
pod7.Labels = map[string]string{"app": "topo-app"}
|
||||
pod7.OwnerReferences = ownerRefs
|
||||
pod7.Spec.NodeName = "topo-n3"
|
||||
pod7.Spec.TopologySpreadConstraints = []apiv1.TopologySpreadConstraint{topoConstraint}
|
||||
|
||||
blocker1 := BuildTestPod("blocker1", 100, 100000)
|
||||
blocker1.Spec.NodeName = "topo-n2"
|
||||
blocker2 := BuildTestPod("blocker2", 100, 100000)
|
||||
blocker2.Spec.NodeName = "topo-n3"
|
||||
|
||||
topoNodeToRemove := NodeToBeRemoved{
|
||||
Node: topoNode1,
|
||||
PodsToReschedule: []*apiv1.Pod{pod5},
|
||||
}
|
||||
|
||||
tests := []findNodesToRemoveTestConfig{
|
||||
{
|
||||
name: "just an empty node, should be removed",
|
||||
candidates: []string{emptyNode.Name},
|
||||
allNodes: []*apiv1.Node{emptyNode},
|
||||
toRemove: []NodeToBeRemoved{emptyNodeToRemove},
|
||||
},
|
||||
{
|
||||
name: "just a drainable node, but nowhere for pods to go to",
|
||||
pods: []*apiv1.Pod{pod1, pod2},
|
||||
candidates: []string{drainableNode.Name},
|
||||
allNodes: []*apiv1.Node{drainableNode},
|
||||
unremovable: []*UnremovableNode{{Node: drainableNode, Reason: NoPlaceToMovePods}},
|
||||
},
|
||||
{
|
||||
name: "drainable node, and a mostly empty node that can take its pods",
|
||||
pods: []*apiv1.Pod{pod1, pod2, pod3},
|
||||
candidates: []string{drainableNode.Name, nonDrainableNode.Name},
|
||||
allNodes: []*apiv1.Node{drainableNode, nonDrainableNode},
|
||||
toRemove: []NodeToBeRemoved{drainableNodeToRemove},
|
||||
unremovable: []*UnremovableNode{{Node: nonDrainableNode, Reason: BlockedByPod, BlockingPod: &drain.BlockingPod{Pod: pod3, Reason: drain.NotReplicated}}},
|
||||
},
|
||||
{
|
||||
name: "drainable node, and a full node that cannot fit anymore pods",
|
||||
pods: []*apiv1.Pod{pod1, pod2, pod4},
|
||||
candidates: []string{drainableNode.Name},
|
||||
allNodes: []*apiv1.Node{drainableNode, fullNode},
|
||||
unremovable: []*UnremovableNode{{Node: drainableNode, Reason: NoPlaceToMovePods}},
|
||||
},
|
||||
{
|
||||
name: "4 nodes, 1 empty, 1 drainable",
|
||||
pods: []*apiv1.Pod{pod1, pod2, pod3, pod4},
|
||||
candidates: []string{emptyNode.Name, drainableNode.Name},
|
||||
allNodes: []*apiv1.Node{emptyNode, drainableNode, fullNode, nonDrainableNode},
|
||||
toRemove: []NodeToBeRemoved{emptyNodeToRemove, drainableNodeToRemove},
|
||||
},
|
||||
{
|
||||
name: "topology spread constraint test - one node should be removable",
|
||||
pods: []*apiv1.Pod{pod5, pod6, pod7, blocker1, blocker2},
|
||||
allNodes: []*apiv1.Node{topoNode1, topoNode2, topoNode3},
|
||||
candidates: []string{topoNode1.Name, topoNode2.Name, topoNode3.Name},
|
||||
toRemove: []NodeToBeRemoved{topoNodeToRemove},
|
||||
unremovable: []*UnremovableNode{
|
||||
{Node: topoNode2, Reason: BlockedByPod, BlockingPod: &drain.BlockingPod{Pod: blocker1, Reason: drain.NotReplicated}},
|
||||
{Node: topoNode3, Reason: BlockedByPod, BlockingPod: &drain.BlockingPod{Pod: blocker2, Reason: drain.NotReplicated}},
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
for _, test := range tests {
|
||||
t.Run(test.name, func(t *testing.T) {
|
||||
destinations := make([]string, 0, len(test.allNodes))
|
||||
for _, node := range test.allNodes {
|
||||
destinations = append(destinations, node.Name)
|
||||
}
|
||||
clustersnapshot.InitializeClusterSnapshotOrDie(t, clusterSnapshot, test.allNodes, test.pods)
|
||||
r := NewRemovalSimulator(registry, clusterSnapshot, testDeleteOptions(), nil, false)
|
||||
toRemove, unremovable := r.FindNodesToRemove(test.candidates, destinations, time.Now(), nil)
|
||||
fmt.Printf("Test scenario: %s, found len(toRemove)=%v, expected len(test.toRemove)=%v\n", test.name, len(toRemove), len(test.toRemove))
|
||||
assert.Equal(t, test.toRemove, toRemove)
|
||||
assert.Equal(t, test.unremovable, unremovable)
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func testDeleteOptions() options.NodeDeleteOptions {
|
||||
return options.NodeDeleteOptions{
|
||||
SkipNodesWithSystemPods: true,
|
||||
SkipNodesWithLocalStorage: true,
|
||||
SkipNodesWithCustomControllerPods: true,
|
||||
}
|
||||
}
|
||||
Loading…
Reference in New Issue