86 lines
3.1 KiB
Go
86 lines
3.1 KiB
Go
/*
|
|
Copyright 2019 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package utils
|
|
|
|
import (
|
|
"reflect"
|
|
|
|
apiv1 "k8s.io/api/core/v1"
|
|
apiequality "k8s.io/apimachinery/pkg/api/equality"
|
|
"k8s.io/autoscaler/cluster-autoscaler/simulator"
|
|
"k8s.io/autoscaler/cluster-autoscaler/utils/drain"
|
|
)
|
|
|
|
// PodSchedulableInfo data structure is used to avoid running predicates #pending_pods * #nodes
|
|
// times (which turned out to be very expensive if there are thousands of pending pods).
|
|
// This optimization is based on the assumption that if there are that many pods they're
|
|
// likely created by controllers (deployment, replication controller, ...).
|
|
// So instead of running all predicates for every pod we first check whether we've
|
|
// already seen identical pod (in this step we're not binpacking, just checking if
|
|
// the pod would fit anywhere right now) and if so we use the result we already
|
|
// calculated.
|
|
// To decide if two pods are similar enough we check if they have identical label
|
|
// and spec and are owned by the same controller. The problem is the whole
|
|
// PodSchedulableInfo struct is not hashable and keeping a list and running deep
|
|
// equality checks would likely also be expensive. So instead we use controller
|
|
// UID as a key in initial lookup and only run full comparison on a set of
|
|
// podSchedulableInfos created for pods owned by this controller.
|
|
type PodSchedulableInfo struct {
|
|
spec apiv1.PodSpec
|
|
labels map[string]string
|
|
schedulingError *simulator.PredicateError
|
|
}
|
|
|
|
// PodSchedulableMap stores mapping from controller ref to PodSchedulableInfo
|
|
type PodSchedulableMap map[string][]PodSchedulableInfo
|
|
|
|
// Match tests if given pod matches PodSchedulableInfo
|
|
func (psi *PodSchedulableInfo) Match(pod *apiv1.Pod) bool {
|
|
return reflect.DeepEqual(pod.Labels, psi.labels) && apiequality.Semantic.DeepEqual(pod.Spec, psi.spec)
|
|
}
|
|
|
|
// Get returns scheduling info for given pod if matching one exists in PodSchedulableMap
|
|
func (podMap PodSchedulableMap) Get(pod *apiv1.Pod) (*simulator.PredicateError, bool) {
|
|
ref := drain.ControllerRef(pod)
|
|
if ref == nil {
|
|
return nil, false
|
|
}
|
|
uid := string(ref.UID)
|
|
if infos, found := podMap[uid]; found {
|
|
for _, info := range infos {
|
|
if info.Match(pod) {
|
|
return info.schedulingError, true
|
|
}
|
|
}
|
|
}
|
|
return nil, false
|
|
}
|
|
|
|
// Set sets scheduling info for given pod in PodSchedulableMap
|
|
func (podMap PodSchedulableMap) Set(pod *apiv1.Pod, err *simulator.PredicateError) {
|
|
ref := drain.ControllerRef(pod)
|
|
if ref == nil {
|
|
return
|
|
}
|
|
uid := string(ref.UID)
|
|
podMap[uid] = append(podMap[uid], PodSchedulableInfo{
|
|
spec: pod.Spec,
|
|
labels: pod.Labels,
|
|
schedulingError: err,
|
|
})
|
|
}
|