Remove --unregistered-node-removal-time flag
This commit is contained in:
parent
362b0319de
commit
2ff962e53e
|
|
@ -77,8 +77,6 @@ type AutoscalingOptions struct {
|
||||||
MinMemoryTotal int64
|
MinMemoryTotal int64
|
||||||
// NodeGroupAutoDiscovery represents one or more definition(s) of node group auto-discovery
|
// NodeGroupAutoDiscovery represents one or more definition(s) of node group auto-discovery
|
||||||
NodeGroupAutoDiscovery string
|
NodeGroupAutoDiscovery string
|
||||||
// UnregisteredNodeRemovalTime represents how long CA waits before removing nodes that are not registered in Kubernetes")
|
|
||||||
UnregisteredNodeRemovalTime time.Duration
|
|
||||||
// EstimatorName is the estimator used to estimate the number of needed nodes in scale up.
|
// EstimatorName is the estimator used to estimate the number of needed nodes in scale up.
|
||||||
EstimatorName string
|
EstimatorName string
|
||||||
// ExpanderName sets the type of node group expander to be used in scale up
|
// ExpanderName sets the type of node group expander to be used in scale up
|
||||||
|
|
|
||||||
|
|
@ -329,7 +329,7 @@ func removeOldUnregisteredNodes(unregisteredNodes []clusterstate.UnregisteredNod
|
||||||
currentTime time.Time, logRecorder *utils.LogEventRecorder) (bool, error) {
|
currentTime time.Time, logRecorder *utils.LogEventRecorder) (bool, error) {
|
||||||
removedAny := false
|
removedAny := false
|
||||||
for _, unregisteredNode := range unregisteredNodes {
|
for _, unregisteredNode := range unregisteredNodes {
|
||||||
if unregisteredNode.UnregisteredSince.Add(context.UnregisteredNodeRemovalTime).Before(currentTime) {
|
if unregisteredNode.UnregisteredSince.Add(context.MaxNodeProvisionTime).Before(currentTime) {
|
||||||
glog.V(0).Infof("Removing unregistered node %v", unregisteredNode.Node.Name)
|
glog.V(0).Infof("Removing unregistered node %v", unregisteredNode.Node.Name)
|
||||||
nodeGroup, err := context.CloudProvider.NodeGroupForNode(unregisteredNode.Node)
|
nodeGroup, err := context.CloudProvider.NodeGroupForNode(unregisteredNode.Node)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
@ -372,7 +372,7 @@ func fixNodeGroupSize(context *AutoscalingContext, currentTime time.Time) (bool,
|
||||||
if incorrectSize == nil {
|
if incorrectSize == nil {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if incorrectSize.FirstObserved.Add(context.UnregisteredNodeRemovalTime).Before(currentTime) {
|
if incorrectSize.FirstObserved.Add(context.MaxNodeProvisionTime).Before(currentTime) {
|
||||||
delta := incorrectSize.CurrentSize - incorrectSize.ExpectedSize
|
delta := incorrectSize.CurrentSize - incorrectSize.ExpectedSize
|
||||||
if delta < 0 {
|
if delta < 0 {
|
||||||
glog.V(0).Infof("Decreasing size of %s, expected=%d current=%d delta=%d", nodeGroup.Id(),
|
glog.V(0).Infof("Decreasing size of %s, expected=%d current=%d delta=%d", nodeGroup.Id(),
|
||||||
|
|
|
||||||
|
|
@ -332,7 +332,7 @@ func TestRemoveOldUnregisteredNodes(t *testing.T) {
|
||||||
|
|
||||||
context := &AutoscalingContext{
|
context := &AutoscalingContext{
|
||||||
AutoscalingOptions: AutoscalingOptions{
|
AutoscalingOptions: AutoscalingOptions{
|
||||||
UnregisteredNodeRemovalTime: 45 * time.Minute,
|
MaxNodeProvisionTime: 45 * time.Minute,
|
||||||
},
|
},
|
||||||
CloudProvider: provider,
|
CloudProvider: provider,
|
||||||
ClusterStateRegistry: clusterState,
|
ClusterStateRegistry: clusterState,
|
||||||
|
|
@ -430,7 +430,7 @@ func TestRemoveFixNodeTargetSize(t *testing.T) {
|
||||||
|
|
||||||
context := &AutoscalingContext{
|
context := &AutoscalingContext{
|
||||||
AutoscalingOptions: AutoscalingOptions{
|
AutoscalingOptions: AutoscalingOptions{
|
||||||
UnregisteredNodeRemovalTime: 45 * time.Minute,
|
MaxNodeProvisionTime: 45 * time.Minute,
|
||||||
},
|
},
|
||||||
CloudProvider: provider,
|
CloudProvider: provider,
|
||||||
ClusterStateRegistry: clusterState,
|
ClusterStateRegistry: clusterState,
|
||||||
|
|
|
||||||
|
|
@ -113,7 +113,6 @@ var (
|
||||||
maxTotalUnreadyPercentage = flag.Float64("max-total-unready-percentage", 33, "Maximum percentage of unready nodes after which CA halts operations")
|
maxTotalUnreadyPercentage = flag.Float64("max-total-unready-percentage", 33, "Maximum percentage of unready nodes after which CA halts operations")
|
||||||
okTotalUnreadyCount = flag.Int("ok-total-unready-count", 3, "Number of allowed unready nodes, irrespective of max-total-unready-percentage")
|
okTotalUnreadyCount = flag.Int("ok-total-unready-count", 3, "Number of allowed unready nodes, irrespective of max-total-unready-percentage")
|
||||||
maxNodeProvisionTime = flag.Duration("max-node-provision-time", 15*time.Minute, "Maximum time CA waits for node to be provisioned")
|
maxNodeProvisionTime = flag.Duration("max-node-provision-time", 15*time.Minute, "Maximum time CA waits for node to be provisioned")
|
||||||
unregisteredNodeRemovalTime = flag.Duration("unregistered-node-removal-time", 15*time.Minute, "Time that CA waits before removing nodes that are not registered in Kubernetes")
|
|
||||||
|
|
||||||
estimatorFlag = flag.String("estimator", estimator.BinpackingEstimatorName,
|
estimatorFlag = flag.String("estimator", estimator.BinpackingEstimatorName,
|
||||||
"Type of resource estimator to be used in scale up. Available values: ["+strings.Join(estimator.AvailableEstimators, ",")+"]")
|
"Type of resource estimator to be used in scale up. Available values: ["+strings.Join(estimator.AvailableEstimators, ",")+"]")
|
||||||
|
|
@ -161,7 +160,6 @@ func createAutoscalerOptions() core.AutoscalerOptions {
|
||||||
MaxMemoryTotal: maxMemoryTotal,
|
MaxMemoryTotal: maxMemoryTotal,
|
||||||
MinMemoryTotal: minMemoryTotal,
|
MinMemoryTotal: minMemoryTotal,
|
||||||
NodeGroups: nodeGroupsFlag,
|
NodeGroups: nodeGroupsFlag,
|
||||||
UnregisteredNodeRemovalTime: *unregisteredNodeRemovalTime,
|
|
||||||
ScaleDownDelayAfterAdd: *scaleDownDelayAfterAdd,
|
ScaleDownDelayAfterAdd: *scaleDownDelayAfterAdd,
|
||||||
ScaleDownDelayAfterDelete: *scaleDownDelayAfterDelete,
|
ScaleDownDelayAfterDelete: *scaleDownDelayAfterDelete,
|
||||||
ScaleDownDelayAfterFailure: *scaleDownDelayAfterFailure,
|
ScaleDownDelayAfterFailure: *scaleDownDelayAfterFailure,
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue