The profile helper methods were attached to the cluster model.
Unfortunately, the cluster isn't available when creating a new rke
template.
To resolve this I moved all of the cis helpers out of the cluster model
and utils and moved them into a cisHelpers service so they could be
used without access to the cluster itself.
- Added Set Alert button
- This will set the appropriate options for cis
- Added Set Schedule button
- This will scroll the settings into view
- Added a modal so profiles can be picked
Was previously using the presence of nodeDrainInput to determine
the value of drain. Drain is now a part of the backend so I'm using
that value instead of inferring.
rancher/rancher#25732
Adds an empty name to the cluster model creation so it isn't missing if the user
opens yaml editor and knows they should input it
Removes incorrect next usage
Adds logic to handle overriding the name in name-desc when updateYaml is called
Updates form-name-description model observer to watch the two props it actually
cares about
Fixes bug in removeEmpty util which would remove excludedKeys during filter phase
rancher/rancher#24971
This required refactoring the drain modal into a reusable component
since these fields were going to be used in more than one place.
rancher/rancher#24110
I originally tried to fixrancher/rancher#24704 without completely
special casing. Unfortunately that lead to other issues:
rancher/rancher#24745rancher/rancher#24794rancher/rancher#24814
I decided to revert all of the related changes and to just special
case this one instance. Ultimately I think the removeEmpty is the
culprit but it requires backend changes in order to properly fix
and those changes are not happening right now.
We were erroneously adding cloud_provider.awsCloudProvider on
digital ocean etc due to rancher/rancher#24515.
This change assumes that the presence of
onfig.rancher_kubernetes_engine_config.cloud_provider.name
implies that the cloud_provider should be present. If that
nested field isn't present we remove cloud_provider.
rancher/rancher#24745
While working on a ticket to provide the ability to de-select subnetworks when
using the create subnetwork option I discovered the options were all messed up
and allowed you to misconfigure yourself into a hole.
I've moved ipalias and related networks settings out of advanced becuase
depending on what you select for your subnetwork the ability to choose ipalias
and the other settings changes.
This change allows you deselect a node subnet so you can create a subnetwork
automatically.
rancher/rancher#21079
After pressing the 'create' button of the EKS driver the user was being
transitioned back to step 3 (Vpc & Subnet) rather than waiting on the
final page until the save is complete and returning to the cluster page.
An observer was being triggered by the save process which subsequently
set the step back to 3. To resolve this we will only enter the branch if
there are initialized values that need to be set back to default.
rancher/rancher#23493
While editing a cluster properly support .x kube version comparisons when
filtering out cluster template revisions.
Coercing a .x version converts it to a .0 which made the revision look like
it was a kube downgrade. By making use of .satisfies when the revision
kube version ends with a '.x' we're now better able to check if the
kube version is a downgrade and filter appropriately.
rancher/rancher#23489
the check-override-allowed component did not know how to deal with the k8s
version question because of its tri state and how we deal with the patch version
that is an override but not really an override. I added a check to verify the
mode is view and we have the param then display param so we don't initialize the
form-version component which has logic to inject the current version into its
versions dropdown but only if we're new, editing, or cloning.
rancher/rancher#23478rancher/rancher#23465