mirror of https://github.com/kubernetes/kops.git
Merge pull request #9418 from olemarkus/arm64-kubedns
Make dns pods work on arm64 clusters
This commit is contained in:
commit
da3ba758e4
|
|
@ -1265,7 +1265,7 @@ spec:
|
|||
spec:
|
||||
containers:
|
||||
- name: autoscaler
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-{{Arch}}:1.4.0
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-amd64:1.4.0
|
||||
resources:
|
||||
requests:
|
||||
cpu: "20m"
|
||||
|
|
@ -1280,10 +1280,14 @@ spec:
|
|||
- --default-params={"linear":{"coresPerReplica":256,"nodesPerReplica":16,"preventSinglePointFailure":true}}
|
||||
- --logtostderr=true
|
||||
- --v=2
|
||||
nodeSelector:
|
||||
kubernetes.io/arch: amd64
|
||||
priorityClassName: system-cluster-critical
|
||||
tolerations:
|
||||
- key: "CriticalAddonsOnly"
|
||||
operator: "Exists"
|
||||
- key: node-role.kubernetes.io/master
|
||||
operator: Exists
|
||||
serviceAccountName: coredns-autoscaler
|
||||
---
|
||||
apiVersion: apps/v1
|
||||
|
|
@ -2685,7 +2689,7 @@ spec:
|
|||
spec:
|
||||
containers:
|
||||
- name: autoscaler
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-{{Arch}}:1.4.0
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-amd64:1.4.0
|
||||
resources:
|
||||
requests:
|
||||
cpu: "20m"
|
||||
|
|
@ -2701,10 +2705,14 @@ spec:
|
|||
- --default-params={"linear":{"coresPerReplica":256,"nodesPerReplica":16,"preventSinglePointFailure":true}}
|
||||
- --logtostderr=true
|
||||
- --v=2
|
||||
nodeSelector:
|
||||
kubernetes.io/arch: amd64
|
||||
priorityClassName: system-cluster-critical
|
||||
tolerations:
|
||||
- key: "CriticalAddonsOnly"
|
||||
operator: "Exists"
|
||||
- key: node-role.kubernetes.io/master
|
||||
operator: Exists
|
||||
serviceAccountName: kube-dns-autoscaler
|
||||
|
||||
---
|
||||
|
|
@ -2762,7 +2770,7 @@ spec:
|
|||
|
||||
containers:
|
||||
- name: kubedns
|
||||
image: k8s.gcr.io/k8s-dns-kube-dns-{{Arch}}:1.14.13
|
||||
image: k8s.gcr.io/k8s-dns-kube-dns:1.14.13
|
||||
resources:
|
||||
# TODO: Set memory limits when we've profiled the container for large
|
||||
# clusters, then set request = limit to keep this container in
|
||||
|
|
@ -2814,7 +2822,7 @@ spec:
|
|||
mountPath: /kube-dns-config
|
||||
|
||||
- name: dnsmasq
|
||||
image: k8s.gcr.io/k8s-dns-dnsmasq-nanny-{{Arch}}:1.14.13
|
||||
image: k8s.gcr.io/k8s-dns-dnsmasq-nanny:1.15.13
|
||||
livenessProbe:
|
||||
httpGet:
|
||||
path: /healthcheck/dnsmasq
|
||||
|
|
@ -2856,7 +2864,7 @@ spec:
|
|||
mountPath: /etc/k8s/dns/dnsmasq-nanny
|
||||
|
||||
- name: sidecar
|
||||
image: k8s.gcr.io/k8s-dns-sidecar-amd64:1.14.13
|
||||
image: k8s.gcr.io/k8s-dns-sidecar:1.14.13
|
||||
livenessProbe:
|
||||
httpGet:
|
||||
path: /metrics
|
||||
|
|
@ -3038,7 +3046,7 @@ spec:
|
|||
spec:
|
||||
containers:
|
||||
- name: autoscaler
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-{{Arch}}:1.1.2-r2
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-amd64:1.1.2-r2
|
||||
resources:
|
||||
requests:
|
||||
cpu: "20m"
|
||||
|
|
@ -3103,7 +3111,7 @@ spec:
|
|||
|
||||
containers:
|
||||
- name: kubedns
|
||||
image: k8s.gcr.io/k8s-dns-kube-dns-{{Arch}}:1.14.10
|
||||
image: k8s.gcr.io/k8s-dns-kube-dns:1.14.10
|
||||
resources:
|
||||
# TODO: Set memory limits when we've profiled the container for large
|
||||
# clusters, then set request = limit to keep this container in
|
||||
|
|
@ -3155,7 +3163,7 @@ spec:
|
|||
mountPath: /kube-dns-config
|
||||
|
||||
- name: dnsmasq
|
||||
image: k8s.gcr.io/k8s-dns-dnsmasq-nanny-{{Arch}}:1.14.10
|
||||
image: k8s.gcr.io/k8s-dns-dnsmasq-nanny:1.14.10
|
||||
livenessProbe:
|
||||
httpGet:
|
||||
path: /healthcheck/dnsmasq
|
||||
|
|
@ -3197,7 +3205,7 @@ spec:
|
|||
mountPath: /etc/k8s/dns/dnsmasq-nanny
|
||||
|
||||
- name: sidecar
|
||||
image: k8s.gcr.io/k8s-dns-sidecar-amd64:1.14.10
|
||||
image: k8s.gcr.io/k8s-dns-sidecar:1.14.10
|
||||
livenessProbe:
|
||||
httpGet:
|
||||
path: /metrics
|
||||
|
|
|
|||
|
|
@ -103,7 +103,7 @@ spec:
|
|||
spec:
|
||||
containers:
|
||||
- name: autoscaler
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-{{Arch}}:1.4.0
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-amd64:1.4.0
|
||||
resources:
|
||||
requests:
|
||||
cpu: "20m"
|
||||
|
|
@ -118,10 +118,14 @@ spec:
|
|||
- --default-params={"linear":{"coresPerReplica":256,"nodesPerReplica":16,"preventSinglePointFailure":true}}
|
||||
- --logtostderr=true
|
||||
- --v=2
|
||||
nodeSelector:
|
||||
kubernetes.io/arch: amd64
|
||||
priorityClassName: system-cluster-critical
|
||||
tolerations:
|
||||
- key: "CriticalAddonsOnly"
|
||||
operator: "Exists"
|
||||
- key: node-role.kubernetes.io/master
|
||||
operator: Exists
|
||||
serviceAccountName: coredns-autoscaler
|
||||
---
|
||||
apiVersion: apps/v1
|
||||
|
|
|
|||
|
|
@ -53,7 +53,7 @@ spec:
|
|||
spec:
|
||||
containers:
|
||||
- name: autoscaler
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-{{Arch}}:1.4.0
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-amd64:1.4.0
|
||||
resources:
|
||||
requests:
|
||||
cpu: "20m"
|
||||
|
|
@ -69,10 +69,14 @@ spec:
|
|||
- --default-params={"linear":{"coresPerReplica":256,"nodesPerReplica":16,"preventSinglePointFailure":true}}
|
||||
- --logtostderr=true
|
||||
- --v=2
|
||||
nodeSelector:
|
||||
kubernetes.io/arch: amd64
|
||||
priorityClassName: system-cluster-critical
|
||||
tolerations:
|
||||
- key: "CriticalAddonsOnly"
|
||||
operator: "Exists"
|
||||
- key: node-role.kubernetes.io/master
|
||||
operator: Exists
|
||||
serviceAccountName: kube-dns-autoscaler
|
||||
|
||||
---
|
||||
|
|
@ -130,7 +134,7 @@ spec:
|
|||
|
||||
containers:
|
||||
- name: kubedns
|
||||
image: k8s.gcr.io/k8s-dns-kube-dns-{{Arch}}:1.14.13
|
||||
image: k8s.gcr.io/k8s-dns-kube-dns:1.14.13
|
||||
resources:
|
||||
# TODO: Set memory limits when we've profiled the container for large
|
||||
# clusters, then set request = limit to keep this container in
|
||||
|
|
@ -182,7 +186,7 @@ spec:
|
|||
mountPath: /kube-dns-config
|
||||
|
||||
- name: dnsmasq
|
||||
image: k8s.gcr.io/k8s-dns-dnsmasq-nanny-{{Arch}}:1.14.13
|
||||
image: k8s.gcr.io/k8s-dns-dnsmasq-nanny:1.15.13
|
||||
livenessProbe:
|
||||
httpGet:
|
||||
path: /healthcheck/dnsmasq
|
||||
|
|
@ -224,7 +228,7 @@ spec:
|
|||
mountPath: /etc/k8s/dns/dnsmasq-nanny
|
||||
|
||||
- name: sidecar
|
||||
image: k8s.gcr.io/k8s-dns-sidecar-amd64:1.14.13
|
||||
image: k8s.gcr.io/k8s-dns-sidecar:1.14.13
|
||||
livenessProbe:
|
||||
httpGet:
|
||||
path: /metrics
|
||||
|
|
|
|||
|
|
@ -52,7 +52,7 @@ spec:
|
|||
spec:
|
||||
containers:
|
||||
- name: autoscaler
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-{{Arch}}:1.1.2-r2
|
||||
image: k8s.gcr.io/cluster-proportional-autoscaler-amd64:1.1.2-r2
|
||||
resources:
|
||||
requests:
|
||||
cpu: "20m"
|
||||
|
|
@ -117,7 +117,7 @@ spec:
|
|||
|
||||
containers:
|
||||
- name: kubedns
|
||||
image: k8s.gcr.io/k8s-dns-kube-dns-{{Arch}}:1.14.10
|
||||
image: k8s.gcr.io/k8s-dns-kube-dns:1.14.10
|
||||
resources:
|
||||
# TODO: Set memory limits when we've profiled the container for large
|
||||
# clusters, then set request = limit to keep this container in
|
||||
|
|
@ -169,7 +169,7 @@ spec:
|
|||
mountPath: /kube-dns-config
|
||||
|
||||
- name: dnsmasq
|
||||
image: k8s.gcr.io/k8s-dns-dnsmasq-nanny-{{Arch}}:1.14.10
|
||||
image: k8s.gcr.io/k8s-dns-dnsmasq-nanny:1.14.10
|
||||
livenessProbe:
|
||||
httpGet:
|
||||
path: /healthcheck/dnsmasq
|
||||
|
|
@ -211,7 +211,7 @@ spec:
|
|||
mountPath: /etc/k8s/dns/dnsmasq-nanny
|
||||
|
||||
- name: sidecar
|
||||
image: k8s.gcr.io/k8s-dns-sidecar-amd64:1.14.10
|
||||
image: k8s.gcr.io/k8s-dns-sidecar:1.14.10
|
||||
livenessProbe:
|
||||
httpGet:
|
||||
path: /metrics
|
||||
|
|
|
|||
|
|
@ -70,7 +70,6 @@ func (tf *TemplateFunctions) AddTo(dest template.FuncMap, secretStore fi.SecretS
|
|||
dest["UseBootstrapTokens"] = tf.UseBootstrapTokens
|
||||
dest["UseEtcdTLS"] = tf.UseEtcdTLS
|
||||
// Remember that we may be on a different arch from the target. Hard-code for now.
|
||||
dest["Arch"] = func() string { return "amd64" }
|
||||
dest["replace"] = func(s, find, replace string) string {
|
||||
return strings.Replace(s, find, replace, -1)
|
||||
}
|
||||
|
|
|
|||
|
|
@ -21,7 +21,7 @@ spec:
|
|||
- id: k8s-1.6
|
||||
kubernetesVersion: <1.12.0
|
||||
manifest: kube-dns.addons.k8s.io/k8s-1.6.yaml
|
||||
manifestHash: 555f952a8b955ce7a5dd0bcd06a5be9e72bd2895
|
||||
manifestHash: 7778a47fa6bda0bae9e0eeea1593a117a9b8bba4
|
||||
name: kube-dns.addons.k8s.io
|
||||
selector:
|
||||
k8s-addon: kube-dns.addons.k8s.io
|
||||
|
|
@ -29,7 +29,7 @@ spec:
|
|||
- id: k8s-1.12
|
||||
kubernetesVersion: '>=1.12.0'
|
||||
manifest: kube-dns.addons.k8s.io/k8s-1.12.yaml
|
||||
manifestHash: 15ade04df128488a534141bd5b8593d078f4953f
|
||||
manifestHash: c1d396ba1997d0eb54586cc14da24fb07548e215
|
||||
name: kube-dns.addons.k8s.io
|
||||
selector:
|
||||
k8s-addon: kube-dns.addons.k8s.io
|
||||
|
|
|
|||
|
|
@ -21,7 +21,7 @@ spec:
|
|||
- id: k8s-1.6
|
||||
kubernetesVersion: <1.12.0
|
||||
manifest: kube-dns.addons.k8s.io/k8s-1.6.yaml
|
||||
manifestHash: 555f952a8b955ce7a5dd0bcd06a5be9e72bd2895
|
||||
manifestHash: 7778a47fa6bda0bae9e0eeea1593a117a9b8bba4
|
||||
name: kube-dns.addons.k8s.io
|
||||
selector:
|
||||
k8s-addon: kube-dns.addons.k8s.io
|
||||
|
|
@ -29,7 +29,7 @@ spec:
|
|||
- id: k8s-1.12
|
||||
kubernetesVersion: '>=1.12.0'
|
||||
manifest: kube-dns.addons.k8s.io/k8s-1.12.yaml
|
||||
manifestHash: 15ade04df128488a534141bd5b8593d078f4953f
|
||||
manifestHash: c1d396ba1997d0eb54586cc14da24fb07548e215
|
||||
name: kube-dns.addons.k8s.io
|
||||
selector:
|
||||
k8s-addon: kube-dns.addons.k8s.io
|
||||
|
|
|
|||
|
|
@ -21,7 +21,7 @@ spec:
|
|||
- id: k8s-1.6
|
||||
kubernetesVersion: <1.12.0
|
||||
manifest: kube-dns.addons.k8s.io/k8s-1.6.yaml
|
||||
manifestHash: 555f952a8b955ce7a5dd0bcd06a5be9e72bd2895
|
||||
manifestHash: 7778a47fa6bda0bae9e0eeea1593a117a9b8bba4
|
||||
name: kube-dns.addons.k8s.io
|
||||
selector:
|
||||
k8s-addon: kube-dns.addons.k8s.io
|
||||
|
|
@ -29,7 +29,7 @@ spec:
|
|||
- id: k8s-1.12
|
||||
kubernetesVersion: '>=1.12.0'
|
||||
manifest: kube-dns.addons.k8s.io/k8s-1.12.yaml
|
||||
manifestHash: 15ade04df128488a534141bd5b8593d078f4953f
|
||||
manifestHash: c1d396ba1997d0eb54586cc14da24fb07548e215
|
||||
name: kube-dns.addons.k8s.io
|
||||
selector:
|
||||
k8s-addon: kube-dns.addons.k8s.io
|
||||
|
|
|
|||
|
|
@ -21,7 +21,7 @@ spec:
|
|||
- id: k8s-1.6
|
||||
kubernetesVersion: <1.12.0
|
||||
manifest: kube-dns.addons.k8s.io/k8s-1.6.yaml
|
||||
manifestHash: 555f952a8b955ce7a5dd0bcd06a5be9e72bd2895
|
||||
manifestHash: 7778a47fa6bda0bae9e0eeea1593a117a9b8bba4
|
||||
name: kube-dns.addons.k8s.io
|
||||
selector:
|
||||
k8s-addon: kube-dns.addons.k8s.io
|
||||
|
|
@ -29,7 +29,7 @@ spec:
|
|||
- id: k8s-1.12
|
||||
kubernetesVersion: '>=1.12.0'
|
||||
manifest: kube-dns.addons.k8s.io/k8s-1.12.yaml
|
||||
manifestHash: 15ade04df128488a534141bd5b8593d078f4953f
|
||||
manifestHash: c1d396ba1997d0eb54586cc14da24fb07548e215
|
||||
name: kube-dns.addons.k8s.io
|
||||
selector:
|
||||
k8s-addon: kube-dns.addons.k8s.io
|
||||
|
|
|
|||
Loading…
Reference in New Issue