Merge pull request #1388 from zhxcai/typos
Automatic merge from submit-queue. Update a lot of typos
This commit is contained in:
commit
f6b7936fae
|
|
@ -100,13 +100,13 @@ We'd like to propose following structure in Events object in the new events API
|
||||||
|
|
||||||
```golang
|
```golang
|
||||||
type Event struct {
|
type Event struct {
|
||||||
// <type and object metadata>
|
// <type and object metadata>
|
||||||
|
|
||||||
// Time when this Event was first observed.
|
// Time when this Event was first observed.
|
||||||
EventTime metav1.MicroTime
|
EventTime metav1.MicroTime
|
||||||
|
|
||||||
// Data about the Event series this event represents or nil if it's
|
// Data about the Event series this event represents or nil if it's
|
||||||
// a singleton Event.
|
// a singleton Event.
|
||||||
// +optional
|
// +optional
|
||||||
Series *EventSeries
|
Series *EventSeries
|
||||||
|
|
||||||
|
|
@ -123,7 +123,7 @@ type Event struct {
|
||||||
Reason string
|
Reason string
|
||||||
|
|
||||||
// The object this Event is “about”. In most cases it's the object that the
|
// The object this Event is “about”. In most cases it's the object that the
|
||||||
// given controller implements.
|
// given controller implements.
|
||||||
// +optional
|
// +optional
|
||||||
Regarding ObjectReference
|
Regarding ObjectReference
|
||||||
|
|
||||||
|
|
@ -132,24 +132,24 @@ type Event struct {
|
||||||
Related *ObjectReference
|
Related *ObjectReference
|
||||||
|
|
||||||
// Human readable description of the Event. Possibly discarded when and
|
// Human readable description of the Event. Possibly discarded when and
|
||||||
// Event series is being deduplicated.
|
// Event series is being deduplicated.
|
||||||
// +optional
|
// +optional
|
||||||
Note string
|
Note string
|
||||||
|
|
||||||
// Type of this event (Normal, Warning), new types could be added in the
|
// Type of this event (Normal, Warning), new types could be added in the
|
||||||
// future.
|
// future.
|
||||||
// +optional
|
// +optional
|
||||||
Type string
|
Type string
|
||||||
}
|
}
|
||||||
|
|
||||||
type EventSeries struct {
|
type EventSeries struct {
|
||||||
Count int32
|
Count int32
|
||||||
LastObservedTime MicroTime
|
LastObservedTime MicroTime
|
||||||
State EventSeriesState
|
State EventSeriesState
|
||||||
}
|
}
|
||||||
|
|
||||||
const (
|
const (
|
||||||
EventSeriesStateOngoing = "Ongoing"
|
EventSeriesStateOngoing = "Ongoing"
|
||||||
EventSeriesStateFinished = "Finished"
|
EventSeriesStateFinished = "Finished"
|
||||||
EventSeriesStateUnknown = "Unknown"
|
EventSeriesStateUnknown = "Unknown"
|
||||||
)
|
)
|
||||||
|
|
@ -161,7 +161,7 @@ EventSeriesStateOngoing = "Ongoing"
|
||||||
| ----------| -------| -------| --------------------|---------|
|
| ----------| -------| -------| --------------------|---------|
|
||||||
| Node X | BecameUnreachable | HeartbeatTooOld | kubernetes.io/node-ctrl | <nil> |
|
| Node X | BecameUnreachable | HeartbeatTooOld | kubernetes.io/node-ctrl | <nil> |
|
||||||
| Node Y | FailedToAttachVolume | Unknown | kubernetes.io/pv-attach-ctrl | PVC X |
|
| Node Y | FailedToAttachVolume | Unknown | kubernetes.io/pv-attach-ctrl | PVC X |
|
||||||
| ReplicaSet X | FailedToInstantiantePod | QuotaExceeded | kubernetes.io/replica-set-ctrl | <nil> |
|
| ReplicaSet X | FailedToInstantiatePod | QuotaExceeded | kubernetes.io/replica-set-ctrl | <nil> |
|
||||||
| ReplicaSet X | InstantiatedPod | | kubernetes.io/replica-set-ctrl | Pod Y |
|
| ReplicaSet X | InstantiatedPod | | kubernetes.io/replica-set-ctrl | Pod Y |
|
||||||
| Ingress X | CreatedLoadBalancer | | kubernetes.io/ingress-ctrl | <nil> |
|
| Ingress X | CreatedLoadBalancer | | kubernetes.io/ingress-ctrl | <nil> |
|
||||||
| Pod X | ScheduledOn | | kubernetes.io/scheduler | Node Y |
|
| Pod X | ScheduledOn | | kubernetes.io/scheduler | Node Y |
|
||||||
|
|
|
||||||
|
|
@ -32,7 +32,7 @@ how careful we need to be.
|
||||||
### Huge number of handshakes slows down API server
|
### Huge number of handshakes slows down API server
|
||||||
|
|
||||||
It was a long standing issue for performance and is/was an important bottleneck for scalability (https://github.com/kubernetes/kubernetes/issues/13671). The bug directly
|
It was a long standing issue for performance and is/was an important bottleneck for scalability (https://github.com/kubernetes/kubernetes/issues/13671). The bug directly
|
||||||
causing this problem was incorrect (from the golangs standpoint) handling of TCP connections. Secondary issue was that elliptic curve encryption (only one available in go 1.4)
|
causing this problem was incorrect (from the golang's standpoint) handling of TCP connections. Secondary issue was that elliptic curve encryption (only one available in go 1.4)
|
||||||
is unbelievably slow.
|
is unbelievably slow.
|
||||||
|
|
||||||
## Proposed metrics/statistics to gather/compute to avoid problems
|
## Proposed metrics/statistics to gather/compute to avoid problems
|
||||||
|
|
@ -42,7 +42,7 @@ is unbelievably slow.
|
||||||
Basic ideas:
|
Basic ideas:
|
||||||
- number of Pods/ReplicationControllers/Services in the cluster
|
- number of Pods/ReplicationControllers/Services in the cluster
|
||||||
- number of running replicas of master components (if they are replicated)
|
- number of running replicas of master components (if they are replicated)
|
||||||
- current elected master of ectd cluster (if running distributed version)
|
- current elected master of etcd cluster (if running distributed version)
|
||||||
- number of master component restarts
|
- number of master component restarts
|
||||||
- number of lost Nodes
|
- number of lost Nodes
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -20,7 +20,7 @@ Use cases which are not listed below are out of the scope of MVP version of Reso
|
||||||
|
|
||||||
HPA uses the latest value of cpu usage as an average aggregated across 1 minute
|
HPA uses the latest value of cpu usage as an average aggregated across 1 minute
|
||||||
(the window may change in the future). The data for a given set of pods
|
(the window may change in the future). The data for a given set of pods
|
||||||
(defined either by pod list or label selector) should be accesible in one request
|
(defined either by pod list or label selector) should be accessible in one request
|
||||||
due to performance issues.
|
due to performance issues.
|
||||||
|
|
||||||
#### Scheduler
|
#### Scheduler
|
||||||
|
|
|
||||||
|
|
@ -86,7 +86,7 @@ service of type LoadBalancer. The native cloud load-balancers on both AWS &
|
||||||
GCE are region-level, and support load-balancing across instances in multiple
|
GCE are region-level, and support load-balancing across instances in multiple
|
||||||
zones (in the same region). For both clouds, the behaviour of the native cloud
|
zones (in the same region). For both clouds, the behaviour of the native cloud
|
||||||
load-balancer is reasonable in the face of failures (indeed, this is why clouds
|
load-balancer is reasonable in the face of failures (indeed, this is why clouds
|
||||||
provide load-balancing as a primitve).
|
provide load-balancing as a primitive).
|
||||||
|
|
||||||
For multi-AZ clusters we will therefore simply rely on the native cloud provider
|
For multi-AZ clusters we will therefore simply rely on the native cloud provider
|
||||||
load balancer behaviour, and we do not anticipate substantial code changes.
|
load balancer behaviour, and we do not anticipate substantial code changes.
|
||||||
|
|
@ -170,8 +170,8 @@ GCE. If you had two volumes both named `myvolume` in two different GCE zones,
|
||||||
this would not be ambiguous when Kubernetes is operating only in a single zone.
|
this would not be ambiguous when Kubernetes is operating only in a single zone.
|
||||||
But, when operating a cluster across multiple zones, `myvolume` is no longer
|
But, when operating a cluster across multiple zones, `myvolume` is no longer
|
||||||
sufficient to specify a volume uniquely. Worse, the fact that a volume happens
|
sufficient to specify a volume uniquely. Worse, the fact that a volume happens
|
||||||
to be unambigious at a particular time is no guarantee that it will continue to
|
to be unambiguous at a particular time is no guarantee that it will continue to
|
||||||
be unambigious in future, because a volume with the same name could
|
be unambiguous in future, because a volume with the same name could
|
||||||
subsequently be created in a second zone. While perhaps unlikely in practice,
|
subsequently be created in a second zone. While perhaps unlikely in practice,
|
||||||
we cannot automatically enable multi-AZ clusters for GCE users if this then causes
|
we cannot automatically enable multi-AZ clusters for GCE users if this then causes
|
||||||
volume mounts to stop working.
|
volume mounts to stop working.
|
||||||
|
|
|
||||||
|
|
@ -162,7 +162,7 @@ For the 1.4 release, this feature will be implemented for the GCE cloud provider
|
||||||
|
|
||||||
- Node: On the node, we expect to see the real source IP of the client. Destination IP will be the Service Virtual External IP.
|
- Node: On the node, we expect to see the real source IP of the client. Destination IP will be the Service Virtual External IP.
|
||||||
|
|
||||||
- Pod: For processes running inside the Pod network namepsace, the source IP will be the real client source IP. The destination address will the be Pod IP.
|
- Pod: For processes running inside the Pod network namespace, the source IP will be the real client source IP. The destination address will the be Pod IP.
|
||||||
|
|
||||||
#### GCE Expected Packet Destination IP (HealthCheck path)
|
#### GCE Expected Packet Destination IP (HealthCheck path)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -75,7 +75,7 @@ There are two types of configuration that are stored on disk:
|
||||||
- cached configurations from a remote source, e.g. `ConfigMaps` from etcd.
|
- cached configurations from a remote source, e.g. `ConfigMaps` from etcd.
|
||||||
- the local "init" configuration, e.g. the set of config files the node is provisioned with.
|
- the local "init" configuration, e.g. the set of config files the node is provisioned with.
|
||||||
|
|
||||||
The Kubelet should accept a `--dynamic-config-dir` flag, which specifies a directory for storing all of the information necessary for dynamic configuraiton from remote sources; e.g. the cached configurations, which configuration is currently in use, which configurations are known to be bad, etc.
|
The Kubelet should accept a `--dynamic-config-dir` flag, which specifies a directory for storing all of the information necessary for dynamic configuration from remote sources; e.g. the cached configurations, which configuration is currently in use, which configurations are known to be bad, etc.
|
||||||
- When the Kubelet downloads a `ConfigMap`, it will checkpoint a serialization of the `ConfigMap` object to a file at `{dynamic-config-dir}/checkpoints/{UID}`.
|
- When the Kubelet downloads a `ConfigMap`, it will checkpoint a serialization of the `ConfigMap` object to a file at `{dynamic-config-dir}/checkpoints/{UID}`.
|
||||||
- We checkpoint the entire object, rather than unpacking the contents to disk, because the former is less complex and reduces chance for errors during the checkpoint process.
|
- We checkpoint the entire object, rather than unpacking the contents to disk, because the former is less complex and reduces chance for errors during the checkpoint process.
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -199,7 +199,7 @@ clients attaching as well.
|
||||||
|
|
||||||
There are ad-hoc solutions/discussions that addresses one or two of the
|
There are ad-hoc solutions/discussions that addresses one or two of the
|
||||||
requirements, but no comprehensive solution for CRI specifically has been
|
requirements, but no comprehensive solution for CRI specifically has been
|
||||||
proposed so far (with the excpetion of @tmrtfs's proposal
|
proposed so far (with the exception of @tmrtfs's proposal
|
||||||
[#33111](https://github.com/kubernetes/kubernetes/pull/33111), which has a much
|
[#33111](https://github.com/kubernetes/kubernetes/pull/33111), which has a much
|
||||||
wider scope). It has come up in discussions that kubelet can delegate all the
|
wider scope). It has come up in discussions that kubelet can delegate all the
|
||||||
logging management to the runtime to allow maximum flexibility. However, it is
|
logging management to the runtime to allow maximum flexibility. However, it is
|
||||||
|
|
|
||||||
|
|
@ -74,7 +74,7 @@ In addition, the rkt cli has historically been the primary interface to the rkt
|
||||||
|
|
||||||
The initial integration will execute the rkt binary directly for app creation/start/stop/removal, as well as image pulling/removal.
|
The initial integration will execute the rkt binary directly for app creation/start/stop/removal, as well as image pulling/removal.
|
||||||
|
|
||||||
The creation of pod sanbox is also done via rkt command line, but it will run under `systemd-run` so it's monitored by the init process.
|
The creation of pod sandbox is also done via rkt command line, but it will run under `systemd-run` so it's monitored by the init process.
|
||||||
|
|
||||||
In the future, some of these decisions are expected to be changed such that rkt is vendored as a library dependency for all operations, and other init systems will be supported as well.
|
In the future, some of these decisions are expected to be changed such that rkt is vendored as a library dependency for all operations, and other init systems will be supported as well.
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue