- CI issues were due to v1alpha2 being added, created Calico there
- Added some tests around Calico, similar to Weave
- Added conversion functions similar to Weave in zz_generated_conversion
Adding the option to install Calico with the `--networking calico`
argument. This will currently deploy Calico v2.0 to the cluster.
Documentation has also been updated with information about Calico and
where one can find more information or help.
By testing with data from various schema versions, we effectively check
that they are equivalent.
Also this uncovered a few places where we were not strictly ordering
things - add some sorts in there.
* Zones are now subnets
* Utility subnet is no longer part of Zone
* Bastion InstanceGroup type added instead
* Etcd clusters defined in terms of InstanceGroups, not zones
* AdminAccess split into SSHAccess & APIAccess
* Dropped unused Multizone flag
* kopsrepo/master: (29 commits)
Add verify-boilerplate target
Add logging of AWS retries
adding hack/verify-boilerplate.sh to make ci target
Print time remaining to succeed as a positive value
adding hack/verify-boilerplate.sh to make ci target
updating headers, OMG we need this in the ci
Format resource diffs
Include error in message when we fail to query AZs
Import tidying
Apply gofmt
Update cmd/kops/validate_cluster for refactor
Move to pkg/validation and tidy up
Update command building pattern, a few tweaks
adds more machine types
fix path to adding feature doc
Update dns-controller README
bug in my fix header script
updating header
bumping weave version
Remove old file
...
* kopsrepo/master:
gcs-upload: Use a no-clobber copy instead
gcs-upload: Fix cache-control on other files as well
changes from code review
doc updates
unit tests with fakes
it is working in alpha
working on the start of validate
Starting work on node lookup and validation
starting porting node code
Fix retries for AutoScalingGroup pending delete
Apply gofmt to pkg directory
Avoid tests hitting kubernetes stable.txt HTTP file
Fix printing of max size on instance group
Disable kubelet from starting until after volume mounts
Fix Cluster parsing error message
bumping stable channel to k8s 1.4.6
support more zones(cn-north-1a/b) for cloud provider guess
Without this fix, last generated private subnet address overlaps with main CIDR range provided via --network-cidr= option, which causes error.
For example before change, with --network-cidr=10.0.0.0/22, the list of subnets generated by the code was:
```
I1117 07:34:24.720380 47964 cluster.go:503] Assigned CIDR 10.0.1.128/25 to zone us-east-1c
I1117 07:34:24.720397 47964 cluster.go:514] Assigned Private CIDR 10.0.3.0/25 to zone us-east-1c
I1117 07:34:24.720404 47964 cluster.go:503] Assigned CIDR 10.0.2.0/25 to zone us-east-1d
I1117 07:34:24.720409 47964 cluster.go:514] Assigned Private CIDR 10.0.3.128/25 to zone us-east-1d
I1117 07:34:24.720415 47964 cluster.go:503] Assigned CIDR 10.0.2.128/25 to zone us-east-1e
I1117 07:34:24.720420 47964 cluster.go:514] Assigned Private CIDR 10.0.4.0/25 to zone us-east-1e
```
The last CIDR 10.0.4.0/25 is beyond 10.0.0.0/22 boundaries, which causes the error:
```
W1117 07:39:29.240474 48009 executor.go:100] error running task "subnet/private-us-east-1e.kubpriv.pink-ptdevcloud.com": error creating subnet: InvalidSubnet.Range: The CIDR '10.0.4.0/25' is invalid.
status code: 400, request id: b195c64b-0a35-413c-b6ec-d7ee40d49adb
```
With a code fix, subnets get generated in a correct way:
```
I1118 07:22:31.466899 55710 cluster.go:503] Assigned CIDR 10.0.1.0/25 to zone us-east-1c
I1118 07:22:31.466908 55710 cluster.go:514] Assigned Private CIDR 10.0.2.128/25 to zone us-east-1c
I1118 07:22:31.466913 55710 cluster.go:503] Assigned CIDR 10.0.1.128/25 to zone us-east-1d
I1118 07:22:31.466917 55710 cluster.go:514] Assigned Private CIDR 10.0.3.0/25 to zone us-east-1d
I1118 07:22:31.466922 55710 cluster.go:503] Assigned CIDR 10.0.2.0/25 to zone us-east-1e
I1118 07:22:31.466925 55710 cluster.go:514] Assigned Private CIDR 10.0.3.128/25 to zone us-east-1e
```