If you are using a released version of Kubernetes, you should refer to the docs that go with that version.
The latest release of this document can be found [here](http://releases.k8s.io/release-1.1/docs/getting-started-guides/rackspace.md).Documentation for other releases can be found at releases.k8s.io.
Table of Contents
- Supported Version: v0.18.1
In general, the dev-build-and-up.sh workflow for Rackspace is the similar to Google Compute Engine. The specific implementation is different due to the use of CoreOS, Rackspace Cloud Files and the overall network design.
These scripts should be used to deploy development environments for Kubernetes. If your account leverages RackConnect or non-standard networking, these scripts will most likely not work without modification.
NOTE: The rackspace scripts do NOT rely on saltstack
and instead rely on cloud-init for configuration.
The current cluster design is inspired by:
- Python2.7
- You need to have both
nova
andswiftly
installed. It's recommended to use a python virtualenv to install these packages into. - Make sure you have the appropriate environment variables set to interact with the OpenStack APIs. See Rackspace Documentation for more details.
- To build your own released version from source use
export KUBERNETES_PROVIDER=rackspace
and run thebash hack/dev-build-and-up.sh
- Note: The get.k8s.io install method is not working yet for our scripts.
- To install the latest released version of Kubernetes use
export KUBERNETES_PROVIDER=rackspace; wget -q -O - https://get.k8s.io | bash
- To install the latest released version of Kubernetes use
- The Kubernetes binaries will be built via the common build scripts in
build/
. - If you've set the ENV
KUBERNETES_PROVIDER=rackspace
, the scripts will uploadkubernetes-server-linux-amd64.tar.gz
to Cloud Files. - A cloud files container will be created via the
swiftly
CLI and a temp URL will be enabled on the object. - The built
kubernetes-server-linux-amd64.tar.gz
will be uploaded to this container and the URL will be passed to master/nodes when booted.
There is a specific cluster/rackspace
directory with the scripts for the following steps:
- A cloud network will be created and all instances will be attached to this network.
- flanneld uses this network for next hop routing. These routes allow the containers running on each node to communicate with one another on this private network.
- A SSH key will be created and uploaded if needed. This key must be used to ssh into the machines (we do not capture the password).
- The master server and additional nodes will be created via the
nova
CLI. Acloud-config.yaml
is generated and provided as user-data with the entire configuration for the systems. - We then boot as many nodes as defined via
$NUM_NODES
.
- The scripts expect
eth2
to be the cloud network that the containers will communicate across. - A number of the items in
config-default.sh
are overridable via environment variables. - For older versions please either:
- Sync back to
v0.9
withgit checkout v0.9
- Download a snapshot of
v0.9
- Sync back to
v0.3
withgit checkout v0.3
- Download a snapshot of
v0.3
- eth0 - Public Interface used for servers/containers to reach the internet
- eth1 - ServiceNet - Intra-cluster communication (k8s, etcd, etc) communicate via this interface. The
cloud-config
files use the special CoreOS identifier$private_ipv4
to configure the services. - eth2 - Cloud Network - Used for k8s pods to communicate with one another. The proxy service will pass traffic via this interface.