Documentation for Kubernetes v1.7 is no longer actively maintained. The version you are currently viewing is a static snapshot. For up-to-date documentation, see the latest version.

Edit This Page

kubeadm Setup Tool Reference Guide

This document provides information on how to use kubeadm’s advanced options.

Running kubeadm init bootstraps a Kubernetes cluster. This consists of the following steps:

  1. kubeadm runs a series of pre-flight checks to validate the system state before making changes. Some checks only trigger warnings, others are considered errors and will exit kubeadm until the problem is corrected or the user specifies --skip-preflight-checks.

  2. kubeadm generates a token that additional nodes can use to register themselves with the master in future. Optionally, the user can provide a token.

  3. kubeadm generates a self-signed CA to provision identities for each component (including nodes) in the cluster. It also generates client certificates to be used by various components. If the user has provided their own CA by dropping it in the cert directory (configured via --cert-dir, by default /etc/kubernetes/pki), this step is skipped.

  4. Outputting a kubeconfig file for the kubelet to use to connect to the API server, as well as an additional kubeconfig file for administration.

  5. kubeadm generates Kubernetes static Pod manifests for the API server, controller manager and scheduler. It places them in /etc/kubernetes/manifests. The kubelet watches this directory for Pods to create on startup. These are the core components of Kubernetes. Once they are up and running kubeadm can set up and manage any additional components.

  6. kubeadm “taints” the master node so that only control plane components will run there. It also sets up the RBAC authorization system and writes a special ConfigMap that is used to bootstrap trust with the kubelets.

  7. kubeadm installs add-on components via the API server. Right now this is the internal DNS server and the kube-proxy DaemonSet.

Running kubeadm join on each node in the cluster consists of the following steps:

  1. kubeadm downloads root CA information from the API server. It uses the token to verify the authenticity of that data.

  2. kubeadm creates a local key pair. It prepares a certificate signing request (CSR) and sends that off to the API server for signing. The bootstrap token is used to authenticate. The control plane will sign this CSR requested automatically.

  3. kubeadm configures the local kubelet to connect to the API server.

Usage

Fields that support multiple values do so either with comma separation, or by specifying the flag multiple times.

The kubeadm command line interface is currently in beta. We are aiming to not break any scripted use of the main kubeadm init and kubeadm join. The single exception here is the format of the kubeadm config file as detailed below. That format is still considered alpha and may change.

kubeadm init

It is usually sufficient to run kubeadm init without any flags, but in some cases you might like to override the default behaviour. Here we specify all the flags that can be used to customise the Kubernetes installation.

This is the address the API Server will advertise to other members of the cluster. This is also the address used to construct the suggested kubeadm join line at the end of the init process. If not set (or set to 0.0.0.0) then IP for the default interface will be used.

This address is also added to the certificate that the API Server uses.

The port that the API server will bind on. This defaults to 6443.

Additional hostnames or IP addresses that should be added to the Subject Alternate Name section for the certificate that the API Server will use. If you expose the API Server through a load balancer and public DNS you could specify this with

--apiserver-cert-extra-sans=kubernetes.example.com,kube.example.com,10.100.245.1

The path where to save and store the certificates. The default is “/etc/kubernetes/pki”.

A kubeadm specific config file. This can be used to specify an extended set of options including passing arbitrary command line flags to the control plane components.

Note: When providing configuration values using both a configuration file and flags, the file will take precedence. For example, if a file exists with:

apiVersion: kubeadm.k8s.io/v1alpha1
kind: MasterConfiguration
token: 1234

and the user ran kubeadm init --config file.yaml --token 5678, the chosen token value will be 1234.

The v1.6 version of kubeadm only supports building clusters that are at least v1.6.0. There are many reasons for this including kubeadm’s use of RBAC, the Bootstrap Token system, and enhancements to the Certificates API. With this flag you can try any future version of Kubernetes. Check releases page for a full list of available versions.

For certain networking solutions the Kubernetes master can also play a role in allocating network ranges (CIDRs) to each node. This includes many cloud providers and flannel. You can specify a subnet range that will be broken down and handed out to each node with the --pod-network-cidr flag. This should be a minimum of a /16 so controller-manager is able to assign /24 subnets to each node in the cluster. If you are using flannel with this manifest you should use --pod-network-cidr=10.244.0.0/16. Most CNI based networking solutions do not require this flag.

You can use the --service-cidr flag to override the subnet Kubernetes uses to assign pods IP addresses. If you do, you will also need to update the /etc/systemd/system/kubelet.service.d/10-kubeadm.conf file to reflect this change else DNS will not function correctly.

By default, kubeadm init deploys a cluster that assigns services with DNS names <service_name>.<namespace>.svc.cluster.local. You can use the --service-dns-domain to change the DNS name suffix. Again, you will need to update the /etc/systemd/system/kubelet.service.d/10-kubeadm.conf file accordingly else DNS will not function correctly.

Note: This flag has an effect (it’s needed for the kube-dns Deployment manifest and the API Server’s serving certificate) but not as you might expect, since you will have to modify the arguments to the kubelets in the cluster for it to work fully. Specifying DNS parameters using this flag only is not enough. Rewriting the kubelet’s CLI arguments is out of scope for kubeadm as it should be agnostic to how you run the kubelet. However, making all kubelets in the cluster pick up information dynamically via the API is in scope and is a planned feature for upcoming releases.

By default, kubeadm runs a series of preflight checks to validate the system before making any changes. Advanced users can use this flag to bypass these if necessary.

By default, kubeadm init automatically generates the token used to initialise each new node. If you would like to manually specify this token, you can use the --token flag. The token must be of the format [a-z0-9]{6}\.[a-z0-9]{16}. A compatible random token can be generated kubeadm token generate. Tokens can be managed through the API after the cluster is created. See the section on managing tokens below.

This sets an expiration time for the token. This is specified as a duration from the current time. After this time the token will no longer be valid and will be removed. A value of 0 specifies that the token never expires. 0 is the default. See the section on managing tokens below.

kubeadm join

When joining a kubeadm initialized cluster, we need to establish bidirectional trust. This is split into discovery (having the Node trust the Kubernetes master) and TLS bootstrap (having the Kubernetes master trust the Node).

There are 2 main schemes for discovery. The first is to use a shared token along with the IP address of the API server. The second is to provide a file (a subset of the standard kubeconfig file). This file can be a local file or downloaded via an HTTPS URL. The forms are kubeadm join --discovery-token abcdef.1234567890abcdef 1.2.3.4:6443, kubeadm join --discovery-file path/to/file.conf or kubeadm join --discovery-file https://url/file.conf. Only one form can be used. If the discovery information is loaded from a URL, HTTPS must be used and the host installed CA bundle is used to verify the connection.

The TLS bootstrap mechanism is also driven via a shared token. This is used to temporarily authenticate with the Kubernetes master to submit a certificate signing request (CSR) for a locally created key pair. By default kubeadm will set up the Kubernetes master to automatically approve these signing requests. This token is passed in with the --tls-bootstrap-token abcdef.1234567890abcdef flag.

Often times the same token is used for both parts. In this case, the --token flag can be used instead of specifying the each token individually.

Here’s an example on how to use it:

kubeadm join --token=abcdef.1234567890abcdef 192.168.1.1:6443

Specific options:

Extended options a specified in the kubeadm specific config file.

By default, kubeadm runs a series of preflight checks to validate the system before making any changes. Advanced users can use this flag to bypass these if necessary.

A local file path or HTTPS URL. The file specified must be a kubeconfig file with nothing but an unnamed cluster entry. This is used to find both the location of the API server to join along with a root CA bundle to use when talking to that server.

This might look something like this:

apiVersion: v1
clusters:
- cluster:
    certificate-authority-data: <really long certificate data>
    server: https://10.138.0.2:6443
  name: ""
contexts: []
current-context: ""
kind: Config
preferences: {}
users: []

The discovery token is used along with the address of the API server (as an unnamed argument) to download and verify information about the cluster. The most critical part of the cluster information is the root CA bundle used to verify the identity of the server during subsequent TLS connections.

The token used to authenticate to the API server for the purposes of TLS bootstrapping.

Often times the same token is used for both --discovery-token and --tls-bootstrap-token. This option specifies the same token for both. Other flags override this flag if present.

Using kubeadm with a configuration file

WARNING: While kubeadm command line interface is in beta, the config file is still considered alpha and may change in future versions.

It’s possible to configure kubeadm with a configuration file instead of command line flags, and some more advanced features may only be available as configuration file options. This file is passed in to the --config option on both kubeadm init and kubeadm join.

Sample Master Configuration

apiVersion: kubeadm.k8s.io/v1alpha1
kind: MasterConfiguration
api:
  advertiseAddress: <address|string>
  bindPort: <int>
etcd:
  endpoints:
  - <endpoint1|string>
  - <endpoint2|string>
  caFile: <path|string>
  certFile: <path|string>
  keyFile: <path|string>
networking:
  dnsDomain: <string>
  serviceSubnet: <cidr>
  podSubnet: <cidr>
kubernetesVersion: <string>
cloudProvider: <string>
authorizationModes:
- <authorizationMode1|string>
- <authorizationMode2|string>
token: <string>
tokenTTL: <time duration>
selfHosted: <bool>
apiServerExtraArgs:
  <argument>: <value|string>
  <argument>: <value|string>
controllerManagerExtraArgs:
  <argument>: <value|string>
  <argument>: <value|string>
schedulerExtraArgs:
  <argument>: <value|string>
  <argument>: <value|string>
apiServerCertSANs:
- <name1|string>
- <name2|string>
certificatesDir: <string>

In addition, if authorizationMode is set to ABAC, you should write the config to /etc/kubernetes/abac_policy.json. However, if authorizationMode is set to Webhook, you should write the config to /etc/kubernetes/webhook_authz.conf.

Sample Node Configuration

apiVersion: kubeadm.k8s.io/v1alpha1
kind: NodeConfiguration
caCertPath: <path|string>
discoveryFile: <path|string>
discoveryToken: <string>
discoveryTokenAPIServers:
- <address|string>
- <address|string>
tlsBootstrapToken: <string>

Securing your installation even more

The defaults for kubeadm may not work for everyone. This section documents how to tighten up a kubeadm install at the cost of some usability.

Turning off auto-approval of Node Client Certificates

By default, there is an CSR auto-approver enabled that basically approves any client certificate request for a kubelet when a Bootstrap Token was used when authenticating. If you don’t want the cluster to automatically approve kubelet client certs, you can turn it off by executing this command:

$ kubectl delete clusterrole kubeadm:node-autoapprove-bootstrap

After that, kubeadm join will block until the admin has manually approved the CSR in flight:

$ kubectl get csr
NAME                                                   AGE       REQUESTOR                 CONDITION
node-csr-c69HXe7aYcqkS1bKmH4faEnHAWxn6i2bHZ2mD04jZyQ   18s       system:bootstrap:878f07   Pending

$ kubectl certificate approve node-csr-c69HXe7aYcqkS1bKmH4faEnHAWxn6i2bHZ2mD04jZyQ
certificatesigningrequest "node-csr-c69HXe7aYcqkS1bKmH4faEnHAWxn6i2bHZ2mD04jZyQ" approved

$ kubectl get csr
NAME                                                   AGE       REQUESTOR                 CONDITION
node-csr-c69HXe7aYcqkS1bKmH4faEnHAWxn6i2bHZ2mD04jZyQ   1m        system:bootstrap:878f07   Approved,Issued

Only after kubectl certificate approve has been run, kubeadm join can proceed.

Turning off public access to the cluster-info ConfigMap

In order to achieve the joining flow using the token as the only piece of validation information, a public ConfigMap with some data needed for validation of the master’s identity is exposed publicly by default. While there is no private data in this ConfigMap, some users are sensitive and wish to turn it off regardless. Doing so will disable the ability to use the --discovery-token flag of the kubeadm join flow. Here are the steps to do so:

Fetch the cluster-info file from the API Server:

$ kubectl -n kube-public get cm cluster-info -oyaml | grep "kubeconfig:" -A11 | grep "apiVersion" -A10 | sed "s/    //" | tee cluster-info.yaml
apiVersion: v1
clusters:
- cluster:
    certificate-authority-data: <ca-cert>
    server: https://<ip>:<port>
  name: ""
contexts: []
current-context: ""
kind: Config
preferences: {}
users: []

You can then use the cluster-info.yaml file as an argument to kubeadm join --discovery-file.

Turning of public access to the cluster-info ConfigMap:

$ kubectl -n kube-public delete rolebinding kubeadm:bootstrap-signer-clusterinfo

These command should be run after kubeadm init but before kubeadm join.

Managing Tokens

You can use the kubeadm tool to manage tokens on a running cluster. It will automatically grab the default admin credentials on a master from a kubeadm created cluster (/etc/kubernetes/admin.conf). You can specify an alternate kubeconfig file for credentials with the --kubeconfig to the following commands.

In addition, you can use the kubeadm token generate command to locally creates a new token. This token is of the correct form for specifying with the --token argument to kubeadm init.

For the gory details on how the tokens are implemented (including managing them outside of kubeadm) see the Bootstrap Token docs.

Automating kubeadm

Rather than copying the token you obtained from kubeadm init to each node, as in the basic kubeadm tutorial, you can parallelize the token distribution for easier automation. To implement this automation, you must know the IP address that the master will have after it is started.

  1. Generate a token. This token must have the form <6 character string>.<16 character string>. More formally, it must match the regex: [a-z0-9]{6}\.[a-z0-9]{16}.

    kubeadm can generate a token for you:

    kubeadm token generate
    
  2. Start both the master node and the worker nodes concurrently with this token. As they come up they should find each other and form the cluster. The same --token argument can be used on both kubeadm init and kubeadm join.

Once the cluster is up, you can grab the admin credentials from the master node at /etc/kubernetes/admin.conf and use that to talk to the cluster.

Use Kubeadm with other CRI runtimes

Since Kubernetes 1.6 release, Kubernetes container runtimes have been transferred to using CRI by default. Currently, the build-in container runtime is Docker which is enabled by build-in dockershim in kubelet.

Using other CRI based runtimes with kubeadm is very simple, and currently supported runtimes are:

After you have successfully installed kubeadm and kubelet, please follow these two steps:

  1. Install runtime shim on every node. You will need to follow the installation document in the runtime shim project listing above.

  2. Configure kubelet to use remote CRI runtime. Please remember to change RUNTIME_ENDPOINT to your own value like /var/run/{your_runtime}.sock:

  $ cat > /etc/systemd/system/kubelet.service.d/20-cri.conf <<EOF
Environment="KUBELET_EXTRA_ARGS=--container-runtime=remote --container-runtime-endpoint=$RUNTIME_ENDPOINT --feature-gates=AllAlpha=true"
EOF
  $ systemctl daemon-reload

Now kubelet is ready to use the specified CRI runtime, and you can continue with kubeadm init and kubeadm join workflow to deploy Kubernetes cluster.

Running kubeadm without an internet connection

All of the control plane components run in Pods started by the kubelet and the following images are required for the cluster works will be automatically pulled by the kubelet if they don’t exist locally while kubeadm init is initializing your master:

Image Name v1.6 release branch version v1.7 release branch version
gcr.io/google_containers/kube-apiserver-${ARCH} v1.6.x v1.7.x
gcr.io/google_containers/kube-controller-manager-${ARCH} v1.6.x v1.7.x
gcr.io/google_containers/kube-scheduler-${ARCH} v1.6.x v1.7.x
gcr.io/google_containers/kube-proxy-${ARCH} v1.6.x v1.7.x
gcr.io/google_containers/etcd-${ARCH} 3.0.17 3.0.17
gcr.io/google_containers/pause-${ARCH} 3.0 3.0
gcr.io/google_containers/k8s-dns-sidecar-${ARCH} 1.14.1 1.14.4
gcr.io/google_containers/k8s-dns-kube-dns-${ARCH} 1.14.1 1.14.4
gcr.io/google_containers/k8s-dns-dnsmasq-nanny-${ARCH} 1.14.1 1.14.4

Here v1.7.x means the “latest patch release of the v1.7 branch”.

${ARCH} can be one of: amd64, arm, arm64, ppc64le or s390x.

Managing the kubeadm drop-in file for the kubelet

The kubeadm deb package ships with configuration for how the kubelet should be run. Note that the kubeadm CLI command will never touch this drop-in file. This drop-in file belongs to the kubeadm deb/rpm package.

This is what it looks like in v1.7:

[Service]
Environment="KUBELET_KUBECONFIG_ARGS=--kubeconfig=/etc/kubernetes/kubelet.conf --require-kubeconfig=true"
Environment="KUBELET_SYSTEM_PODS_ARGS=--pod-manifest-path=/etc/kubernetes/manifests --allow-privileged=true"
Environment="KUBELET_NETWORK_ARGS=--network-plugin=cni --cni-conf-dir=/etc/cni/net.d --cni-bin-dir=/opt/cni/bin"
Environment="KUBELET_DNS_ARGS=--cluster-dns=10.96.0.10 --cluster-domain=cluster.local"
Environment="KUBELET_AUTHZ_ARGS=--authorization-mode=Webhook --client-ca-file=/etc/kubernetes/pki/ca.crt"
Environment="KUBELET_CADVISOR_ARGS=--cadvisor-port=0"
ExecStart=
ExecStart=/usr/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_SYSTEM_PODS_ARGS $KUBELET_NETWORK_ARGS $KUBELET_DNS_ARGS $KUBELET_AUTHZ_ARGS $KUBELET_CADVISOR_ARGS $KUBELET_EXTRA_ARGS

A breakdown of what/why:

Cloudprovider integrations (experimental)

Enabling specific cloud providers is a common request. This currently requires manual configuration and is therefore not yet fully supported. If you wish to do so, edit the kubeadm drop-in for the kubelet service (/etc/systemd/system/kubelet.service.d/10-kubeadm.conf) on all nodes, including the master. If your cloud provider requires any extra packages installed on the host, for example for volume mounting/unmounting, install those packages.

Specify the --cloud-provider flag for the kubelet and set it to the cloud of your choice. If your cloudprovider requires a configuration file, create the file /etc/kubernetes/cloud-config on every node. The exact format and content of that file depends on the requirements imposed by your cloud provider. If you use the /etc/kubernetes/cloud-config file, you must append it to the kubelet arguments as follows: --cloud-config=/etc/kubernetes/cloud-config

Note that there is most likely other per-provider configuration that may be needed (IAM roles for AWS) that is currently underdocumented.

Next, specify the cloud provider in the kubeadm config file. Create a file called kubeadm.conf with the following contents:

kind: MasterConfiguration
apiVersion: kubeadm.k8s.io/v1alpha1
cloudProvider: <cloud provider>

Lastly, run kubeadm init --config=kubeadm.conf to bootstrap your cluster with the cloud provider.

This workflow is not yet fully supported, however we hope to make it extremely easy to spin up clusters with cloud providers in the future. (See this proposal for more information) The Kubelet Dynamic Settings feature may also help to fully automate this process in the future.

Environment variables

There are some environment variables that modify the way that kubeadm works. Most users will have no need to set these. These environment variables are a short-term solution, eventually they will be integrated in the kubeadm configuration file.

Note: These environment variables are deprecated and will stop functioning in v1.8!

Variable Default Description
KUBE_KUBERNETES_DIR /etc/kubernetes Where most configuration files are written to and read from
KUBE_HYPERKUBE_IMAGE   If set, use a single hyperkube image with this name. If not set, individual images per server component will be used.
KUBE_ETCD_IMAGE gcr.io/google_containers/etcd-<arch>:3.0.17 The etcd container image to use.
KUBE_REPO_PREFIX gcr.io/google_containers The image prefix for all images that are used.

If KUBE_KUBERNETES_DIR is specified, you may need to rewrite the arguments of the kubelet. (e.g. –kubeconfig, –pod-manifest-path)

If KUBE_REPO_PREFIX is specified, you may need to set the kubelet flag --pod-infra-container-image which specifies which pause image to use.

Defaults to gcr.io/google_containers/pause-${ARCH}:3.0 where ${ARCH} can be one of amd64, arm, arm64, ppc64le or s390x.

cat > /etc/systemd/system/kubelet.service.d/20-pod-infra-image.conf <<EOF
[Service]
Environment="KUBELET_EXTRA_ARGS=--pod-infra-container-image=<your-image>"
EOF
systemctl daemon-reload
systemctl restart kubelet

If you want to use kubeadm with an http proxy, you may need to configure it to support http_proxy, https_proxy, or no_proxy.

For example, if your kube master node IP address is 10.18.17.16 and you have a proxy which supports both http/https on 10.18.17.16 port 8080, you can use the following command:

export PROXY_PORT=8080
export PROXY_IP=10.18.17.16
export http_proxy=http://$PROXY_IP:$PROXY_PORT
export HTTP_PROXY=$http_proxy
export https_proxy=$http_proxy
export HTTPS_PROXY=$http_proxy
export no_proxy="localhost,127.0.0.1,localaddress,.localdomain.com,example.com,10.18.17.16"

Remember to change proxy_ip and add a kube master node IP address to no_proxy.

Using custom certificates

By default kubeadm will generate all the certificates needed for a cluster to run. You can override this behaviour by providing your own certificates.

To do so, you must place them in whatever directory is specified by the --cert-dir flag or CertificatesDir configuration file key. By default this is /etc/kubernetes/pki.

If a given certificate and private key pair both exist, kubeadm will skip the generation step and those files will be validated and used for the prescribed use-case.

This means you can, for example, prepopulate /etc/kubernetes/pki/ca.crt and /etc/kubernetes/pki/ca.key with an existing CA, which then will be used for signing the rest of the certs.

Releases and release notes

If you already have kubeadm installed and want to upgrade, run apt-get update && apt-get upgrade or yum update to get the latest version of kubeadm.

Refer to the CHANGELOG.md for more information.

Analytics

Create an Issue Edit this Page