2019-12-04 15:22:57 +00:00
# Getting started
2016-07-04 12:13:18 +00:00
2019-12-04 15:22:57 +00:00
## Building your own inventory
2016-12-09 09:38:38 +00:00
2017-02-14 10:08:27 +00:00
Ansible inventory can be stored in 3 formats: YAML, JSON, or INI-like. There is
2016-12-09 09:38:38 +00:00
an example inventory located
2019-04-20 18:09:45 +00:00
[here ](https://github.com/kubernetes-sigs/kubespray/blob/master/inventory/sample/inventory.ini ).
2016-12-09 09:38:38 +00:00
2017-01-11 15:15:04 +00:00
You can use an
2018-11-28 08:15:25 +00:00
[inventory generator ](https://github.com/kubernetes-sigs/kubespray/blob/master/contrib/inventory_builder/inventory.py )
2016-12-09 09:38:38 +00:00
to create or modify an Ansible inventory. Currently, it is limited in
2017-09-25 11:19:38 +00:00
functionality and is only used for configuring a basic Kubespray cluster inventory, but it does
support creating inventory file for large clusters as well. It now supports
2017-01-11 15:15:04 +00:00
separated ETCD and Kubernetes master roles from node role if the size exceeds a
2017-09-25 11:19:38 +00:00
certain threshold. Run `python3 contrib/inventory_builder/inventory.py help` help for more information.
2017-01-11 15:15:04 +00:00
Example inventory generator usage:
2017-01-11 11:46:44 +00:00
2019-12-04 15:22:57 +00:00
```ShellSession
cp -r inventory/sample inventory/mycluster
declare -a IPS=(10.10.1.3 10.10.1.4 10.10.1.5)
CONFIG_FILE=inventory/mycluster/hosts.yml python3 contrib/inventory_builder/inventory.py ${IPS[@]}
```
2019-04-20 18:09:45 +00:00
Then use `inventory/mycluster/hosts.yml` as inventory file.
2017-01-11 11:46:44 +00:00
2019-12-04 15:22:57 +00:00
## Starting custom deployment
2017-01-11 11:46:44 +00:00
Once you have an inventory, you may want to customize deployment data vars
and start the deployment:
2018-02-16 12:53:35 +00:00
**IMPORTANT**: Edit my\_inventory/groups\_vars/\*.yaml to override data vars:
2017-02-14 10:09:18 +00:00
2019-12-04 15:22:57 +00:00
```ShellSession
ansible-playbook -i inventory/mycluster/hosts.yml cluster.yml -b -v \
--private-key=~/.ssh/private_key
```
2017-01-11 11:46:44 +00:00
2020-03-20 22:40:44 +00:00
See more details in the [ansible guide ](docs/ansible.md ).
2017-05-24 19:49:21 +00:00
2019-12-04 15:22:57 +00:00
### Adding nodes
2017-05-24 19:49:21 +00:00
2018-07-12 15:22:11 +00:00
You may want to add worker, master or etcd nodes to your existing cluster. This can be done by re-running the `cluster.yml` playbook, or you can target the bare minimum needed to get kubelet installed on the worker and talking to your masters. This is especially helpful when doing something like autoscaling your clusters.
2017-05-24 19:49:21 +00:00
2019-12-04 15:22:57 +00:00
- Add the new worker node to your inventory in the appropriate group (or utilize a [dynamic inventory ](https://docs.ansible.com/ansible/intro_dynamic_inventory.html )).
- Run the ansible-playbook command, substituting `cluster.yml` for `scale.yml` :
2018-02-16 12:53:35 +00:00
2019-12-04 15:22:57 +00:00
```ShellSession
ansible-playbook -i inventory/mycluster/hosts.yml scale.yml -b -v \
--private-key=~/.ssh/private_key
```
2017-09-06 17:20:25 +00:00
2019-12-04 15:22:57 +00:00
### Remove nodes
2018-03-13 06:05:03 +00:00
2019-08-07 11:46:50 +00:00
You may want to remove **master** , **worker** , or **etcd** nodes from your
existing cluster. This can be done by re-running the `remove-node.yml`
playbook. First, all specified nodes will be drained, then stop some
kubernetes services and delete some certificates,
and finally execute the kubectl command to delete these nodes.
This can be combined with the add node function. This is generally helpful
when doing something like autoscaling your clusters. Of course, if a node
is not working, you can remove the node and install it again.
Use `--extra-vars "node=<nodename>,<nodename2>"` to select the node(s) you want to delete.
2019-12-04 15:22:57 +00:00
```ShellSession
2019-04-20 18:09:45 +00:00
ansible-playbook -i inventory/mycluster/hosts.yml remove-node.yml -b -v \
2019-12-04 15:22:57 +00:00
--private-key=~/.ssh/private_key \
--extra-vars "node=nodename,nodename2"
2018-06-29 05:51:04 +00:00
```
2018-03-13 06:05:03 +00:00
2019-08-07 11:46:50 +00:00
If a node is completely unreachable by ssh, add `--extra-vars reset_nodes=no`
to skip the node reset step. If one node is unavailable, but others you wish
to remove are able to connect via SSH, you could set reset_nodes=no as a host
var in inventory.
2019-12-04 15:22:57 +00:00
## Connecting to Kubernetes
2018-02-16 12:53:35 +00:00
2017-09-06 17:20:25 +00:00
By default, Kubespray configures kube-master hosts with insecure access to
kube-apiserver via port 8080. A kubeconfig file is not necessary in this case,
2018-02-16 12:53:35 +00:00
because kubectl will use < http: / / localhost:8080 > to connect. The kubeconfig files
2017-09-06 17:20:25 +00:00
generated will point to localhost (on kube-masters) and kube-node hosts will
connect either to a localhost nginx proxy or to a loadbalancer if configured.
2020-03-20 22:40:44 +00:00
More details on this process are in the [HA guide ](docs/ha-mode.md ).
2017-09-06 17:20:25 +00:00
2018-02-16 12:53:35 +00:00
Kubespray permits connecting to the cluster remotely on any IP of any
kube-master host on port 6443 by default. However, this requires
2020-02-11 09:36:00 +00:00
authentication. One can get a kubeconfig from kube-master hosts
(see [below ](#accessing-kubernetes-api )) or connect with a [username and password ](vars.md#user-accounts ).
2017-09-06 17:20:25 +00:00
For more information on kubeconfig and accessing a Kubernetes cluster, refer to
the Kubernetes [documentation ](https://kubernetes.io/docs/tasks/access-application-cluster/configure-access-multiple-clusters/ ).
2017-09-09 20:38:03 +00:00
2019-12-04 15:22:57 +00:00
## Accessing Kubernetes Dashboard
2017-09-09 20:38:03 +00:00
2020-04-29 14:20:25 +00:00
Supported version is kubernetes-dashboard v2.0.x :
2018-02-16 12:53:35 +00:00
2020-04-29 14:20:25 +00:00
- Login options are : token/kubeconfig by default, basic can be enabled with `kube_basic_auth: true` inventory variable - not recommended because this requires ABAC api-server which is not tested by kubespray team
- Deployed by default in "kube-system" namespace, can be overriden with `dashboard_namespace: kubernetes-dashboard` in inventory,
2019-12-04 15:22:57 +00:00
- Only serves over https
2017-09-09 20:38:03 +00:00
2020-04-29 14:20:25 +00:00
Access is described in [dashboard docs ](https://github.com/kubernetes/dashboard/blob/master/docs/user/accessing-dashboard/1.7.x-and-above.md ). With kubespray's default deployment in kube-system namespace, instead of kuberntes-dashboard :
2017-09-09 20:38:03 +00:00
2020-04-29 14:20:25 +00:00
- Proxy URL is < http: // localhost:8001 / api / v1 / namespaces / kube-system / services / https:kubernetes-dashboard: / proxy /#/ login >
- kubectl commands must be run with "-n kube-system"
2017-09-18 12:30:57 +00:00
2020-04-29 14:20:25 +00:00
Accessing through Ingress is highly recommended. For proxy access, please note that proxy must listen to [localhost ](https://github.com/kubernetes/dashboard/issues/692#issuecomment-220492484 ) (`proxy --address="x.x.x.x"` will not work)
For token authentication, guide to create Service Account is provided in [dashboard sample user ](https://github.com/kubernetes/dashboard/blob/master/docs/user/access-control/creating-sample-user.md ) doc. Still take care of default namespace.
Access can also by achieved via ssh tunnel on a master :
```bash
# localhost:8081 will be sent to master-1's own localhost:8081
ssh -L8001:localhost:8001 user@master-1
sudo -i
kubectl proxy
```
2017-10-15 19:41:17 +00:00
2019-12-04 15:22:57 +00:00
## Accessing Kubernetes API
2017-09-18 12:30:57 +00:00
The main client of Kubernetes is `kubectl` . It is installed on each kube-master
host and can optionally be configured on your ansible host by setting
2018-02-16 12:53:35 +00:00
`kubectl_localhost: true` and `kubeconfig_localhost: true` in the configuration:
2019-12-04 15:22:57 +00:00
- If `kubectl_localhost` enabled, `kubectl` will download onto `/usr/local/bin/` and setup with bash completion. A helper script `inventory/mycluster/artifacts/kubectl.sh` also created for setup with below `admin.conf` .
- If `kubeconfig_localhost` enabled `admin.conf` will appear in the `inventory/mycluster/artifacts/` directory after deployment.
2020-02-11 09:36:00 +00:00
- The location where these files are downloaded to can be configured via the `artifacts_dir` variable.
2018-02-16 12:53:35 +00:00
You can see a list of nodes by running the following commands:
2017-09-18 12:30:57 +00:00
2019-12-04 15:22:57 +00:00
```ShellSession
cd inventory/mycluster/artifacts
./kubectl.sh get nodes
```
2017-09-18 12:30:57 +00:00
2018-02-16 12:53:35 +00:00
If desired, copy admin.conf to ~/.kube/config.