89 lines
3.6 KiB
YAML
89 lines
3.6 KiB
YAML
## Valid bootstrap options (required): ubuntu, coreos, centos, none
|
|
## If the OS is not listed here, it means it doesn't require extra/bootstrap steps.
|
|
## In example, python is not available on 'coreos' so it must be installed before
|
|
## anything else. In the opposite, Debian has already all its dependencies fullfiled, then bootstrap_os should be set to `none`.
|
|
bootstrap_os: none
|
|
|
|
## Directory where etcd data stored
|
|
etcd_data_dir: /var/lib/etcd
|
|
|
|
## Directory where the binaries will be installed
|
|
bin_dir: /usr/local/bin
|
|
|
|
## The access_ip variable is used to define how other nodes should access
|
|
## the node. This is used in flannel to allow other flannel nodes to see
|
|
## this node for example. The access_ip is really useful AWS and Google
|
|
## environments where the nodes are accessed remotely by the "public" ip,
|
|
## but don't know about that address themselves.
|
|
#access_ip: 1.1.1.1
|
|
|
|
|
|
## External LB example config
|
|
## apiserver_loadbalancer_domain_name: "elb.some.domain"
|
|
#loadbalancer_apiserver:
|
|
# address: 1.2.3.4
|
|
# port: 1234
|
|
|
|
## Internal loadbalancers for apiservers
|
|
#loadbalancer_apiserver_localhost: true
|
|
|
|
## Local loadbalancer should use this port instead, if defined.
|
|
## Defaults to kube_apiserver_port (6443)
|
|
#nginx_kube_apiserver_port: 8443
|
|
|
|
### OTHER OPTIONAL VARIABLES
|
|
## For some things, kubelet needs to load kernel modules. For example, dynamic kernel services are needed
|
|
## for mounting persistent volumes into containers. These may not be loaded by preinstall kubernetes
|
|
## processes. For example, ceph and rbd backed volumes. Set to true to allow kubelet to load kernel
|
|
## modules.
|
|
#kubelet_load_modules: false
|
|
|
|
## With calico it is possible to distributed routes with border routers of the datacenter.
|
|
## Warning : enabling router peering will disable calico's default behavior ('node mesh').
|
|
## The subnets of each nodes will be distributed by the datacenter router
|
|
#peer_with_router: false
|
|
|
|
## With contiv, L3 BGP mode is possible by setting contiv_fwd_mode to "routing".
|
|
## In this case, you may need to peer with an uplink
|
|
## NB: The hostvars must contain a key "contiv" of which value is a dict containing "router_ip", "as"(defaults to contiv_global_as), "neighbor_as" (defaults to contiv_global_neighbor_as), "neighbor"
|
|
#contiv_peer_with_uplink_leaf: false
|
|
#contiv_global_as: "65002"
|
|
#contiv_global_neighbor_as: "500"
|
|
|
|
## Upstream dns servers used by dnsmasq
|
|
#upstream_dns_servers:
|
|
# - 8.8.8.8
|
|
# - 8.8.4.4
|
|
|
|
## There are some changes specific to the cloud providers
|
|
## for instance we need to encapsulate packets with some network plugins
|
|
## If set the possible values are either 'gce', 'aws', 'azure', 'openstack', 'vsphere', 'oci', or 'external'
|
|
## When openstack is used make sure to source in the openstack credentials
|
|
## like you would do when using nova-client before starting the playbook.
|
|
#cloud_provider:
|
|
|
|
|
|
## Uncomment to enable experimental kubeadm deployment mode
|
|
#kubeadm_enabled: false
|
|
|
|
## Set these proxy values in order to update package manager and docker daemon to use proxies
|
|
#http_proxy: ""
|
|
#https_proxy: ""
|
|
|
|
## Refer to roles/kubespray-defaults/defaults/main.yml before modifying no_proxy
|
|
#no_proxy: ""
|
|
|
|
## Certificate Management
|
|
## This setting determines whether certs are generated via scripts or whether a
|
|
## cluster of Hashicorp's Vault is started to issue certificates (using etcd
|
|
## as a backend). Options are "script" or "vault"
|
|
#cert_management: script
|
|
|
|
## Set to true to allow pre-checks to fail and continue deployment
|
|
#ignore_assert_errors: false
|
|
|
|
## The read-only port for the Kubelet to serve on with no authentication/authorization. Uncomment to enable.
|
|
#kube_read_only_port: 10255
|
|
|
|
## Set true to download and cache container
|
|
#download_container: true
|