ab6f0012cc
* Make local volume provisioner dir mode a variable I need to change this for Nagios monitoring. Others may need to as well. Had to close previous commits, sorry for the spam. * Make local volume provisioner dir mode a variable I need to change this for Nagios monitoring. Others may need to as well. Had to close previous commits, sorry for the spam.
506 lines
20 KiB
YAML
506 lines
20 KiB
YAML
---
|
||
# Use proxycommand if bastion host is in group all
|
||
# This change obseletes editing ansible.cfg file depending on bastion existance
|
||
ansible_ssh_common_args: "{% if 'bastion' in groups['all'] %} -o ProxyCommand='ssh -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no -W %h:%p {{ hostvars['bastion']['ansible_user'] }}@{{ hostvars['bastion']['ansible_host'] }} {% if ansible_ssh_private_key_file is defined %}-i {{ ansible_ssh_private_key_file }}{% endif %} ' {% endif %}"
|
||
|
||
kube_api_anonymous_auth: true
|
||
|
||
# Default value, but will be set to true automatically if detected
|
||
is_atomic: false
|
||
|
||
# optional disable the swap
|
||
disable_swap: true
|
||
|
||
## Change this to use another Kubernetes version, e.g. a current beta release
|
||
kube_version: v1.14.1
|
||
|
||
## Kube Proxy mode One of ['iptables','ipvs']
|
||
kube_proxy_mode: ipvs
|
||
|
||
## Delete kube-proxy daemonset if kube_proxy_remove set, e.g. kube_network_plugin providing proxy services
|
||
kube_proxy_remove: "{{ (kube_network_plugin == 'kube-router') and (kube_router_run_service_proxy is defined and kube_router_run_service_proxy)| bool }}"
|
||
|
||
# A string slice of values which specify the addresses to use for NodePorts.
|
||
# Values may be valid IP blocks (e.g. 1.2.3.0/24, 1.2.3.4/32).
|
||
# The default empty string slice ([]) means to use all local addresses.
|
||
# kube_proxy_nodeport_addresses_cidr is retained for legacy config
|
||
kube_proxy_nodeport_addresses: >-
|
||
{%- if kube_proxy_nodeport_addresses_cidr is defined -%}
|
||
[{{ kube_proxy_nodeport_addresses_cidr }}]
|
||
{%- else -%}
|
||
[]
|
||
{%- endif -%}
|
||
|
||
# Set to true to allow pre-checks to fail and continue deployment
|
||
ignore_assert_errors: false
|
||
|
||
# nginx-proxy configure
|
||
nginx_config_dir: "/etc/nginx"
|
||
|
||
# haproxy configure
|
||
haproxy_config_dir: "/etc/haproxy"
|
||
|
||
# Directory where the binaries will be installed
|
||
bin_dir: /usr/local/bin
|
||
docker_bin_dir: /usr/bin
|
||
etcd_data_dir: /var/lib/etcd
|
||
# Where the binaries will be downloaded.
|
||
# Note: ensure that you've enough disk space (about 1G)
|
||
local_release_dir: "/tmp/releases"
|
||
# Random shifts for retrying failed ops like pushing/downloading
|
||
retry_stagger: 5
|
||
|
||
# Install epel repo on Centos/RHEL
|
||
epel_enabled: false
|
||
|
||
# DNS configuration.
|
||
# Kubernetes cluster name, also will be used as DNS domain
|
||
cluster_name: cluster.local
|
||
# Subdomains of DNS domain to be resolved via /etc/resolv.conf for hostnet pods
|
||
ndots: 2
|
||
# Can be coredns, coredns_dual, manual, or none
|
||
dns_mode: coredns
|
||
|
||
# Enable nodelocal dns cache
|
||
enable_nodelocaldns: true
|
||
nodelocaldns_ip: 169.254.25.10
|
||
|
||
# Should be set to a cluster IP if using a custom cluster DNS
|
||
manual_dns_server: ""
|
||
|
||
# Can be docker_dns, host_resolvconf or none
|
||
resolvconf_mode: docker_dns
|
||
# Deploy netchecker app to verify DNS resolve as an HTTP service
|
||
deploy_netchecker: false
|
||
# Ip address of the kubernetes DNS service (called skydns for historical reasons)
|
||
skydns_server: "{{ kube_service_addresses|ipaddr('net')|ipaddr(3)|ipaddr('address') }}"
|
||
skydns_server_secondary: "{{ kube_service_addresses|ipaddr('net')|ipaddr(4)|ipaddr('address') }}"
|
||
dns_domain: "{{ cluster_name }}"
|
||
|
||
kube_dns_servers:
|
||
coredns: ["{{skydns_server}}"]
|
||
coredns_dual: "{{[skydns_server] + [ skydns_server_secondary ]}}"
|
||
manual: ["{{manual_dns_server}}"]
|
||
|
||
dns_servers: "{{kube_dns_servers[dns_mode]}}"
|
||
|
||
|
||
# Kubernetes configuration dirs and system namespace.
|
||
# Those are where all the additional config stuff goes
|
||
# the kubernetes normally puts in /srv/kubernetes.
|
||
# This puts them in a sane location and namespace.
|
||
# Editing those values will almost surely break something.
|
||
kube_config_dir: /etc/kubernetes
|
||
kube_script_dir: "{{ bin_dir }}/kubernetes-scripts"
|
||
kube_manifest_dir: "{{ kube_config_dir }}/manifests"
|
||
|
||
# This is where all the cert scripts and certs will be located
|
||
kube_cert_dir: "{{ kube_config_dir }}/ssl"
|
||
|
||
# compatibility directory for kubeadm
|
||
kube_cert_compat_dir: "/etc/kubernetes/pki"
|
||
|
||
# This is where all of the bearer tokens will be stored
|
||
kube_token_dir: "{{ kube_config_dir }}/tokens"
|
||
|
||
# This is where to save basic auth file
|
||
kube_users_dir: "{{ kube_config_dir }}/users"
|
||
|
||
|
||
# This is the group that the cert creation scripts chgrp the
|
||
# cert files to. Not really changeable...
|
||
kube_cert_group: kube-cert
|
||
|
||
# Cluster Loglevel configuration
|
||
kube_log_level: 2
|
||
|
||
# Users to create for basic auth in Kubernetes API via HTTP
|
||
kube_api_pwd: "changeme"
|
||
kube_users:
|
||
kube:
|
||
pass: "{{kube_api_pwd}}"
|
||
role: admin
|
||
|
||
# Choose network plugin (cilium, calico, weave or flannel)
|
||
# Can also be set to 'cloud', which lets the cloud provider setup appropriate routing
|
||
kube_network_plugin: calico
|
||
kube_network_plugin_multus: false
|
||
|
||
# Determines if calico-rr group exists
|
||
peer_with_calico_rr: "{{ 'calico-rr' in groups and groups['calico-rr']|length > 0 }}"
|
||
|
||
# Set to false to disable calico-upgrade
|
||
calico_upgrade_enabled: true
|
||
|
||
# Kubernetes internal network for services, unused block of space.
|
||
kube_service_addresses: 10.233.0.0/18
|
||
|
||
# internal network. When used, it will assign IP
|
||
# addresses from this range to individual pods.
|
||
# This network must be unused in your network infrastructure!
|
||
kube_pods_subnet: 10.233.64.0/18
|
||
|
||
# internal network node size allocation (optional). This is the size allocated
|
||
# to each node on your network. With these defaults you should have
|
||
# room for 64 nodes with 254 pods per node.
|
||
# Example: Up to 256 nodes, 100 pods per node (/16 network):
|
||
# - kube_service_addresses: 10.233.0.0/17
|
||
# - kube_pods_subnet: 10.233.128.0/17
|
||
# - kube_network_node_prefix: 25
|
||
# Example: Up to 4096 nodes, 100 pods per node (/12 network):
|
||
# - kube_service_addresses: 10.192.0.0/13
|
||
# - kube_pods_subnet: 10.200.0.0/13
|
||
# - kube_network_node_prefix: 25
|
||
kube_network_node_prefix: 24
|
||
|
||
# The virtual cluster IP, real host IPs and ports the API Server will be
|
||
# listening on.
|
||
# NOTE: loadbalancer_apiserver_localhost somewhat alters the final API enpdoint
|
||
# access IP value (automatically evaluated below)
|
||
kube_apiserver_ip: "{{ kube_service_addresses|ipaddr('net')|ipaddr(1)|ipaddr('address') }}"
|
||
kube_apiserver_bind_address: 0.0.0.0
|
||
# https
|
||
kube_apiserver_port: 6443
|
||
# http
|
||
kube_apiserver_insecure_bind_address: 127.0.0.1
|
||
kube_apiserver_insecure_port: 0
|
||
|
||
# If non-empty, will use this string as identification instead of the actual hostname
|
||
kube_override_hostname: >-
|
||
{%- if cloud_provider is defined and cloud_provider in [ 'aws' ] -%}
|
||
{%- else -%}
|
||
{{ inventory_hostname }}
|
||
{%- endif -%}
|
||
|
||
# dynamic kubelet configuration
|
||
dynamic_kubelet_configuration: false
|
||
|
||
# define kubelet config dir for dynamic kubelet
|
||
# kubelet_config_dir:
|
||
default_kubelet_config_dir: "{{ kube_config_dir }}/dynamic_kubelet_dir"
|
||
dynamic_kubelet_configuration_dir: "{{ kubelet_config_dir | default(default_kubelet_config_dir) }}"
|
||
|
||
# Aggregator
|
||
kube_api_aggregator_routing: false
|
||
|
||
# Profiling
|
||
kube_profiling: false
|
||
|
||
# Container for runtime
|
||
container_manager: docker
|
||
|
||
## Uncomment this if you want to force overlay/overlay2 as docker storage driver
|
||
## Please note that overlay2 is only supported on newer kernels
|
||
# docker_storage_options: -s overlay2
|
||
|
||
## Enable docker_container_storage_setup, it will configure devicemapper driver on Centos7 or RedHat7.
|
||
docker_container_storage_setup: false
|
||
|
||
## It must be define a disk path for docker_container_storage_setup_devs.
|
||
## Otherwise docker-storage-setup will be executed incorrectly.
|
||
# docker_container_storage_setup_devs: /dev/vdb
|
||
|
||
## Uncomment this if you have more than 3 nameservers, then we'll only use the first 3.
|
||
docker_dns_servers_strict: false
|
||
|
||
# Path used to store Docker data
|
||
docker_daemon_graph: "/var/lib/docker"
|
||
|
||
## Used to set docker daemon iptables options to true
|
||
docker_iptables_enabled: "false"
|
||
|
||
# Docker log options
|
||
# Rotate container stderr/stdout logs at 50m and keep last 5
|
||
docker_log_opts: "--log-opt max-size=50m --log-opt max-file=5"
|
||
|
||
## An obvious use case is allowing insecure-registry access to self hosted registries.
|
||
## Can be ipaddress and domain_name.
|
||
## example define 172.19.16.11 or mirror.registry.io
|
||
# docker_insecure_registries:
|
||
# - mirror.registry.io
|
||
# - 172.19.16.11
|
||
|
||
## Add other registry,example China registry mirror.
|
||
# docker_registry_mirrors:
|
||
# - https://registry.docker-cn.com
|
||
# - https://mirror.aliyuncs.com
|
||
|
||
## If non-empty will override default system MounFlags value.
|
||
## This option takes a mount propagation flag: shared, slave
|
||
## or private, which control whether mounts in the file system
|
||
## namespace set up for docker will receive or propagate mounts
|
||
## and unmounts. Leave empty for system default
|
||
# docker_mount_flags:
|
||
|
||
## A string of extra options to pass to the docker daemon.
|
||
## This string should be exactly as you wish it to appear.
|
||
docker_options: >-
|
||
{%- if docker_insecure_registries is defined %}
|
||
{{ docker_insecure_registries | map('regex_replace', '^(.*)$', '--insecure-registry=\1' ) | list | join(' ') }}
|
||
{%- endif %}
|
||
{% if docker_registry_mirrors is defined %}
|
||
{{ docker_registry_mirrors | map('regex_replace', '^(.*)$', '--registry-mirror=\1' ) | list | join(' ') }}
|
||
{%- endif %}
|
||
{%- if docker_version != "latest" and docker_version is version('17.05', '<') %}
|
||
--graph={{ docker_daemon_graph }} {% if ansible_os_family not in ["openSUSE Leap", "openSUSE Tumbleweed", "Suse"] %}{{ docker_log_opts }}{% endif %}
|
||
{%- else %}
|
||
--data-root={{ docker_daemon_graph }} {% if ansible_os_family not in ["openSUSE Leap", "openSUSE Tumbleweed", "Suse"] %}{{ docker_log_opts }}{% endif %}
|
||
{%- endif %}
|
||
{%- if ansible_architecture == "aarch64" and ansible_os_family == "RedHat" %}
|
||
--add-runtime docker-runc=/usr/libexec/docker/docker-runc-current
|
||
--default-runtime=docker-runc --exec-opt native.cgroupdriver=systemd
|
||
--userland-proxy-path=/usr/libexec/docker/docker-proxy-current --signature-verification=false
|
||
{%- endif -%}
|
||
|
||
# Settings for containerized control plane (etcd/kubelet/secrets)
|
||
etcd_deployment_type: docker
|
||
cert_management: script
|
||
|
||
helm_deployment_type: host
|
||
|
||
# Make a copy of kubeconfig on the host that runs Ansible in {{ inventory_dir }}/artifacts
|
||
kubeconfig_localhost: false
|
||
# Download kubectl onto the host that runs Ansible in {{ bin_dir }}
|
||
kubectl_localhost: false
|
||
|
||
# Define credentials_dir here so it can be overridden
|
||
credentials_dir: "{{ inventory_dir }}/credentials"
|
||
|
||
# K8s image pull policy (imagePullPolicy)
|
||
k8s_image_pull_policy: IfNotPresent
|
||
|
||
# Kubernetes dashboard
|
||
# RBAC required. see docs/getting-started.md for access details.
|
||
dashboard_enabled: true
|
||
|
||
# Addons which can be enabled
|
||
helm_enabled: false
|
||
registry_enabled: false
|
||
metrics_server_enabled: false
|
||
enable_network_policy: true
|
||
local_volume_provisioner_enabled: "{{ local_volumes_enabled | default('false') }}"
|
||
local_volume_provisioner_directory_mode: 0700
|
||
persistent_volumes_enabled: false
|
||
cephfs_provisioner_enabled: false
|
||
rbd_provisioner_enabled: false
|
||
ingress_nginx_enabled: false
|
||
cert_manager_enabled: false
|
||
expand_persistent_volumes: false
|
||
|
||
## When OpenStack is used, Cinder version can be explicitly specified if autodetection fails (Fixed in 1.9: https://github.com/kubernetes/kubernetes/issues/50461)
|
||
# openstack_blockstorage_version: "v1/v2/auto (default)"
|
||
openstack_blockstorage_ignore_volume_az: "{{ volume_cross_zone_attachment | default('false') }}"
|
||
## When OpenStack is used, if LBaaSv2 is available you can enable it with the following 2 variables.
|
||
openstack_lbaas_enabled: false
|
||
# openstack_lbaas_subnet_id: "Neutron subnet ID (not network ID) to create LBaaS VIP"
|
||
## To enable automatic floating ip provisioning, specify a subnet.
|
||
# openstack_lbaas_floating_network_id: "Neutron network ID (not subnet ID) to get floating IP from, disabled by default"
|
||
## Override default LBaaS behavior
|
||
# openstack_lbaas_use_octavia: False
|
||
# openstack_lbaas_method: "ROUND_ROBIN"
|
||
# openstack_lbaas_provider: "haproxy"
|
||
openstack_lbaas_create_monitor: "yes"
|
||
openstack_lbaas_monitor_delay: "1m"
|
||
openstack_lbaas_monitor_timeout: "30s"
|
||
openstack_lbaas_monitor_max_retries: "3"
|
||
openstack_cacert: "{{ lookup('env','OS_CACERT') }}"
|
||
|
||
## List of authorization modes that must be configured for
|
||
## the k8s cluster. Only 'AlwaysAllow', 'AlwaysDeny', 'Node' and
|
||
## 'RBAC' modes are tested. Order is important.
|
||
authorization_modes: ['Node', 'RBAC']
|
||
rbac_enabled: "{{ 'RBAC' in authorization_modes }}"
|
||
|
||
# When enabled, API bearer tokens (including service account tokens) can be used to authenticate to the kubelet’s HTTPS endpoint
|
||
kubelet_authentication_token_webhook: true
|
||
|
||
# When enabled, access to the kubelet API requires authorization by delegation to the API server
|
||
kubelet_authorization_mode_webhook: true
|
||
|
||
# kubelet uses certificates for authenticating to the Kubernetes API
|
||
# Automatically generate a new key and request a new certificate from the Kubernetes API as the current certificate approaches expiration
|
||
kubelet_rotate_certificates: true
|
||
|
||
## v1.11 feature
|
||
feature_gate_v1_11:
|
||
- "PersistentLocalVolumes={{ local_volume_provisioner_enabled | string }}"
|
||
- "VolumeScheduling={{ local_volume_provisioner_enabled | string }}"
|
||
- "MountPropagation={{ local_volume_provisioner_enabled | string }}"
|
||
|
||
## v1.12 feature
|
||
feature_gate_v1_12: []
|
||
|
||
## List of key=value pairs that describe feature gates for
|
||
## the k8s cluster.
|
||
kube_feature_gates: |-
|
||
{%- if kube_version is version('v1.12.0', '<') -%}
|
||
{{ feature_gate_v1_11 }}
|
||
{%- else -%}
|
||
{{ feature_gate_v1_12 }}
|
||
{%- endif %}
|
||
|
||
# Enable kubeadm experimental control plane
|
||
kubeadm_control_plane: false
|
||
|
||
# Local volume provisioner storage classes
|
||
# Levarages Ansibles string to Python datatype casting. Otherwise the dict_key isn't substituted
|
||
# see https://github.com/ansible/ansible/issues/17324
|
||
local_volume_provisioner_storage_classes: |
|
||
{
|
||
"{{ local_volume_provisioner_storage_class | default('local-storage') }}": {
|
||
"host_dir": "{{ local_volume_provisioner_base_dir | default ('/mnt/disks') }}",
|
||
"mount_dir": "{{ local_volume_provisioner_mount_dir | default('/mnt/disks') }}"
|
||
}
|
||
}
|
||
|
||
# weave's network password for encryption
|
||
# if null then no network encryption
|
||
# you can use --extra-vars to pass the password in command line
|
||
weave_password: EnterPasswordHere
|
||
|
||
# Weave uses consensus mode by default
|
||
# Enabling seed mode allow to dynamically add or remove hosts
|
||
# https://www.weave.works/docs/net/latest/ipam/
|
||
weave_mode_seed: false
|
||
|
||
# This two variable are automatically changed by the weave's role in group_vars/k8s-cluster.yml.
|
||
# Do not manually change these values
|
||
weave_seed: uninitialized
|
||
weave_peers: uninitialized
|
||
|
||
# Contiv L3 BGP Mode
|
||
contiv_peer_with_uplink_leaf: false
|
||
contiv_global_as: "65002"
|
||
contiv_global_neighbor_as: "500"
|
||
|
||
# Set 127.0.0.1 as fallback IP if we do not have host facts for host
|
||
fallback_ips_base: |
|
||
---
|
||
{% for item in groups['k8s-cluster'] + groups['etcd'] + groups['calico-rr']|default([])|unique %}
|
||
{{ item }}: "{{ hostvars[item].get('ansible_default_ipv4', {'address': '127.0.0.1'})['address'] }}"
|
||
{% endfor %}
|
||
fallback_ips: "{{ fallback_ips_base | from_yaml }}"
|
||
|
||
## Set no_proxy to all assigned cluster IPs and hostnames
|
||
no_proxy: >-
|
||
{%- if http_proxy is defined or https_proxy is defined %}
|
||
{%- if loadbalancer_apiserver is defined -%}
|
||
{{ apiserver_loadbalancer_domain_name| default('') }},
|
||
{{ loadbalancer_apiserver.address | default('') }},
|
||
{%- endif -%}
|
||
{%- for item in (groups['k8s-cluster'] + groups['etcd'] + groups['calico-rr']|default([]))|unique -%}
|
||
{{ hostvars[item]['access_ip'] | default(hostvars[item]['ip'] | default(fallback_ips[item])) }},
|
||
{%- if item != hostvars[item].get('ansible_hostname', '') -%}
|
||
{{ hostvars[item]['ansible_hostname'] }},
|
||
{{ hostvars[item]['ansible_hostname'] }}.{{ dns_domain }},
|
||
{%- endif -%}
|
||
{{ item }},{{ item }}.{{ dns_domain }},
|
||
{%- endfor -%}
|
||
{%- if additional_no_proxy is defined -%}
|
||
{{ additional_no_proxy }},
|
||
{%- endif -%}
|
||
127.0.0.1,localhost
|
||
{%- endif %}
|
||
|
||
proxy_env:
|
||
http_proxy: "{{ http_proxy| default ('') }}"
|
||
https_proxy: "{{ https_proxy| default ('') }}"
|
||
no_proxy: "{{ no_proxy| default ('') }}"
|
||
|
||
ssl_ca_dirs: >-
|
||
[
|
||
{% if ansible_os_family in ['CoreOS', 'Container Linux by CoreOS'] -%}
|
||
'/usr/share/ca-certificates',
|
||
{% elif ansible_os_family == 'RedHat' -%}
|
||
'/etc/pki/tls',
|
||
'/etc/pki/ca-trust',
|
||
{% elif ansible_os_family == 'Debian' -%}
|
||
'/usr/share/ca-certificates',
|
||
{% endif -%}
|
||
]
|
||
|
||
# Vars for pointing to kubernetes api endpoints
|
||
is_kube_master: "{{ inventory_hostname in groups['kube-master'] }}"
|
||
kube_apiserver_count: "{{ groups['kube-master'] | length }}"
|
||
kube_apiserver_address: "{{ ip | default(fallback_ips[inventory_hostname]) }}"
|
||
kube_apiserver_access_address: "{{ access_ip | default(kube_apiserver_address) }}"
|
||
first_kube_master: "{{ hostvars[groups['kube-master'][0]]['access_ip'] | default(hostvars[groups['kube-master'][0]]['ip'] | default(fallback_ips[groups['kube-master'][0]])) }}"
|
||
loadbalancer_apiserver_localhost: "{{ loadbalancer_apiserver is not defined }}"
|
||
loadbalancer_apiserver_type: "nginx"
|
||
# applied if only external loadbalancer_apiserver is defined, otherwise ignored
|
||
apiserver_loadbalancer_domain_name: "lb-apiserver.kubernetes.local"
|
||
kube_apiserver_endpoint: |-
|
||
{% if loadbalancer_apiserver is defined and loadbalancer_apiserver.port is defined -%}
|
||
https://{{ apiserver_loadbalancer_domain_name|default('lb-apiserver.kubernetes.local') }}:{{ loadbalancer_apiserver.port|default(kube_apiserver_port) }}
|
||
{%- elif not is_kube_master and loadbalancer_apiserver_localhost -%}
|
||
https://localhost:{{ loadbalancer_apiserver_port|default(kube_apiserver_port) }}
|
||
{%- elif is_kube_master -%}
|
||
https://{{ kube_apiserver_bind_address | regex_replace('0\.0\.0\.0','127.0.0.1') }}:{{ kube_apiserver_port }}
|
||
{%- else -%}
|
||
https://{{ first_kube_master }}:{{ kube_apiserver_port }}
|
||
{%- endif %}
|
||
kube_apiserver_insecure_endpoint: >-
|
||
http://{{ kube_apiserver_insecure_bind_address | regex_replace('0\.0\.0\.0','127.0.0.1') }}:{{ kube_apiserver_insecure_port }}
|
||
kube_apiserver_client_cert: "{{ kube_cert_dir }}/ca.crt"
|
||
kube_apiserver_client_key: "{{ kube_cert_dir }}/ca.key"
|
||
|
||
# Set to true to deploy etcd-events cluster
|
||
etcd_events_cluster_enabled: false
|
||
|
||
# Vars for pointing to etcd endpoints
|
||
is_etcd_master: "{{ inventory_hostname in groups['etcd'] }}"
|
||
etcd_address: "{{ ip | default(fallback_ips[inventory_hostname]) }}"
|
||
etcd_access_address: "{{ access_ip | default(etcd_address) }}"
|
||
etcd_events_access_address: "{{ access_ip | default(etcd_address) }}"
|
||
etcd_peer_url: "https://{{ etcd_access_address }}:2380"
|
||
etcd_client_url: "https://{{ etcd_access_address }}:2379"
|
||
etcd_events_peer_url: "https://{{ etcd_events_access_address }}:2382"
|
||
etcd_events_client_url: "https://{{ etcd_events_access_address }}:2381"
|
||
etcd_access_addresses: |-
|
||
{% for item in groups['etcd'] -%}
|
||
https://{{ hostvars[item]['access_ip'] | default(hostvars[item]['ip'] | default(fallback_ips[item])) }}:2379{% if not loop.last %},{% endif %}
|
||
{%- endfor %}
|
||
etcd_events_access_addresses_list: |-
|
||
[
|
||
{% for item in groups['etcd'] -%}
|
||
'https://{{ hostvars[item]['access_ip'] | default(hostvars[item]['ip'] | default(fallback_ips[item])) }}:2381'{% if not loop.last %},{% endif %}
|
||
{%- endfor %}
|
||
]
|
||
etcd_events_access_addresses: "{{etcd_events_access_addresses_list | join(',')}}"
|
||
etcd_events_access_addresses_semicolon: "{{etcd_events_access_addresses_list | join(';')}}"
|
||
# user should set etcd_member_name in inventory/mycluster/hosts.ini
|
||
etcd_member_name: |-
|
||
{% for host in groups['etcd'] %}
|
||
{% if inventory_hostname == host %}{{ hostvars[host].etcd_member_name | default("etcd" + loop.index|string) }}{% endif %}
|
||
{% endfor %}
|
||
etcd_peer_addresses: |-
|
||
{% for item in groups['etcd'] -%}
|
||
{{ hostvars[item].etcd_member_name | default("etcd" + loop.index|string) }}=https://{{ hostvars[item].access_ip | default(hostvars[item].ip | default(fallback_ips[item])) }}:2380{% if not loop.last %},{% endif %}
|
||
{%- endfor %}
|
||
etcd_events_peer_addresses: |-
|
||
{% for item in groups['etcd'] -%}
|
||
{{ hostvars[item].etcd_member_name | default("etcd" + loop.index|string) }}-events=https://{{ hostvars[item].access_ip | default(hostvars[item].ip | default(fallback_ips[item])) }}:2382{% if not loop.last %},{% endif %}
|
||
{%- endfor %}
|
||
|
||
podsecuritypolicy_enabled: false
|
||
etcd_heartbeat_interval: "250"
|
||
etcd_election_timeout: "5000"
|
||
etcd_snapshot_count: "10000"
|
||
|
||
certificates_key_size: 2048
|
||
certificates_duration: 36500
|
||
|
||
pip_extra_args: |-
|
||
{%- set pip_extra_args_list = [] -%}
|
||
{%- if pyrepo_index is defined -%}
|
||
{%- set DO = pip_extra_args_list.append('--index-url %s' | format(pyrepo_index)) -%}
|
||
{%- if pyrepo_cert is defined -%}
|
||
{%- set DO = pip_extra_args_list.append('--cert %s' | format(pyrepo_cert)) -%}
|
||
{%- endif -%}
|
||
{%- endif -%}
|
||
{{ pip_extra_args_list|join(' ') }}
|
||
|
||
etcd_config_dir: /etc/ssl/etcd
|
||
etcd_cert_dir: "{{ etcd_config_dir }}/ssl"
|
||
|
||
typha_enabled: false
|