fd83ec6526
* Enforce a etcd-proxy role to a k8s-cluster group members. This provides an HA layout for all of the k8s cluster internal clients. * Proxies to be run on each node in the group as a separate etcd instances with a readwrite proxy mode and listen the given endpoint, which is either the access_ip:2379 or the localhost:2379. * A notion for the 'kube_etcd_multiaccess' is: ignore endpoints and loadbalancers and use the etcd members IPs as a comma-separated list. Otherwise, clients shall use the local endpoint provided by a etcd-proxy instances on each etcd node. A Netwroking plugins always use that access mode. * Fix apiserver's etcd servers args to use the etcd_access_endpoint. * Fix networking plugins flannel/calico to use the etcd_endpoint. * Fix name env var for non masters to be set as well. * Fix etcd_client_url was not used anywhere and other etcd_* facts evaluation was duplicated in a few places. * Define proxy modes only in the env file, if not a master. Del an automatic proxy mode decisions for etcd nodes in init/unit scripts. * Use Wants= instead of Requires= as "This is the recommended way to hook start-up of one unit to the start-up of another unit" * Make apiserver/calico Wants= etcd-proxy to keep it always up Signed-off-by: Bogdan Dobrelya <bdobrelia@mirantis.com> Co-authored-by: Matthew Mosesohn <mmosesohn@mirantis.com>
161 lines
6.2 KiB
YAML
161 lines
6.2 KiB
YAML
# Directory where the binaries will be installed
|
|
bin_dir: /usr/local/bin
|
|
|
|
# Where the binaries will be downloaded.
|
|
# Note: ensure that you've enough disk space (about 1G)
|
|
local_release_dir: "/tmp/releases"
|
|
|
|
# Uncomment this line for CoreOS only.
|
|
# Directory where python binary is installed
|
|
# ansible_python_interpreter: "/opt/bin/python"
|
|
|
|
# This is the group that the cert creation scripts chgrp the
|
|
# cert files to. Not really changable...
|
|
kube_cert_group: kube-cert
|
|
|
|
# Cluster Loglevel configuration
|
|
kube_log_level: 2
|
|
|
|
# Users to create for basic auth in Kubernetes API via HTTP
|
|
kube_api_pwd: "changeme"
|
|
kube_users:
|
|
kube:
|
|
pass: "{{kube_api_pwd}}"
|
|
role: admin
|
|
root:
|
|
pass: "changeme"
|
|
role: admin
|
|
|
|
# Kubernetes cluster name, also will be used as DNS domain
|
|
cluster_name: cluster.local
|
|
|
|
# For some environments, each node has a pubilcally accessible
|
|
# address and an address it should bind services to. These are
|
|
# really inventory level variables, but described here for consistency.
|
|
#
|
|
# When advertising access, the access_ip will be used, but will defer to
|
|
# ip and then the default ansible ip when unspecified.
|
|
#
|
|
# When binding to restrict access, the ip variable will be used, but will
|
|
# defer to the default ansible ip when unspecified.
|
|
#
|
|
# The ip variable is used for specific address binding, e.g. listen address
|
|
# for etcd. This is use to help with environments like Vagrant or multi-nic
|
|
# systems where one address should be preferred over another.
|
|
# ip: 10.2.2.2
|
|
#
|
|
# The access_ip variable is used to define how other nodes should access
|
|
# the node. This is used in flannel to allow other flannel nodes to see
|
|
# this node for example. The access_ip is really useful AWS and Google
|
|
# environments where the nodes are accessed remotely by the "public" ip,
|
|
# but don't know about that address themselves.
|
|
# access_ip: 1.1.1.1
|
|
|
|
# Service endpoints. May be a VIP or a load balanced frontend IP, like one
|
|
# that a HAProxy or Nginx provides, or just a local service endpoint.
|
|
#
|
|
# Etcd endpoints use a local etcd-proxies to reach the etcd cluster via
|
|
# auto-evaluated endpoints. Those will reuse the access_ip for etcd cluster,
|
|
# if specified, or defer to the localhost:2379 as well.
|
|
|
|
# Etcd access modes:
|
|
# Enable multiaccess to configure clients to access all of the etcd members directly
|
|
# as the "http://hostX:port, http://hostY:port, ..." and ignore the proxy loadbalancers.
|
|
# This may be the case if clients support and loadbalance multiple etcd servers natively.
|
|
etcd_multiaccess: false
|
|
|
|
#
|
|
# TODO apiserver localhost:8080 and localhost:443 endpoints for kubelets and
|
|
# (hyper)kube-* and networking components.
|
|
|
|
# Choose network plugin (calico, weave or flannel)
|
|
kube_network_plugin: flannel
|
|
|
|
# Kubernetes internal network for services, unused block of space.
|
|
kube_service_addresses: 10.233.0.0/18
|
|
|
|
# internal network. When used, it will assign IP
|
|
# addresses from this range to individual pods.
|
|
# This network must be unused in your network infrastructure!
|
|
kube_pods_subnet: 10.233.64.0/18
|
|
|
|
# internal network total size (optional). This is the prefix of the
|
|
# entire network. Must be unused in your environment.
|
|
# kube_network_prefix: 18
|
|
|
|
# internal network node size allocation (optional). This is the size allocated
|
|
# to each node on your network. With these defaults you should have
|
|
# room for 4096 nodes with 254 pods per node.
|
|
kube_network_node_prefix: 24
|
|
|
|
# With calico it is possible to distributed routes with border routers of the datacenter.
|
|
peer_with_router: false
|
|
# Warning : enabling router peering will disable calico's default behavior ('node mesh').
|
|
# The subnets of each nodes will be distributed by the datacenter router
|
|
|
|
# The port the API Server will be listening on.
|
|
kube_apiserver_ip: "{{ kube_service_addresses|ipaddr('net')|ipaddr(1)|ipaddr('address') }}"
|
|
kube_apiserver_port: 443 # (https)
|
|
kube_apiserver_insecure_port: 8080 # (http)
|
|
|
|
# Internal DNS configuration.
|
|
# Kubernetes can create and mainatain its own DNS server to resolve service names
|
|
# into appropriate IP addresses. It's highly advisable to run such DNS server,
|
|
# as it greatly simplifies configuration of your applications - you can use
|
|
# service names instead of magic environment variables.
|
|
# You still must manually configure all your containers to use this DNS server,
|
|
# Kubernetes won't do this for you (yet).
|
|
|
|
# Upstream dns servers used by dnsmasq
|
|
upstream_dns_servers:
|
|
- 8.8.8.8
|
|
- 8.8.4.4
|
|
#
|
|
# # Use dns server : https://github.com/ansibl8s/k8s-skydns/blob/master/skydns-README.md
|
|
dns_setup: true
|
|
dns_domain: "{{ cluster_name }}"
|
|
#
|
|
# # Ip address of the kubernetes skydns service
|
|
skydns_server: "{{ kube_service_addresses|ipaddr('net')|ipaddr(3)|ipaddr('address') }}"
|
|
dns_server: "{{ kube_service_addresses|ipaddr('net')|ipaddr(2)|ipaddr('address') }}"
|
|
|
|
# There are some changes specific to the cloud providers
|
|
# for instance we need to encapsulate packets with some network plugins
|
|
# If set the possible values are either 'gce', 'aws' or 'openstack'
|
|
# When openstack is used make sure to source in the openstack credentials
|
|
# like you would do when using nova-client before starting the playbook.
|
|
# cloud_provider:
|
|
|
|
# For multi masters architecture:
|
|
# kube-proxy doesn't support multiple apiservers for the time being so you'll need to configure your own loadbalancer
|
|
# This domain name will be inserted into the /etc/hosts file of all servers
|
|
# configuration example with haproxy :
|
|
# listen kubernetes-apiserver-https
|
|
# bind 10.99.0.21:8383
|
|
# option ssl-hello-chk
|
|
# mode tcp
|
|
# timeout client 3h
|
|
# timeout server 3h
|
|
# server master1 10.99.0.26:443
|
|
# server master2 10.99.0.27:443
|
|
# balance roundrobin
|
|
# apiserver_loadbalancer_domain_name: "lb-apiserver.kubernetes.local"
|
|
|
|
## Set these proxy values in order to update docker daemon to use proxies
|
|
# http_proxy: ""
|
|
# https_proxy: ""
|
|
# no_proxy: ""
|
|
|
|
## A string of extra options to pass to the docker daemon.
|
|
## This string should be exactly as you wish it to appear.
|
|
## An obvious use case is allowing insecure-registry access
|
|
## to self hosted registries like so:
|
|
docker_options: "--insecure-registry={{ kube_service_addresses }}"
|
|
|
|
# default packages to install within the cluster
|
|
kpm_packages:
|
|
- name: kube-system/kubedns
|
|
namespace: kube-system
|
|
variables:
|
|
cluster_ip: "{{skydns_server}}"
|
|
# - name: kube-system/grafana
|