Add native Vagrant support
This allows you to simply run `vagrant up` to get a 3 node HA cluster. * Creates a dynamic inventory and uses the inventory/group_vars/all.yml * commented lines in inventory.example so that ansible doesn't try to use it. * added requirements.txt to give easy way to install ansible/ipaddr * added gitignore files to stop attempts to save unwated files * changed `Check if kube-system exists` to `failed_when: false` instead of `ignore_errors`
This commit is contained in:
parent
3914d51a7e
commit
8f4e879ca7
7 changed files with 165 additions and 30 deletions
3
.gitignore
vendored
Normal file
3
.gitignore
vendored
Normal file
|
@ -0,0 +1,3 @@
|
||||||
|
.vagrant
|
||||||
|
*.retry
|
||||||
|
inventory/vagrant_ansible_inventory
|
|
@ -9,6 +9,8 @@
|
||||||
- Support most popular **Linux distributions**
|
- Support most popular **Linux distributions**
|
||||||
- **Continuous integration tests**
|
- **Continuous integration tests**
|
||||||
|
|
||||||
|
To create a cluster in vagrant simply run `vagrant up`
|
||||||
|
|
||||||
For an easy way to use it, check out [**kargo-cli**](https://github.com/kubespray/kargo-cli) </br>
|
For an easy way to use it, check out [**kargo-cli**](https://github.com/kubespray/kargo-cli) </br>
|
||||||
A complete **documentation** can be found [**here**](https://docs.kubespray.io)
|
A complete **documentation** can be found [**here**](https://docs.kubespray.io)
|
||||||
|
|
||||||
|
|
124
Vagrantfile
vendored
Normal file
124
Vagrantfile
vendored
Normal file
|
@ -0,0 +1,124 @@
|
||||||
|
# -*- mode: ruby -*-
|
||||||
|
# # vi: set ft=ruby :
|
||||||
|
|
||||||
|
require 'fileutils'
|
||||||
|
|
||||||
|
Vagrant.require_version ">= 1.8.0"
|
||||||
|
|
||||||
|
CONFIG = File.join(File.dirname(__FILE__), "vagrant/config.rb")
|
||||||
|
|
||||||
|
# Defaults for config options defined in CONFIG
|
||||||
|
$num_instances = 3
|
||||||
|
$instance_name_prefix = "k8s"
|
||||||
|
$vm_gui = false
|
||||||
|
$vm_memory = 1024
|
||||||
|
$vm_cpus = 1
|
||||||
|
$shared_folders = {}
|
||||||
|
$forwarded_ports = {}
|
||||||
|
$subnet = "172.17.8"
|
||||||
|
|
||||||
|
host_vars = {}
|
||||||
|
|
||||||
|
if File.exist?(CONFIG)
|
||||||
|
require CONFIG
|
||||||
|
end
|
||||||
|
|
||||||
|
# if $inventory is not set, try to use example
|
||||||
|
$inventory = File.join(File.dirname(__FILE__), "inventory") if ! $inventory
|
||||||
|
|
||||||
|
# if $inventory has a hosts file use it, otherwise copy over vars etc
|
||||||
|
# to where vagrant expects dynamic inventory to be.
|
||||||
|
if ! File.exist?(File.join(File.dirname($inventory), "hosts"))
|
||||||
|
$vagrant_ansible = File.join(File.dirname(__FILE__), ".vagrant",
|
||||||
|
"provisioners", "ansible")
|
||||||
|
FileUtils.mkdir_p($vagrant_ansible) if ! File.exist?($vagrant_ansible)
|
||||||
|
if ! File.exist?(File.join($vagrant_ansible,"inventory"))
|
||||||
|
FileUtils.ln_s($inventory, $vagrant_ansible)
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
Vagrant.configure("2") do |config|
|
||||||
|
# always use Vagrants insecure key
|
||||||
|
config.ssh.insert_key = false
|
||||||
|
|
||||||
|
config.vm.box = "ubuntu-14.04"
|
||||||
|
config.vm.box_url = "https://storage.googleapis.com/%s.release.core-os.net/amd64-usr/%s/coreos_production_vagrant.json" % [$update_channel, $image_version]
|
||||||
|
|
||||||
|
["vmware_fusion", "vmware_workstation"].each do |vmware|
|
||||||
|
config.vm.provider vmware do |v, override|
|
||||||
|
override.vm.box_url = "https://storage.googleapis.com/%s.release.core-os.net/amd64-usr/%s/coreos_production_vagrant_vmware_fusion.json" % [$update_channel, $image_version]
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
config.vm.provider :virtualbox do |v|
|
||||||
|
# On VirtualBox, we don't have guest additions or a functional vboxsf
|
||||||
|
# in CoreOS, so tell Vagrant that so it can be smarter.
|
||||||
|
v.check_guest_additions = false
|
||||||
|
v.functional_vboxsf = false
|
||||||
|
end
|
||||||
|
|
||||||
|
# plugin conflict
|
||||||
|
if Vagrant.has_plugin?("vagrant-vbguest") then
|
||||||
|
config.vbguest.auto_update = false
|
||||||
|
end
|
||||||
|
|
||||||
|
(1..$num_instances).each do |i|
|
||||||
|
config.vm.define vm_name = "%s-%02d" % [$instance_name_prefix, i] do |config|
|
||||||
|
config.vm.hostname = vm_name
|
||||||
|
|
||||||
|
if $expose_docker_tcp
|
||||||
|
config.vm.network "forwarded_port", guest: 2375, host: ($expose_docker_tcp + i - 1), auto_correct: true
|
||||||
|
end
|
||||||
|
|
||||||
|
$forwarded_ports.each do |guest, host|
|
||||||
|
config.vm.network "forwarded_port", guest: guest, host: host, auto_correct: true
|
||||||
|
end
|
||||||
|
|
||||||
|
["vmware_fusion", "vmware_workstation"].each do |vmware|
|
||||||
|
config.vm.provider vmware do |v|
|
||||||
|
v.vmx['memsize'] = $vm_memory
|
||||||
|
v.vmx['numvcpus'] = $vm_cpus
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
config.vm.provider :virtualbox do |vb|
|
||||||
|
vb.gui = $vm_gui
|
||||||
|
vb.memory = $vm_memory
|
||||||
|
vb.cpus = $vm_cpus
|
||||||
|
end
|
||||||
|
|
||||||
|
ip = "#{$subnet}.#{i+100}"
|
||||||
|
host_vars[vm_name] = {
|
||||||
|
"ip" => ip,
|
||||||
|
"access_ip" => ip
|
||||||
|
}
|
||||||
|
config.vm.network :private_network, ip: ip
|
||||||
|
|
||||||
|
# Only execute once the Ansible provisioner,
|
||||||
|
# when all the machines are up and ready.
|
||||||
|
if i == $num_instances
|
||||||
|
config.vm.provision "ansible" do |ansible|
|
||||||
|
ansible.playbook = "cluster.yml"
|
||||||
|
if File.exist?(File.join(File.dirname($inventory), "hosts"))
|
||||||
|
ansible.inventory_path = $inventory
|
||||||
|
end
|
||||||
|
ansible.sudo = true
|
||||||
|
ansible.limit = "all"
|
||||||
|
ansible.host_key_checking = false
|
||||||
|
ansible.raw_arguments = ["--forks=#{$num_instances}"]
|
||||||
|
ansible.host_vars = host_vars
|
||||||
|
ansible.groups = {
|
||||||
|
# The first three nodes should be etcd servers
|
||||||
|
"etcd" => ["k8s-0[1:3]"],
|
||||||
|
# The first two nodes should be masters
|
||||||
|
"kube-master" => ["k8s-0[1:2]"],
|
||||||
|
# all nodes should be kube nodes
|
||||||
|
"kube-node" => ["k8s-0[1:#{$num_instances}]"],
|
||||||
|
"k8s-cluster:children" => ["kube-master", "kube-node"],
|
||||||
|
}
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
end
|
||||||
|
end
|
||||||
|
end
|
|
@ -1,29 +1,29 @@
|
||||||
[kube-master]
|
#[kube-master]
|
||||||
node1 ansible_ssh_host=10.99.0.26
|
#node1 ansible_ssh_host=10.99.0.26
|
||||||
node2 ansible_ssh_host=10.99.0.27
|
#node2 ansible_ssh_host=10.99.0.27
|
||||||
|
#
|
||||||
[etcd]
|
#[etcd]
|
||||||
node1 ansible_ssh_host=10.99.0.26
|
#node1 ansible_ssh_host=10.99.0.26
|
||||||
node2 ansible_ssh_host=10.99.0.27
|
#node2 ansible_ssh_host=10.99.0.27
|
||||||
node3 ansible_ssh_host=10.99.0.4
|
#node3 ansible_ssh_host=10.99.0.4
|
||||||
|
#
|
||||||
[kube-node]
|
#[kube-node]
|
||||||
node2 ansible_ssh_host=10.99.0.27
|
#node2 ansible_ssh_host=10.99.0.27
|
||||||
node3 ansible_ssh_host=10.99.0.4
|
#node3 ansible_ssh_host=10.99.0.4
|
||||||
node4 ansible_ssh_host=10.99.0.5
|
#node4 ansible_ssh_host=10.99.0.5
|
||||||
node5 ansible_ssh_host=10.99.0.36
|
#node5 ansible_ssh_host=10.99.0.36
|
||||||
node6 ansible_ssh_host=10.99.0.37
|
#node6 ansible_ssh_host=10.99.0.37
|
||||||
|
#
|
||||||
[paris]
|
#[paris]
|
||||||
node1 ansible_ssh_host=10.99.0.26
|
#node1 ansible_ssh_host=10.99.0.26
|
||||||
node3 ansible_ssh_host=10.99.0.4 local_as=xxxxxxxx
|
#node3 ansible_ssh_host=10.99.0.4 local_as=xxxxxxxx
|
||||||
node4 ansible_ssh_host=10.99.0.5 local_as=xxxxxxxx
|
#node4 ansible_ssh_host=10.99.0.5 local_as=xxxxxxxx
|
||||||
|
#
|
||||||
[new-york]
|
#[new-york]
|
||||||
node2 ansible_ssh_host=10.99.0.27
|
#node2 ansible_ssh_host=10.99.0.27
|
||||||
node5 ansible_ssh_host=10.99.0.36 local_as=xxxxxxxx
|
#node5 ansible_ssh_host=10.99.0.36 local_as=xxxxxxxx
|
||||||
node6 ansible_ssh_host=10.99.0.37 local_as=xxxxxxxx
|
#node6 ansible_ssh_host=10.99.0.37 local_as=xxxxxxxx
|
||||||
|
#
|
||||||
[k8s-cluster:children]
|
#[k8s-cluster:children]
|
||||||
kube-node
|
#kube-node
|
||||||
kube-master
|
#kube-master
|
||||||
|
|
2
requirements.txt
Normal file
2
requirements.txt
Normal file
|
@ -0,0 +1,2 @@
|
||||||
|
ansible
|
||||||
|
netaddr
|
|
@ -58,7 +58,7 @@
|
||||||
command: "{{ bin_dir }}/kubectl get ns kube-system"
|
command: "{{ bin_dir }}/kubectl get ns kube-system"
|
||||||
register: 'kubesystem'
|
register: 'kubesystem'
|
||||||
changed_when: False
|
changed_when: False
|
||||||
ignore_errors: yes
|
failed_when: False
|
||||||
run_once: yes
|
run_once: yes
|
||||||
|
|
||||||
- name: wait for the apiserver to be running
|
- name: wait for the apiserver to be running
|
||||||
|
|
4
roles/kubernetes/secrets/files/.gitignore
vendored
Normal file
4
roles/kubernetes/secrets/files/.gitignore
vendored
Normal file
|
@ -0,0 +1,4 @@
|
||||||
|
tokens/*.token
|
||||||
|
tokens/known_tokens.csv
|
||||||
|
certs/*.pem
|
||||||
|
openssl.conf
|
Loading…
Reference in a new issue