Skip to content
Snippets Groups Projects
Select Git revision
  • master default protected
  • v2.28.0
  • v2.27.0
  • v2.25.1
  • v2.24.3
  • v2.26.0
  • v2.24.2
  • v2.25.0
  • v2.24.1
  • v2.22.2
  • v2.23.3
  • v2.24.0
  • v2.23.2
  • v2.23.1
  • v2.23.0
  • v2.22.1
  • v2.22.0
  • v2.21.0
  • v2.20.0
  • v2.19.1
  • v2.18.2
21 results

Vagrantfile

Blame
    • Cristian Calin's avatar
      360aff4a
      Rename ansible groups to use _ instead of - (#7552) · 360aff4a
      Cristian Calin authored
      * rename ansible groups to use _ instead of -
      
      k8s-cluster -> k8s_cluster
      k8s-node -> k8s_node
      calico-rr -> calico_rr
      no-floating -> no_floating
      
      Note: kube-node,k8s-cluster groups in upgrade CI
            need clean-up after v2.16 is tagged
      
      * ensure old groups are mapped to the new ones
      Rename ansible groups to use _ instead of - (#7552)
      Cristian Calin authored
      * rename ansible groups to use _ instead of -
      
      k8s-cluster -> k8s_cluster
      k8s-node -> k8s_node
      calico-rr -> calico_rr
      no-floating -> no_floating
      
      Note: kube-node,k8s-cluster groups in upgrade CI
            need clean-up after v2.16 is tagged
      
      * ensure old groups are mapped to the new ones
    Vagrantfile 10.69 KiB
    # -*- mode: ruby -*-
    # # vi: set ft=ruby :
    
    # For help on using kubespray with vagrant, check out docs/vagrant.md
    
    require 'fileutils'
    
    Vagrant.require_version ">= 2.0.0"
    
    CONFIG = File.join(File.dirname(__FILE__), ENV['KUBESPRAY_VAGRANT_CONFIG'] || 'vagrant/config.rb')
    
    FLATCAR_URL_TEMPLATE = "https://%s.release.flatcar-linux.net/amd64-usr/current/flatcar_production_vagrant.json"
    
    # Uniq disk UUID for libvirt
    DISK_UUID = Time.now.utc.to_i
    
    SUPPORTED_OS = {
      "flatcar-stable"      => {box: "flatcar-stable",             user: "core", box_url: FLATCAR_URL_TEMPLATE % ["stable"]},
      "flatcar-beta"        => {box: "flatcar-beta",               user: "core", box_url: FLATCAR_URL_TEMPLATE % ["beta"]},
      "flatcar-alpha"       => {box: "flatcar-alpha",              user: "core", box_url: FLATCAR_URL_TEMPLATE % ["alpha"]},
      "flatcar-edge"        => {box: "flatcar-edge",               user: "core", box_url: FLATCAR_URL_TEMPLATE % ["edge"]},
      "ubuntu1604"          => {box: "generic/ubuntu1604",         user: "vagrant"},
      "ubuntu1804"          => {box: "generic/ubuntu1804",         user: "vagrant"},
      "ubuntu2004"          => {box: "generic/ubuntu2004",         user: "vagrant"},
      "centos"              => {box: "centos/7",                   user: "vagrant"},
      "centos-bento"        => {box: "bento/centos-7.6",           user: "vagrant"},
      "centos8"             => {box: "centos/8",                   user: "vagrant"},
      "centos8-bento"       => {box: "bento/centos-8",             user: "vagrant"},
      "fedora32"            => {box: "fedora/32-cloud-base",       user: "vagrant"},
      "fedora33"            => {box: "fedora/33-cloud-base",       user: "vagrant"},
      "opensuse"            => {box: "bento/opensuse-leap-15.2",   user: "vagrant"},
      "opensuse-tumbleweed" => {box: "opensuse/Tumbleweed.x86_64", user: "vagrant"},
      "oraclelinux"         => {box: "generic/oracle7",            user: "vagrant"},
      "oraclelinux8"        => {box: "generic/oracle8",            user: "vagrant"},
      "rhel7"               => {box: "generic/rhel7",              user: "vagrant"},
      "rhel8"               => {box: "generic/rhel8",              user: "vagrant"},
    }
    
    if File.exist?(CONFIG)
      require CONFIG
    end
    
    # Defaults for config options defined in CONFIG
    $num_instances ||= 3
    $instance_name_prefix ||= "k8s"
    $vm_gui ||= false
    $vm_memory ||= 2048
    $vm_cpus ||= 2
    $shared_folders ||= {}
    $forwarded_ports ||= {}
    $subnet ||= "172.18.8"
    $subnet_ipv6 ||= "fd3c:b398:0698:0756"
    $os ||= "ubuntu1804"
    $network_plugin ||= "flannel"
    # Setting multi_networking to true will install Multus: https://github.com/intel/multus-cni
    $multi_networking ||= false
    $download_run_once ||= "True"
    $download_force_cache ||= "True"
    # The first three nodes are etcd servers
    $etcd_instances ||= $num_instances
    # The first two nodes are kube masters
    $kube_master_instances ||= $num_instances == 1 ? $num_instances : ($num_instances - 1)
    # All nodes are kube nodes
    $kube_node_instances ||= $num_instances
    # The following only works when using the libvirt provider
    $kube_node_instances_with_disks ||= false
    $kube_node_instances_with_disks_size ||= "20G"
    $kube_node_instances_with_disks_number ||= 2
    $override_disk_size ||= false
    $disk_size ||= "20GB"
    $local_path_provisioner_enabled ||= false
    $local_path_provisioner_claim_root ||= "/opt/local-path-provisioner/"
    $libvirt_nested ||= false
    
    $playbook ||= "cluster.yml"
    
    host_vars = {}
    
    $box = SUPPORTED_OS[$os][:box]
    # if $inventory is not set, try to use example
    $inventory = "inventory/sample" if ! $inventory
    $inventory = File.absolute_path($inventory, File.dirname(__FILE__))
    
    # if $inventory has a hosts.ini file use it, otherwise copy over
    # vars etc to where vagrant expects dynamic inventory to be
    if ! File.exist?(File.join(File.dirname($inventory), "hosts.ini"))
      $vagrant_ansible = File.join(File.dirname(__FILE__), ".vagrant", "provisioners", "ansible")
      FileUtils.mkdir_p($vagrant_ansible) if ! File.exist?($vagrant_ansible)
      $vagrant_inventory = File.join($vagrant_ansible,"inventory")
      FileUtils.rm_f($vagrant_inventory)
      FileUtils.ln_s($inventory, $vagrant_inventory)
    end
    
    if Vagrant.has_plugin?("vagrant-proxyconf")
      $no_proxy = ENV['NO_PROXY'] || ENV['no_proxy'] || "127.0.0.1,localhost"
      (1..$num_instances).each do |i|
          $no_proxy += ",#{$subnet}.#{i+100}"
      end
    end
    
    Vagrant.configure("2") do |config|
    
      config.vm.box = $box
      if SUPPORTED_OS[$os].has_key? :box_url
        config.vm.box_url = SUPPORTED_OS[$os][:box_url]
      end
      config.ssh.username = SUPPORTED_OS[$os][:user]
    
      # plugin conflict
      if Vagrant.has_plugin?("vagrant-vbguest") then
        config.vbguest.auto_update = false
      end
    
      # always use Vagrants insecure key
      config.ssh.insert_key = false
    
      if ($override_disk_size)
        unless Vagrant.has_plugin?("vagrant-disksize")
          system "vagrant plugin install vagrant-disksize"
        end
        config.disksize.size = $disk_size
      end
    
      (1..$num_instances).each do |i|
        config.vm.define vm_name = "%s-%01d" % [$instance_name_prefix, i] do |node|
    
          node.vm.hostname = vm_name
    
          if Vagrant.has_plugin?("vagrant-proxyconf")
            node.proxy.http     = ENV['HTTP_PROXY'] || ENV['http_proxy'] || ""
            node.proxy.https    = ENV['HTTPS_PROXY'] || ENV['https_proxy'] ||  ""
            node.proxy.no_proxy = $no_proxy
          end
    
          ["vmware_fusion", "vmware_workstation"].each do |vmware|
            node.vm.provider vmware do |v|
              v.vmx['memsize'] = $vm_memory
              v.vmx['numvcpus'] = $vm_cpus
            end
          end
    
          node.vm.provider :virtualbox do |vb|
            vb.memory = $vm_memory
            vb.cpus = $vm_cpus
            vb.gui = $vm_gui
            vb.linked_clone = true
            vb.customize ["modifyvm", :id, "--vram", "8"] # ubuntu defaults to 256 MB which is a waste of precious RAM
            vb.customize ["modifyvm", :id, "--audio", "none"]
          end
    
          node.vm.provider :libvirt do |lv|
            lv.nested = $libvirt_nested
            lv.cpu_mode = "host-model"
            lv.memory = $vm_memory
            lv.cpus = $vm_cpus
            lv.default_prefix = 'kubespray'
            # Fix kernel panic on fedora 28
            if $os == "fedora"
              lv.cpu_mode = "host-passthrough"
            end
          end
    
          if $kube_node_instances_with_disks
            # Libvirt
            driverletters = ('a'..'z').to_a
            node.vm.provider :libvirt do |lv|
              # always make /dev/sd{a/b/c} so that CI can ensure that
              # virtualbox and libvirt will have the same devices to use for OSDs
              (1..$kube_node_instances_with_disks_number).each do |d|
                lv.storage :file, :device => "hd#{driverletters[d]}", :path => "disk-#{i}-#{d}-#{DISK_UUID}.disk", :size => $kube_node_instances_with_disks_size, :bus => "ide"
              end
            end
          end
    
          if $expose_docker_tcp
            node.vm.network "forwarded_port", guest: 2375, host: ($expose_docker_tcp + i - 1), auto_correct: true
          end
    
          $forwarded_ports.each do |guest, host|
            node.vm.network "forwarded_port", guest: guest, host: host, auto_correct: true
          end
    
          if ["rhel7","rhel8"].include? $os
            # Vagrant synced_folder rsync options cannot be used for RHEL boxes as Rsync package cannot
            # be installed until the host is registered with a valid Red Hat support subscription
            node.vm.synced_folder ".", "/vagrant", disabled: false
            $shared_folders.each do |src, dst|
              node.vm.synced_folder src, dst
            end
          else
            node.vm.synced_folder ".", "/vagrant", disabled: false, type: "rsync", rsync__args: ['--verbose', '--archive', '--delete', '-z'] , rsync__exclude: ['.git','venv']
            $shared_folders.each do |src, dst|
              node.vm.synced_folder src, dst, type: "rsync", rsync__args: ['--verbose', '--archive', '--delete', '-z']
            end
          end
    
          ip = "#{$subnet}.#{i+100}"
          node.vm.network :private_network, ip: ip,
            :libvirt__guest_ipv6 => 'yes',
            :libvirt__ipv6_address => "#{$subnet_ipv6}::#{i+100}",
            :libvirt__ipv6_prefix => "64",
            :libvirt__forward_mode => "none",
            :libvirt__dhcp_enabled => false
    
          # Disable swap for each vm
          node.vm.provision "shell", inline: "swapoff -a"
    
          # ubuntu1804 and ubuntu2004 have IPv6 explicitly disabled. This undoes that.
          if ["ubuntu1804", "ubuntu2004"].include? $os
            node.vm.provision "shell", inline: "rm -f /etc/modprobe.d/local.conf"
            node.vm.provision "shell", inline: "sed -i '/net.ipv6.conf.all.disable_ipv6/d' /etc/sysctl.d/99-sysctl.conf /etc/sysctl.conf"
          end
    
          # Disable firewalld on oraclelinux/redhat vms
          if ["oraclelinux","oraclelinux8","rhel7","rhel8"].include? $os
            node.vm.provision "shell", inline: "systemctl stop firewalld; systemctl disable firewalld"
          end
    
          host_vars[vm_name] = {
            "ip": ip,
            "flannel_interface": "eth1",
            "kube_network_plugin": $network_plugin,
            "kube_network_plugin_multus": $multi_networking,
            "download_run_once": $download_run_once,
            "download_localhost": "False",
            "download_cache_dir": ENV['HOME'] + "/kubespray_cache",
            # Make kubespray cache even when download_run_once is false
            "download_force_cache": $download_force_cache,
            # Keeping the cache on the nodes can improve provisioning speed while debugging kubespray
            "download_keep_remote_cache": "False",
            "docker_rpm_keepcache": "1",
            # These two settings will put kubectl and admin.config in $inventory/artifacts
            "kubeconfig_localhost": "True",
            "kubectl_localhost": "True",
            "local_path_provisioner_enabled": "#{$local_path_provisioner_enabled}",
            "local_path_provisioner_claim_root": "#{$local_path_provisioner_claim_root}",
            "ansible_ssh_user": SUPPORTED_OS[$os][:user]
          }
    
          # Only execute the Ansible provisioner once, when all the machines are up and ready.
          if i == $num_instances
            node.vm.provision "ansible" do |ansible|
              ansible.playbook = $playbook
              $ansible_inventory_path = File.join( $inventory, "hosts.ini")
              if File.exist?($ansible_inventory_path)
                ansible.inventory_path = $ansible_inventory_path
              end
              ansible.become = true
              ansible.limit = "all,localhost"
              ansible.host_key_checking = false
              ansible.raw_arguments = ["--forks=#{$num_instances}", "--flush-cache", "-e ansible_become_pass=vagrant"]
              ansible.host_vars = host_vars
              #ansible.tags = ['download']
              ansible.groups = {
                "etcd" => ["#{$instance_name_prefix}-[1:#{$etcd_instances}]"],
                "kube_control_plane" => ["#{$instance_name_prefix}-[1:#{$kube_master_instances}]"],
                "kube_node" => ["#{$instance_name_prefix}-[1:#{$kube_node_instances}]"],
                "k8s_cluster:children" => ["kube_control_plane", "kube_node"],
              }
            end
          end
    
        end
      end
    end