kube-router.md
-
lijin-union authored
* remove a useless parenthesis in the _sidebar file
lijin-union authored* remove a useless parenthesis in the _sidebar file
Getting started
Building your own inventory
Ansible inventory can be stored in 3 formats: YAML, JSON, or INI-like. There is an example inventory located here.
You can use an
inventory generator
to create or modify an Ansible inventory. Currently, it is limited in
functionality and is only used for configuring a basic Kubespray cluster inventory, but it does
support creating inventory file for large clusters as well. It now supports
separated ETCD and Kubernetes master roles from node role if the size exceeds a
certain threshold. Run python3 contrib/inventory_builder/inventory.py help
help for more information.
Example inventory generator usage:
cp -r inventory/sample inventory/mycluster
declare -a IPS=(10.10.1.3 10.10.1.4 10.10.1.5)
CONFIG_FILE=inventory/mycluster/hosts.ini python3 contrib/inventory_builder/inventory.py ${IPS[@]}
Starting custom deployment
Once you have an inventory, you may want to customize deployment data vars and start the deployment:
IMPORTANT: Edit my_inventory/groups_vars/*.yaml to override data vars:
ansible-playbook -i inventory/mycluster/hosts.ini cluster.yml -b -v \
--private-key=~/.ssh/private_key
See more details in the ansible guide.
Adding nodes
You may want to add worker, master or etcd nodes to your existing cluster. This can be done by re-running the cluster.yml
playbook, or you can target the bare minimum needed to get kubelet installed on the worker and talking to your masters. This is especially helpful when doing something like autoscaling your clusters.
-
Add the new worker node to your inventory in the appropriate group (or utilize a dynamic inventory).
-
Run the ansible-playbook command, substituting
cluster.yml
forscale.yml
:ansible-playbook -i inventory/mycluster/hosts.ini scale.yml -b -v \ --private-key=~/.ssh/private_key
Remove nodes
You may want to remove worker nodes to your existing cluster. This can be done by re-running the remove-node.yml
playbook. First, all nodes will be drained, then stop some kubernetes services and delete some certificates, and finally execute the kubectl command to delete these nodes. This can be combined with the add node function, This is generally helpful when doing something like autoscaling your clusters. Of course if a node is not working, you can remove the node and install it again.
Add worker nodes to the list under kube-node if you want to delete them (or utilize a dynamic inventory).
ansible-playbook -i inventory/mycluster/hosts.ini remove-node.yml -b -v \
--private-key=~/.ssh/private_key
We support two ways to select the nodes:
- Use
--extra-vars "node=<nodename>,<nodename2>"
to select the node you want to delete.