diff --git a/.yamllint b/.yamllint
index eb061917ee407ef18bfd0f0bddb36f99d9d3efd9..56786e0a9887e54358d80ac33b6747cfce9e8190 100644
--- a/.yamllint
+++ b/.yamllint
@@ -26,4 +26,3 @@ rules:
   octal-values:
     forbid-implicit-octal: true # yamllint defaults to false
     forbid-explicit-octal: true # yamllint defaults to false
-  truthy: disable
diff --git a/contrib/azurerm/generate-inventory.yml b/contrib/azurerm/generate-inventory.yml
index 01ee386626ee4cbeab0ffb6ea16bd61b3cb33029..59e1e90b6a4ec449777aa262688f956a3b0dd35f 100644
--- a/contrib/azurerm/generate-inventory.yml
+++ b/contrib/azurerm/generate-inventory.yml
@@ -1,6 +1,6 @@
 ---
 - name: Generate Azure inventory
   hosts: localhost
-  gather_facts: False
+  gather_facts: false
   roles:
     - generate-inventory
diff --git a/contrib/azurerm/generate-inventory_2.yml b/contrib/azurerm/generate-inventory_2.yml
index 9173e1d8204f5bc5ace05fab6cffe815cb614f3a..8c2cbff86b58e4c0c40e0512476df09c1d2a4dbe 100644
--- a/contrib/azurerm/generate-inventory_2.yml
+++ b/contrib/azurerm/generate-inventory_2.yml
@@ -1,6 +1,6 @@
 ---
 - name: Generate Azure inventory
   hosts: localhost
-  gather_facts: False
+  gather_facts: false
   roles:
     - generate-inventory_2
diff --git a/contrib/azurerm/generate-templates.yml b/contrib/azurerm/generate-templates.yml
index f1fcb626f8b27838187d090a69992416ea1c24b5..f2cf231bc4d2d801acf83c25feb11bc0b4033ea3 100644
--- a/contrib/azurerm/generate-templates.yml
+++ b/contrib/azurerm/generate-templates.yml
@@ -1,6 +1,6 @@
 ---
 - name: Generate Azure templates
   hosts: localhost
-  gather_facts: False
+  gather_facts: false
   roles:
     - generate-templates
diff --git a/contrib/dind/dind-cluster.yaml b/contrib/dind/dind-cluster.yaml
index 258837d083c53d60a7b09560cb198b6363c2fbfd..0c61c3f2d4810e4d2a0e53d3966bb69a467aa52c 100644
--- a/contrib/dind/dind-cluster.yaml
+++ b/contrib/dind/dind-cluster.yaml
@@ -1,7 +1,7 @@
 ---
 - name: Create nodes as docker containers
   hosts: localhost
-  gather_facts: False
+  gather_facts: false
   roles:
     - { role: dind-host }
 
diff --git a/contrib/dind/kubespray-dind.yaml b/contrib/dind/kubespray-dind.yaml
index ecfb5573a7032bf010429d65463c59f8e35bb738..6d57cf00d1874ea7a0833a363d4207ed035a6ba5 100644
--- a/contrib/dind/kubespray-dind.yaml
+++ b/contrib/dind/kubespray-dind.yaml
@@ -15,7 +15,7 @@ docker_storage_options: -s overlay2 --storage-opt overlay2.override_kernel_check
 
 dns_mode: coredns
 
-deploy_netchecker: True
+deploy_netchecker: true
 netcheck_agent_image_repo: quay.io/l23network/k8s-netchecker-agent
 netcheck_server_image_repo: quay.io/l23network/k8s-netchecker-server
 netcheck_agent_image_tag: v1.0
diff --git a/contrib/dind/roles/dind-cluster/tasks/main.yaml b/contrib/dind/roles/dind-cluster/tasks/main.yaml
index dcb086c644782b3cd53cc92ad37432bb1c995c1a..1a3630f9ff65a4233feb22960614e7bccd3c0610 100644
--- a/contrib/dind/roles/dind-cluster/tasks/main.yaml
+++ b/contrib/dind/roles/dind-cluster/tasks/main.yaml
@@ -14,7 +14,7 @@
     src: "/bin/true"
     dest: "{{ item }}"
     state: link
-    force: yes
+    force: true
   with_items:
     # DIND box may have swap enable, don't bother
     - /sbin/swapoff
@@ -58,7 +58,7 @@
     name: "{{ distro_user }}"
     uid: 1000
     # groups: sudo
-    append: yes
+    append: true
 
 - name: Allow password-less sudo to "{{ distro_user }}"
   copy:
diff --git a/contrib/dind/roles/dind-host/tasks/main.yaml b/contrib/dind/roles/dind-host/tasks/main.yaml
index 56c8ff4c513c1482e99bbd59857b68d05c4ca74b..e0dc71008d5909253d805e2196e3061c9f10d945 100644
--- a/contrib/dind/roles/dind-host/tasks/main.yaml
+++ b/contrib/dind/roles/dind-host/tasks/main.yaml
@@ -19,7 +19,7 @@
     state: started
     hostname: "{{ item }}"
     command: "{{ distro_init }}"
-    # recreate: yes
+    # recreate: true
     privileged: true
     tmpfs:
       - /sys/module/nf_conntrack/parameters
diff --git a/contrib/kvm-setup/kvm-setup.yml b/contrib/kvm-setup/kvm-setup.yml
index b8d44058767d75533252c3aa578ab47682624d47..73b81978969eeb156921b72f38741f927cd9b310 100644
--- a/contrib/kvm-setup/kvm-setup.yml
+++ b/contrib/kvm-setup/kvm-setup.yml
@@ -1,8 +1,8 @@
 ---
 - name: Prepare Hypervisor to later install kubespray VMs
   hosts: localhost
-  gather_facts: False
-  become: yes
+  gather_facts: false
+  become: true
   vars:
     bootstrap_os: none
   roles:
diff --git a/contrib/kvm-setup/roles/kvm-setup/tasks/main.yml b/contrib/kvm-setup/roles/kvm-setup/tasks/main.yml
index 3e8ade645bbfee9e4bbe673dfa68f31230015e05..dfcd3be7367c97a593cdc5e696f45c1b494f430c 100644
--- a/contrib/kvm-setup/roles/kvm-setup/tasks/main.yml
+++ b/contrib/kvm-setup/roles/kvm-setup/tasks/main.yml
@@ -11,12 +11,12 @@
 
 - name: Install required packages
   apt:
-    upgrade: yes
-    update_cache: yes
+    upgrade: true
+    update_cache: true
     cache_valid_time: 3600
     name: "{{ item }}"
     state: present
-    install_recommends: no
+    install_recommends: false
   with_items:
     - dnsutils
     - ntp
diff --git a/contrib/kvm-setup/roles/kvm-setup/tasks/sysctl.yml b/contrib/kvm-setup/roles/kvm-setup/tasks/sysctl.yml
index 6934eccf3b9616a0af24ab6ee2bba7b04b123e62..75b7ff8fd229d7762d02ee9e7fc02ec2df795922 100644
--- a/contrib/kvm-setup/roles/kvm-setup/tasks/sysctl.yml
+++ b/contrib/kvm-setup/roles/kvm-setup/tasks/sysctl.yml
@@ -30,7 +30,7 @@
     value: 1
     sysctl_file: "{{ sysctl_file_path }}"
     state: present
-    reload: yes
+    reload: true
 
 - name: Set bridge-nf-call-{arptables,iptables} to 0
   ansible.posix.sysctl:
@@ -38,7 +38,7 @@
     state: present
     value: 0
     sysctl_file: "{{ sysctl_file_path }}"
-    reload: yes
+    reload: true
   with_items:
     - net.bridge.bridge-nf-call-arptables
     - net.bridge.bridge-nf-call-ip6tables
diff --git a/contrib/network-storage/glusterfs/roles/glusterfs/README.md b/contrib/network-storage/glusterfs/roles/glusterfs/README.md
index dda243df04fd23c2978367e3c6a585b1ef4619f2..9e5bf5dcfbc682aafcca4b78ab40bf9c9bec9cde 100644
--- a/contrib/network-storage/glusterfs/roles/glusterfs/README.md
+++ b/contrib/network-storage/glusterfs/roles/glusterfs/README.md
@@ -21,7 +21,7 @@ glusterfs_default_release: ""
 You can specify a `default_release` for apt on Debian/Ubuntu by overriding this variable. This is helpful if you need a different package or version for the main GlusterFS packages (e.g. GlusterFS 3.5.x instead of 3.2.x with the `wheezy-backports` default release on Debian Wheezy).
 
 ```yaml
-glusterfs_ppa_use: yes
+glusterfs_ppa_use: true
 glusterfs_ppa_version: "3.5"
 ```
 
diff --git a/contrib/network-storage/glusterfs/roles/glusterfs/client/defaults/main.yml b/contrib/network-storage/glusterfs/roles/glusterfs/client/defaults/main.yml
index b9f0d2d1d3b76adc37524340fe09c756f0bc4156..c3fff2e63248ce6f4473b47e297d82d9a85c1551 100644
--- a/contrib/network-storage/glusterfs/roles/glusterfs/client/defaults/main.yml
+++ b/contrib/network-storage/glusterfs/roles/glusterfs/client/defaults/main.yml
@@ -1,7 +1,7 @@
 ---
 # For Ubuntu.
 glusterfs_default_release: ""
-glusterfs_ppa_use: yes
+glusterfs_ppa_use: true
 glusterfs_ppa_version: "4.1"
 
 # Gluster configuration.
diff --git a/contrib/network-storage/glusterfs/roles/glusterfs/client/tasks/setup-Debian.yml b/contrib/network-storage/glusterfs/roles/glusterfs/client/tasks/setup-Debian.yml
index da7a4d8decc6a7c768f28460dccae4928f049387..0d7cc18747ac39b0bdabec3a146ba613836395a3 100644
--- a/contrib/network-storage/glusterfs/roles/glusterfs/client/tasks/setup-Debian.yml
+++ b/contrib/network-storage/glusterfs/roles/glusterfs/client/tasks/setup-Debian.yml
@@ -3,7 +3,7 @@
   apt_repository:
     repo: 'ppa:gluster/glusterfs-{{ glusterfs_ppa_version }}'
     state: present
-    update_cache: yes
+    update_cache: true
   register: glusterfs_ppa_added
   when: glusterfs_ppa_use
 
diff --git a/contrib/network-storage/glusterfs/roles/glusterfs/server/defaults/main.yml b/contrib/network-storage/glusterfs/roles/glusterfs/server/defaults/main.yml
index ef9a71eba401947cdc71a782193ac1f606e0cf41..7d6e1025b1f0513c4901937a29f2fa7d56552111 100644
--- a/contrib/network-storage/glusterfs/roles/glusterfs/server/defaults/main.yml
+++ b/contrib/network-storage/glusterfs/roles/glusterfs/server/defaults/main.yml
@@ -1,7 +1,7 @@
 ---
 # For Ubuntu.
 glusterfs_default_release: ""
-glusterfs_ppa_use: yes
+glusterfs_ppa_use: true
 glusterfs_ppa_version: "3.12"
 
 # Gluster configuration.
diff --git a/contrib/network-storage/glusterfs/roles/glusterfs/server/tasks/main.yml b/contrib/network-storage/glusterfs/roles/glusterfs/server/tasks/main.yml
index 6bdc41420ed1598785e478c8a16469f43df1781a..a9f7698a37e829878b9755c0811fac948c6ffc3a 100644
--- a/contrib/network-storage/glusterfs/roles/glusterfs/server/tasks/main.yml
+++ b/contrib/network-storage/glusterfs/roles/glusterfs/server/tasks/main.yml
@@ -43,7 +43,7 @@
   service:
     name: "{{ glusterfs_daemon }}"
     state: started
-    enabled: yes
+    enabled: true
 
 - name: Ensure Gluster brick and mount directories exist.
   file:
@@ -62,7 +62,7 @@
     replicas: "{{ groups['gfs-cluster'] | length }}"
     cluster: "{% for item in groups['gfs-cluster'] -%}{{ hostvars[item]['ip'] | default(hostvars[item].ansible_default_ipv4['address']) }}{% if not loop.last %},{% endif %}{%- endfor %}"
     host: "{{ inventory_hostname }}"
-    force: yes
+    force: true
   run_once: true
   when: groups['gfs-cluster'] | length > 1
 
@@ -73,7 +73,7 @@
     brick: "{{ gluster_brick_dir }}"
     cluster: "{% for item in groups['gfs-cluster'] -%}{{ hostvars[item]['ip'] | default(hostvars[item].ansible_default_ipv4['address']) }}{% if not loop.last %},{% endif %}{%- endfor %}"
     host: "{{ inventory_hostname }}"
-    force: yes
+    force: true
   run_once: true
   when: groups['gfs-cluster'] | length <= 1
 
diff --git a/contrib/network-storage/glusterfs/roles/glusterfs/server/tasks/setup-Debian.yml b/contrib/network-storage/glusterfs/roles/glusterfs/server/tasks/setup-Debian.yml
index 104735903eec01361225be07b41e3a4efbc6d06a..4d4b1b4b80d5997a890a164ca7854431c03ab491 100644
--- a/contrib/network-storage/glusterfs/roles/glusterfs/server/tasks/setup-Debian.yml
+++ b/contrib/network-storage/glusterfs/roles/glusterfs/server/tasks/setup-Debian.yml
@@ -3,7 +3,7 @@
   apt_repository:
     repo: 'ppa:gluster/glusterfs-{{ glusterfs_ppa_version }}'
     state: present
-    update_cache: yes
+    update_cache: true
   register: glusterfs_ppa_added
   when: glusterfs_ppa_use
 
diff --git a/contrib/network-storage/heketi/heketi-tear-down.yml b/contrib/network-storage/heketi/heketi-tear-down.yml
index e64f085cb6701effb09c075f4df20c823bdac3d4..8c9d1c3a000b1aa74374d2ced32d7944ecec78e9 100644
--- a/contrib/network-storage/heketi/heketi-tear-down.yml
+++ b/contrib/network-storage/heketi/heketi-tear-down.yml
@@ -6,6 +6,6 @@
 
 - name: Teardown disks in heketi
   hosts: heketi-node
-  become: yes
+  become: true
   roles:
     - { role: tear-down-disks }
diff --git a/contrib/offline/generate_list.yml b/contrib/offline/generate_list.yml
index 6b2bcf8067f48cabdd56e8722a2d8a09606a0a50..f103d031f725f7e362cf7a4363d88c7e0a2e2832 100644
--- a/contrib/offline/generate_list.yml
+++ b/contrib/offline/generate_list.yml
@@ -1,7 +1,7 @@
 ---
 - name: Collect container images for offline deployment
   hosts: localhost
-  become: no
+  become: false
 
   roles:
     # Just load default variables from roles.
diff --git a/contrib/os-services/roles/prepare/tasks/main.yml b/contrib/os-services/roles/prepare/tasks/main.yml
index 177712e42d9c1cee1501cf29d7d8df784fb10ba0..487b3b6f1cef84e1295a13bf9e9d563048b8f5a1 100644
--- a/contrib/os-services/roles/prepare/tasks/main.yml
+++ b/contrib/os-services/roles/prepare/tasks/main.yml
@@ -10,7 +10,7 @@
     systemd_service:
       name: firewalld
       state: stopped
-      enabled: no
+      enabled: false
     when:
       "'firewalld.service' in services and services['firewalld.service'].status != 'not-found'"
 
@@ -18,6 +18,6 @@
     systemd_service:
       name: ufw
       state: stopped
-      enabled: no
+      enabled: false
     when:
       "'ufw.service' in services and services['ufw.service'].status != 'not-found'"
diff --git a/extra_playbooks/upgrade-only-k8s.yml b/extra_playbooks/upgrade-only-k8s.yml
index 4207f8d28cb3ade65263e019d243a629cec5e413..5f396fa28ebf64554eb033f7676a9ab1eac6efa2 100644
--- a/extra_playbooks/upgrade-only-k8s.yml
+++ b/extra_playbooks/upgrade-only-k8s.yml
@@ -12,7 +12,7 @@
 
 - name: Setup ssh config to use the bastion
   hosts: localhost
-  gather_facts: False
+  gather_facts: false
   roles:
     - { role: kubespray-defaults}
     - { role: bastion-ssh-config, tags: ["localhost", "bastion"]}
diff --git a/playbooks/ansible_version.yml b/playbooks/ansible_version.yml
index aa2d6b47642bcb61c03c02a50fb7b4fb81464194..2c8bac63c2818aa123dce08622c53c94364e327a 100644
--- a/playbooks/ansible_version.yml
+++ b/playbooks/ansible_version.yml
@@ -2,7 +2,7 @@
 - name: Check Ansible version
   hosts: all
   gather_facts: false
-  become: no
+  become: false
   run_once: true
   vars:
     minimal_ansible_version: 2.16.4
diff --git a/playbooks/boilerplate.yml b/playbooks/boilerplate.yml
index 137a4c2c5b42ae8403a1930513699a5b274477bb..eafa9b42f126af4b2d71245fa82578e2d75fe613 100644
--- a/playbooks/boilerplate.yml
+++ b/playbooks/boilerplate.yml
@@ -51,7 +51,7 @@
 
 - name: Install bastion ssh config
   hosts: bastion[0]
-  gather_facts: False
+  gather_facts: false
   environment: "{{ proxy_disable_env }}"
   roles:
     - { role: kubespray-defaults }
diff --git a/playbooks/cluster.yml b/playbooks/cluster.yml
index c433a8c69284e167cf6912c80e9af434edaf8461..ca67a28d33388fc46c3112899e9798b34643df28 100644
--- a/playbooks/cluster.yml
+++ b/playbooks/cluster.yml
@@ -7,7 +7,7 @@
 
 - name: Prepare for etcd install
   hosts: k8s_cluster:etcd
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -21,7 +21,7 @@
 
 - name: Install Kubernetes nodes
   hosts: k8s_cluster
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -30,7 +30,7 @@
 
 - name: Install the control plane
   hosts: kube_control_plane
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -41,7 +41,7 @@
 
 - name: Invoke kubeadm and install a CNI
   hosts: k8s_cluster
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -54,7 +54,7 @@
 
 - name: Install Calico Route Reflector
   hosts: calico_rr
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -63,7 +63,7 @@
 
 - name: Patch Kubernetes for Windows
   hosts: kube_control_plane[0]
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -72,7 +72,7 @@
 
 - name: Install Kubernetes apps
   hosts: kube_control_plane
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -86,7 +86,7 @@
 
 - name: Apply resolv.conf changes now that cluster DNS is up
   hosts: k8s_cluster
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
diff --git a/playbooks/facts.yml b/playbooks/facts.yml
index 77823aca40f2f3c23345662b585c3bb5c766ff86..d35eea80c28ae5547282be654fc2e72abc3ba630 100644
--- a/playbooks/facts.yml
+++ b/playbooks/facts.yml
@@ -15,7 +15,7 @@
 
 - name: Gather facts
   hosts: k8s_cluster:etcd:calico_rr
-  gather_facts: False
+  gather_facts: false
   tags: always
   tasks:
     - name: Gather minimal facts
diff --git a/playbooks/install_etcd.yml b/playbooks/install_etcd.yml
index b8e4d1d89b040900f60ef89d5a956fd6714a1a73..1f585119ca4670c3df5fb7dc29c40f89fc2dbf2c 100644
--- a/playbooks/install_etcd.yml
+++ b/playbooks/install_etcd.yml
@@ -16,7 +16,7 @@
 
 - name: Install etcd
   hosts: etcd:kube_control_plane:_kubespray_needs_etcd
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
diff --git a/playbooks/remove_node.yml b/playbooks/remove_node.yml
index e01338965e14022bbdd66c6dcc2e062169cea3a3..f994dae43bb50283b00b81590d9beba5d04066d2 100644
--- a/playbooks/remove_node.yml
+++ b/playbooks/remove_node.yml
@@ -4,13 +4,13 @@
 
 - name: Confirm node removal
   hosts: "{{ node | default('etcd:k8s_cluster:calico_rr') }}"
-  gather_facts: no
+  gather_facts: false
   tasks:
     - name: Confirm Execution
       pause:
         prompt: "Are you sure you want to delete nodes state? Type 'yes' to delete nodes."
       register: pause_result
-      run_once: True
+      run_once: true
       when:
         - not (skip_confirmation | default(false) | bool)
 
@@ -25,7 +25,7 @@
 
 - name: Reset node
   hosts: "{{ node | default('kube_node') }}"
-  gather_facts: no
+  gather_facts: false
   environment: "{{ proxy_disable_env }}"
   roles:
     - { role: kubespray-defaults, when: reset_nodes | default(True) | bool }
@@ -36,7 +36,7 @@
 # Currently cannot remove first master or etcd
 - name: Post node removal
   hosts: "{{ node | default('kube_control_plane[1:]:etcd[1:]') }}"
-  gather_facts: no
+  gather_facts: false
   environment: "{{ proxy_disable_env }}"
   roles:
     - { role: kubespray-defaults, when: reset_nodes | default(True) | bool }
diff --git a/playbooks/reset.yml b/playbooks/reset.yml
index 5742bd844e49fca23f29995a88302f38df7ecc9d..6fa18c0aca19313e0ee0bfa64e65fb6afae68edc 100644
--- a/playbooks/reset.yml
+++ b/playbooks/reset.yml
@@ -7,13 +7,13 @@
 
 - name: Reset cluster
   hosts: etcd:k8s_cluster:calico_rr
-  gather_facts: False
+  gather_facts: false
   pre_tasks:
     - name: Reset Confirmation
       pause:
         prompt: "Are you sure you want to reset cluster state? Type 'yes' to reset your cluster."
       register: reset_confirmation_prompt
-      run_once: True
+      run_once: true
       when:
         - not (skip_confirmation | default(false) | bool)
         - reset_confirmation is not defined
diff --git a/playbooks/scale.yml b/playbooks/scale.yml
index 171e378328d8f59e60cce91d0379fd5f6e5faf1c..ef0936607850c0d1912297b03ba00da5479cda8e 100644
--- a/playbooks/scale.yml
+++ b/playbooks/scale.yml
@@ -7,7 +7,7 @@
 
 - name: Generate the etcd certificates beforehand
   hosts: etcd:kube_control_plane
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -24,7 +24,7 @@
 
 - name: Download images to ansible host cache via first kube_control_plane node
   hosts: kube_control_plane[0]
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -34,7 +34,7 @@
 
 - name: Target only workers to get kubelet installed and checking in on any new nodes(engine)
   hosts: kube_node
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -53,7 +53,7 @@
 
 - name: Target only workers to get kubelet installed and checking in on any new nodes(node)
   hosts: kube_node
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -63,7 +63,7 @@
 - name: Upload control plane certs and retrieve encryption key
   hosts: kube_control_plane | first
   environment: "{{ proxy_disable_env }}"
-  gather_facts: False
+  gather_facts: false
   tags: kubeadm
   roles:
     - { role: kubespray-defaults }
@@ -84,7 +84,7 @@
 
 - name: Target only workers to get kubelet installed and checking in on any new nodes(network)
   hosts: kube_node
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -96,7 +96,7 @@
 
 - name: Apply resolv.conf changes now that cluster DNS is up
   hosts: k8s_cluster
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
diff --git a/playbooks/upgrade_cluster.yml b/playbooks/upgrade_cluster.yml
index 3180fec9310e309d9c97e6858631e11738eb864e..99511a8206f78dd8281ecf1553f1ad214a873f4d 100644
--- a/playbooks/upgrade_cluster.yml
+++ b/playbooks/upgrade_cluster.yml
@@ -7,7 +7,7 @@
 
 - name: Download images to ansible host cache via first kube_control_plane node
   hosts: kube_control_plane[0]
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -17,7 +17,7 @@
 
 - name: Prepare nodes for upgrade
   hosts: k8s_cluster:etcd:calico_rr
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -27,7 +27,7 @@
 
 - name: Upgrade container engine on non-cluster nodes
   hosts: etcd:calico_rr:!k8s_cluster
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   serial: "{{ serial | default('20%') }}"
@@ -39,7 +39,7 @@
   import_playbook: install_etcd.yml
 
 - name: Handle upgrades to master components first to maintain backwards compat.
-  gather_facts: False
+  gather_facts: false
   hosts: kube_control_plane
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
@@ -62,7 +62,7 @@
 
 - name: Upgrade calico and external cloud provider on all masters, calico-rrs, and nodes
   hosts: kube_control_plane:calico_rr:kube_node
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   serial: "{{ serial | default('20%') }}"
   environment: "{{ proxy_disable_env }}"
@@ -75,7 +75,7 @@
 
 - name: Finally handle worker upgrades, based on given batch size
   hosts: kube_node:calico_rr:!kube_control_plane
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   serial: "{{ serial | default('20%') }}"
@@ -93,7 +93,7 @@
 
 - name: Patch Kubernetes for Windows
   hosts: kube_control_plane[0]
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: true
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -102,7 +102,7 @@
 
 - name: Install Calico Route Reflector
   hosts: calico_rr
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -111,7 +111,7 @@
 
 - name: Install Kubernetes apps
   hosts: kube_control_plane
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
@@ -122,7 +122,7 @@
 
 - name: Apply resolv.conf changes now that cluster DNS is up
   hosts: k8s_cluster
-  gather_facts: False
+  gather_facts: false
   any_errors_fatal: "{{ any_errors_fatal | default(true) }}"
   environment: "{{ proxy_disable_env }}"
   roles:
diff --git a/roles/adduser/defaults/main.yml b/roles/adduser/defaults/main.yml
index df3fc2d0284a9ef360e54c943ed9dbed6400d538..3307032e05843a68210a3024003cf73f199ac0e6 100644
--- a/roles/adduser/defaults/main.yml
+++ b/roles/adduser/defaults/main.yml
@@ -7,14 +7,14 @@ addusers:
   etcd:
     name: etcd
     comment: "Etcd user"
-    create_home: no
-    system: yes
+    create_home: false
+    system: true
     shell: /sbin/nologin
   kube:
     name: kube
     comment: "Kubernetes user"
-    create_home: no
-    system: yes
+    create_home: false
+    system: true
     shell: /sbin/nologin
     group: "{{ kube_cert_group }}"
 
diff --git a/roles/adduser/vars/coreos.yml b/roles/adduser/vars/coreos.yml
index 5c258df6e5623fdb77cc508d3b40aa27c95bf90c..60fb05cf0dc7273170a30bf2eeb7cc00017b9128 100644
--- a/roles/adduser/vars/coreos.yml
+++ b/roles/adduser/vars/coreos.yml
@@ -3,6 +3,6 @@ addusers:
   - name: kube
     comment: "Kubernetes user"
     shell: /sbin/nologin
-    system: yes
+    system: true
     group: "{{ kube_cert_group }}"
-    create_home: no
+    create_home: false
diff --git a/roles/adduser/vars/debian.yml b/roles/adduser/vars/debian.yml
index 99e5b382175965a12628241b3138ffe574e82166..b14b8612e5140e5cb093d77f63737ae53bfca893 100644
--- a/roles/adduser/vars/debian.yml
+++ b/roles/adduser/vars/debian.yml
@@ -2,14 +2,14 @@
 addusers:
   - name: etcd
     comment: "Etcd user"
-    create_home: yes
+    create_home: true
     home: "{{ etcd_data_dir }}"
-    system: yes
+    system: true
     shell: /sbin/nologin
 
   - name: kube
     comment: "Kubernetes user"
-    create_home: no
-    system: yes
+    create_home: false
+    system: true
     shell: /sbin/nologin
     group: "{{ kube_cert_group }}"
diff --git a/roles/adduser/vars/redhat.yml b/roles/adduser/vars/redhat.yml
index 99e5b382175965a12628241b3138ffe574e82166..b14b8612e5140e5cb093d77f63737ae53bfca893 100644
--- a/roles/adduser/vars/redhat.yml
+++ b/roles/adduser/vars/redhat.yml
@@ -2,14 +2,14 @@
 addusers:
   - name: etcd
     comment: "Etcd user"
-    create_home: yes
+    create_home: true
     home: "{{ etcd_data_dir }}"
-    system: yes
+    system: true
     shell: /sbin/nologin
 
   - name: kube
     comment: "Kubernetes user"
-    create_home: no
-    system: yes
+    create_home: false
+    system: true
     shell: /sbin/nologin
     group: "{{ kube_cert_group }}"
diff --git a/roles/bootstrap-os/molecule/default/converge.yml b/roles/bootstrap-os/molecule/default/converge.yml
index 1f44ec9ca2c8dbc9060cc7f4aaad412040debf3c..89a83255944488d7a2d2a1f1a6e64f3caa7d91fc 100644
--- a/roles/bootstrap-os/molecule/default/converge.yml
+++ b/roles/bootstrap-os/molecule/default/converge.yml
@@ -1,6 +1,6 @@
 ---
 - name: Converge
   hosts: all
-  gather_facts: no
+  gather_facts: false
   roles:
     - role: bootstrap-os
diff --git a/roles/bootstrap-os/tasks/amzn.yml b/roles/bootstrap-os/tasks/amzn.yml
index 0da5591caa33667c04ed3cbeb08842c73e841f81..8a473a07f87f9feeaceec4f450505ab081c99efc 100644
--- a/roles/bootstrap-os/tasks/amzn.yml
+++ b/roles/bootstrap-os/tasks/amzn.yml
@@ -8,9 +8,9 @@
     file: epel
     description: Extra Packages for Enterprise Linux 7 - $basearch
     baseurl: http://download.fedoraproject.org/pub/epel/7/$basearch
-    gpgcheck: yes
+    gpgcheck: true
     gpgkey: http://download.fedoraproject.org/pub/epel/RPM-GPG-KEY-EPEL-7
-    skip_if_unavailable: yes
-    enabled: yes
-    repo_gpgcheck: no
+    skip_if_unavailable: true
+    enabled: true
+    repo_gpgcheck: false
   when: epel_enabled
diff --git a/roles/bootstrap-os/tasks/centos.yml b/roles/bootstrap-os/tasks/centos.yml
index fc9a3cf0cff795e9c066482509cf1ae13677e5e6..304a37b079e16a643f1c994798ab9a1598ac15a5 100644
--- a/roles/bootstrap-os/tasks/centos.yml
+++ b/roles/bootstrap-os/tasks/centos.yml
@@ -119,9 +119,9 @@
 - name: Check presence of fastestmirror.conf
   stat:
     path: /etc/yum/pluginconf.d/fastestmirror.conf
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: fastestmirror
 
 # the fastestmirror plugin can actually slow down Ansible deployments
diff --git a/roles/bootstrap-os/tasks/fedora-coreos.yml b/roles/bootstrap-os/tasks/fedora-coreos.yml
index b8c0f3fe7b65acef67b9eb1e18a0ff38bdbd78e3..3062a5a88343b00ef49f25c621a1e21cb34266e6 100644
--- a/roles/bootstrap-os/tasks/fedora-coreos.yml
+++ b/roles/bootstrap-os/tasks/fedora-coreos.yml
@@ -28,7 +28,7 @@
   raw: "nohup bash -c 'sleep 5s && shutdown -r now'"
   become: true
   ignore_errors: true  # noqa ignore-errors
-  ignore_unreachable: yes
+  ignore_unreachable: true
   when: need_bootstrap.rc != 0
 
 - name: Wait for the reboot to complete
diff --git a/roles/bootstrap-os/tasks/main.yml b/roles/bootstrap-os/tasks/main.yml
index e62fbf49654e30bb2bfa525dd46d63826f293b01..c16fe1bec63d309487b009480d7899cc6ac984dd 100644
--- a/roles/bootstrap-os/tasks/main.yml
+++ b/roles/bootstrap-os/tasks/main.yml
@@ -22,7 +22,7 @@
       - "{{ os_release_dict['ID'] }}.yml"
       paths:
       - vars/
-      skip: True
+      skip: true
   - name: Include tasks
     include_tasks: "{{ included_tasks_file }}"
     with_first_found:
diff --git a/roles/bootstrap-os/tasks/opensuse.yml b/roles/bootstrap-os/tasks/opensuse.yml
index 9b69dcd8916289b2379c5631cd2465f72c67d812..5a4f9dead52e2ac85f0df3f770f3951d593dd052 100644
--- a/roles/bootstrap-os/tasks/opensuse.yml
+++ b/roles/bootstrap-os/tasks/opensuse.yml
@@ -8,9 +8,9 @@
 - name: Check that /etc/sysconfig/proxy file exists
   stat:
     path: /etc/sysconfig/proxy
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: stat_result
 
 - name: Create the /etc/sysconfig/proxy empty file
diff --git a/roles/bootstrap-os/tasks/redhat.yml b/roles/bootstrap-os/tasks/redhat.yml
index 0aae5a0d6ccb68fd866dee3b5fe6816b78eb1d64..76a39b2f515e41d530e12b46883c1d136c86e028 100644
--- a/roles/bootstrap-os/tasks/redhat.yml
+++ b/roles/bootstrap-os/tasks/redhat.yml
@@ -87,9 +87,9 @@
 - name: Check presence of fastestmirror.conf
   stat:
     path: /etc/yum/pluginconf.d/fastestmirror.conf
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: fastestmirror
 
 # the fastestmirror plugin can actually slow down Ansible deployments
diff --git a/roles/bootstrap-os/vars/fedora-coreos.yml b/roles/bootstrap-os/vars/fedora-coreos.yml
index e0bb069f9e26c2fa5a063d34265552c3a0315d23..37e4c46e8e394518c9eb92c7bd9a1815892de6e2 100644
--- a/roles/bootstrap-os/vars/fedora-coreos.yml
+++ b/roles/bootstrap-os/vars/fedora-coreos.yml
@@ -1,2 +1,2 @@
 ---
-is_fedora_coreos: True
+is_fedora_coreos: true
diff --git a/roles/container-engine/containerd-common/tasks/main.yml b/roles/container-engine/containerd-common/tasks/main.yml
index d0cf1f139428407fff763e0e885093a022506dc2..c5b896808694c2b6d17a13ecfd1557317379d362 100644
--- a/roles/container-engine/containerd-common/tasks/main.yml
+++ b/roles/container-engine/containerd-common/tasks/main.yml
@@ -2,9 +2,9 @@
 - name: Containerd-common | check if fedora coreos
   stat:
     path: /run/ostree-booted
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: ostree
 
 - name: Containerd-common | set is_ostree
diff --git a/roles/container-engine/containerd/handlers/main.yml b/roles/container-engine/containerd/handlers/main.yml
index 4e7722f4fb29520a0cc70b53dba62d06c5e21012..6a024fd59c625a05f81fd762b6e79c205958f5b2 100644
--- a/roles/container-engine/containerd/handlers/main.yml
+++ b/roles/container-engine/containerd/handlers/main.yml
@@ -3,9 +3,9 @@
   systemd_service:
     name: containerd
     state: restarted
-    enabled: yes
-    daemon-reload: yes
-    masked: no
+    enabled: true
+    daemon-reload: true
+    masked: false
   listen: Restart containerd
 
 - name: Containerd | wait for containerd
diff --git a/roles/container-engine/containerd/molecule/default/prepare.yml b/roles/container-engine/containerd/molecule/default/prepare.yml
index ddc9c045396882233b34f3f530da0becd5b01ed9..a3d09ad80054e694269b25246f303422580a5967 100644
--- a/roles/container-engine/containerd/molecule/default/prepare.yml
+++ b/roles/container-engine/containerd/molecule/default/prepare.yml
@@ -1,7 +1,7 @@
 ---
 - name: Prepare
   hosts: all
-  gather_facts: False
+  gather_facts: false
   become: true
   vars:
     ignore_assert_errors: true
@@ -19,7 +19,7 @@
 
 - name: Prepare CNI
   hosts: all
-  gather_facts: False
+  gather_facts: false
   become: true
   vars:
     ignore_assert_errors: true
diff --git a/roles/container-engine/containerd/tasks/main.yml b/roles/container-engine/containerd/tasks/main.yml
index 073412cd0a0d6ab1024bb3e4d49c3cf29c071c4b..8b8c12cbbf733104cb1100f4235e1da0e2b8e674 100644
--- a/roles/container-engine/containerd/tasks/main.yml
+++ b/roles/container-engine/containerd/tasks/main.yml
@@ -36,7 +36,7 @@
     src: "{{ downloads.containerd.dest }}"
     dest: "{{ containerd_bin_dir }}"
     mode: "0755"
-    remote_src: yes
+    remote_src: true
     extra_opts:
       - --strip-components=1
   notify: Restart containerd
@@ -138,6 +138,6 @@
 - name: Containerd | Ensure containerd is started and enabled
   systemd_service:
     name: containerd
-    daemon_reload: yes
-    enabled: yes
+    daemon_reload: true
+    enabled: true
     state: started
diff --git a/roles/container-engine/cri-dockerd/handlers/main.yml b/roles/container-engine/cri-dockerd/handlers/main.yml
index 00d00e7b2b9eb620ddae8eb3b889572f032aa43d..f60f28fce298dcba904b38ec9b1038e0351547a2 100644
--- a/roles/container-engine/cri-dockerd/handlers/main.yml
+++ b/roles/container-engine/cri-dockerd/handlers/main.yml
@@ -3,7 +3,7 @@
   systemd_service:
     name: cri-dockerd
     daemon_reload: true
-    masked: no
+    masked: false
   listen: Restart and enable cri-dockerd
 
 - name: Cri-dockerd | restart docker.service
@@ -27,5 +27,5 @@
 - name: Cri-dockerd | enable cri-dockerd service
   service:
     name: cri-dockerd.service
-    enabled: yes
+    enabled: true
   listen: Restart and enable cri-dockerd
diff --git a/roles/container-engine/cri-o/handlers/main.yml b/roles/container-engine/cri-o/handlers/main.yml
index d173ce41b00e0046c5f9075dce6c9c1de86e014a..0595e4f94ea3f297f4b4097ad6d4d75bc49d26c8 100644
--- a/roles/container-engine/cri-o/handlers/main.yml
+++ b/roles/container-engine/cri-o/handlers/main.yml
@@ -8,5 +8,5 @@
   service:
     name: crio
     state: restarted
-    enabled: yes
+    enabled: true
   listen: Restart crio
diff --git a/roles/container-engine/cri-o/molecule/default/prepare.yml b/roles/container-engine/cri-o/molecule/default/prepare.yml
index c769d7cd2d34924998a9d235a790ba9a069a2024..55ad5174d706b305a29acc060afb3151b98162a7 100644
--- a/roles/container-engine/cri-o/molecule/default/prepare.yml
+++ b/roles/container-engine/cri-o/molecule/default/prepare.yml
@@ -1,7 +1,7 @@
 ---
 - name: Prepare
   hosts: all
-  gather_facts: False
+  gather_facts: false
   become: true
   vars:
     ignore_assert_errors: true
@@ -19,7 +19,7 @@
 
 - name: Prepare CNI
   hosts: all
-  gather_facts: False
+  gather_facts: false
   become: true
   vars:
     ignore_assert_errors: true
diff --git a/roles/container-engine/cri-o/tasks/main.yaml b/roles/container-engine/cri-o/tasks/main.yaml
index a7b234563dcd75eef8d2f332e1dc259ca15255db..bde2e075623471c9ae507960ab5f1e877d315160 100644
--- a/roles/container-engine/cri-o/tasks/main.yaml
+++ b/roles/container-engine/cri-o/tasks/main.yaml
@@ -5,9 +5,9 @@
 - name: Cri-o | check if fedora coreos
   stat:
     path: /run/ostree-booted
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: ostree
 
 - name: Cri-o | set is_ostree
diff --git a/roles/container-engine/cri-o/tasks/setup-amazon.yaml b/roles/container-engine/cri-o/tasks/setup-amazon.yaml
index 2462c30fdfd17a76170e7e076a8e517cae080d2a..cef0112b29376a1f692891dbc5dabc4c11544761 100644
--- a/roles/container-engine/cri-o/tasks/setup-amazon.yaml
+++ b/roles/container-engine/cri-o/tasks/setup-amazon.yaml
@@ -8,7 +8,7 @@
   lineinfile:
     dest: /etc/yum.repos.d/amzn2-extras.repo
     line: "[amzn2extra-docker]"
-  check_mode: yes
+  check_mode: true
   register: amzn2_extras_docker_repo
   when:
     - amzn2_extras_file_stat.stat.exists
@@ -19,7 +19,7 @@
     section: amzn2extra-docker
     option: enabled
     value: "0"
-    backup: yes
+    backup: true
     mode: "0644"
   when:
     - amzn2_extras_file_stat.stat.exists
diff --git a/roles/container-engine/crictl/handlers/main.yml b/roles/container-engine/crictl/handlers/main.yml
index 785823fc4b86aaf7a37bfcba05b4d9a9dcf70da6..d6ffe169a50001548028c61969195f89c3dde263 100644
--- a/roles/container-engine/crictl/handlers/main.yml
+++ b/roles/container-engine/crictl/handlers/main.yml
@@ -1,7 +1,7 @@
 ---
 - name: Get crictl completion
   command: "{{ bin_dir }}/crictl completion"
-  changed_when: False
+  changed_when: false
   register: cri_completion
   check_mode: false
 
diff --git a/roles/container-engine/docker-storage/tasks/main.yml b/roles/container-engine/docker-storage/tasks/main.yml
index e3c713db22cc021eb64fda0915519896e38f4344..d90dcb775c36dd6535a43ef8cd2b739411782207 100644
--- a/roles/container-engine/docker-storage/tasks/main.yml
+++ b/roles/container-engine/docker-storage/tasks/main.yml
@@ -39,7 +39,7 @@
     state: present
 
 - name: Docker-storage-setup | install and run container-storage-setup
-  become: yes
+  become: true
   script: |
     install_container_storage_setup.sh \
       {{ docker_container_storage_setup_repository }} \
diff --git a/roles/container-engine/docker/handlers/main.yml b/roles/container-engine/docker/handlers/main.yml
index 72e95e6bf5b1041cd3afb0f55095b23a72078cb6..76d16f58927d6f363c81858d94f4a6ed1ddcda83 100644
--- a/roles/container-engine/docker/handlers/main.yml
+++ b/roles/container-engine/docker/handlers/main.yml
@@ -3,7 +3,7 @@
   systemd_service:
     name: docker
     daemon_reload: true
-    masked: no
+    masked: false
   listen: Restart docker
 
 - name: Docker | reload docker.socket
diff --git a/roles/container-engine/docker/tasks/main.yml b/roles/container-engine/docker/tasks/main.yml
index 55b3a0be6c5a821dfd0fbc7c9e871e5498534620..e56fe4ca9945440f4140f0ea9fa479fb83f52432 100644
--- a/roles/container-engine/docker/tasks/main.yml
+++ b/roles/container-engine/docker/tasks/main.yml
@@ -2,9 +2,9 @@
 - name: Check if fedora coreos
   stat:
     path: /run/ostree-booted
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: ostree
 
 - name: Set is_ostree
@@ -66,7 +66,7 @@
     path: /etc/apt/sources.list
     regexp: 'buster-backports'
     state: absent
-    backup: yes
+    backup: true
   when:
     - ansible_os_family == 'Debian'
     - ansible_distribution_release == "buster"
@@ -183,7 +183,7 @@
 - name: Ensure docker service is started and enabled
   service:
     name: "{{ item }}"
-    enabled: yes
+    enabled: true
     state: started
   with_items:
     - docker
diff --git a/roles/container-engine/docker/tasks/set_facts_dns.yml b/roles/container-engine/docker/tasks/set_facts_dns.yml
index d7c10392e85daa74d16679862fb0d3e2a3edba38..d0ccd745a70951c12b09a1856d7801bd0d1ab820 100644
--- a/roles/container-engine/docker/tasks/set_facts_dns.yml
+++ b/roles/container-engine/docker/tasks/set_facts_dns.yml
@@ -21,9 +21,9 @@
   shell: set -o pipefail && grep "^nameserver" /etc/resolv.conf | sed -r 's/^nameserver\s*([^#\s]+)\s*(#.*)?/\1/'
   args:
     executable: /bin/bash
-  changed_when: False
+  changed_when: false
   register: system_nameservers
-  check_mode: no
+  check_mode: false
 
 - name: Check system search domains
   # noqa risky-shell-pipe - if resolf.conf has no search domain, grep will exit 1 which would force us to add failed_when: false
@@ -31,9 +31,9 @@
   shell: grep "^search" /etc/resolv.conf | sed -r 's/^search\s*([^#]+)\s*(#.*)?/\1/'
   args:
     executable: /bin/bash
-  changed_when: False
+  changed_when: false
   register: system_search_domains
-  check_mode: no
+  check_mode: false
 
 - name: Add system nameservers to docker options
   set_fact:
diff --git a/roles/container-engine/gvisor/tasks/main.yml b/roles/container-engine/gvisor/tasks/main.yml
index 13b19a2f6fadc0a31813485ada4f48ab86a1085c..4bab9a99671c4f8bafba0f560097b0e576f74d4e 100644
--- a/roles/container-engine/gvisor/tasks/main.yml
+++ b/roles/container-engine/gvisor/tasks/main.yml
@@ -14,7 +14,7 @@
     src: "{{ item.src }}"
     dest: "{{ bin_dir }}/{{ item.dest }}"
     mode: "0755"
-    remote_src: yes
+    remote_src: true
   with_items:
     - { src: "{{ downloads.gvisor_runsc.dest }}", dest: "runsc" }
     - { src: "{{ downloads.gvisor_containerd_shim.dest }}", dest: "containerd-shim-runsc-v1" }
diff --git a/roles/container-engine/kata-containers/tasks/main.yml b/roles/container-engine/kata-containers/tasks/main.yml
index 38778987d4421eecec5889e7b900c3da629242d7..5014c214a494c703105d3ca71cab5055b44502f7 100644
--- a/roles/container-engine/kata-containers/tasks/main.yml
+++ b/roles/container-engine/kata-containers/tasks/main.yml
@@ -11,7 +11,7 @@
     mode: "0755"
     owner: root
     group: root
-    remote_src: yes
+    remote_src: true
 
 - name: Kata-containers | Create config directory
   file:
diff --git a/roles/container-engine/nerdctl/handlers/main.yml b/roles/container-engine/nerdctl/handlers/main.yml
index 98de60c1c3ebcf52d5fa95132f71d0c25ffaffc2..1744706075c722666869d337a02f3e894329c88b 100644
--- a/roles/container-engine/nerdctl/handlers/main.yml
+++ b/roles/container-engine/nerdctl/handlers/main.yml
@@ -1,7 +1,7 @@
 ---
 - name: Get nerdctl completion
   command: "{{ bin_dir }}/nerdctl completion bash"
-  changed_when: False
+  changed_when: false
   register: nerdctl_completion
   check_mode: false
 
diff --git a/roles/container-engine/runc/tasks/main.yml b/roles/container-engine/runc/tasks/main.yml
index 3ee3fdae05f106403527cc85921a7700e9b65b5f..1d388768d47ecab1187633064e101264a68d8bb6 100644
--- a/roles/container-engine/runc/tasks/main.yml
+++ b/roles/container-engine/runc/tasks/main.yml
@@ -2,9 +2,9 @@
 - name: Runc | check if fedora coreos
   stat:
     path: /run/ostree-booted
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: ostree
 
 - name: Runc | set is_ostree
diff --git a/roles/container-engine/skopeo/tasks/main.yml b/roles/container-engine/skopeo/tasks/main.yml
index 95bb9697fb16d50a26af3e674df4dad08d70ca5b..8f21e3f1c3be2a515a452fc335fdca840f591bf3 100644
--- a/roles/container-engine/skopeo/tasks/main.yml
+++ b/roles/container-engine/skopeo/tasks/main.yml
@@ -2,9 +2,9 @@
 - name: Skopeo | check if fedora coreos
   stat:
     path: /run/ostree-booted
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: ostree
 
 - name: Skopeo | set is_ostree
diff --git a/roles/container-engine/validate-container-engine/tasks/main.yml b/roles/container-engine/validate-container-engine/tasks/main.yml
index 08ea1e5ca3859ad9e2b8ad3d6aeb4c1beef6f9a5..ffb541c24805e62f9ab448593f258faf1be15519 100644
--- a/roles/container-engine/validate-container-engine/tasks/main.yml
+++ b/roles/container-engine/validate-container-engine/tasks/main.yml
@@ -2,9 +2,9 @@
 - name: Validate-container-engine | check if fedora coreos
   stat:
     path: /run/ostree-booted
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: ostree
   tags:
     - facts
@@ -30,8 +30,8 @@
 - name: Check if containerd is installed
   find:
     file_type: file
-    recurse: yes
-    use_regex: yes
+    recurse: true
+    use_regex: true
     patterns:
       - containerd.service$
     paths:
@@ -45,8 +45,8 @@
 - name: Check if docker is installed
   find:
     file_type: file
-    recurse: yes
-    use_regex: yes
+    recurse: true
+    use_regex: true
     patterns:
       - docker.service$
     paths:
@@ -60,8 +60,8 @@
 - name: Check if crio is installed
   find:
     file_type: file
-    recurse: yes
-    use_regex: yes
+    recurse: true
+    use_regex: true
     patterns:
       - crio.service$
     paths:
diff --git a/roles/download/tasks/check_pull_required.yml b/roles/download/tasks/check_pull_required.yml
index e5ae1dcf349efabffd0fef78f5379e8855bb12bf..585013125726dfd1280b812e18d10e3f5c2cfd89 100644
--- a/roles/download/tasks/check_pull_required.yml
+++ b/roles/download/tasks/check_pull_required.yml
@@ -5,7 +5,7 @@
   shell: "{{ image_info_command }}"
   register: docker_images
   changed_when: false
-  check_mode: no
+  check_mode: false
   when: not download_always_pull
 
 - name: Check_pull_required | Set pull_required if the desired image is not yet loaded
diff --git a/roles/download/tasks/download_container.yml b/roles/download/tasks/download_container.yml
index f98adfa3f9fd68e292bb5d8d600d8b26df6e8562..5e67fe8c55049cecb53ff30939f30f64411f8348 100644
--- a/roles/download/tasks/download_container.yml
+++ b/roles/download/tasks/download_container.yml
@@ -26,12 +26,12 @@
     - name: Download_container | Determine if image is in cache
       stat:
         path: "{{ image_path_cached }}"
-        get_attributes: no
-        get_checksum: no
-        get_mime: no
+        get_attributes: false
+        get_checksum: false
+        get_mime: false
       delegate_to: localhost
       connection: local
-      delegate_facts: no
+      delegate_facts: false
       register: cache_image
       changed_when: false
       become: false
@@ -57,7 +57,7 @@
     - name: Download_container | Download image if required
       command: "{{ image_pull_command_on_localhost if download_localhost else image_pull_command }} {{ image_reponame }}"
       delegate_to: "{{ download_delegate if download_run_once else inventory_hostname }}"
-      delegate_facts: yes
+      delegate_facts: true
       run_once: "{{ download_run_once }}"
       register: pull_task_result
       until: pull_task_result is succeeded
@@ -72,7 +72,7 @@
     - name: Download_container | Save and compress image
       shell: "{{ image_save_command_on_localhost if download_localhost else image_save_command }}"  # noqa command-instead-of-shell - image_save_command_on_localhost contains a pipe, therefore requires shell
       delegate_to: "{{ download_delegate }}"
-      delegate_facts: no
+      delegate_facts: false
       register: container_save_status
       failed_when: container_save_status.stderr
       run_once: true
@@ -99,7 +99,7 @@
         dest: "{{ image_path_final }}"
         use_ssh_args: true
         mode: push
-      delegate_facts: no
+      delegate_facts: false
       register: upload_image
       failed_when: not upload_image
       until: upload_image is succeeded
diff --git a/roles/download/tasks/download_file.yml b/roles/download/tasks/download_file.yml
index 00dd33a28ebefab2e69601a03e1005f8860a7320..53a7a819798166f9a7f7906104817e50b8c9ecf1 100644
--- a/roles/download/tasks/download_file.yml
+++ b/roles/download/tasks/download_file.yml
@@ -24,13 +24,13 @@
       owner: "{{ download.owner | default(omit) }}"
       mode: "0755"
       state: directory
-      recurse: yes
+      recurse: true
 
   - name: Download_file | Create local cache directory
     file:
       path: "{{ file_path_cached | dirname }}"
       state: directory
-      recurse: yes
+      recurse: true
     delegate_to: localhost
     connection: local
     delegate_facts: false
@@ -45,7 +45,7 @@
     file:
       path: "{{ file_path_cached | dirname }}"
       state: directory
-      recurse: yes
+      recurse: true
     delegate_to: "{{ download_delegate }}"
     delegate_facts: false
     run_once: true
diff --git a/roles/download/tasks/extract_file.yml b/roles/download/tasks/extract_file.yml
index 59d0531f6132ff4371b20ecba7b60739b846318d..ce7536f4fae443a1b01e0427fc0bfa1b08ab60aa 100644
--- a/roles/download/tasks/extract_file.yml
+++ b/roles/download/tasks/extract_file.yml
@@ -5,7 +5,7 @@
     dest: "{{ download.dest | dirname }}"
     owner: "{{ download.owner | default(omit) }}"
     mode: "{{ download.mode | default(omit) }}"
-    copy: no
+    copy: false
     extra_opts: "{{ download.unarchive_extra_opts | default(omit) }}"
   when:
     - download.unarchive | default(false)
diff --git a/roles/download/tasks/prep_download.yml b/roles/download/tasks/prep_download.yml
index a8a79d711152505955f8e6fa6e050b45c032c0bd..15f3a91daf1d70cfcd5c2d36acbc513bd8fb9aed 100644
--- a/roles/download/tasks/prep_download.yml
+++ b/roles/download/tasks/prep_download.yml
@@ -62,7 +62,7 @@
   register: docker_images
   failed_when: false
   changed_when: false
-  check_mode: no
+  check_mode: false
   when: download_container
 
 - name: Prep_download | Create staging directory on remote node
@@ -81,7 +81,7 @@
     mode: "0755"
   delegate_to: localhost
   connection: local
-  delegate_facts: no
+  delegate_facts: false
   run_once: true
   become: false
   when:
diff --git a/roles/etcd/handlers/backup.yml b/roles/etcd/handlers/backup.yml
index 9c05a3ad00f1e6bf80a81ce531383102a3d20fbb..84e03accd0b21c2a4aa88d5dff12a530559636b3 100644
--- a/roles/etcd/handlers/backup.yml
+++ b/roles/etcd/handlers/backup.yml
@@ -23,9 +23,9 @@
 - name: Stat etcd v2 data directory
   stat:
     path: "{{ etcd_data_dir }}/member"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: etcd_data_dir_member
   listen: Restart etcd
   when: etcd_cluster_is_healthy.rc == 0
diff --git a/roles/etcd/handlers/main.yml b/roles/etcd/handlers/main.yml
index 62a8999456e54ea71838b97430de957bc99d47d9..9c8b8a82f165d86c86216ba26389e5eb75888fc6 100644
--- a/roles/etcd/handlers/main.yml
+++ b/roles/etcd/handlers/main.yml
@@ -26,7 +26,7 @@
 - name: Wait for etcd up
   uri:
     url: "https://{% if is_etcd_master %}{{ etcd_address }}{% else %}127.0.0.1{% endif %}:2379/health"
-    validate_certs: no
+    validate_certs: false
     client_cert: "{{ etcd_cert_dir }}/member-{{ inventory_hostname }}.pem"
     client_key: "{{ etcd_cert_dir }}/member-{{ inventory_hostname }}-key.pem"
   register: result
@@ -41,7 +41,7 @@
 - name: Wait for etcd-events up
   uri:
     url: "https://{% if is_etcd_master %}{{ etcd_address }}{% else %}127.0.0.1{% endif %}:2383/health"
-    validate_certs: no
+    validate_certs: false
     client_cert: "{{ etcd_cert_dir }}/member-{{ inventory_hostname }}.pem"
     client_key: "{{ etcd_cert_dir }}/member-{{ inventory_hostname }}-key.pem"
   register: result
diff --git a/roles/etcd/tasks/check_certs.yml b/roles/etcd/tasks/check_certs.yml
index 1611f9ec1d95f38faff84962c142f7e132ec3f57..440685aa739bad66f40e11a481b57fba0f8ae46f 100644
--- a/roles/etcd/tasks/check_certs.yml
+++ b/roles/etcd/tasks/check_certs.yml
@@ -17,9 +17,9 @@
 - name: "Check certs | Register ca and etcd admin/member certs on etcd hosts"
   stat:
     path: "{{ etcd_cert_dir }}/{{ item }}"
-    get_attributes: no
-    get_checksum: yes
-    get_mime: no
+    get_attributes: false
+    get_checksum: true
+    get_mime: false
   register: etcd_member_certs
   when: inventory_hostname in groups['etcd']
   with_items:
diff --git a/roles/etcd/tasks/configure.yml b/roles/etcd/tasks/configure.yml
index 4cf5387a09d6ec16dc80f1f8649352b1a9fc7f94..b7b943f0d10696d984818d7af0a948d11f0dd162 100644
--- a/roles/etcd/tasks/configure.yml
+++ b/roles/etcd/tasks/configure.yml
@@ -6,8 +6,8 @@
   register: etcd_cluster_is_healthy
   failed_when: false
   changed_when: false
-  check_mode: no
-  run_once: yes
+  check_mode: false
+  run_once: true
   when:
     - is_etcd_master
     - etcd_cluster_setup
@@ -27,8 +27,8 @@
   register: etcd_events_cluster_is_healthy
   failed_when: false
   changed_when: false
-  check_mode: no
-  run_once: yes
+  check_mode: false
+  run_once: true
   when:
     - is_etcd_master
     - etcd_events_cluster_setup
@@ -49,7 +49,7 @@
   template:
     src: "etcd-{{ etcd_deployment_type }}.service.j2"
     dest: /etc/systemd/system/etcd.service
-    backup: yes
+    backup: true
     mode: "0644"
     # FIXME: check that systemd version >= 250 (factory-reset.target was introduced in that release)
     # Remove once we drop support for systemd < 250
@@ -60,7 +60,7 @@
   template:
     src: "etcd-events-{{ etcd_deployment_type }}.service.j2"
     dest: /etc/systemd/system/etcd-events.service
-    backup: yes
+    backup: true
     mode: "0644"
     validate: "sh -c '[ -f /usr/bin/systemd/system/factory-reset.target ] || exit 0 && systemd-analyze verify %s:etcd-events-{{ etcd_deployment_type }}.service'"
     # FIXME: check that systemd version >= 250 (factory-reset.target was introduced in that release)
@@ -77,7 +77,7 @@
   service:
     name: etcd
     state: started
-    enabled: yes
+    enabled: true
   ignore_errors: "{{ etcd_cluster_is_healthy.rc == 0 }}"  # noqa ignore-errors
   when: is_etcd_master and etcd_cluster_setup
 
@@ -86,7 +86,7 @@
   service:
     name: etcd-events
     state: started
-    enabled: yes
+    enabled: true
   ignore_errors: "{{ etcd_events_cluster_is_healthy.rc != 0 }}"  # noqa ignore-errors
   when: is_etcd_master and etcd_events_cluster_setup
 
@@ -99,8 +99,8 @@
   retries: "{{ etcd_retries }}"
   delay: "{{ retry_stagger | random + 3 }}"
   changed_when: false
-  check_mode: no
-  run_once: yes
+  check_mode: false
+  run_once: true
   when:
     - is_etcd_master
     - etcd_cluster_setup
@@ -122,8 +122,8 @@
   retries: "{{ etcd_retries }}"
   delay: "{{ retry_stagger | random + 3 }}"
   changed_when: false
-  check_mode: no
-  run_once: yes
+  check_mode: false
+  run_once: true
   when:
     - is_etcd_master
     - etcd_events_cluster_setup
@@ -141,7 +141,7 @@
   register: etcd_member_in_cluster
   ignore_errors: true  # noqa ignore-errors
   changed_when: false
-  check_mode: no
+  check_mode: false
   when: is_etcd_master and etcd_cluster_setup
   tags:
     - facts
@@ -157,7 +157,7 @@
   register: etcd_events_member_in_cluster
   ignore_errors: true  # noqa ignore-errors
   changed_when: false
-  check_mode: no
+  check_mode: false
   when: is_etcd_master and etcd_events_cluster_setup
   tags:
     - facts
diff --git a/roles/etcd/tasks/gen_certs_script.yml b/roles/etcd/tasks/gen_certs_script.yml
index 711c14d64795cb83bb1cf22b8db2cd316eaf9f50..934b5eb370db26a87efa7023d4bc74248e63b687 100644
--- a/roles/etcd/tasks/gen_certs_script.yml
+++ b/roles/etcd/tasks/gen_certs_script.yml
@@ -6,7 +6,7 @@
     state: directory
     owner: "{{ etcd_owner }}"
     mode: "{{ etcd_cert_dir_mode }}"
-    recurse: yes
+    recurse: true
 
 - name: "Gen_certs | create etcd script dir (on {{ groups['etcd'][0] }})"
   file:
@@ -14,7 +14,7 @@
     state: directory
     owner: root
     mode: "0700"
-  run_once: yes
+  run_once: true
   when: inventory_hostname == groups['etcd'][0]
 
 - name: Gen_certs | write openssl config
@@ -22,7 +22,7 @@
     src: "openssl.conf.j2"
     dest: "{{ etcd_config_dir }}/openssl.conf"
     mode: "0640"
-  run_once: yes
+  run_once: true
   delegate_to: "{{ groups['etcd'][0] }}"
   when:
     - gen_certs | default(false)
@@ -33,7 +33,7 @@
     src: "make-ssl-etcd.sh.j2"
     dest: "{{ etcd_script_dir }}/make-ssl-etcd.sh"
     mode: "0700"
-  run_once: yes
+  run_once: true
   when:
     - gen_certs | default(false)
     - inventory_hostname == groups['etcd'][0]
@@ -43,7 +43,7 @@
   environment:
     MASTERS: "{{ groups['gen_master_certs_True'] | ansible.builtin.intersect(groups['etcd']) | join(' ') }}"
     HOSTS: "{{ groups['gen_node_certs_True'] | ansible.builtin.intersect(groups['kube_control_plane']) | join(' ') }}"
-  run_once: yes
+  run_once: true
   delegate_to: "{{ groups['etcd'][0] }}"
   when: gen_certs | default(false)
   notify: Set etcd_secret_changed
@@ -52,7 +52,7 @@
   command: "bash -x {{ etcd_script_dir }}/make-ssl-etcd.sh -f {{ etcd_config_dir }}/openssl.conf -d {{ etcd_cert_dir }}"
   environment:
     HOSTS: "{{ groups['gen_node_certs_True'] | ansible.builtin.intersect(groups['k8s_cluster']) | join(' ') }}"
-  run_once: yes
+  run_once: true
   delegate_to: "{{ groups['etcd'][0] }}"
   when:
     - kube_network_plugin in ["calico", "flannel", "cilium"] or cilium_deploy_additionally | default(false) | bool
@@ -153,4 +153,4 @@
     state: directory
     owner: "{{ etcd_owner }}"
     mode: "{{ etcd_cert_dir_mode }}"
-    recurse: yes
+    recurse: true
diff --git a/roles/etcd/tasks/gen_nodes_certs_script.yml b/roles/etcd/tasks/gen_nodes_certs_script.yml
index 2093bf8807fe3f79c31b50f6432c8f2d8a5d03d5..e074d0c01a731bf77f0dfff38677838c9889a48e 100644
--- a/roles/etcd/tasks/gen_nodes_certs_script.yml
+++ b/roles/etcd/tasks/gen_nodes_certs_script.yml
@@ -21,7 +21,7 @@
     executable: /bin/bash
   no_log: "{{ not (unsafe_show_logs | bool) }}"
   register: etcd_node_certs
-  check_mode: no
+  check_mode: false
   delegate_to: "{{ groups['etcd'][0] }}"
   changed_when: false
 
diff --git a/roles/etcd/tasks/install_docker.yml b/roles/etcd/tasks/install_docker.yml
index a7aba5094c589ca22c62a281699eb7fc188a5122..f393bd4eb661102419920e030b66529d199398b2 100644
--- a/roles/etcd/tasks/install_docker.yml
+++ b/roles/etcd/tasks/install_docker.yml
@@ -29,7 +29,7 @@
     dest: "{{ bin_dir }}/etcd"
     owner: 'root'
     mode: "0750"
-    backup: yes
+    backup: true
   when: etcd_cluster_setup
 
 - name: Install etcd-events launch script
@@ -38,5 +38,5 @@
     dest: "{{ bin_dir }}/etcd-events"
     owner: 'root'
     mode: "0750"
-    backup: yes
+    backup: true
   when: etcd_events_cluster_setup
diff --git a/roles/etcd/tasks/install_host.yml b/roles/etcd/tasks/install_host.yml
index 7bfc7e2ab021fda58c325340a9a480d57d58b2a0..eb67952ea3c4cdf6193b89af9f49d2fe94fb550f 100644
--- a/roles/etcd/tasks/install_host.yml
+++ b/roles/etcd/tasks/install_host.yml
@@ -25,7 +25,7 @@
     src: "{{ local_release_dir }}/etcd-{{ etcd_version }}-linux-{{ host_architecture }}/{{ item }}"
     dest: "{{ bin_dir }}/{{ item }}"
     mode: "0755"
-    remote_src: yes
+    remote_src: true
   with_items:
     - etcd
   when: etcd_cluster_setup
diff --git a/roles/etcd/tasks/join_etcd-events_member.yml b/roles/etcd/tasks/join_etcd-events_member.yml
index 0fad331e3891b613a0d006e6eca6a5655d3e3899..106f06e03368881498c5791fd6b108e9401f5bbb 100644
--- a/roles/etcd/tasks/join_etcd-events_member.yml
+++ b/roles/etcd/tasks/join_etcd-events_member.yml
@@ -32,7 +32,7 @@
     executable: /bin/bash
   register: etcd_events_member_in_cluster
   changed_when: false
-  check_mode: no
+  check_mode: false
   tags:
     - facts
   environment:
@@ -46,4 +46,4 @@
   service:
     name: etcd-events
     state: started
-    enabled: yes
+    enabled: true
diff --git a/roles/etcd/tasks/join_etcd_member.yml b/roles/etcd/tasks/join_etcd_member.yml
index ee77d4b2631fdb6f0226ed2090adff294ccd0e13..a2e37714d6896f26b021bcbc82abe8cc6aad80ef 100644
--- a/roles/etcd/tasks/join_etcd_member.yml
+++ b/roles/etcd/tasks/join_etcd_member.yml
@@ -33,7 +33,7 @@
     executable: /bin/bash
   register: etcd_member_in_cluster
   changed_when: false
-  check_mode: no
+  check_mode: false
   retries: "{{ etcd_retries }}"
   delay: "{{ retry_stagger | random + 3 }}"
   until: etcd_member_in_cluster.rc == 0
@@ -50,4 +50,4 @@
   service:
     name: etcd
     state: started
-    enabled: yes
+    enabled: true
diff --git a/roles/etcd/tasks/main.yml b/roles/etcd/tasks/main.yml
index 40ca3de5f12eeb6fe05c6293f2916d6c85af491a..74d5f16d3992c6644cb34ed09fb2f8e2060d6e6f 100644
--- a/roles/etcd/tasks/main.yml
+++ b/roles/etcd/tasks/main.yml
@@ -33,7 +33,7 @@
   command: "openssl x509 -in {{ etcd_cert_dir }}/node-{{ inventory_hostname }}.pem -noout -serial"
   register: "etcd_client_cert_serial_result"
   changed_when: false
-  check_mode: no
+  check_mode: false
   when:
     - kube_network_plugin in ["calico", "flannel", "cilium"] or cilium_deploy_additionally | default(false) | bool
     - kube_network_plugin != "calico" or calico_datastore == "etcd"
diff --git a/roles/etcdctl_etcdutl/tasks/main.yml b/roles/etcdctl_etcdutl/tasks/main.yml
index b9e6832f549841c2205107a01098f9dfbc65991c..053e14295f4e02df24a2ca1e0891d5c7399d7d5a 100644
--- a/roles/etcdctl_etcdutl/tasks/main.yml
+++ b/roles/etcdctl_etcdutl/tasks/main.yml
@@ -24,7 +24,7 @@
   unarchive:
     src: "{{ downloads.etcd.dest }}"
     dest: "{{ local_release_dir }}/"
-    remote_src: yes
+    remote_src: true
   when: container_manager in ['crio', 'containerd']
 
 - name: Copy etcdctl and etcdutl binary from download dir
@@ -32,7 +32,7 @@
     src: "{{ local_release_dir }}/etcd-{{ etcd_version }}-linux-{{ host_architecture }}/{{ item }}"
     dest: "{{ bin_dir }}/{{ item }}"
     mode: "0755"
-    remote_src: yes
+    remote_src: true
   with_items:
     - etcdctl
     - etcdutl
diff --git a/roles/kubernetes-apps/ansible/tasks/main.yml b/roles/kubernetes-apps/ansible/tasks/main.yml
index e3d82f10699dba962b9d6fbe32b29c3c85e52a44..18deee8053e6514e70693bf5aeaab990af5010fb 100644
--- a/roles/kubernetes-apps/ansible/tasks/main.yml
+++ b/roles/kubernetes-apps/ansible/tasks/main.yml
@@ -2,7 +2,7 @@
 - name: Kubernetes Apps | Wait for kube-apiserver
   uri:
     url: "{{ kube_apiserver_endpoint }}/healthz"
-    validate_certs: no
+    validate_certs: false
     client_cert: "{{ kube_apiserver_client_cert }}"
     client_key: "{{ kube_apiserver_client_key }}"
   register: result
diff --git a/roles/kubernetes-apps/argocd/tasks/main.yml b/roles/kubernetes-apps/argocd/tasks/main.yml
index 3cfe06fc7ba8c483c054f9e56e1b04ee9ffecb75..05c63337bed06fa528aa66a5fde661d2c03b18e3 100644
--- a/roles/kubernetes-apps/argocd/tasks/main.yml
+++ b/roles/kubernetes-apps/argocd/tasks/main.yml
@@ -8,10 +8,10 @@
   ansible.posix.synchronize:
     src: "{{ downloads.yq.dest }}"
     dest: "{{ bin_dir }}/yq"
-    compress: no
-    perms: yes
-    owner: no
-    group: no
+    compress: false
+    perms: true
+    owner: false
+    group: false
   delegate_to: "{{ inventory_hostname }}"
 
 - name: Kubernetes Apps | Set ArgoCD template list
@@ -49,17 +49,17 @@
   ansible.posix.synchronize:
     src: "{{ local_release_dir }}/{{ item.file }}"
     dest: "{{ kube_config_dir }}/{{ item.file }}"
-    compress: no
-    perms: yes
-    owner: no
-    group: no
+    compress: false
+    perms: true
+    owner: false
+    group: false
   delegate_to: "{{ inventory_hostname }}"
   with_items: "{{ argocd_templates | selectattr('url', 'defined') | list }}"
   when:
     - "inventory_hostname == groups['kube_control_plane'][0]"
 
 - name: Kubernetes Apps | Set ArgoCD namespace for remote manifests
-  become: yes
+  become: true
   command: |
     {{ bin_dir }}/yq eval-all -i '.metadata.namespace="{{ argocd_namespace }}"' {{ kube_config_dir }}/{{ item.file }}
   with_items: "{{ argocd_templates | selectattr('url', 'defined') | list }}"
@@ -69,7 +69,7 @@
     - "inventory_hostname == groups['kube_control_plane'][0]"
 
 - name: Kubernetes Apps | Create ArgoCD manifests from templates
-  become: yes
+  become: true
   template:
     src: "{{ item.file }}.j2"
     dest: "{{ kube_config_dir }}/{{ item.file }}"
@@ -81,7 +81,7 @@
     - "inventory_hostname == groups['kube_control_plane'][0]"
 
 - name: Kubernetes Apps | Install ArgoCD
-  become: yes
+  become: true
   kube:
     name: ArgoCD
     kubectl: "{{ bin_dir }}/kubectl"
@@ -93,7 +93,7 @@
 
 # https://github.com/argoproj/argo-cd/blob/master/docs/faq.md#i-forgot-the-admin-password-how-do-i-reset-it
 - name: Kubernetes Apps | Set ArgoCD custom admin password
-  become: yes
+  become: true
   shell: |
     {{ bin_dir }}/kubectl --kubeconfig /etc/kubernetes/admin.conf -n {{ argocd_namespace }} patch secret argocd-secret -p \
       '{
diff --git a/roles/kubernetes-apps/cluster_roles/tasks/main.yml b/roles/kubernetes-apps/cluster_roles/tasks/main.yml
index 8d7230e0af028517bfb717d99c56c5568a9d662c..ef4737eac4f7fe233d4f147b31fbbcde7915e0a9 100644
--- a/roles/kubernetes-apps/cluster_roles/tasks/main.yml
+++ b/roles/kubernetes-apps/cluster_roles/tasks/main.yml
@@ -2,7 +2,7 @@
 - name: Kubernetes Apps | Wait for kube-apiserver
   uri:
     url: "{{ kube_apiserver_endpoint }}/healthz"
-    validate_certs: no
+    validate_certs: false
     client_cert: "{{ kube_apiserver_client_cert }}"
     client_key: "{{ kube_apiserver_client_key }}"
   register: result
diff --git a/roles/kubernetes-apps/csi_driver/vsphere/defaults/main.yml b/roles/kubernetes-apps/csi_driver/vsphere/defaults/main.yml
index 0d4144141badca7e967fcfdb0a8513b797263d0b..325e3cb7d89497321154a972c09bff99beef78b8 100644
--- a/roles/kubernetes-apps/csi_driver/vsphere/defaults/main.yml
+++ b/roles/kubernetes-apps/csi_driver/vsphere/defaults/main.yml
@@ -21,7 +21,7 @@ vsphere_csi_controller_replicas: 1
 
 csi_endpoint: '{% if external_vsphere_version >= "7.0u1" %}/csi{% else %}/var/lib/csi/sockets/pluginproxy{% endif %}'
 
-vsphere_csi_aggressive_node_drain: False
+vsphere_csi_aggressive_node_drain: false
 vsphere_csi_aggressive_node_unreachable_timeout: 300
 vsphere_csi_aggressive_node_not_ready_timeout: 300
 
diff --git a/roles/kubernetes-apps/helm/tasks/main.yml b/roles/kubernetes-apps/helm/tasks/main.yml
index 61596aefb02858d969e019529c4a227b0d45d27c..5951381a2cac6ad7fa81de2d8474bc6345043e63 100644
--- a/roles/kubernetes-apps/helm/tasks/main.yml
+++ b/roles/kubernetes-apps/helm/tasks/main.yml
@@ -37,13 +37,13 @@
 
 - name: Helm | Get helm completion
   command: "{{ bin_dir }}/helm completion bash"
-  changed_when: False
+  changed_when: false
   register: helm_completion
-  check_mode: False
+  check_mode: false
 
 - name: Helm | Install helm completion
   copy:
     dest: /etc/bash_completion.d/helm.sh
     content: "{{ helm_completion.stdout }}"
     mode: "0755"
-  become: True
+  become: true
diff --git a/roles/kubernetes-apps/helm/tasks/pyyaml-flatcar.yml b/roles/kubernetes-apps/helm/tasks/pyyaml-flatcar.yml
index ea0d63a08d00d8ce9926b0f43f84e592d4653f05..72f0e2182aadaedce43cdf0709fef52e445a28a0 100644
--- a/roles/kubernetes-apps/helm/tasks/pyyaml-flatcar.yml
+++ b/roles/kubernetes-apps/helm/tasks/pyyaml-flatcar.yml
@@ -2,13 +2,13 @@
 - name: Get installed pip version
   command: "{{ ansible_python_interpreter if ansible_python_interpreter is defined else 'python' }} -m pip --version"
   register: pip_version_output
-  ignore_errors: yes
+  ignore_errors: true
   changed_when: false
 
 - name: Get installed PyYAML version
   command: "{{ ansible_python_interpreter if ansible_python_interpreter is defined else 'python' }} -m pip show PyYAML"
   register: pyyaml_version_output
-  ignore_errors: yes
+  ignore_errors: true
   changed_when: false
 
 - name: Install pip
diff --git a/roles/kubernetes-apps/krew/tasks/krew.yml b/roles/kubernetes-apps/krew/tasks/krew.yml
index e46dbb48dcc8f94af3f2727c2bda386d7c6b2ae9..3308aef85701ea3ae4e96ea62b1e7e16faa0dff8 100644
--- a/roles/kubernetes-apps/krew/tasks/krew.yml
+++ b/roles/kubernetes-apps/krew/tasks/krew.yml
@@ -24,15 +24,15 @@
 
 - name: Krew | Get krew completion
   command: "{{ local_release_dir }}/krew-{{ host_os }}_{{ image_arch }} completion bash"
-  changed_when: False
+  changed_when: false
   register: krew_completion
-  check_mode: False
-  ignore_errors: yes  # noqa ignore-errors
+  check_mode: false
+  ignore_errors: true  # noqa ignore-errors
 
 - name: Krew | Install krew completion
   copy:
     dest: /etc/bash_completion.d/krew.sh
     content: "{{ krew_completion.stdout }}"
     mode: "0755"
-  become: True
+  become: true
   when: krew_completion.rc == 0
diff --git a/roles/kubernetes-apps/network_plugin/weave/tasks/main.yml b/roles/kubernetes-apps/network_plugin/weave/tasks/main.yml
index bc0f932d8d66cc7f084b74af215a2117243b8a08..587b652a5bfdaf609583838cebc7d3c94e39e594 100644
--- a/roles/kubernetes-apps/network_plugin/weave/tasks/main.yml
+++ b/roles/kubernetes-apps/network_plugin/weave/tasks/main.yml
@@ -13,7 +13,7 @@
 - name: Weave | Wait for Weave to become available
   uri:
     url: http://127.0.0.1:6784/status
-    return_content: yes
+    return_content: true
   register: weave_status
   retries: 180
   delay: 5
diff --git a/roles/kubernetes/client/tasks/main.yml b/roles/kubernetes/client/tasks/main.yml
index cc788775048794e4ffe443c984c273c1df64caee..9202051981f5cf4943e77c848718b51f4ca5f51b 100644
--- a/roles/kubernetes/client/tasks/main.yml
+++ b/roles/kubernetes/client/tasks/main.yml
@@ -30,9 +30,9 @@
   copy:
     src: "{{ kube_config_dir }}/admin.conf"
     dest: "{{ ansible_env.HOME | default('/root') }}/.kube/config"
-    remote_src: yes
+    remote_src: true
     mode: "0600"
-    backup: yes
+    backup: true
 
 - name: Create kube artifacts dir
   file:
@@ -41,8 +41,8 @@
     state: directory
   delegate_to: localhost
   connection: local
-  become: no
-  run_once: yes
+  become: false
+  run_once: true
   when: kubeconfig_localhost
 
 - name: Wait for k8s apiserver
@@ -54,7 +54,7 @@
 - name: Get admin kubeconfig from remote host
   slurp:
     src: "{{ kube_config_dir }}/admin.conf"
-  run_once: yes
+  run_once: true
   register: raw_admin_kubeconfig
   when: kubeconfig_localhost
 
@@ -83,21 +83,21 @@
     mode: "0600"
   delegate_to: localhost
   connection: local
-  become: no
-  run_once: yes
+  become: false
+  run_once: true
   when: kubeconfig_localhost
 
 - name: Copy kubectl binary to ansible host
   fetch:
     src: "{{ bin_dir }}/kubectl"
     dest: "{{ artifacts_dir }}/kubectl"
-    flat: yes
-    validate_checksum: no
+    flat: true
+    validate_checksum: false
   register: copy_binary_result
   until: copy_binary_result is not failed
   retries: 20
-  become: no
-  run_once: yes
+  become: false
+  run_once: true
   when: kubectl_localhost
 
 - name: Create helper script kubectl.sh on ansible host
@@ -107,8 +107,8 @@
       ${BASH_SOURCE%/*}/kubectl --kubeconfig=${BASH_SOURCE%/*}/admin.conf "$@"
     dest: "{{ artifacts_dir }}/kubectl.sh"
     mode: "0755"
-  become: no
-  run_once: yes
+  become: false
+  run_once: true
   delegate_to: localhost
   connection: local
   when: kubectl_localhost and kubeconfig_localhost
diff --git a/roles/kubernetes/control-plane/handlers/main.yml b/roles/kubernetes/control-plane/handlers/main.yml
index be5fdffb125717b01f0a14fef20c83063117bfc5..3d7f3e074096ce1d6d09c45442b957496dda58fe 100644
--- a/roles/kubernetes/control-plane/handlers/main.yml
+++ b/roles/kubernetes/control-plane/handlers/main.yml
@@ -81,7 +81,7 @@
     endpoint: "{{ kube_scheduler_bind_address if kube_scheduler_bind_address != '0.0.0.0' else 'localhost' }}"
   uri:
     url: https://{{ endpoint }}:10259/healthz
-    validate_certs: no
+    validate_certs: false
   register: scheduler_result
   until: scheduler_result.status == 200
   retries: 60
@@ -95,7 +95,7 @@
     endpoint: "{{ kube_controller_manager_bind_address if kube_controller_manager_bind_address != '0.0.0.0' else 'localhost' }}"
   uri:
     url: https://{{ endpoint }}:10257/healthz
-    validate_certs: no
+    validate_certs: false
   register: controller_manager_result
   until: controller_manager_result.status == 200
   retries: 60
@@ -107,7 +107,7 @@
 - name: Master | wait for the apiserver to be running
   uri:
     url: "{{ kube_apiserver_endpoint }}/healthz"
-    validate_certs: no
+    validate_certs: false
   register: result
   until: result.status == 200
   retries: 60
diff --git a/roles/kubernetes/control-plane/tasks/define-first-kube-control.yml b/roles/kubernetes/control-plane/tasks/define-first-kube-control.yml
index ce5894d11b14622d0176142b95c8aea069f6075e..5faa184858fc5d6f749c5c381c996627b461c1d7 100644
--- a/roles/kubernetes/control-plane/tasks/define-first-kube-control.yml
+++ b/roles/kubernetes/control-plane/tasks/define-first-kube-control.yml
@@ -3,7 +3,7 @@
 - name: Check which kube-control nodes are already members of the cluster
   command: "{{ bin_dir }}/kubectl get nodes --selector=node-role.kubernetes.io/control-plane -o json"
   register: kube_control_planes_raw
-  ignore_errors: yes
+  ignore_errors: true
   changed_when: false
 
 - name: Set fact joined_control_planes
@@ -12,7 +12,7 @@
   delegate_to: "{{ item }}"
   loop: "{{ groups['kube_control_plane'] }}"
   when: kube_control_planes_raw is succeeded
-  run_once: yes
+  run_once: true
 
 - name: Set fact first_kube_control_plane
   set_fact:
diff --git a/roles/kubernetes/control-plane/tasks/encrypt-at-rest.yml b/roles/kubernetes/control-plane/tasks/encrypt-at-rest.yml
index 9b998c52bc7293657273973433fd0af11515e679..2950c76e27d8adb1a910647056ef36a9b2e22009 100644
--- a/roles/kubernetes/control-plane/tasks/encrypt-at-rest.yml
+++ b/roles/kubernetes/control-plane/tasks/encrypt-at-rest.yml
@@ -2,9 +2,9 @@
 - name: Check if secret for encrypting data at rest already exist
   stat:
     path: "{{ kube_cert_dir }}/secrets_encryption.yaml"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: secrets_encryption_file
 
 - name: Slurp secrets_encryption file if it exists
diff --git a/roles/kubernetes/control-plane/tasks/kubeadm-backup.yml b/roles/kubernetes/control-plane/tasks/kubeadm-backup.yml
index 36bb62798ecd7010c7aa538ab9a494e9b7232839..918f7cf47dadecf5635ad44e533ad5f210d8de7d 100644
--- a/roles/kubernetes/control-plane/tasks/kubeadm-backup.yml
+++ b/roles/kubernetes/control-plane/tasks/kubeadm-backup.yml
@@ -4,7 +4,7 @@
     src: "{{ kube_cert_dir }}/{{ item }}"
     dest: "{{ kube_cert_dir }}/{{ item }}.old"
     mode: preserve
-    remote_src: yes
+    remote_src: true
   with_items:
     - apiserver.crt
     - apiserver.key
@@ -19,7 +19,7 @@
     src: "{{ kube_config_dir }}/{{ item }}"
     dest: "{{ kube_config_dir }}/{{ item }}.old"
     mode: preserve
-    remote_src: yes
+    remote_src: true
   with_items:
     - admin.conf
     - controller-manager.conf
diff --git a/roles/kubernetes/control-plane/tasks/kubeadm-fix-apiserver.yml b/roles/kubernetes/control-plane/tasks/kubeadm-fix-apiserver.yml
index 5376aba81e68fec96edaab8ef1d9c317df1f488b..e47f571d339e96a3288941ff174ba7b12bd4ae9e 100644
--- a/roles/kubernetes/control-plane/tasks/kubeadm-fix-apiserver.yml
+++ b/roles/kubernetes/control-plane/tasks/kubeadm-fix-apiserver.yml
@@ -5,7 +5,7 @@
     dest: "{{ kube_config_dir }}/{{ item }}"
     regexp: '^    server: https'
     line: '    server: {{ kube_apiserver_endpoint }}'
-    backup: yes
+    backup: true
   with_items:
     - admin.conf
     - controller-manager.conf
diff --git a/roles/kubernetes/control-plane/tasks/kubeadm-secondary.yml b/roles/kubernetes/control-plane/tasks/kubeadm-secondary.yml
index 128e93f3662293c985060884b18a3e7ac6631af9..413d4946cb86c5526ce205e3fe8249713858e669 100644
--- a/roles/kubernetes/control-plane/tasks/kubeadm-secondary.yml
+++ b/roles/kubernetes/control-plane/tasks/kubeadm-secondary.yml
@@ -25,7 +25,7 @@
 - name: Parse certificate key if not set
   set_fact:
     kubeadm_certificate_key: "{{ hostvars[groups['kube_control_plane'][0]]['kubeadm_upload_cert'].stdout_lines[-1] | trim }}"
-  run_once: yes
+  run_once: true
   when:
     - hostvars[groups['kube_control_plane'][0]]['kubeadm_upload_cert'] is defined
     - hostvars[groups['kube_control_plane'][0]]['kubeadm_upload_cert'] is not skipped
@@ -35,7 +35,7 @@
     src: "kubeadm-controlplane.{{ kubeadmConfig_api_version }}.yaml.j2"
     dest: "{{ kube_config_dir }}/kubeadm-controlplane.yaml"
     mode: "0640"
-    backup: yes
+    backup: true
   when:
     - inventory_hostname != first_kube_control_plane
     - not kubeadm_already_run.stat.exists
diff --git a/roles/kubernetes/control-plane/tasks/kubeadm-setup.yml b/roles/kubernetes/control-plane/tasks/kubeadm-setup.yml
index dfbe604a4c87651335f99a168f82bc6be4a1f2e8..52700af2e5927b0e87ae68453f33f0303a3fb8f5 100644
--- a/roles/kubernetes/control-plane/tasks/kubeadm-setup.yml
+++ b/roles/kubernetes/control-plane/tasks/kubeadm-setup.yml
@@ -13,9 +13,9 @@
 - name: Kubeadm | Check if kubeadm has already run
   stat:
     path: "/var/lib/kubelet/config.yaml"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kubeadm_already_run
 
 - name: Kubeadm | Backup kubeadm certs / kubeconfig
diff --git a/roles/kubernetes/control-plane/tasks/kubelet-fix-client-cert-rotation.yml b/roles/kubernetes/control-plane/tasks/kubelet-fix-client-cert-rotation.yml
index 7d0c1a0d59e88f0bbd5659006cf4b675c7ae2a97..409ecb043a089a5619818c12c9ac4cdbc40afc6d 100644
--- a/roles/kubernetes/control-plane/tasks/kubelet-fix-client-cert-rotation.yml
+++ b/roles/kubernetes/control-plane/tasks/kubelet-fix-client-cert-rotation.yml
@@ -4,7 +4,7 @@
     path: "{{ kube_config_dir }}/kubelet.conf"
     regexp: '^    client-certificate-data: '
     line: '    client-certificate: /var/lib/kubelet/pki/kubelet-client-current.pem'
-    backup: yes
+    backup: true
   notify:
     - "Master | reload kubelet"
 
@@ -13,6 +13,6 @@
     path: "{{ kube_config_dir }}/kubelet.conf"
     regexp: '^    client-key-data: '
     line: '    client-key: /var/lib/kubelet/pki/kubelet-client-current.pem'
-    backup: yes
+    backup: true
   notify:
     - "Master | reload kubelet"
diff --git a/roles/kubernetes/control-plane/tasks/main.yml b/roles/kubernetes/control-plane/tasks/main.yml
index 5d58014e80ba5bfc7376f2fd23109bb30c790bb9..518bac9613195b7433cc961aafeb17485ac0ab60 100644
--- a/roles/kubernetes/control-plane/tasks/main.yml
+++ b/roles/kubernetes/control-plane/tasks/main.yml
@@ -120,7 +120,7 @@
 - name: Renew K8S control plane certificates monthly 2/2
   systemd_service:
     name: k8s-certs-renew.timer
-    enabled: yes
+    enabled: true
     state: started
     daemon_reload: "{{ k8s_certs_units is changed }}"
   when: auto_renew_certificates
diff --git a/roles/kubernetes/kubeadm/tasks/main.yml b/roles/kubernetes/kubeadm/tasks/main.yml
index 2b5778726aaa8da8c1246d64e5b599a4ae323cad..ad6ac36beafe7592517163f5ae349dcc3e001483 100644
--- a/roles/kubernetes/kubeadm/tasks/main.yml
+++ b/roles/kubernetes/kubeadm/tasks/main.yml
@@ -14,17 +14,17 @@
 - name: Check if kubelet.conf exists
   stat:
     path: "{{ kube_config_dir }}/kubelet.conf"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kubelet_conf
 
 - name: Check if kubeadm CA cert is accessible
   stat:
     path: "{{ kube_cert_dir }}/ca.crt"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kubeadm_ca_stat
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
   run_once: true
@@ -79,7 +79,7 @@
   template:
     src: "kubeadm-client.conf.{{ kubeadmConfig_api_version }}.j2"
     dest: "{{ kube_config_dir }}/kubeadm-client.conf"
-    backup: yes
+    backup: true
     mode: "0640"
   when: not is_kube_master
 
@@ -140,7 +140,7 @@
     dest: "{{ kube_config_dir }}/kubelet.conf"
     regexp: 'server:'
     line: '    server: {{ kube_apiserver_endpoint }}'
-    backup: yes
+    backup: true
   when:
     - kubeadm_config_api_fqdn is not defined
     - not is_kube_master
@@ -152,7 +152,7 @@
     dest: "{{ kube_config_dir }}/kubelet.conf"
     regexp: '^    server: https'
     line: '    server: {{ kube_apiserver_endpoint }}'
-    backup: yes
+    backup: true
   when:
     - not is_kube_master
     - loadbalancer_apiserver is defined
diff --git a/roles/kubernetes/node-label/tasks/main.yml b/roles/kubernetes/node-label/tasks/main.yml
index 00e87504c6f2a26ded37e0fca757b002a3dc8cba..3ebb64594762f898d9d5ed58f1468d87e73cd2d0 100644
--- a/roles/kubernetes/node-label/tasks/main.yml
+++ b/roles/kubernetes/node-label/tasks/main.yml
@@ -2,7 +2,7 @@
 - name: Kubernetes Apps | Wait for kube-apiserver
   uri:
     url: "{{ kube_apiserver_endpoint }}/healthz"
-    validate_certs: no
+    validate_certs: false
     client_cert: "{{ kube_apiserver_client_cert }}"
     client_key: "{{ kube_apiserver_client_key }}"
   register: result
diff --git a/roles/kubernetes/node/tasks/facts.yml b/roles/kubernetes/node/tasks/facts.yml
index 0aaa11d60a78c7d7c2baa2482c155404cba5be2d..6e8995274cb16a3abded773dd473555616edbf7e 100644
--- a/roles/kubernetes/node/tasks/facts.yml
+++ b/roles/kubernetes/node/tasks/facts.yml
@@ -8,7 +8,7 @@
       executable: /bin/bash
     register: docker_cgroup_driver_result
     changed_when: false
-    check_mode: no
+    check_mode: false
 
   - name: Set kubelet_cgroup_driver_detected fact for docker
     set_fact:
diff --git a/roles/kubernetes/node/tasks/kubelet.yml b/roles/kubernetes/node/tasks/kubelet.yml
index b63aefe1f18f643417df5b5e1f2ec508f87cc327..1f27bd072d18b2eb828374c52f40164901b55411 100644
--- a/roles/kubernetes/node/tasks/kubelet.yml
+++ b/roles/kubernetes/node/tasks/kubelet.yml
@@ -11,7 +11,7 @@
     src: "kubelet.env.{{ kubeletConfig_api_version }}.j2"
     dest: "{{ kube_config_dir }}/kubelet.env"
     setype: "{{ (preinstall_selinux_state != 'disabled') | ternary('etc_t', omit) }}"
-    backup: yes
+    backup: true
     mode: "0600"
   notify: Node | restart kubelet
   tags:
@@ -32,7 +32,7 @@
   template:
     src: "kubelet.service.j2"
     dest: "/etc/systemd/system/kubelet.service"
-    backup: "yes"
+    backup: true
     mode: "0600"
     validate: "sh -c '[ -f /usr/bin/systemd/system/factory-reset.target ] || exit 0 && systemd-analyze verify %s:kubelet.service'"
     # FIXME: check that systemd version >= 250 (factory-reset.target was introduced in that release)
@@ -48,7 +48,7 @@
 - name: Enable kubelet
   service:
     name: kubelet
-    enabled: yes
+    enabled: true
     state: started
   tags:
     - kubelet
diff --git a/roles/kubernetes/node/tasks/loadbalancer/haproxy.yml b/roles/kubernetes/node/tasks/loadbalancer/haproxy.yml
index 2d3454e5a2f4df28efccd13e4c73061a4e398f7a..b4c58126e481fc66805df723e66f4ecfcbfa06cd 100644
--- a/roles/kubernetes/node/tasks/loadbalancer/haproxy.yml
+++ b/roles/kubernetes/node/tasks/loadbalancer/haproxy.yml
@@ -17,14 +17,14 @@
     dest: "{{ haproxy_config_dir }}/haproxy.cfg"
     owner: root
     mode: "0755"
-    backup: yes
+    backup: true
 
 - name: Haproxy | Get checksum from config
   stat:
     path: "{{ haproxy_config_dir }}/haproxy.cfg"
-    get_attributes: no
-    get_checksum: yes
-    get_mime: no
+    get_attributes: false
+    get_checksum: true
+    get_mime: false
   register: haproxy_stat
 
 - name: Haproxy | Write static pod
diff --git a/roles/kubernetes/node/tasks/loadbalancer/kube-vip.yml b/roles/kubernetes/node/tasks/loadbalancer/kube-vip.yml
index b210cfe16364033d0aba7bb51ac4edd86c126d0c..b52261a62c0bc3737c7b31b8c3d4a0a5993503e2 100644
--- a/roles/kubernetes/node/tasks/loadbalancer/kube-vip.yml
+++ b/roles/kubernetes/node/tasks/loadbalancer/kube-vip.yml
@@ -16,9 +16,9 @@
 - name: Kube-vip | Check if kubeadm has already run
   stat:
     path: "/var/lib/kubelet/config.yaml"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kubeadm_already_run
 
 - name: Kube-vip | Set admin.conf
diff --git a/roles/kubernetes/node/tasks/loadbalancer/nginx-proxy.yml b/roles/kubernetes/node/tasks/loadbalancer/nginx-proxy.yml
index aeeacc80d15faa27d4e9d62ff620ce6b445ee6dc..66ebe55e0cb4e5419c97dbd470339c9d02bdc732 100644
--- a/roles/kubernetes/node/tasks/loadbalancer/nginx-proxy.yml
+++ b/roles/kubernetes/node/tasks/loadbalancer/nginx-proxy.yml
@@ -17,14 +17,14 @@
     dest: "{{ nginx_config_dir }}/nginx.conf"
     owner: root
     mode: "0755"
-    backup: yes
+    backup: true
 
 - name: Nginx-proxy | Get checksum from config
   stat:
     path: "{{ nginx_config_dir }}/nginx.conf"
-    get_attributes: no
-    get_checksum: yes
-    get_mime: no
+    get_attributes: false
+    get_checksum: true
+    get_mime: false
   register: nginx_stat
 
 - name: Nginx-proxy | Write static pod
diff --git a/roles/kubernetes/node/tasks/main.yml b/roles/kubernetes/node/tasks/main.yml
index 7dc2114057e2dce400f64f1fe5e700fe49de41c3..56117bc3a99d08ec2972820414d115f2679a5a25 100644
--- a/roles/kubernetes/node/tasks/main.yml
+++ b/roles/kubernetes/node/tasks/main.yml
@@ -51,10 +51,10 @@
   ansible.posix.sysctl:
     name: net.ipv4.ip_local_reserved_ports
     value: "{{ kube_apiserver_node_port_range }}"
-    sysctl_set: yes
+    sysctl_set: true
     sysctl_file: "{{ sysctl_file_path }}"
     state: present
-    reload: yes
+    reload: true
   when: kube_apiserver_node_port_range is defined
   tags:
     - kube-proxy
@@ -66,7 +66,7 @@
   register: modinfo_br_netfilter
   failed_when: modinfo_br_netfilter.rc not in [0, 1]
   changed_when: false
-  check_mode: no
+  check_mode: false
 
 # TODO: Remove once upstream issue is fixed
 # https://github.com/ansible-collections/community.general/issues/7717
@@ -97,7 +97,7 @@
   command: "sysctl net.bridge.bridge-nf-call-iptables"
   failed_when: false
   changed_when: false
-  check_mode: no
+  check_mode: false
   register: sysctl_bridge_nf_call_iptables
 
 - name: Enable bridge-nf-call tables
@@ -106,7 +106,7 @@
     state: present
     sysctl_file: "{{ sysctl_file_path }}"
     value: "1"
-    reload: yes
+    reload: true
   when: sysctl_bridge_nf_call_iptables.rc == 0
   with_items:
     - net.bridge.bridge-nf-call-iptables
diff --git a/roles/kubernetes/node/tasks/pre_upgrade.yml b/roles/kubernetes/node/tasks/pre_upgrade.yml
index d9c2d07ef17e2efcd5da83816b9cdafbf4c4a2c9..e4bbf6b747b6aef4834f3bfaf95ff86a59c4f16b 100644
--- a/roles/kubernetes/node/tasks/pre_upgrade.yml
+++ b/roles/kubernetes/node/tasks/pre_upgrade.yml
@@ -11,7 +11,7 @@
     executable: /bin/bash
   failed_when: false
   changed_when: false
-  check_mode: no
+  check_mode: false
   register: kubelet_container_check
 
 - name: "Pre-upgrade | copy /var/lib/cni from kubelet"
diff --git a/roles/kubernetes/preinstall/handlers/main.yml b/roles/kubernetes/preinstall/handlers/main.yml
index 35140ab42fc5b3f08bd074f9fb971e025388b8f4..cc69fe42c8dd07d6bab4fd3147e3a4dc349d40bd 100644
--- a/roles/kubernetes/preinstall/handlers/main.yml
+++ b/roles/kubernetes/preinstall/handlers/main.yml
@@ -31,9 +31,9 @@
 - name: Preinstall | kube-apiserver configured
   stat:
     path: "{{ kube_manifest_dir }}/kube-apiserver.yaml"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kube_apiserver_set
   when: inventory_hostname in groups['kube_control_plane'] and dns_mode != 'none' and resolvconf_mode == 'host_resolvconf'
   listen: Preinstall | propagate resolvconf to k8s components
@@ -42,9 +42,9 @@
 - name: Preinstall | kube-controller configured
   stat:
     path: "{{ kube_manifest_dir }}/kube-controller-manager.yaml"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kube_controller_set
   when: inventory_hostname in groups['kube_control_plane'] and dns_mode != 'none' and resolvconf_mode == 'host_resolvconf'
   listen: Preinstall | propagate resolvconf to k8s components
@@ -109,7 +109,7 @@
 - name: Preinstall | wait for the apiserver to be running
   uri:
     url: "{{ kube_apiserver_endpoint }}/healthz"
-    validate_certs: no
+    validate_certs: false
   register: result
   until: result.status == 200
   retries: 60
diff --git a/roles/kubernetes/preinstall/tasks/0010-swapoff.yml b/roles/kubernetes/preinstall/tasks/0010-swapoff.yml
index 45474c844d48d3ce1ed891f22ce909caa8b205fa..76d95d11d900a98e34ded549bbb4496729bf8d2d 100644
--- a/roles/kubernetes/preinstall/tasks/0010-swapoff.yml
+++ b/roles/kubernetes/preinstall/tasks/0010-swapoff.yml
@@ -2,9 +2,9 @@
 - name: Check if /etc/fstab exists
   stat:
     path: "/etc/fstab"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: fstab_file
 
 - name: Remove swapfile from /etc/fstab
diff --git a/roles/kubernetes/preinstall/tasks/0020-set_facts.yml b/roles/kubernetes/preinstall/tasks/0020-set_facts.yml
index 18ad6569468ff1a86d8bfdc08308dd0f3354d7d5..263bca400a1e24e04c610064c12553b0748cd474 100644
--- a/roles/kubernetes/preinstall/tasks/0020-set_facts.yml
+++ b/roles/kubernetes/preinstall/tasks/0020-set_facts.yml
@@ -12,24 +12,24 @@
   register: resolvconf
   failed_when: false
   changed_when: false
-  check_mode: no
+  check_mode: false
 
 - name: Check existence of /etc/resolvconf/resolv.conf.d
   stat:
     path: /etc/resolvconf/resolv.conf.d
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   failed_when: false
   register: resolvconfd_path
 
 - name: Check status of /etc/resolv.conf
   stat:
     path: /etc/resolv.conf
-    follow: no
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    follow: false
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   failed_when: false
   register: resolvconf_stat
 
@@ -72,7 +72,7 @@
   register: systemd_resolved_enabled
   failed_when: false
   changed_when: false
-  check_mode: no
+  check_mode: false
 
 - name: Set default dns if remove_default_searchdomains is false
   set_fact:
@@ -94,9 +94,9 @@
 - name: Check if kubelet is configured
   stat:
     path: "{{ kube_config_dir }}/kubelet.env"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kubelet_configured
   changed_when: false
 
@@ -121,9 +121,9 @@
 - name: Check if /etc/dhclient.conf exists
   stat:
     path: /etc/dhclient.conf
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: dhclient_stat
 
 - name: Target dhclient conf file for /etc/dhclient.conf
@@ -134,9 +134,9 @@
 - name: Check if /etc/dhcp/dhclient.conf exists
   stat:
     path: /etc/dhcp/dhclient.conf
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: dhcp_dhclient_stat
 
 - name: Target dhclient conf file for /etc/dhcp/dhclient.conf
@@ -218,9 +218,9 @@
 - name: Check /usr readonly
   stat:
     path: "/usr"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: usr
 
 - name: Set alternate flexvolume path
diff --git a/roles/kubernetes/preinstall/tasks/0040-verify-settings.yml b/roles/kubernetes/preinstall/tasks/0040-verify-settings.yml
index 1bb0f4856f496207799449f02e1741d1d6c8bfcb..d609ab574c8b14dcb59dc3215d1b7729fa00c2b0 100644
--- a/roles/kubernetes/preinstall/tasks/0040-verify-settings.yml
+++ b/roles/kubernetes/preinstall/tasks/0040-verify-settings.yml
@@ -44,7 +44,7 @@
   assert:
     that: item.value | type_debug == 'bool'
     msg: "{{ item.value }} isn't a bool"
-  run_once: yes
+  run_once: true
   with_items:
     - { name: download_run_once, value: "{{ download_run_once }}" }
     - { name: deploy_netchecker, value: "{{ deploy_netchecker }}" }
@@ -172,21 +172,21 @@
     that:
       - kube_service_addresses | ansible.utils.ipaddr('net')
     msg: "kube_service_addresses = '{{ kube_service_addresses }}' is not a valid network range"
-  run_once: yes
+  run_once: true
 
 - name: "Check that kube_pods_subnet is a network range"
   assert:
     that:
       - kube_pods_subnet | ansible.utils.ipaddr('net')
     msg: "kube_pods_subnet = '{{ kube_pods_subnet }}' is not a valid network range"
-  run_once: yes
+  run_once: true
 
 - name: "Check that kube_pods_subnet does not collide with kube_service_addresses"
   assert:
     that:
       - kube_pods_subnet | ansible.utils.ipaddr(kube_service_addresses) | string == 'None'
     msg: "kube_pods_subnet cannot be the same network segment as kube_service_addresses"
-  run_once: yes
+  run_once: true
 
 - name: "Check that IP range is enough for the nodes"
   assert:
@@ -194,7 +194,7 @@
       - 2 ** (kube_network_node_prefix - kube_pods_subnet | ansible.utils.ipaddr('prefix')) >= groups['k8s_cluster'] | length
     msg: "Not enough IPs are available for the desired node count."
   when: kube_network_plugin != 'calico'
-  run_once: yes
+  run_once: true
 
 - name: Stop if unknown dns mode
   assert:
@@ -246,7 +246,7 @@
 
 # TODO: Clean this task up when we drop backward compatibility support for `etcd_kubeadm_enabled`
 - name: Stop if etcd deployment type is not host or kubeadm when container_manager != docker and etcd_kubeadm_enabled is not defined
-  run_once: yes
+  run_once: true
   when: etcd_kubeadm_enabled is defined
   block:
     - name: Warn the user if they are still using `etcd_kubeadm_enabled`
@@ -292,7 +292,7 @@
   assert:
     that: containerd_version is version(containerd_min_version_required, '>=')
     msg: "containerd_version is too low. Minimum version {{ containerd_min_version_required }}"
-  run_once: yes
+  run_once: true
   when:
     - containerd_version not in ['latest', 'edge', 'stable']
     - container_manager == 'containerd'
diff --git a/roles/kubernetes/preinstall/tasks/0050-create_directories.yml b/roles/kubernetes/preinstall/tasks/0050-create_directories.yml
index 2fff8ef560892c7694bd1008450d6b9ef4e0b0c4..507a72d7817a0203db0306c2e74620a0e3b2946e 100644
--- a/roles/kubernetes/preinstall/tasks/0050-create_directories.yml
+++ b/roles/kubernetes/preinstall/tasks/0050-create_directories.yml
@@ -48,9 +48,9 @@
 - name: Check if kubernetes kubeadm compat cert dir exists
   stat:
     path: "{{ kube_cert_compat_dir }}"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kube_cert_compat_dir_check
   when:
     - inventory_hostname in groups['k8s_cluster']
diff --git a/roles/kubernetes/preinstall/tasks/0060-resolvconf.yml b/roles/kubernetes/preinstall/tasks/0060-resolvconf.yml
index 6219161fa4f2a8c68d2e28d9e3843d6f5761acf7..9aad0dba8e3c4b79eb84f2cc5f04d5f4b4239307 100644
--- a/roles/kubernetes/preinstall/tasks/0060-resolvconf.yml
+++ b/roles/kubernetes/preinstall/tasks/0060-resolvconf.yml
@@ -16,7 +16,7 @@
       options ndots:{{ ndots }} timeout:{{ dns_timeout | default('2') }} attempts:{{ dns_attempts | default('2') }}
     state: present
     insertbefore: BOF
-    create: yes
+    create: true
     backup: "{{ not resolvconf_stat.stat.islnk }}"
     marker: "# Ansible entries {mark}"
     mode: "0644"
diff --git a/roles/kubernetes/preinstall/tasks/0062-networkmanager-unmanaged-devices.yml b/roles/kubernetes/preinstall/tasks/0062-networkmanager-unmanaged-devices.yml
index ca51e88b910808607941b15d5a3af75922546ba7..6ebed25535ab03e8e26a0baba0e8e38c42cd7904 100644
--- a/roles/kubernetes/preinstall/tasks/0062-networkmanager-unmanaged-devices.yml
+++ b/roles/kubernetes/preinstall/tasks/0062-networkmanager-unmanaged-devices.yml
@@ -3,7 +3,7 @@
   file:
     path: "/etc/NetworkManager/conf.d"
     state: directory
-    recurse: yes
+    recurse: true
 
 - name: NetworkManager | Prevent NetworkManager from managing Calico interfaces (cali*/tunl*/vxlan.calico)
   copy:
diff --git a/roles/kubernetes/preinstall/tasks/0063-networkmanager-dns.yml b/roles/kubernetes/preinstall/tasks/0063-networkmanager-dns.yml
index e155f0a180541c9ad7b29427ba8806ffa049b4aa..6dfa7242643d793a25a432b3711abd891d82503b 100644
--- a/roles/kubernetes/preinstall/tasks/0063-networkmanager-dns.yml
+++ b/roles/kubernetes/preinstall/tasks/0063-networkmanager-dns.yml
@@ -6,7 +6,7 @@
     option: servers
     value: "{{ nameserverentries }}"
     mode: '0600'
-    backup: yes
+    backup: true
   when:
     - nameserverentries != "127.0.0.53" or systemd_resolved_enabled.rc != 0
   notify: Preinstall | update resolvconf for networkmanager
@@ -23,7 +23,7 @@
     option: searches
     value: "{{ (default_searchdomains | default([]) + searchdomains | default([])) | join(',') }}"
     mode: '0600'
-    backup: yes
+    backup: true
   notify: Preinstall | update resolvconf for networkmanager
 
 - name: NetworkManager | Add DNS options to NM configuration
@@ -33,5 +33,5 @@
     option: options
     value: "ndots:{{ ndots }},timeout:{{ dns_timeout | default('2') }},attempts:{{ dns_attempts | default('2') }}"
     mode: '0600'
-    backup: yes
+    backup: true
   notify: Preinstall | update resolvconf for networkmanager
diff --git a/roles/kubernetes/preinstall/tasks/0070-system-packages.yml b/roles/kubernetes/preinstall/tasks/0070-system-packages.yml
index cddbe1ecfec725626839500c44beee266bb16aec..c8b480c84494e0d913e4bf7bba439812ffd5166b 100644
--- a/roles/kubernetes/preinstall/tasks/0070-system-packages.yml
+++ b/roles/kubernetes/preinstall/tasks/0070-system-packages.yml
@@ -34,7 +34,7 @@
 
 - name: Update package management cache (APT)
   apt:
-    update_cache: yes
+    update_cache: true
     cache_valid_time: 3600
   when: ansible_os_family == "Debian"
   tags:
diff --git a/roles/kubernetes/preinstall/tasks/0080-system-configurations.yml b/roles/kubernetes/preinstall/tasks/0080-system-configurations.yml
index 5b2c7d10af93661d0ff3b4c8383262e7a282b352..8941a649a4c25d2c01b6cbb3eaab0c31482f6516 100644
--- a/roles/kubernetes/preinstall/tasks/0080-system-configurations.yml
+++ b/roles/kubernetes/preinstall/tasks/0080-system-configurations.yml
@@ -3,9 +3,9 @@
 - name: Confirm selinux deployed
   stat:
     path: /etc/selinux/config
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   when:
     - ansible_os_family == "RedHat"
     - "'Amazon' not in ansible_distribution"
@@ -27,8 +27,8 @@
     dest: /etc/gai.conf
     line: "precedence ::ffff:0:0/96  100"
     state: present
-    create: yes
-    backup: yes
+    create: true
+    backup: true
     mode: "0644"
   when:
     - disable_ipv6_dns
@@ -47,9 +47,9 @@
 - name: Stat sysctl file configuration
   stat:
     path: "{{ sysctl_file_path }}"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: sysctl_file_stat
   tags:
     - bootstrap-os
@@ -75,7 +75,7 @@
     name: net.ipv4.ip_forward
     value: "1"
     state: present
-    reload: yes
+    reload: true
 
 - name: Enable ipv6 forwarding
   ansible.posix.sysctl:
@@ -83,15 +83,15 @@
     name: net.ipv6.conf.all.forwarding
     value: "1"
     state: present
-    reload: yes
+    reload: true
   when: enable_dual_stack_networks | bool
 
 - name: Check if we need to set fs.may_detach_mounts
   stat:
     path: /proc/sys/fs/may_detach_mounts
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: fs_may_detach_mounts
   ignore_errors: true  # noqa ignore-errors
 
@@ -101,7 +101,7 @@
     name: fs.may_detach_mounts
     value: 1
     state: present
-    reload: yes
+    reload: true
   when: fs_may_detach_mounts.stat.exists | d(false)
 
 - name: Ensure kubelet expected parameters are set
@@ -110,7 +110,7 @@
     name: "{{ item.name }}"
     value: "{{ item.value }}"
     state: present
-    reload: yes
+    reload: true
   with_items:
     - { name: kernel.keys.root_maxbytes, value: 25000000 }
     - { name: kernel.keys.root_maxkeys, value: 1000000 }
@@ -133,7 +133,7 @@
     name: "{{ item.name }}"
     value: "{{ item.value }}"
     state: present
-    reload: yes
+    reload: true
   with_items: "{{ additional_sysctl }}"
 
 - name: Disable fapolicyd service
diff --git a/roles/kubernetes/preinstall/tasks/0090-etchosts.yml b/roles/kubernetes/preinstall/tasks/0090-etchosts.yml
index 4ec9a69e6e2c45d657d9931bb3d7b1b884f36085..0b44d26adc6a4cff7343b80772c58e4ae2b59959 100644
--- a/roles/kubernetes/preinstall/tasks/0090-etchosts.yml
+++ b/roles/kubernetes/preinstall/tasks/0090-etchosts.yml
@@ -11,17 +11,17 @@
       {% endfor %}
   delegate_to: localhost
   connection: local
-  delegate_facts: yes
-  run_once: yes
+  delegate_facts: true
+  run_once: true
 
 - name: Hosts | populate inventory into hosts file
   blockinfile:
     path: /etc/hosts
     block: "{{ hostvars.localhost.etc_hosts_inventory_block }}"
     state: "{{ 'present' if populate_inventory_to_hosts_file else 'absent' }}"
-    create: yes
-    backup: yes
-    unsafe_writes: yes
+    create: true
+    backup: true
+    unsafe_writes: true
     marker: "# Ansible inventory hosts {mark}"
     mode: "0644"
 
@@ -31,8 +31,8 @@
     regexp: ".*{{ apiserver_loadbalancer_domain_name }}$"
     line: "{{ loadbalancer_apiserver.address }} {{ apiserver_loadbalancer_domain_name }}"
     state: present
-    backup: yes
-    unsafe_writes: yes
+    backup: true
+    unsafe_writes: true
   when:
     - populate_loadbalancer_apiserver_to_hosts_file
     - loadbalancer_apiserver is defined
@@ -69,8 +69,8 @@
         line: "{{ item.key }} {{ item.value | join(' ') }}"
         regexp: "^{{ item.key }}.*$"
         state: present
-        backup: yes
-        unsafe_writes: yes
+        backup: true
+        unsafe_writes: true
       loop: "{{ etc_hosts_localhosts_dict_target | default({}) | dict2items }}"
 
 # gather facts to update ansible_fqdn
diff --git a/roles/kubernetes/preinstall/tasks/0100-dhclient-hooks.yml b/roles/kubernetes/preinstall/tasks/0100-dhclient-hooks.yml
index 6276034d3d73d04fec0cb7e7909b871d8f6c6fa0..9745ab261bc83e927c3691ca00f2a0f05353f00c 100644
--- a/roles/kubernetes/preinstall/tasks/0100-dhclient-hooks.yml
+++ b/roles/kubernetes/preinstall/tasks/0100-dhclient-hooks.yml
@@ -6,10 +6,10 @@
       {{ item }}
       {% endfor %}
     path: "{{ dhclientconffile }}"
-    create: yes
+    create: true
     state: present
     insertbefore: BOF
-    backup: yes
+    backup: true
     marker: "# Ansible entries {mark}"
     mode: "0644"
   notify: Preinstall | propagate resolvconf to k8s components
diff --git a/roles/kubernetes/preinstall/tasks/0110-dhclient-hooks-undo.yml b/roles/kubernetes/preinstall/tasks/0110-dhclient-hooks-undo.yml
index 024e39f9fd44371a04cd2544ec869895760c2e40..dd320d50a96e53839a35b5b11be8d16bc86db268 100644
--- a/roles/kubernetes/preinstall/tasks/0110-dhclient-hooks-undo.yml
+++ b/roles/kubernetes/preinstall/tasks/0110-dhclient-hooks-undo.yml
@@ -7,7 +7,7 @@
   blockinfile:
     path: "{{ dhclientconffile }}"
     state: absent
-    backup: yes
+    backup: true
     marker: "# Ansible entries {mark}"
   notify: Preinstall | propagate resolvconf to k8s components
 
diff --git a/roles/kubernetes/preinstall/tasks/0120-growpart-azure-centos-7.yml b/roles/kubernetes/preinstall/tasks/0120-growpart-azure-centos-7.yml
index 621629f6adefed11b905787c837fc6f6d3089a35..b9c35875e03738367b9965a5fca7391179c1f149 100644
--- a/roles/kubernetes/preinstall/tasks/0120-growpart-azure-centos-7.yml
+++ b/roles/kubernetes/preinstall/tasks/0120-growpart-azure-centos-7.yml
@@ -22,7 +22,7 @@
 
 - name: Check if growpart needs to be run
   command: growpart -N {{ device }} {{ partition }}
-  failed_when: False
+  failed_when: false
   changed_when: "'NOCHANGE:' not in growpart_needed.stdout"
   register: growpart_needed
   environment:
@@ -30,7 +30,7 @@
 
 - name: Check fs type
   command: file -Ls {{ root_device }}
-  changed_when: False
+  changed_when: false
   register: fs_type
 
 - name: Run growpart  # noqa no-handler
diff --git a/roles/kubernetes/preinstall/tasks/main.yml b/roles/kubernetes/preinstall/tasks/main.yml
index ee846f8ba02c8cbf80160100a6224270e013cbf2..722beecd3a85b2185e5320bb0c72afcd530b3d16 100644
--- a/roles/kubernetes/preinstall/tasks/main.yml
+++ b/roles/kubernetes/preinstall/tasks/main.yml
@@ -121,9 +121,9 @@
 - name: Check if we are running inside a Azure VM
   stat:
     path: /var/lib/waagent/
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: azure_check
   when:
     - not dns_late
diff --git a/roles/kubernetes/tokens/tasks/check-tokens.yml b/roles/kubernetes/tokens/tasks/check-tokens.yml
index a157a0597ee3f61d9dd41521577474511dd24b01..d8bb203e91ce5348b095b5a1eff63499feceb250 100644
--- a/roles/kubernetes/tokens/tasks/check-tokens.yml
+++ b/roles/kubernetes/tokens/tasks/check-tokens.yml
@@ -2,9 +2,9 @@
 - name: "Check_tokens | check if the tokens have already been generated on first master"
   stat:
     path: "{{ kube_token_dir }}/known_tokens.csv"
-    get_attributes: no
-    get_checksum: yes
-    get_mime: no
+    get_attributes: false
+    get_checksum: true
+    get_mime: false
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
   register: known_tokens_master
   run_once: true
@@ -23,9 +23,9 @@
 - name: "Check tokens | check if a cert already exists"
   stat:
     path: "{{ kube_token_dir }}/known_tokens.csv"
-    get_attributes: no
-    get_checksum: yes
-    get_mime: no
+    get_attributes: false
+    get_checksum: true
+    get_mime: false
   register: known_tokens
 
 - name: "Check_tokens | Set 'sync_tokens' to true"
diff --git a/roles/kubernetes/tokens/tasks/gen_tokens.yml b/roles/kubernetes/tokens/tasks/gen_tokens.yml
index 1dabf9657554189ab0838952f724d1266e05b68e..a64aea9e022f4fb0256cbe4ea33f94cc131ed787 100644
--- a/roles/kubernetes/tokens/tasks/gen_tokens.yml
+++ b/roles/kubernetes/tokens/tasks/gen_tokens.yml
@@ -4,7 +4,7 @@
     src: "kube-gen-token.sh"
     dest: "{{ kube_script_dir }}/kube-gen-token.sh"
     mode: "0700"
-  run_once: yes
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
   when: gen_tokens | default(false)
 
@@ -17,7 +17,7 @@
     - "{{ groups['kube_control_plane'] }}"
   register: gentoken_master
   changed_when: "'Added' in gentoken_master.stdout"
-  run_once: yes
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
   when: gen_tokens | default(false)
 
@@ -30,14 +30,14 @@
     - "{{ groups['kube_node'] }}"
   register: gentoken_node
   changed_when: "'Added' in gentoken_node.stdout"
-  run_once: yes
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
   when: gen_tokens | default(false)
 
 - name: Gen_tokens | Get list of tokens from first master
   command: "find {{ kube_token_dir }} -maxdepth 1 -type f"
   register: tokens_list
-  check_mode: no
+  check_mode: false
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
   run_once: true
   when: sync_tokens | default(false)
@@ -47,7 +47,7 @@
   args:
     executable: /bin/bash
   register: tokens_data
-  check_mode: no
+  check_mode: false
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
   run_once: true
   when: sync_tokens | default(false)
diff --git a/roles/kubespray-defaults/tasks/fallback_ips.yml b/roles/kubespray-defaults/tasks/fallback_ips.yml
index a1aff37eefad528c43d266cd5c2efef90b35c7fd..ae3b15150541b56aea6de498c70aed12f97f29ad 100644
--- a/roles/kubespray-defaults/tasks/fallback_ips.yml
+++ b/roles/kubespray-defaults/tasks/fallback_ips.yml
@@ -8,10 +8,10 @@
     gather_subset: '!all,network'
     filter: "ansible_default_ipv4"
   delegate_to: "{{ item }}"
-  delegate_facts: yes
+  delegate_facts: true
   when: hostvars[item].ansible_default_ipv4 is not defined
   loop: "{{ (ansible_play_hosts_all + [groups['kube_control_plane'][0]]) | unique if ansible_limit is defined else (groups['k8s_cluster'] | default([]) + groups['etcd'] | default([]) + groups['calico_rr'] | default([])) | unique }}"
-  run_once: yes
+  run_once: true
   ignore_unreachable: true
   tags: always
 
@@ -26,9 +26,9 @@
       {% endfor %}
   delegate_to: localhost
   connection: local
-  delegate_facts: yes
-  become: no
-  run_once: yes
+  delegate_facts: true
+  become: false
+  run_once: true
 
 - name: Set fallback_ips
   set_fact:
diff --git a/roles/kubespray-defaults/tasks/no_proxy.yml b/roles/kubespray-defaults/tasks/no_proxy.yml
index d2d5cc6d1e1c955a26a52480668b4087e13620c2..adec886f4a3b99f45363247280a33f1038ae3c04 100644
--- a/roles/kubespray-defaults/tasks/no_proxy.yml
+++ b/roles/kubespray-defaults/tasks/no_proxy.yml
@@ -26,9 +26,9 @@
       127.0.0.1,localhost,{{ kube_service_addresses }},{{ kube_pods_subnet }},svc,svc.{{ dns_domain }}
   delegate_to: localhost
   connection: local
-  delegate_facts: yes
-  become: no
-  run_once: yes
+  delegate_facts: true
+  become: false
+  run_once: true
 
 - name: Populates no_proxy to all hosts
   set_fact:
diff --git a/roles/network_plugin/calico/rr/tasks/pre.yml b/roles/network_plugin/calico/rr/tasks/pre.yml
index d8dbd8072803f2d47de45e6fb9de8f6cca706cce..f8a9de6118b9dbbd3b9d813da3dd173cb8ee2d96 100644
--- a/roles/network_plugin/calico/rr/tasks/pre.yml
+++ b/roles/network_plugin/calico/rr/tasks/pre.yml
@@ -3,7 +3,7 @@
   service:
     name: calico-rr
     state: stopped
-    enabled: no
+    enabled: false
   failed_when: false
 
 - name: Calico-rr | Delete obsolete files
diff --git a/roles/network_plugin/calico/tasks/check.yml b/roles/network_plugin/calico/tasks/check.yml
index 95dcfa6731abb6bf94601be72942cdaa3c855baf..7f73a08c4cb8916f8ae296ad29873914e467d620 100644
--- a/roles/network_plugin/calico/tasks/check.yml
+++ b/roles/network_plugin/calico/tasks/check.yml
@@ -4,7 +4,7 @@
     that:
       - ipip is not defined
     msg: "'ipip' configuration variable is deprecated, please configure your inventory with 'calico_ipip_mode' set to 'Always' or 'CrossSubnet' according to your specific needs"
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: Stop if legacy encapsulation variables are detected (ipip_mode)
@@ -12,7 +12,7 @@
     that:
       - ipip_mode is not defined
     msg: "'ipip_mode' configuration variable is deprecated, please configure your inventory with 'calico_ipip_mode' set to 'Always' or 'CrossSubnet' according to your specific needs"
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: Stop if legacy encapsulation variables are detected (calcio_ipam_autoallocateblocks)
@@ -20,7 +20,7 @@
     that:
       - calcio_ipam_autoallocateblocks is not defined
     msg: "'calcio_ipam_autoallocateblocks' configuration variable is deprecated, it's a typo, please configure your inventory with 'calico_ipam_autoallocateblocks' set to 'true' or 'false' according to your specific needs"
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 
@@ -32,7 +32,7 @@
     msg: "When using cloud_provider azure and network_plugin calico calico_ipip_mode must be 'Never' and calico_vxlan_mode 'Always' or 'CrossSubnet'"
   when:
     - cloud_provider is defined and cloud_provider == 'azure'
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: Stop if supported Calico versions
@@ -40,21 +40,21 @@
     that:
       - "calico_version in calico_crds_archive_checksums.keys()"
     msg: "Calico version not supported {{ calico_version }} not in {{ calico_crds_archive_checksums.keys() }}"
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: Check if calicoctl.sh exists
   stat:
     path: "{{ bin_dir }}/calicoctl.sh"
   register: calicoctl_sh_exists
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: Check if calico ready
   command: "{{ bin_dir }}/calicoctl.sh get ClusterInformation default"
   register: calico_ready
-  run_once: True
-  ignore_errors: True
+  run_once: true
+  ignore_errors: true
   retries: 5
   delay: 10
   until: calico_ready.rc == 0
@@ -62,7 +62,7 @@
   when: calicoctl_sh_exists.stat.exists
 
 - name: Check that current calico version is enough for upgrade
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
   when: calicoctl_sh_exists.stat.exists and calico_ready.rc == 0
   block:
@@ -91,7 +91,7 @@
   when:
     - peer_with_calico_rr
     - inventory_hostname == groups['kube_control_plane'][0]
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Check that calico_rr nodes are in k8s_cluster group"
@@ -101,7 +101,7 @@
     msg: "calico_rr must be a child group of k8s_cluster group"
   when:
     - '"calico_rr" in group_names'
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Check vars defined correctly"
@@ -110,7 +110,7 @@
       - "calico_pool_name is defined"
       - "calico_pool_name is match('^[a-zA-Z0-9-_\\\\.]{2,63}$')"
     msg: "calico_pool_name contains invalid characters"
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Check calico network backend defined correctly"
@@ -118,11 +118,11 @@
     that:
       - "calico_network_backend in ['bird', 'vxlan', 'none']"
     msg: "calico network backend is not 'bird', 'vxlan' or 'none'"
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Check ipip and vxlan mode defined correctly"
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
   assert:
     that:
@@ -137,7 +137,7 @@
     msg: "IP in IP and VXLAN mode is mutualy exclusive modes"
   when:
     - "calico_ipip_mode in ['Always', 'CrossSubnet']"
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Check ipip and vxlan mode if simultaneously enabled"
@@ -147,23 +147,23 @@
     msg: "IP in IP and VXLAN mode is mutualy exclusive modes"
   when:
     - "calico_vxlan_mode in ['Always', 'CrossSubnet']"
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Get Calico {{ calico_pool_name }} configuration"
   command: "{{ bin_dir }}/calicoctl.sh get ipPool {{ calico_pool_name }} -o json"
-  failed_when: False
-  changed_when: False
-  check_mode: no
+  failed_when: false
+  changed_when: false
+  check_mode: false
   register: calico
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Set calico_pool_conf"
   set_fact:
     calico_pool_conf: '{{ calico.stdout | from_json }}'
   when: calico.rc == 0 and calico.stdout
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Check if inventory match current cluster configuration"
@@ -176,7 +176,7 @@
     msg: "Your inventory doesn't match the current cluster configuration"
   when:
     - calico_pool_conf is defined
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Check kdd calico_datastore if calico_apiserver_enabled"
@@ -185,7 +185,7 @@
     msg: "When using calico apiserver you need to use the kubernetes datastore"
   when:
     - calico_apiserver_enabled
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Check kdd calico_datastore if typha_enabled"
@@ -194,7 +194,7 @@
     msg: "When using typha you need to use the kubernetes datastore"
   when:
     - typha_enabled
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
 
 - name: "Check ipip mode is Never for calico ipv6"
@@ -204,5 +204,5 @@
     msg: "Calico doesn't support ipip tunneling for the IPv6"
   when:
     - enable_dual_stack_networks
-  run_once: True
+  run_once: true
   delegate_to: "{{ groups['kube_control_plane'][0] }}"
diff --git a/roles/network_plugin/calico/tasks/install.yml b/roles/network_plugin/calico/tasks/install.yml
index 7f895b55502bce67ac3b18c18ab73da2fd6ff91f..1d3b02339c2e85d81c1cb972ff27cd3e24275054 100644
--- a/roles/network_plugin/calico/tasks/install.yml
+++ b/roles/network_plugin/calico/tasks/install.yml
@@ -14,7 +14,7 @@
     src: "{{ downloads.calicoctl.dest }}"
     dest: "{{ bin_dir }}/calicoctl"
     mode: "0755"
-    remote_src: yes
+    remote_src: true
 
 - name: Calico | Create calico certs directory
   file:
@@ -31,7 +31,7 @@
     dest: "{{ calico_cert_dir }}/{{ item.d }}"
     state: hard
     mode: "0640"
-    force: yes
+    force: true
   with_items:
     - {s: "{{ kube_etcd_cacert_file }}", d: "ca_cert.crt"}
     - {s: "{{ kube_etcd_cert_file }}", d: "cert.crt"}
@@ -61,7 +61,7 @@
 - name: Calico | wait for etcd
   uri:
     url: "{{ etcd_access_addresses.split(',') | first }}/health"
-    validate_certs: no
+    validate_certs: false
     client_cert: "{{ calico_cert_dir }}/cert.crt"
     client_key: "{{ calico_cert_dir }}/key.pem"
   register: result
@@ -165,8 +165,8 @@
     - name: Calico | Get existing FelixConfiguration
       command: "{{ bin_dir }}/calicoctl.sh get felixconfig default -o json"
       register: _felix_cmd
-      ignore_errors: True
-      changed_when: False
+      ignore_errors: true
+      changed_when: false
 
     - name: Calico | Set kubespray FelixConfiguration
       set_fact:
@@ -201,7 +201,7 @@
       command:
         cmd: "{{ bin_dir }}/calicoctl.sh apply -f -"
         stdin: "{{ _felix_config is string | ternary(_felix_config, _felix_config | to_json) }}"
-      changed_when: False
+      changed_when: false
 
 - name: Calico | Configure Calico IP Pool
   when:
@@ -210,8 +210,8 @@
     - name: Calico | Get existing calico network pool
       command: "{{ bin_dir }}/calicoctl.sh get ippool {{ calico_pool_name }} -o json"
       register: _calico_pool_cmd
-      ignore_errors: True
-      changed_when: False
+      ignore_errors: true
+      changed_when: false
 
     - name: Calico | Set kubespray calico network pool
       set_fact:
@@ -251,7 +251,7 @@
       command:
         cmd: "{{ bin_dir }}/calicoctl.sh apply -f -"
         stdin: "{{ _calico_pool is string | ternary(_calico_pool, _calico_pool | to_json) }}"
-      changed_when: False
+      changed_when: false
 
 - name: Calico | Configure Calico IPv6 Pool
   when:
@@ -261,8 +261,8 @@
     - name: Calico | Get existing calico ipv6 network pool
       command: "{{ bin_dir }}/calicoctl.sh get ippool {{ calico_pool_name }}-ipv6 -o json"
       register: _calico_pool_ipv6_cmd
-      ignore_errors: True
-      changed_when: False
+      ignore_errors: true
+      changed_when: false
 
     - name: Calico | Set kubespray calico network pool
       set_fact:
@@ -302,19 +302,19 @@
       command:
         cmd: "{{ bin_dir }}/calicoctl.sh apply -f -"
         stdin: "{{ _calico_pool_ipv6 is string | ternary(_calico_pool_ipv6, _calico_pool_ipv6 | to_json) }}"
-      changed_when: False
+      changed_when: false
 
 - name: Populate Service External IPs
   set_fact:
     _service_external_ips: "{{ _service_external_ips | default([]) + [{'cidr': item}] }}"
   with_items: "{{ calico_advertise_service_external_ips }}"
-  run_once: yes
+  run_once: true
 
 - name: Populate Service LoadBalancer IPs
   set_fact:
     _service_loadbalancer_ips: "{{ _service_loadbalancer_ips | default([]) + [{'cidr': item}] }}"
   with_items: "{{ calico_advertise_service_loadbalancer_ips }}"
-  run_once: yes
+  run_once: true
 
 - name: "Determine nodeToNodeMesh needed state"
   set_fact:
@@ -322,7 +322,7 @@
   when:
     - peer_with_router | default(false) or peer_with_calico_rr | default(false)
     - inventory_hostname in groups['k8s_cluster']
-  run_once: yes
+  run_once: true
 
 - name: Calico | Configure Calico BGP
   when:
@@ -331,8 +331,8 @@
     - name: Calico | Get existing BGP Configuration
       command: "{{ bin_dir }}/calicoctl.sh get bgpconfig default -o json"
       register: _bgp_config_cmd
-      ignore_errors: True
-      changed_when: False
+      ignore_errors: true
+      changed_when: false
 
     - name: Calico | Set kubespray BGP Configuration
       set_fact:
@@ -366,7 +366,7 @@
       command:
         cmd: "{{ bin_dir }}/calicoctl.sh apply -f -"
         stdin: "{{ _bgp_config is string | ternary(_bgp_config, _bgp_config | to_json) }}"
-      changed_when: False
+      changed_when: false
 
 - name: Calico | Create calico manifests
   template:
diff --git a/roles/network_plugin/calico/tasks/repos.yml b/roles/network_plugin/calico/tasks/repos.yml
index dd29f45207247cdc2e76bb7897eb563223d5cba1..7eba916bbab93ea3853a91b78c56a5a6a90ecf7b 100644
--- a/roles/network_plugin/calico/tasks/repos.yml
+++ b/roles/network_plugin/calico/tasks/repos.yml
@@ -10,11 +10,11 @@
         file: _copr:copr.fedorainfracloud.org:jdoss:wireguard
         description: Copr repo for wireguard owned by jdoss
         baseurl: "{{ calico_wireguard_repo }}"
-        gpgcheck: yes
+        gpgcheck: true
         gpgkey: https://download.copr.fedorainfracloud.org/results/jdoss/wireguard/pubkey.gpg
-        skip_if_unavailable: yes
-        enabled: yes
-        repo_gpgcheck: no
+        skip_if_unavailable: true
+        enabled: true
+        repo_gpgcheck: false
       when:
         - ansible_os_family in ['RedHat']
         - ansible_distribution not in ['Fedora']
diff --git a/roles/network_plugin/calico/tasks/reset.yml b/roles/network_plugin/calico/tasks/reset.yml
index 8dab21462d7eba2e2749e40ab08f567f620f560e..16c85097710df88741fd5e57e9e6a96a64f4955b 100644
--- a/roles/network_plugin/calico/tasks/reset.yml
+++ b/roles/network_plugin/calico/tasks/reset.yml
@@ -2,9 +2,9 @@
 - name: Reset | check vxlan.calico network device
   stat:
     path: /sys/class/net/vxlan.calico
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: vxlan
 
 - name: Reset | remove the network vxlan.calico device created by calico
@@ -14,9 +14,9 @@
 - name: Reset | check dummy0 network device
   stat:
     path: /sys/class/net/dummy0
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: dummy0
 
 - name: Reset | remove the network device created by calico
diff --git a/roles/network_plugin/cilium/tasks/install.yml b/roles/network_plugin/cilium/tasks/install.yml
index 1039953a00759432e45246540b83ae23d9757bad..7da39644b721b6992d804cbad8d822cd7a9b670c 100644
--- a/roles/network_plugin/cilium/tasks/install.yml
+++ b/roles/network_plugin/cilium/tasks/install.yml
@@ -22,7 +22,7 @@
     dest: "{{ cilium_cert_dir }}/{{ item.d }}"
     mode: "0644"
     state: hard
-    force: yes
+    force: true
   loop:
     - {s: "{{ kube_etcd_cacert_file }}", d: "ca_cert.crt"}
     - {s: "{{ kube_etcd_cert_file }}", d: "cert.crt"}
@@ -94,4 +94,4 @@
     src: "{{ local_release_dir }}/cilium"
     dest: "{{ bin_dir }}/cilium"
     mode: "0755"
-    remote_src: yes
+    remote_src: true
diff --git a/roles/network_plugin/cilium/tasks/reset_iface.yml b/roles/network_plugin/cilium/tasks/reset_iface.yml
index e2f7c14af51ea15344fc302e68d79eed23f6c86d..57a2d5765e4ca3061c0f6aa40db548d371d5d485 100644
--- a/roles/network_plugin/cilium/tasks/reset_iface.yml
+++ b/roles/network_plugin/cilium/tasks/reset_iface.yml
@@ -2,9 +2,9 @@
 - name: "Reset | check if network device {{ iface }} is present"
   stat:
     path: "/sys/class/net/{{ iface }}"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: device_remains
 
 - name: "Reset | remove network device {{ iface }}"
diff --git a/roles/network_plugin/cni/tasks/main.yml b/roles/network_plugin/cni/tasks/main.yml
index 8ac0dc53a3e9d3b28c73a7e0fdce76bf63cffc06..28376bd76056c4ad0b01dc1ae23671302dfdb98b 100644
--- a/roles/network_plugin/cni/tasks/main.yml
+++ b/roles/network_plugin/cni/tasks/main.yml
@@ -13,4 +13,4 @@
     dest: "/opt/cni/bin"
     mode: "0755"
     owner: "{{ cni_bin_owner }}"
-    remote_src: yes
+    remote_src: true
diff --git a/roles/network_plugin/flannel/tasks/reset.yml b/roles/network_plugin/flannel/tasks/reset.yml
index 03d40a0c13adbcebd74ec227821b4ce9c2d05c88..c4b1b8815818378b5bc644682795e8b564394d57 100644
--- a/roles/network_plugin/flannel/tasks/reset.yml
+++ b/roles/network_plugin/flannel/tasks/reset.yml
@@ -2,9 +2,9 @@
 - name: Reset | check cni network device
   stat:
     path: /sys/class/net/cni0
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: cni
 
 - name: Reset | remove the network device created by the flannel
@@ -14,9 +14,9 @@
 - name: Reset | check flannel network device
   stat:
     path: /sys/class/net/flannel.1
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: flannel
 
 - name: Reset | remove the network device created by the flannel
diff --git a/roles/network_plugin/kube-router/tasks/reset.yml b/roles/network_plugin/kube-router/tasks/reset.yml
index ae9ee55c7b779c6b0678ba20a4362159924f3aba..32f707591e40b358621e7e1a4c73c387b55b2ef6 100644
--- a/roles/network_plugin/kube-router/tasks/reset.yml
+++ b/roles/network_plugin/kube-router/tasks/reset.yml
@@ -2,9 +2,9 @@
 - name: Reset | check kube-dummy-if network device
   stat:
     path: /sys/class/net/kube-dummy-if
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kube_dummy_if
 
 - name: Reset | remove the network device created by kube-router
@@ -14,9 +14,9 @@
 - name: Check kube-bridge exists
   stat:
     path: /sys/class/net/kube-bridge
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kube_bridge_if
 
 - name: Reset | donw the network bridge create by kube-router
diff --git a/roles/network_plugin/macvlan/tasks/main.yml b/roles/network_plugin/macvlan/tasks/main.yml
index 165030d592cf342516ec242fb6da0ce241d66d6f..6ffe3348cd0b7d13e5eccc7de8549b02fa43c689 100644
--- a/roles/network_plugin/macvlan/tasks/main.yml
+++ b/roles/network_plugin/macvlan/tasks/main.yml
@@ -104,7 +104,7 @@
   ansible.posix.sysctl:
     name: net.ipv4.conf.all.arp_notify
     value: 1
-    sysctl_set: yes
+    sysctl_set: true
     sysctl_file: "{{ sysctl_file_path }}"
     state: present
-    reload: yes
+    reload: true
diff --git a/roles/recover_control_plane/etcd/tasks/main.yml b/roles/recover_control_plane/etcd/tasks/main.yml
index 599f56b15060611d1098369db234e9af52ffe1dd..6291ea36f793d293cd61261b8d17dced2ac6c72f 100644
--- a/roles/recover_control_plane/etcd/tasks/main.yml
+++ b/roles/recover_control_plane/etcd/tasks/main.yml
@@ -4,7 +4,7 @@
   register: etcd_endpoint_health
   ignore_errors: true  # noqa ignore-errors
   changed_when: false
-  check_mode: no
+  check_mode: false
   environment:
     ETCDCTL_API: "3"
     ETCDCTL_ENDPOINTS: "{{ etcd_access_addresses }}"
@@ -64,7 +64,7 @@
   command: "{{ bin_dir }}/etcdctl member list"
   register: member_list
   changed_when: false
-  check_mode: no
+  check_mode: false
   environment:
     ETCDCTL_API: "3"
     ETCDCTL_ENDPOINTS: "{{ etcd_access_addresses }}"
diff --git a/roles/remove-node/remove-etcd-node/tasks/main.yml b/roles/remove-node/remove-etcd-node/tasks/main.yml
index 0279018d4fd6cb0883775c917f1fcddd29e47048..70e33251a0f3d58298dd3d1520435b9485488329 100644
--- a/roles/remove-node/remove-etcd-node/tasks/main.yml
+++ b/roles/remove-node/remove-etcd-node/tasks/main.yml
@@ -32,7 +32,7 @@
   register: etcd_member_id
   ignore_errors: true  # noqa ignore-errors
   changed_when: false
-  check_mode: no
+  check_mode: false
   tags:
     - facts
   environment:
diff --git a/roles/reset/tasks/main.yml b/roles/reset/tasks/main.yml
index 1d00913e1b26dfa5f0342b94f05f541ba9118e4a..acafb5a760e36b018c03ebc8aeaf039dc9d70a45 100644
--- a/roles/reset/tasks/main.yml
+++ b/roles/reset/tasks/main.yml
@@ -50,9 +50,9 @@
 - name: Reset | check if crictl is present
   stat:
     path: "{{ bin_dir }}/crictl"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: crictl
 
 - name: Reset | stop all cri containers
@@ -161,7 +161,7 @@
   shell: set -o pipefail && mount | grep /var/lib/kubelet/ | awk '{print $3}' | tac
   args:
     executable: /bin/bash
-  check_mode: no
+  check_mode: false
   register: mounted_dirs
   failed_when: false
   changed_when: false
@@ -182,7 +182,7 @@
 - name: Flush iptables
   iptables:
     table: "{{ item }}"
-    flush: yes
+    flush: true
   with_items:
     - filter
     - nat
@@ -195,7 +195,7 @@
 - name: Flush ip6tables
   iptables:
     table: "{{ item }}"
-    flush: yes
+    flush: true
     ip_version: ipv6
   with_items:
     - filter
@@ -215,9 +215,9 @@
 - name: Reset | check kube-ipvs0 network device
   stat:
     path: /sys/class/net/kube-ipvs0
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: kube_ipvs0
 
 - name: Reset | Remove kube-ipvs0
@@ -229,9 +229,9 @@
 - name: Reset | check nodelocaldns network device
   stat:
     path: /sys/class/net/nodelocaldns
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: nodelocaldns_device
 
 - name: Reset | Remove nodelocaldns
@@ -243,9 +243,9 @@
 - name: Reset | Check whether /var/lib/kubelet directory exists
   stat:
     path: /var/lib/kubelet
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: var_lib_kubelet_directory
 
 - name: Reset | Find files/dirs with immutable flag in /var/lib/kubelet
diff --git a/roles/upgrade/post-upgrade/tasks/main.yml b/roles/upgrade/post-upgrade/tasks/main.yml
index d82fe8d33c4d0a1b9e24aec7b3f88d13f75e8c26..467ff05e18435cfabb9d918b682f65e055b61c5f 100644
--- a/roles/upgrade/post-upgrade/tasks/main.yml
+++ b/roles/upgrade/post-upgrade/tasks/main.yml
@@ -13,7 +13,7 @@
 
 - name: Confirm node uncordon
   pause:
-    echo: yes
+    echo: true
     prompt: "Ready to uncordon node?"
   when:
     - upgrade_node_post_upgrade_confirm
diff --git a/roles/upgrade/pre-upgrade/tasks/main.yml b/roles/upgrade/pre-upgrade/tasks/main.yml
index 8d5d99c937d4d0949256033b6bbd59512dbb5511..6e3cdd2b85943317f4159f5e7ad9b90b8587cea0 100644
--- a/roles/upgrade/pre-upgrade/tasks/main.yml
+++ b/roles/upgrade/pre-upgrade/tasks/main.yml
@@ -2,7 +2,7 @@
 # Wait for upgrade
 - name: Confirm node upgrade
   pause:
-    echo: yes
+    echo: true
     prompt: "Ready to upgrade node? (Press Enter to continue or Ctrl+C for other options)"
   when:
     - upgrade_node_confirm
diff --git a/roles/win_nodes/kubernetes_patch/tasks/main.yml b/roles/win_nodes/kubernetes_patch/tasks/main.yml
index 880c58cf8e287be25ee01c8d5ab257e1f4ac829b..1dd504e1673c5877ae03fdce8aa5e95c363f1b7e 100644
--- a/roles/win_nodes/kubernetes_patch/tasks/main.yml
+++ b/roles/win_nodes/kubernetes_patch/tasks/main.yml
@@ -4,7 +4,7 @@
   file:
     path: "{{ kubernetes_user_manifests_path }}/kubernetes"
     state: directory
-    recurse: yes
+    recurse: true
   tags: [init, cni]
 
 - name: Apply kube-proxy nodeselector
diff --git a/scripts/collect-info.yaml b/scripts/collect-info.yaml
index 0234c073361e5f329d3a82ec014d619fbda379ac..272fb724b24e0e7d539fb10f1257567e8aa72c62 100644
--- a/scripts/collect-info.yaml
+++ b/scripts/collect-info.yaml
@@ -2,7 +2,7 @@
 - name: Collect debug info
   hosts: all
   become: true
-  gather_facts: no
+  gather_facts: false
 
   vars:
     docker_bin_dir: /usr/bin
@@ -118,7 +118,7 @@
       failed_when: false
       with_items: "{{ commands }}"
       when: item.when | default(True)
-      no_log: True
+      no_log: true
 
     - name: Fetch results
       fetch:
diff --git a/tests/cloud_playbooks/cleanup-packet.yml b/tests/cloud_playbooks/cleanup-packet.yml
index 009071ec2b92510ea036fb4fc90ea5bb86e234f7..2ba5e30210ae9dd5a8c61a985c1dc68731ed1503 100644
--- a/tests/cloud_playbooks/cleanup-packet.yml
+++ b/tests/cloud_playbooks/cleanup-packet.yml
@@ -2,7 +2,7 @@
 
 - name: Cleanup packet vms
   hosts: localhost
-  gather_facts: no
+  gather_facts: false
   become: true
   roles:
     - { role: cleanup-packet-ci }
diff --git a/tests/cloud_playbooks/create-packet.yml b/tests/cloud_playbooks/create-packet.yml
index 8212fb6c836e478d01ed23adfeb7902a14a4e9e2..2cd08b54d18dde857197f1091062d9b96fa82ec0 100644
--- a/tests/cloud_playbooks/create-packet.yml
+++ b/tests/cloud_playbooks/create-packet.yml
@@ -2,7 +2,7 @@
 
 - name: Provision Packet VMs
   hosts: localhost
-  gather_facts: no
+  gather_facts: false
   become: true
   vars:
     ci_job_name: "{{ lookup('env', 'CI_JOB_NAME') }}"
diff --git a/tests/cloud_playbooks/delete-packet.yml b/tests/cloud_playbooks/delete-packet.yml
index 7d0c9003c7a2660a61441ed3b4a8e727fcdf80aa..7320da62238fa329872abb54afc03e0619938192 100644
--- a/tests/cloud_playbooks/delete-packet.yml
+++ b/tests/cloud_playbooks/delete-packet.yml
@@ -2,7 +2,7 @@
 
 - name: Terminate Packet VMs
   hosts: localhost
-  gather_facts: no
+  gather_facts: false
   become: true
   vars:
     ci_job_name: "{{ lookup('env', 'CI_JOB_NAME') }}"
diff --git a/tests/cloud_playbooks/roles/packet-ci/tasks/delete-vms.yml b/tests/cloud_playbooks/roles/packet-ci/tasks/delete-vms.yml
index 98bd05a61d4f6b7cb298095b5381b8058a37924e..75156584a89cccd19400cb10a5953f64798e4fdb 100644
--- a/tests/cloud_playbooks/roles/packet-ci/tasks/delete-vms.yml
+++ b/tests/cloud_playbooks/roles/packet-ci/tasks/delete-vms.yml
@@ -3,9 +3,9 @@
 - name: Check if temp directory for {{ test_name }} exists
   stat:
     path: "/tmp/{{ test_name }}"
-    get_attributes: no
-    get_checksum: no
-    get_mime: no
+    get_attributes: false
+    get_checksum: false
+    get_mime: false
   register: temp_dir_details
 
 - name: "Cleanup temp directory for {{ test_name }}"
diff --git a/tests/cloud_playbooks/wait-for-ssh.yml b/tests/cloud_playbooks/wait-for-ssh.yml
index 0e09c9f04cd8213514eacebddae704a4b22d379a..54b2682737b1a6ee2ccf6e1094b4b68378291651 100644
--- a/tests/cloud_playbooks/wait-for-ssh.yml
+++ b/tests/cloud_playbooks/wait-for-ssh.yml
@@ -1,8 +1,8 @@
 ---
 - name: Wait until SSH is available
   hosts: all
-  become: False
-  gather_facts: False
+  become: false
+  gather_facts: false
 
   tasks:
   - name: Wait until SSH is available
diff --git a/tests/files/packet_ubuntu20-all-in-one-docker.yml b/tests/files/packet_ubuntu20-all-in-one-docker.yml
index 2ed6307d8c21695027d2a434aff6054434264bf5..0116eae42c887a9a3af033f108c9d4af2a46f55b 100644
--- a/tests/files/packet_ubuntu20-all-in-one-docker.yml
+++ b/tests/files/packet_ubuntu20-all-in-one-docker.yml
@@ -8,7 +8,7 @@ auto_renew_certificates: true
 
 # Currently ipvs not available on KVM: https://packages.ubuntu.com/search?suite=focal&arch=amd64&mode=exactfilename&searchon=contents&keywords=ip_vs_sh.ko
 kube_proxy_mode: iptables
-enable_nodelocaldns: False
+enable_nodelocaldns: false
 
 # Use docker
 container_manager: docker
diff --git a/tests/files/packet_ubuntu20-calico-all-in-one-hardening.yml b/tests/files/packet_ubuntu20-calico-all-in-one-hardening.yml
index c494810cf0719ee8dc090d8f2c37ad1ced3e8da6..5dafe23bfc8f88899b457937b8dac02fd064c434 100644
--- a/tests/files/packet_ubuntu20-calico-all-in-one-hardening.yml
+++ b/tests/files/packet_ubuntu20-calico-all-in-one-hardening.yml
@@ -8,7 +8,7 @@ auto_renew_certificates: true
 
 # Currently ipvs not available on KVM: https://packages.ubuntu.com/search?suite=focal&arch=amd64&mode=exactfilename&searchon=contents&keywords=ip_vs_sh.ko
 kube_proxy_mode: iptables
-enable_nodelocaldns: False
+enable_nodelocaldns: false
 
 # The followings are for hardening
 ## kube-apiserver
diff --git a/tests/files/packet_ubuntu20-calico-all-in-one.yml b/tests/files/packet_ubuntu20-calico-all-in-one.yml
index 3cfc99c96d5085b5af078b16f29835a841aaed80..f59e72a3ca157b062c42674a877e6beef87bd542 100644
--- a/tests/files/packet_ubuntu20-calico-all-in-one.yml
+++ b/tests/files/packet_ubuntu20-calico-all-in-one.yml
@@ -8,4 +8,4 @@ auto_renew_certificates: true
 
 # Currently ipvs not available on KVM: https://packages.ubuntu.com/search?suite=focal&arch=amd64&mode=exactfilename&searchon=contents&keywords=ip_vs_sh.ko
 kube_proxy_mode: iptables
-enable_nodelocaldns: False
+enable_nodelocaldns: false
diff --git a/tests/files/packet_ubuntu20-calico-etcd-kubeadm-upgrade-ha.yml b/tests/files/packet_ubuntu20-calico-etcd-kubeadm-upgrade-ha.yml
index 57187a8dd46ba80aa9158616cb48be8fdb9c758b..425ce75b81689c0d4f2915886f4ffcb38e8ac0f2 100644
--- a/tests/files/packet_ubuntu20-calico-etcd-kubeadm-upgrade-ha.yml
+++ b/tests/files/packet_ubuntu20-calico-etcd-kubeadm-upgrade-ha.yml
@@ -10,7 +10,7 @@ upgrade_cluster_setup: true
 
 # Currently ipvs not available on KVM: https://packages.ubuntu.com/search?suite=focal&arch=amd64&mode=exactfilename&searchon=contents&keywords=ip_vs_sh.ko
 kube_proxy_mode: iptables
-enable_nodelocaldns: False
+enable_nodelocaldns: false
 
 # Pin disabling ipip mode to ensure proper upgrade
 ipip: false
diff --git a/tests/files/packet_ubuntu20-calico-etcd-kubeadm.yml b/tests/files/packet_ubuntu20-calico-etcd-kubeadm.yml
index ba9d7b34b8d0765fe61193bc40c377798057aba2..ddc5cb5561e06da02a86cc2a2ddd14ee16a4d05d 100644
--- a/tests/files/packet_ubuntu20-calico-etcd-kubeadm.yml
+++ b/tests/files/packet_ubuntu20-calico-etcd-kubeadm.yml
@@ -8,7 +8,7 @@ etcd_deployment_type: kubeadm
 
 # Currently ipvs not available on KVM: https://packages.ubuntu.com/search?suite=focal&arch=amd64&mode=exactfilename&searchon=contents&keywords=ip_vs_sh.ko
 kube_proxy_mode: iptables
-enable_nodelocaldns: False
+enable_nodelocaldns: false
 
 # Remove anonymous access to cluster
 remove_anonymous_access: true
diff --git a/tests/files/packet_ubuntu22-all-in-one-docker.yml b/tests/files/packet_ubuntu22-all-in-one-docker.yml
index 16ae459862932e0a887547a9ac6b84eff6db9fdb..fcdd8f3ccfa684a3c601703884c6730d1d910d37 100644
--- a/tests/files/packet_ubuntu22-all-in-one-docker.yml
+++ b/tests/files/packet_ubuntu22-all-in-one-docker.yml
@@ -9,7 +9,7 @@ auto_renew_certificates: true
 
 # Currently ipvs not available on KVM: https://packages.ubuntu.com/search?suite=focal&arch=amd64&mode=exactfilename&searchon=contents&keywords=ip_vs_sh.ko
 kube_proxy_mode: iptables
-enable_nodelocaldns: False
+enable_nodelocaldns: false
 
 # Use docker
 container_manager: docker
diff --git a/tests/files/packet_ubuntu22-calico-all-in-one.yml b/tests/files/packet_ubuntu22-calico-all-in-one.yml
index 2c666f8e33d06e5ddcb0998a92ba925e7967cf6a..61553010753b08310de5b92d701fd623dbae05e9 100644
--- a/tests/files/packet_ubuntu22-calico-all-in-one.yml
+++ b/tests/files/packet_ubuntu22-calico-all-in-one.yml
@@ -9,7 +9,7 @@ auto_renew_certificates: true
 
 # Currently ipvs not available on KVM: https://packages.ubuntu.com/search?suite=focal&arch=amd64&mode=exactfilename&searchon=contents&keywords=ip_vs_sh.ko
 kube_proxy_mode: iptables
-enable_nodelocaldns: False
+enable_nodelocaldns: false
 
 containerd_registries_mirrors:
   - prefix: docker.io
diff --git a/tests/files/packet_ubuntu24-all-in-one-docker.yml b/tests/files/packet_ubuntu24-all-in-one-docker.yml
index d4a0adccb64fb2b8109f742b35724e0efb7d22fe..8b1da4ac0cbd1ae4d3c76fe6f4c102d02fd07331 100644
--- a/tests/files/packet_ubuntu24-all-in-one-docker.yml
+++ b/tests/files/packet_ubuntu24-all-in-one-docker.yml
@@ -9,7 +9,7 @@ auto_renew_certificates: true
 
 # Currently ipvs not available on KVM: https://packages.ubuntu.com/search?suite=noble&arch=amd64&mode=exactfilename&searchon=contents&keywords=ip_vs_sh.ko
 kube_proxy_mode: iptables
-enable_nodelocaldns: False
+enable_nodelocaldns: false
 
 # Use docker
 container_manager: docker
diff --git a/tests/files/packet_ubuntu24-calico-all-in-one.yml b/tests/files/packet_ubuntu24-calico-all-in-one.yml
index 4b9e403ca12e1ee0cc89968c0315927106690001..5d7f55878ad2a3c028316c666d6f20d2e6af6ded 100644
--- a/tests/files/packet_ubuntu24-calico-all-in-one.yml
+++ b/tests/files/packet_ubuntu24-calico-all-in-one.yml
@@ -9,7 +9,7 @@ auto_renew_certificates: true
 
 # Currently ipvs not available on KVM: https://packages.ubuntu.com/search?suite=noble&arch=amd64&mode=exactfilename&searchon=contents&keywords=ip_vs_sh.ko
 kube_proxy_mode: iptables
-enable_nodelocaldns: False
+enable_nodelocaldns: false
 
 containerd_registries_mirrors:
   - prefix: docker.io
diff --git a/tests/files/packet_ubuntu24-calico-etcd-datastore.yml b/tests/files/packet_ubuntu24-calico-etcd-datastore.yml
index 2805fa731a3604666d6e9694c22f04492bd84858..4f35d2f8763c15a6f3a52b21b4cc25307af0b0dd 100644
--- a/tests/files/packet_ubuntu24-calico-etcd-datastore.yml
+++ b/tests/files/packet_ubuntu24-calico-etcd-datastore.yml
@@ -9,7 +9,7 @@ auto_renew_certificates: true
 
 # Currently ipvs not available on KVM: https://packages.ubuntu.com/search?suite=noble&arch=amd64&mode=exactfilename&searchon=contents&keywords=ip_vs_sh.ko
 kube_proxy_mode: iptables
-enable_nodelocaldns: False
+enable_nodelocaldns: false
 
 containerd_registries:
   "docker.io": "https://mirror.gcr.io"
diff --git a/tests/testcases/010_check-apiserver.yml b/tests/testcases/010_check-apiserver.yml
index 0d20bda029833833aad85de06d9e9a2abb0503f5..081a2a31ed2007ffe8667fdd000771b7b78a116f 100644
--- a/tests/testcases/010_check-apiserver.yml
+++ b/tests/testcases/010_check-apiserver.yml
@@ -6,7 +6,7 @@
   - name: Check the API servers are responding
     uri:
       url: "https://{{ access_ip | default(ansible_default_ipv4.address) }}:{{ kube_apiserver_port | default(6443) }}/version"
-      validate_certs: no
+      validate_certs: false
       status_code: 200
     register: apiserver_response
     retries: 12
diff --git a/tests/testcases/030_check-network.yml b/tests/testcases/030_check-network.yml
index 9b7eacfc1e9a63176f3257a326dbe75007d7a2fa..e13128dd9cef34f55843945d103d6088674319e3 100644
--- a/tests/testcases/030_check-network.yml
+++ b/tests/testcases/030_check-network.yml
@@ -153,13 +153,13 @@
   - name: Get running pods
     command: "{{ bin_dir }}/kubectl get pods -n test -o
             jsonpath='{range .items[?(.status.phase==\"Running\")]}{.metadata.name} {.status.podIP} {.status.containerStatuses} {end}'"
-    changed_when: False
+    changed_when: false
     register: running_pods
     no_log: true
 
   - name: Check kubectl output
     command: "{{ bin_dir }}/kubectl get pods --all-namespaces -owide"
-    changed_when: False
+    changed_when: false
     register: get_pods
     no_log: true
 
diff --git a/tests/testcases/040_check-network-adv.yml b/tests/testcases/040_check-network-adv.yml
index 4fc70eb07ad79ed0c08f4ca43eea8db158284a6f..45cf6db284bf6650ff9b735f1a8523c0e96996c5 100644
--- a/tests/testcases/040_check-network-adv.yml
+++ b/tests/testcases/040_check-network-adv.yml
@@ -69,7 +69,7 @@
     - name: Get netchecker agents
       uri:
         url: "http://{{ ansible_default_ipv4.address }}:{{ netchecker_port }}/api/v1/agents/"
-        return_content: yes
+        return_content: true
       run_once: true
       delegate_to: "{{ groups['kube_control_plane'][0] }}"
       register: agents
@@ -85,7 +85,7 @@
       uri:
         url: "http://{{ ansible_default_ipv4.address }}:{{ netchecker_port }}/api/v1/connectivity_check"
         status_code: 200
-        return_content: yes
+        return_content: true
       delegate_to: "{{ groups['kube_control_plane'][0] }}"
       run_once: true
       register: connectivity_check
diff --git a/tests/testcases/100_check-k8s-conformance.yml b/tests/testcases/100_check-k8s-conformance.yml
index 3c07ffe46da5cca0c34f0001aa3a3e5ce0b40c65..3e0f17109dd2fe2b1c13f08e22d43b3cf805a145 100644
--- a/tests/testcases/100_check-k8s-conformance.yml
+++ b/tests/testcases/100_check-k8s-conformance.yml
@@ -24,7 +24,7 @@
       unarchive:
         src: /tmp/sonobuoy.tar.gz
         dest: /usr/local/bin/
-        copy: no
+        copy: false
 
     - name: Run sonobuoy
       command: "{{ sonobuoy_path }} run --mode {{ sonobuoy_mode }} --e2e-parallel {{ sonobuoy_parallel }} --wait"