From d321e42d9e7720ba9449e41d7cf50d13c1f170ac Mon Sep 17 00:00:00 2001
From: Max Gautier <mg@max.gautier.name>
Date: Mon, 13 May 2024 10:20:48 +0000
Subject: [PATCH] CI: Factorize ansible-playbook flags (#11173)

We have inconsistent sets of options passed to the playbooks during our
CI runs.

Don't run ansible-playbook directly, instead factorize the execution in
a bash function using all the common flags.
Also remove various ENABLE_* variables and instead directly test for the
relevant conditions at execution time, as this makes it more obvious and
does not force one to go back and forth in the script.
---
 tests/scripts/testcases_run.sh | 101 +++++++++++++++++----------------
 1 file changed, 52 insertions(+), 49 deletions(-)

diff --git a/tests/scripts/testcases_run.sh b/tests/scripts/testcases_run.sh
index 7cd4671a7..ec22a4dcb 100755
--- a/tests/scripts/testcases_run.sh
+++ b/tests/scripts/testcases_run.sh
@@ -47,21 +47,6 @@ if [[ "$CI_JOB_NAME" =~ "ubuntu" ]]; then
   CI_TEST_ADDITIONAL_VARS="-e ansible_python_interpreter=/usr/bin/python3"
 fi
 
-ENABLE_020_TEST="true"
-ENABLE_030_TEST="true"
-ENABLE_040_TEST="true"
-if [[ "$CI_JOB_NAME" =~ "macvlan" ]]; then
-  ENABLE_020_TEST="false"
-  ENABLE_030_TEST="false"
-  ENABLE_040_TEST="false"
-fi
-
-if [[ "$CI_JOB_NAME" =~ "hardening" ]]; then
-  # TODO: We need to remove this condition by finding alternative container
-  # image instead of netchecker which doesn't work at hardening environments.
-  ENABLE_040_TEST="false"
-fi
-
 # Check out latest tag if testing upgrade
 test "${UPGRADE_TEST}" != "false" && git fetch --all && git checkout "$KUBESPRAY_VERSION"
 # Checkout the CI vars file so it is available
@@ -69,21 +54,41 @@ test "${UPGRADE_TEST}" != "false" && git checkout "${CI_COMMIT_SHA}" tests/files
 test "${UPGRADE_TEST}" != "false" && git checkout "${CI_COMMIT_SHA}" ${CI_TEST_REGISTRY_MIRROR}
 test "${UPGRADE_TEST}" != "false" && git checkout "${CI_COMMIT_SHA}" ${CI_TEST_SETTING}
 
+
+run_playbook () {
+playbook=$1
+shift
+# We can set --limit here and still pass it as supplemental args because `--limit`  is a 'last one wins' option
+ansible-playbook --limit "all:!fake_hosts" \
+     $ANSIBLE_LOG_LEVEL \
+    -e @${CI_TEST_SETTING} \
+    -e @${CI_TEST_REGISTRY_MIRROR} \
+    -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} \
+    -e local_release_dir=${PWD}/downloads \
+    "$@" \
+    ${playbook}
+}
+
 # Create cluster
-ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_SETTING} -e @${CI_TEST_REGISTRY_MIRROR} -e @${CI_TEST_VARS} -e local_release_dir=${PWD}/downloads --limit "all:!fake_hosts" cluster.yml
+run_playbook cluster.yml
 
 # Repeat deployment if testing upgrade
-if [ "${UPGRADE_TEST}" != "false" ]; then
-  test "${UPGRADE_TEST}" == "basic" && PLAYBOOK="cluster.yml"
-  test "${UPGRADE_TEST}" == "graceful" && PLAYBOOK="upgrade-cluster.yml"
-  git checkout "${CI_COMMIT_SHA}"
-  ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_SETTING} -e @${CI_TEST_REGISTRY_MIRROR} -e @${CI_TEST_VARS} -e local_release_dir=${PWD}/downloads --limit "all:!fake_hosts" $PLAYBOOK
-fi
+case "${UPGRADE_TEST}" in
+
+    "basic")
+        run_playbook cluster.yml
+        ;;
+    "graceful")
+        run_playbook upgrade-cluster.yml
+        ;;
+    *)
+        ;;
+esac
 
 # Test control plane recovery
 if [ "${RECOVER_CONTROL_PLANE_TEST}" != "false" ]; then
-  ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_SETTING} -e @${CI_TEST_REGISTRY_MIRROR} -e @${CI_TEST_VARS} -e local_release_dir=${PWD}/downloads --limit "${RECOVER_CONTROL_PLANE_TEST_GROUPS}:!fake_hosts" -e reset_confirmation=yes reset.yml
-  ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_SETTING} -e @${CI_TEST_REGISTRY_MIRROR} -e @${CI_TEST_VARS} -e local_release_dir=${PWD}/downloads -e etcd_retries=10 --limit "etcd:kube_control_plane:!fake_hosts" recover-control-plane.yml
+    run_playbook reset.yml --limit "${RECOVER_CONTROL_PLANE_TEST_GROUPS}:!fake_hosts" -e reset_confirmation=yes
+    run_playbook recover-control-plane.yml -e etcd_retries=10 --limit "etcd:kube_control_plane:!fake_hosts"
 fi
 
 # Test collection build and install by installing our collection, emptying our repository, adding
@@ -114,61 +119,59 @@ EOF
   # Write remove-node.yml
 cat > remove-node.yml <<EOF
 - name: Remove node from Kubernetes
-  ansible.builtin.import_playbook: kubernetes_sigs.kubespray.remote-node
+  ansible.builtin.import_playbook: kubernetes_sigs.kubespray.remove_node
 EOF
 
 fi
-
 # Tests Cases
 ## Test Master API
-ansible-playbook --limit "all:!fake_hosts" -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} tests/testcases/010_check-apiserver.yml $ANSIBLE_LOG_LEVEL
+run_playbook tests/testcases/010_check-apiserver.yml
+run_playbook tests/testcases/015_check-nodes-ready.yml
 
 ## Test that all nodes are Ready
-ansible-playbook --limit "all:!fake_hosts" -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} tests/testcases/015_check-nodes-ready.yml $ANSIBLE_LOG_LEVEL
-
-## Test that all pods are Running
-if [ "${ENABLE_020_TEST}" = "true" ]; then
-ansible-playbook --limit "all:!fake_hosts" -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} tests/testcases/020_check-pods-running.yml $ANSIBLE_LOG_LEVEL
-fi
 
-## Test pod creation and ping between them
-if [ "${ENABLE_030_TEST}" = "true" ]; then
-ansible-playbook --limit "all:!fake_hosts" -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} tests/testcases/030_check-network.yml $ANSIBLE_LOG_LEVEL
-fi
-
-## Advanced DNS checks
-if [ "${ENABLE_040_TEST}" = "true" ]; then
-  ansible-playbook --limit "all:!fake_hosts" -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} tests/testcases/040_check-network-adv.yml $ANSIBLE_LOG_LEVEL
+if [[ ! ( "$CI_JOB_NAME" =~ "macvlan" ) ]]; then
+    run_playbook tests/testcases/020_check-pods-running.yml
+    run_playbook tests/testcases/030_check-network.yml
+    if [[ ! ( "$CI_JOB_NAME" =~ "hardening" ) ]]; then
+      # TODO: We need to remove this condition by finding alternative container
+      # image instead of netchecker which doesn't work at hardening environments.
+      run_playbook tests/testcases/040_check-network-adv.yml
+    fi
 fi
 
 ## Kubernetes conformance tests
-ansible-playbook -i ${ANSIBLE_INVENTORY} -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} --limit "all:!fake_hosts" tests/testcases/100_check-k8s-conformance.yml $ANSIBLE_LOG_LEVEL
+run_playbook tests/testcases/100_check-k8s-conformance.yml
 
 if [ "${IDEMPOT_CHECK}" = "true" ]; then
   ## Idempotency checks 1/5 (repeat deployment)
-  ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_SETTING} -e @${CI_TEST_REGISTRY_MIRROR} ${CI_TEST_ADDITIONAL_VARS} -e @${CI_TEST_VARS} -e local_release_dir=${PWD}/downloads --limit "all:!fake_hosts" cluster.yml
+  run_playbook cluster.yml
 
   ## Idempotency checks 2/5 (Advanced DNS checks)
-  ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} --limit "all:!fake_hosts" tests/testcases/040_check-network-adv.yml
+  if [[ ! ( "$CI_JOB_NAME" =~ "hardening" ) ]]; then
+      run_playbook tests/testcases/040_check-network-adv.yml
+  fi
 
   if [ "${RESET_CHECK}" = "true" ]; then
     ## Idempotency checks 3/5 (reset deployment)
-    ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_SETTING} -e @${CI_TEST_REGISTRY_MIRROR}  -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} -e reset_confirmation=yes --limit "all:!fake_hosts" reset.yml
+    run_playbook reset.yml -e reset_confirmation=yes
 
     ## Idempotency checks 4/5 (redeploy after reset)
-    ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_SETTING} -e @${CI_TEST_REGISTRY_MIRROR} -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} -e local_release_dir=${PWD}/downloads --limit "all:!fake_hosts" cluster.yml
+    run_playbook cluster.yml
 
     ## Idempotency checks 5/5 (Advanced DNS checks)
-    ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_SETTING} -e @${CI_TEST_REGISTRY_MIRROR} -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} --limit "all:!fake_hosts" tests/testcases/040_check-network-adv.yml
+    if [[ ! ( "$CI_JOB_NAME" =~ "hardening" ) ]]; then
+        run_playbook tests/testcases/040_check-network-adv.yml
+    fi
   fi
 fi
 
 # Test node removal procedure
 if [ "${REMOVE_NODE_CHECK}" = "true" ]; then
-  ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_SETTING} -e @${CI_TEST_REGISTRY_MIRROR}  -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} -e skip_confirmation=yes -e node=${REMOVE_NODE_NAME} --limit "all:!fake_hosts" remove-node.yml
+  run_playbook remove-node.yml -e skip_confirmation=yes -e node=${REMOVE_NODE_NAME}
 fi
 
 # Clean up at the end, this is to allow stage1 tests to include cleanup test
 if [ "${RESET_CHECK}" = "true" ]; then
-  ansible-playbook ${ANSIBLE_LOG_LEVEL} -e @${CI_TEST_SETTING} -e @${CI_TEST_REGISTRY_MIRROR}  -e @${CI_TEST_VARS} ${CI_TEST_ADDITIONAL_VARS} -e reset_confirmation=yes --limit "all:!fake_hosts" reset.yml
+  run_playbook reset.yml -e reset_confirmation=yes
 fi
-- 
GitLab