Переглянути джерело

Issue #173: Removed common packages and nfs-utils validation

Signed-off-by: Blesson James <blesson_james@dellteam.com>
blesson-james 4 роки тому
батько
коміт
fc90f88c1e

+ 79 - 20
omnia.yml

@@ -13,72 +13,124 @@
 # limitations under the License.
 ---
 
-# Omnia playbook. Will be updated later.
+- name: Validate the cluster
+  hosts: localhost
+  connection: local
+  gather_facts: no
+  roles:
+    - cluster_validation
 
 - name: Gather facts from all the nodes
   hosts: all
 
+- name: Prepare the cluster with passwordless ssh from manager to compute
+  hosts: manager
+  gather_facts: false
+  pre_tasks:
+    - name: Set Fact
+      set_fact:
+        ssh_to: "{{ groups['compute'] }}"
+  roles:
+    - cluster_preperation
+
+- name: Prepare the cluster with passwordless ssh from compute to manager
+  hosts: compute
+  gather_facts: false
+  pre_tasks:
+    - name: Set Fact
+      set_fact:
+        ssh_to: "{{ groups['manager'] }}"
+  roles:
+    - cluster_preperation
+    
 - name: Apply common installation and config
   hosts: manager, compute
   gather_facts: false
   roles:
     - common
- 
-#- name: Apply GPU node config
-#  hosts: gpus
-#  gather_facts: false
-#  roles:
-#    - compute_gpu
+  tags: common
+
+- name: Apply common K8s installation and config
+  hosts: manager, compute
+  gather_facts: false
+  roles:
+    - k8s_common
+  tags: kubernetes
+
+- name: Apply GPU node config
+  hosts: gpus
+  gather_facts: false
+  roles:
+    - compute_gpu
 
 - name: Apply K8s manager config
   hosts: manager
   gather_facts: true
   roles:
-    - manager
+    - k8s_manager
+  tags: kubernetes
 
 - name: Apply K8s firewalld config on manager and compute nodes
   hosts: manager, compute
   gather_facts: false
   roles:
-    - firewalld
+    - k8s_firewalld
+  tags: kubernetes
+
+- name: Apply NFS server setup on manager node
+  hosts: manager
+  gather_facts: false
+  roles:
+    - k8s_nfs_server_setup
+  tags: kubernetes
+
+- name: Apply NFS client setup on compute nodes
+  hosts: compute
+  gather_facts: false
+  roles:
+    - k8s_nfs_client_setup
+  tags: kubernetes
 
 - name: Start K8s on manager server
   hosts: manager
   gather_facts: true
   roles:
-    - startmanager
+    - k8s_start_manager
+  tags: kubernetes
 
 - name: Start K8s worker servers on compute nodes
   hosts: compute
   gather_facts: false
   roles:
-    - startworkers
+    - k8s_start_workers
+  tags: kubernetes
 
 - name: Start K8s worker servers on manager nodes
   hosts: manager
   gather_facts: false
   roles:
-    - startservices
+    - k8s_start_services
+  tags: kubernetes
 
-- name: Apply SLURM manager config
-  hosts: manager
+- name: Apply common Slurm installation and config
+  hosts: manager, compute
   gather_facts: false
   roles:
-    - slurm_manager
+    - slurm_common
   tags: slurm
 
-- name: Apply common Slurm installation and config
-  hosts: manager, compute
+- name: Apply Slurm manager config
+  hosts: manager
   gather_facts: false
   roles:
-    - slurm_common
+    - slurm_manager
   tags: slurm
 
-- name: Start slurm workers
+- name: Start Slurm workers
   hosts: compute
   gather_facts: false
   roles:
-    - start_slurm_workers
+    - slurm_workers
   tags: slurm
 
 - name: Start Slurm services
@@ -87,3 +139,10 @@
   roles:
     - slurm_start_services
   tags: slurm
+
+- name: Install slurm exporter
+  hosts: manager
+  gather_facts: false
+  roles:
+    - slurm_exporter
+  tags: slurm

+ 1 - 1
roles/compute_gpu/files/daemon.json

@@ -6,4 +6,4 @@
     }
   },
   "default-runtime": "nvidia"
-}
+}

+ 18 - 0
roles/common/files/inventory.fact

@@ -0,0 +1,18 @@
+#!/bin/bash
+INVENTORY=$(mktemp lspci.XXXXXXXX)
+
+lspci > $INVENTORY
+
+NVIDIA_GPU=$(cat $INVENTORY | grep -i nvidia | wc -l)
+XILINX_FPGA=$(cat $INVENTORY | grep "Processing accelerators: Xilinx Corporation Device" | wc -l)
+INTEL_A10_FPGA=$(cat $INVENTORY | grep "Processing accelerators: Intel Corporation Device" | wc -l)
+
+cat << EOF
+{
+	"xilinx_fpga" : $XILINX_FPGA,
+	"nvidia_gpu" : $NVIDIA_GPU,
+	"intel_a10_fpga" : $INTEL_A10_FPGA
+}
+EOF
+
+rm -f $INVENTORY

+ 22 - 12
roles/compute_gpu/vars/main.yml

@@ -13,18 +13,28 @@
 #  limitations under the License.
 ---
 
-nvidia_docker_repo_url: https://nvidia.github.io/nvidia-docker/centos7/nvidia-docker.repo
+- name: Restart ntpd
+  systemd:
+    name: ntpd
+    state: started
+    enabled: yes
 
-nvidia_docker_repo_dest: /etc/yum.repos.d/nvidia-docker.repo
+- name: Restart chrony
+  service:
+    name: chronyd
+    state: restarted
+    enabled: yes
 
-nvidia_container_repo_url: https://nvidia.github.io/libnvidia-container/centos7/libnvidia-container.repo 
+- name: Sync tp clocks
+  command: ntpdc -np
+  register: ntp_clock
+  until:  ntp_clock.stdout.find('*') > -1
+  retries: "{{ retry_count_one }}"
+  delay: "{{ delay_count_one }}"
 
-nvidia_container_repo_dest: /etc/yum.repos.d/libnvidia-container.repo
-
-nvidia_packages:
-  - kmod-nvidia
-  - nvidia-docker2
-
-daemon_file_dest: /etc/docker/
-
-daemon_file_mode: 0644
+- name: Sync chrony sources
+  command: chronyc sources
+  register: chrony_src
+  until:  chrony_src.stdout.find('^*') > -1
+  retries: "{{ retry_count }}"
+  delay: "{{ delay_count }}"

+ 73 - 0
roles/common/tasks/main.yml

@@ -0,0 +1,73 @@
+#  Copyright 2020 Dell Inc. or its subsidiaries. All Rights Reserved.
+#
+#  Licensed under the Apache License, Version 2.0 (the "License");
+#  you may not use this file except in compliance with the License.
+#  You may obtain a copy of the License at
+#
+#      http://www.apache.org/licenses/LICENSE-2.0
+#
+#  Unless required by applicable law or agreed to in writing, software
+#  distributed under the License is distributed on an "AS IS" BASIS,
+#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+#  See the License for the specific language governing permissions and
+#  limitations under the License.
+---
+
+- name: Create a custom fact directory on each host
+  file:
+    path: "{{ custom_fact_dir }}"
+    state: directory
+    mode: "{{ custom_fact_dir_mode }}"
+
+- name: Install accelerator discovery script
+  copy:
+    src: inventory.fact
+    dest: "{{ accelerator_discovery_script_dest }}"
+    owner: root
+    group: root
+    mode: "{{ accelerator_discovery_script_mode }}"
+
+- name: Add elrepo GPG key
+  rpm_key:
+    state: present
+    key: "{{ elrepo_gpg_key_url }}"
+  tags: install
+
+- name: Add elrepo (nvidia kmod drivers)
+  package:
+    name: "{{ elrepo_rpm_url }}"
+    state: present
+  tags: install
+
+- name: Disable swap
+  command: /sbin/swapoff -a
+  changed_when: true
+  tags: install
+
+- name: Disable selinux
+  selinux:
+    state: disabled
+  tags: install
+
+- name: Install common packages
+  package:
+    name: "{{ common_packages }}"
+    state: present
+  tags: install
+
+- name: Collect host facts (including acclerator information)
+  setup: ~
+
+- name: Install infiniBand support
+  package:
+    name: "@Infiniband Support"
+    state: present
+  tags: install
+
+- name: Deploy time ntp/chrony
+  include_tasks: ntp.yml
+  tags: install
+
+- name: Install Nvidia drivers and software components
+  include_tasks: nvidia.yml
+  when: ansible_local.inventory.nvidia_gpu > 0

+ 56 - 0
roles/common/tasks/ntp.yml

@@ -0,0 +1,56 @@
+#  Copyright 2020 Dell Inc. or its subsidiaries. All Rights Reserved.
+#
+#  Licensed under the Apache License, Version 2.0 (the "License");
+#  you may not use this file except in compliance with the License.
+#  You may obtain a copy of the License at
+#
+#      http://www.apache.org/licenses/LICENSE-2.0
+#
+#  Unless required by applicable law or agreed to in writing, software
+#  distributed under the License is distributed on an "AS IS" BASIS,
+#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+#  See the License for the specific language governing permissions and
+#  limitations under the License.
+---
+
+  - name: Deploy ntp servers
+    block:
+      - name: Deploy ntpd
+        package:
+          name: ntp
+          state: present
+      - name: Deploy ntpdate
+        package:
+          name: ntpdate
+          state: present
+      - name: Update ntp servers
+        template:
+          src: ntp.conf.j2
+          dest: "{{ ntp_path }}"
+          owner: root
+          group: root
+          mode: "{{ ntp_mode }}"
+          backup: yes
+        notify:
+          - Restart ntpd
+          - Sync tp clocks
+    when:  ( ansible_distribution == "CentOS" or   ansible_distribution == "RedHat" ) and ansible_distribution_major_version  < os_higher_version
+
+  - name: Deploy chrony server
+    block:
+      - name: Deploy chrony
+        package:
+            name: chrony
+            state: present
+      - name: Update ntp servers
+        template:
+          src: chrony.conf.j2
+          dest: "{{ chrony_path }}"
+          owner: root
+          group: root
+          mode: "{{ ntp_mode }}"
+          backup: yes
+        notify:
+          - Restart chrony
+          - Sync chrony sources
+    when:  ( ansible_distribution == "CentOS" or   ansible_distribution == "RedHat" ) and ansible_distribution_major_version  > os_version

+ 1 - 1
roles/compute_gpu/tasks/main.yml

@@ -28,7 +28,7 @@
 - name: Install nvidia driver and nvidia-docker2
   package:
     name: "{{ nvidia_packages }}"
-    enablerepo: libnvidia-container, nvidia-docker
+    enablerepo: libnvidia-container,nvidia-docker
     state: present
   tags: install
 

+ 41 - 0
roles/common/templates/chrony.conf.j2

@@ -0,0 +1,41 @@
+# Use public servers from the pool.ntp.org project.
+# Please consider joining the pool (http://www.pool.ntp.org/join.html).
+{% for item in chrony_servers %}
+pool {{ item }} iburst
+{% endfor %}
+
+
+# Record the rate at which the system clock gains/losses time.
+driftfile /var/lib/chrony/drift
+
+# Allow the system clock to be stepped in the first three updates
+# if its offset is larger than 1 second.
+makestep 1.0 3
+
+# Enable kernel synchronization of the real-time clock (RTC).
+rtcsync
+
+# Enable hardware timestamping on all interfaces that support it.
+#hwtimestamp *
+
+# Increase the minimum number of selectable sources required to adjust
+# the system clock.
+#minsources 2
+
+# Allow NTP client access from local network.
+#allow 192.168.0.0/16
+
+# Serve time even if not synchronized to a time source.
+#local stratum 10
+
+# Specify file containing keys for NTP authentication.
+keyfile /etc/chrony.keys
+
+# Get TAI-UTC offset and leap seconds from the system tz database.
+leapsectz right/UTC
+
+# Specify directory for log files.
+logdir /var/log/chrony
+
+# Select which information is logged.
+#log measurements statistics tracking

+ 14 - 0
roles/common/templates/ntp.conf.j2

@@ -0,0 +1,14 @@
+driftfile /var/lib/ntp/drift
+
+restrict default nomodify notrap nopeer noquery
+
+restrict 127.0.0.1
+restrict ::1
+
+{% for item in ntp_servers %}
+server  {{ item }} iburst
+{% endfor %}
+
+includefile /etc/ntp/crypto/pw
+
+keys /etc/ntp/keys

+ 66 - 0
roles/common/vars/main.yml

@@ -0,0 +1,66 @@
+#  Copyright 2020 Dell Inc. or its subsidiaries. All Rights Reserved.
+#
+#  Licensed under the Apache License, Version 2.0 (the "License");
+#  you may not use this file except in compliance with the License.
+#  You may obtain a copy of the License at
+#
+#      http://www.apache.org/licenses/LICENSE-2.0
+#
+#  Unless required by applicable law or agreed to in writing, software
+#  distributed under the License is distributed on an "AS IS" BASIS,
+#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+#  See the License for the specific language governing permissions and
+#  limitations under the License.
+---
+
+common_packages:
+  - epel-release
+  - yum-plugin-versionlock
+  - gcc
+  - nfs-utils
+  - python3-pip
+  - bash-completion
+  - nvidia-detect
+  - chrony
+  - pciutils
+
+custom_fact_dir: /etc/ansible/facts.d
+
+custom_fact_dir_mode: 0755
+
+accelerator_discovery_script_dest: /etc/ansible/facts.d/inventory.fact
+
+accelerator_discovery_script_mode: 0755
+
+elrepo_gpg_key_url: https://www.elrepo.org/RPM-GPG-KEY-elrepo.org
+
+elrepo_rpm_url: https://www.elrepo.org/elrepo-release-7.el7.elrepo.noarch.rpm
+
+chrony_path: "/etc/chrony.conf"
+ntp_path: "/etc/ntp.conf"
+ntp_mode: "0644"
+os_higher_version: "8"
+os_version: "7"
+retry_count_one: "10"
+delay_count_one: "60"
+retry_count: "6"
+delay_count: "10"
+
+ntp_servers: 
+  - 0.centos.pool.ntp.org
+  - 1.centos.pool.ntp.org
+  - 2.centos.pool.ntp.org
+chrony_servers:
+  - 2.centos.pool.ntp.org
+
+nvidia_docker_repo_url: https://nvidia.github.io/nvidia-docker/centos7/nvidia-docker.repo
+nvidia_docker_repo_dest: /etc/yum.repos.d/nvidia-docker.repo
+nvidia_container_repo_url: https://nvidia.github.io/libnvidia-container/centos7/libnvidia-container.repo
+nvidia_container_repo_dest: /etc/yum.repos.d/libnvidia-container.repo
+
+nvidia_packages:
+  - kmod-nvidia
+  - nvidia-docker2
+
+daemon_file_dest: /etc/docker/
+daemon_file_mode: 0644

+ 0 - 3
roles/compute_gpu/files/k8s.conf

@@ -1,3 +0,0 @@
-net.bridge.bridge-nf-call-ip6tables = 1
-net.bridge.bridge-nf-call-iptables = 1
-

+ 0 - 8
roles/compute_gpu/files/kubernetes.repo

@@ -1,8 +0,0 @@
-[kubernetes]
-name=Kubernetes
-baseurl=https://packages.cloud.google.com/yum/repos/kubernetes-el7-x86_64
-enabled=1
-gpgcheck=1
-repo_gpgcheck=1
-gpgkey=https://packages.cloud.google.com/yum/doc/yum-key.gpg https://packages.cloud.google.com/yum/doc/rpm-package-key.gpg
-

+ 2 - 2
roles/k8s_common/handlers/main.yml

@@ -18,11 +18,11 @@
     name: docker
     state: restarted
     enabled: yes
-  #tags: install
+  tags: install
 
 - name: Start and Enable Kubernetes - kubelet
   service:
     name: kubelet
     state: started
     enabled: yes
-  #tags: install
+  tags: install

+ 2 - 30
roles/k8s_common/tasks/main.yml

@@ -22,18 +22,6 @@
     mode: "{{ k8s_repo_file_mode }}"
   tags: install
 
-- name: Add elrepo GPG key
-  rpm_key:
-    state: present
-    key: "{{ elrepo_gpg_key_url }}"
-  tags: install
-
-- name: Add elrepo (nvidia kmod drivers)
-  package:
-    name: "{{ elrepo_rpm_url }}"
-    state: present
-  tags: install
-
 - name: Add docker community edition repository
   get_url:
     url: "{{ docker_repo_url }}"
@@ -54,19 +42,9 @@
   changed_when: true
   tags: install
 
-- name: Disable swap
-  command: /sbin/swapoff -a
-  changed_when: true
-  tags: install
-
-- name: Disable selinux
-  selinux:
-    state: disabled
-  tags: install
-
-- name: Install common packages
+- name: Install docker
   package:
-    name: "{{ common_packages }}"
+    name: docker-ce
     state: present
   tags: install
 
@@ -85,12 +63,6 @@
   changed_when: true
   tags: install
 
-- name: Install infiniBand support
-  package:
-    name: "@Infiniband Support"
-    state: present
-  tags: install
-
 - name: Start and enable docker service
   service:
     name: docker

+ 0 - 14
roles/k8s_common/vars/main.yml

@@ -13,16 +13,6 @@
 #  limitations under the License.
 ---
 
-common_packages:
-  - epel-release
-  - yum-plugin-versionlock
-  - gcc
-  - nfs-utils
-  - python3-pip
-  - docker-ce
-  - bash-completion
-  - nvidia-detect
-
 k8s_packages:
   - kubelet-1.16.7
   - kubeadm-1.16.7
@@ -30,10 +20,6 @@ k8s_packages:
 
 k8s_repo_dest: /etc/yum.repos.d/
 
-elrepo_gpg_key_url: https://www.elrepo.org/RPM-GPG-KEY-elrepo.org
-
-elrepo_rpm_url: https://www.elrepo.org/elrepo-release-7.el7.elrepo.noarch.rpm
-
 docker_repo_url: https://download.docker.com/linux/centos/docker-ce.repo
 
 docker_repo_dest: /etc/yum.repos.d/docker-ce.repo

+ 1 - 1
roles/k8s_start_manager/tasks/main.yml

@@ -146,7 +146,7 @@
   changed_when: true
   tags: init
 
-- name: Edge / Workstation Install allows pods to scheudle on manager
+- name: Edge / Workstation Install allows pods to schedule on manager
   command: kubectl taint nodes --all node-role.kubernetes.io/master-
   when: single_node
   tags: init

+ 1 - 1
roles/k8s_start_services/vars/main.yml

@@ -27,7 +27,7 @@ k8s_dashboard_yaml_url: https://raw.githubusercontent.com/kubernetes/dashboard/v
 
 helm_stable_repo_url: https://charts.helm.sh/stable
 
-nfs_server: {{ ansible_host }}
+nfs_server: "{{ ansible_host }}"
 
 nfs_path: /home/k8snfs
 

+ 0 - 156
test/test_compute_gpu.yml

@@ -1,156 +0,0 @@
-#  Copyright 2020 Dell Inc. or its subsidiaries. All Rights Reserved.
-#
-#  Licensed under the Apache License, Version 2.0 (the "License");
-#  you may not use this file except in compliance with the License.
-#  You may obtain a copy of the License at
-#
-#      http://www.apache.org/licenses/LICENSE-2.0
-#
-#  Unless required by applicable law or agreed to in writing, software
-#  distributed under the License is distributed on an "AS IS" BASIS,
-#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-#  See the License for the specific language governing permissions and
-#  limitations under the License.
----
-
-# OMNIA_UKP_US_VFKP_TC_003
-# Execute compute_gpu role in compute nodes with os installed centos 7.8
-- name: OMNIA_UKP_US_VFKP_TC_003
-  hosts: gpus
-  vars_files:
-    - test_vars/test_compute_gpu_vars.yml
-  tasks:
-    - block:
-        - name: Call compute_gpu role
-          include_role:
-            name: ../roles/compute_gpu
-      tags: TC_003
-
-    - name: Checking Nvidia drivers installation
-      command: whereis nvidia
-      register: nvidia_drivers
-      changed_when: false
-      ignore_errors: True
-      tags: TC_003, VERIFY_003
-
-    - name: Checking nvidia-docker2 installation status
-      command: nvidia-docker version
-      register: nvidia_docker_version
-      changed_when: false
-      ignore_errors: True
-      tags: TC_003, VERIFY_003
-
-    - name: Checking docker service status
-      systemd:
-        name: docker
-      register: docker_service
-      tags: TC_003, VERIFY_003
-
-    - name: Checking K8s service status
-      systemd:
-        name: kubelet
-      register: kubelet_service
-      tags: TC_003, VERIFY_003
-
-    - name: Validating Nvidia drivers
-      assert:
-        that:
-          - "'/usr/lib/nvidia' in nvidia_drivers.stdout"
-        fail_msg: "{{ nvidia_drivers_fail_msg }}"
-        success_msg: "{{ nvidia_drivers_success_msg }}"
-      tags: TC_003, VERIFY_003
-
-    - name: Validating nvidia-docker2 installation
-      assert:
-        that:
-          - "'command not found' not in nvidia_docker_version.stdout"
-        fail_msg: "{{ nvidia_docker_fail_msg }}"
-        success_msg: "{{ nvidia_docker_success_msg }}"
-      tags: TC_003, VERIFY_003
-
-    - name: Validating docker service status
-      assert:
-        that:
-          - docker_service.status.ActiveState == 'active'
-        fail_msg: "{{ docker_service_fail_msg }}"
-        success_msg: "{{ docker_service_success_msg }}"
-      tags: TC_003, VERIFY_003
-
-    - name: Validating K8s service status
-      assert:
-        that:
-          - kubelet_service.status.ActiveState == 'active'
-        fail_msg: "{{ kubelet_service_fail_msg }}"
-        success_msg: "{{ kubelet_service_success_msg }}"
-      tags: TC_003, VERIFY_003
-
-# OMNIA_UKP_US_VFKP_TC_004
-# Execute compute_gpu role in compute nodes with NVIDIA kmod/docker drivers already installed
-- name: OMNIA_UKP_US_VFKP_TC_004
-  hosts: gpus
-  vars_files:
-    - test_vars/test_compute_gpu_vars.yml
-  tasks:
-    - block:
-        - name: Call compute_gpu role
-          include_role:
-            name: ../roles/compute_gpu
-      tags: TC_004
-
-    - name: Checking Nvidia drivers installation
-      command: whereis nvidia
-      register: nvidia_drivers
-      changed_when: false
-      ignore_errors: True
-      tags: TC_004, VERIFY_004
-
-    - name: Checking nvidia-docker2 installation status
-      command: nvidia-docker version
-      register: nvidia_docker_version
-      changed_when: false
-      ignore_errors: True
-      tags: TC_004, VERIFY_004
-
-    - name: Checking docker service status
-      systemd:
-        name: docker
-      register: docker_service
-      tags: TC_004, VERIFY_0042
-
-    - name: Checking K8s service status
-      systemd:
-        name: kubelet
-      register: kubelet_service
-      tags: TC_004, VERIFY_004
-
-    - name: Validating Nvidia drivers
-      assert:
-        that:
-          - "'/usr/lib/nvidia' in nvidia_drivers.stdout"
-        fail_msg: "{{ nvidia_drivers_fail_msg }}"
-        success_msg: "{{ nvidia_drivers_success_msg }}"
-      tags: TC_004, VERIFY_004
-
-    - name: Validating nvidia-docker2 installation
-      assert:
-        that:
-          - "'command not found' not in nvidia_docker_version.stdout"
-        fail_msg: "{{ nvidia_docker_fail_msg }}"
-        success_msg: "{{ nvidia_docker_success_msg }}"
-      tags: TC_004, VERIFY_004
-
-    - name: Validating docker service status
-      assert:
-        that:
-          - docker_service.status.ActiveState == 'active'
-        fail_msg: "{{ docker_service_fail_msg }}"
-        success_msg: "{{ docker_service_success_msg }}"
-      tags: TC_004, VERIFY_004
-
-    - name: Validating K8s service status
-      assert:
-        that:
-          - kubelet_service.status.ActiveState == 'active'
-        fail_msg: "{{ kubelet_service_fail_msg }}"
-        success_msg: "{{ kubelet_service_success_msg }}"
-      tags: TC_004, VERIFY_004

+ 0 - 22
test/test_k8s_common.yml

@@ -44,16 +44,6 @@
       ignore_errors: True
       tags: TC_001, VERIFY_001
 
-    - name: Checking nfs-utils and versionlock status
-      command: "{{ item }}"
-      with_items:
-        - "yum versionlock"
-        - "nfsstat"
-      register: nfs_versionlock_status
-      changed_when: false
-      ignore_errors: True
-      tags: TC_001, VERIFY_001
-
     - name: Checking docker service status
       systemd:
         name: docker
@@ -75,7 +65,6 @@
       with_items:
         - "{{ common_packages_status.results }}"
         - "{{ k8s_packages_status.results }}"
-        - "{{ nfs_versionlock_status.results }}"
       tags: TC_001, VERIFY_001
 
     - name: Validating docker service status
@@ -125,16 +114,6 @@
       ignore_errors: True
       tags: TC_002, VERIFY_002
 
-    - name: Checking nfs-utils and versionlock status
-      command: "{{ item }}"
-      with_items:
-        - "yum versionlock"
-        - "nfsstat"
-      register: nfs_versionlock_status
-      changed_when: false
-      ignore_errors: True
-      tags: TC_002, VERIFY_002
-
     - name: Checking docker service status
       systemd:
         name: docker
@@ -157,7 +136,6 @@
       with_items:
         - "{{ common_packages_status.results }}"
         - "{{ k8s_packages_status.results }}"
-        - "{{ nfs_versionlock_status.results }}"
       tags: TC_002, VERIFY_002
 
     - name: Validating docker service status

+ 1 - 1
test/test_k8s_manager.yml

@@ -14,7 +14,7 @@
 ---
 
 # OMNIA_UKP_US_VFKP_TC_005
-# Execute k8s_manager role in manager nodes with os installed centos 7.8
+# Execute k8s_manager role in manager nodes with os installed centos 7.9
 - name: OMNIA_UKP_US_VFKP_TC_005
   hosts: manager
   vars_files:

+ 2 - 2
test/test_k8s_start_manager_workers.yml

@@ -14,7 +14,7 @@
 ---
 
 # OMNIA_UKP_US_VFKP_TC_009, OMNIA_UKP_US_VFKP_TC_010
-# Execute k8s_start_manager role in manager nodes with os installed centos 7.8 and swap enabled
+# Execute k8s_start_manager role in manager nodes with os installed centos 7.9 and swap enabled
 - name: OMNIA_UKP_US_VFKP_TC_009, OMNIA_UKP_US_VFKP_TC_010
   hosts: manager
   gather_facts: false
@@ -106,7 +106,7 @@
       tags: TC_009, TC_010, VERIFY_009
 
 # OMNIA_UKP_US_VFKP_TC_011, OMNIA_UKP_US_VFKP_TC_012
-# Execute k8s_start_workers role in compute nodes with os installed centos 7.8 and swap enabled
+# Execute k8s_start_workers role in compute nodes with os installed centos 7.9 and swap enabled
 - name: OMNIA_UKP_US_VFKP_TC_011, OMNIA_UKP_US_VFKP_TC_012
   hosts: compute
   gather_facts: false

+ 1 - 1
test/test_k8s_start_workers.yml

@@ -14,7 +14,7 @@
 ---
 
 # OMNIA_UKP_US_VFKP_TC_013, OMNIA_UKP_US_VFKP_TC_014
-# Execute startworkers role in compute nodes with os installed centos 7.8 and swap enabled
+# Execute startworkers role in compute nodes with os installed centos 7.9 and swap enabled
 - name: OMNIA_UKP_US_VFKP_TC_013, OMNIA_UKP_US_VFKP_TC_014
   hosts: compute
   vars_files:

+ 0 - 30
test/test_vars/test_compute_gpu_vars.yml

@@ -1,30 +0,0 @@
-#  Copyright 2020 Dell Inc. or its subsidiaries. All Rights Reserved.
-#
-#  Licensed under the Apache License, Version 2.0 (the "License");
-#  you may not use this file except in compliance with the License.
-#  You may obtain a copy of the License at
-#
-#      http://www.apache.org/licenses/LICENSE-2.0
-#
-#  Unless required by applicable law or agreed to in writing, software
-#  distributed under the License is distributed on an "AS IS" BASIS,
-#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-#  See the License for the specific language governing permissions and
-#  limitations under the License.
----
-
-nvidia_drivers_fail_msg: "Nvidia drivers are not installed"
-
-nvidia_drivers_success_msg: "Nvidia drivers are installed"
-
-nvidia_docker_fail_msg: "nvidia-docker2 package is not installed"
-
-nvidia_docker_success_msg: "nvidia-docker2 package is installed"
-
-docker_service_fail_msg: "Docker service is not running"
-
-docker_service_success_msg: "Docker service is running"
-
-kubelet_service_fail_msg: "K8s service is not running"
-
-kubelet_service_success_msg: "K8s service is running"

+ 1 - 5
test/test_vars/test_k8s_common_vars.yml

@@ -14,12 +14,8 @@
 ---
 
 common_packages:
-  - gcc
-  - python3
-  - pip3
   - docker
-  - nvidia-detect
-  - kubelet
+  - kubelet 
 
 k8_packages:
   - kubeadm