Group 1 Project
Group 1 Project
on
Deploying HPC Cluster
On Containers
Guided by:
HOD ACTS
Mr.
PG-DHPCSA
ACKNOWLEDGEMENT
This project “ Deploying HPC Cluster on Containers ” was a great learning experience for us and
we are submitting this work toAdvanced Computing Training School (CDAC ACTS).
We all are very glad to mention the name of Mr. Ashutosh Das for his valuable guidance to work on this
project. Overcome various obstacles and intricacies during the course of project work.
We are highly grateful to HPC tech team (ACTS training Centre, C-DAC),
For his valuable guidance and support whenever necessary while doing this course Post
Graduate Diploma in High Performance Computing System Administration (PG- DHPCSA)
Through CDAC ACTS , Pune.
Our most heartfelt thank goes to Ms. Swati salunkhe (Course Coordinator, PG- DHPCSA)
who gave all the required support and kind coordination to provide all the necessities like
required hardware, internet facility and extra Lab hours to complete the project and
throughout the course up to the last day here in C-DAC ACTS, Pune.
From:
1. Introduction
2. Workflow
3. System Requirements
a. Software
b. Hardware
4. Setting up the Master
5. Installation of Warewulf
7. Installation of Ganglia
8. Troubleshooting
10.Project Link
P a g e 4 | 44
Introduction
Resources and accounting. This HPC stacks uses Rocky 8.8 Linux as an
P a g e 5 | 44
Workflow
P a g e 6 | 44
System Requirements
• RAM: 16 Gb
• HDD: 100 Gb
• Processors: 4 cores
• Network adapters
o NAT
o Host Only
Software Requirements
P a g e 7 | 44
Setting up the master
Step 2: Click next & choose `I will install Operating System Later`
P a g e 8 | 44
Step 3: Select Guest Operating System Type
P a g e 9 | 44
Step 5: Select no of processors for master
P a g e 10 | 44
Step 8: Next step is to select disk & it’s size
P a g e 11 | 44
Step 9: Next Step is to add another network adapter & change its type to Host-
only
P a g e 12 | 44
Step 10: Final step is to confirm the configuration and click finish
P a g e 13 | 44
Installing OS on new VM
Step 1: select boot device and choose rocky Linux 8.8 iso & click open. Then
start the virtual machine
Step 2: After vm is started select the first option & press enter key
P a g e 14 | 44
Step 3: Next step is to select language & press Continue button.
Step 4: setup the root password, Installation destination & click Begin Installation
P a g e 15 | 44
Step 5: After system has been installed click Reboot System
P a g e 16 | 44
Step 7: Next step is to accept the agreement & click finish Configuration
Step 8: in Next step, Enter the username & password on the login screen
P a g e 17 | 44
Step 9: Finally, we land the Rock Linux desktop Environment
P a g e 18 | 44
Warewulf Installation
Step 1: First, we need to disable the SELINUX, firewall and change hostname
P a g e 19 | 44
Step 3: Next step is to install warewulf
P a g e 20 | 44
Step 5: Next step is to configure the warewulf
P a g e 21 | 44
Step 7: Next step is to import container from docker hub & set password for it.
P a g e 22 | 44
Step 8: Next step is to create a new node and test if it is booting or not.
Warning : RAM given to Node VM must be greater than the Container Size
P a g e 23 | 44
P a g e 24 | 44
Step 9: add this node to warewulf
P a g e 25 | 44
P a g e 26 | 44
Installation of MUNGE
Step 3: installing random number generator tools, this will help in generating the
munge key
P a g e 27 | 44
Step 4: Starting the rngd service
P a g e 28 | 44
Step 7: installing the munge on the container
Step 8: copying the munge key from shared folder to munge path & changing it’s
ownership. Finally, we have to rebuild the container
P a g e 29 | 44
Step 9: checking status of munge service on Compute Node & Master
P a g e 30 | 44
Installation of SLURM
P a g e 31 | 44
Step 4 : Next step is to build rpm from tar file
Step 6 : Next step is to check rpmbuild packages after rpmbuild packages has
been build
P a g e 32 | 44
Step 7 : Next step is to install rpmbuild packages
Step 9 : Next step is to make directory , change ownership and give permission
to Slurm User
P a g e 33 | 44
Step 10 : Next step is to install Slurm on container
P a g e 34 | 44
Note: to get this slurm configuration from node, node has to be in a booted state.
Otherwise, this command will show master machine configuration.
Step 12: Next step is to change ownership to slurm and restart the service
Step 13 : Next step is to create spool & log directories for Slurmd on Container
P a g e 35 | 44
Step 15 : Next step is to reboot the Container & check all services
P a g e 36 | 44
Installation of Ganglia
P a g e 37 | 44
Step 7 : Next step is to start the services
P a g e 38 | 44
Step 8 : Next step is to install ganglia on container
P a g e 39 | 44
Step 10 : Next step is to add the gmond service in bashrc file
Step 11 : Next step is to reboot the node & check the status of the gmond service
Step 12 : Next step is to open browser on master & check the ganglia cluster output
https://localhost/ganglia
Cluster Output:
P a g e 40 | 44
Node Result:
Master result:
P a g e 41 | 44
Troubleshooting
P a g e 42 | 44
Error 2: Failed to set Locale, defaulting to C.UTF-8
P a g e 43 | 44
References & Bibliography
1. Warewulf Documentation
https://warewulf.org/docs/development/
2. Slurm Documentation
https://slurm.schedmd.com/documentation.html
Project Link
Github: https://github.com/ravi30flash/HPC-project/tree/master
P a g e 44 | 44