This article was automatically translated from the original Turkish version.
High Performance Computing (HPC) systems play a critical role in fields such as scientific research simulations and data-intensive engineering applications. Efficient utilization of computational resources in these systems requires advanced job schedulers and resource managers. Slurm (Simple Linux Utility for Resource Management) is a leading open-source job scheduling and resource management system in this domain.
Slurm is an open-source resource management and job scheduling software originally developed by Hewlett-Packard and currently maintained by SchedMD. Slurm manages job execution on Linux-based cluster systems. It is used to fairly and efficiently queue schedule and allocate resources for users on supercomputers consisting of thousands of nodes.
Although designed for large-scale systems Slurm is also widely used in medium-sized research clusters.
The Slurm architecture consists of the following main components:
Slurm supports various scheduling policies to ensure fair resource sharing among jobs. For example jobs are ordered using strategies such as Priority Backfill and FairShare. Users can submit jobs using commands like sbatch srun and salloc and monitor job status with squeue and sacct.
Slurm supports multi-core and multi-node jobs and facilitates the coordinated execution of distributed workloads such as those using MPI (Message Passing Interface).
Slurm is extensible through plugins. For example:
In addition hardware resources can be precisely defined through Slurm configuration files such as slurm.conf gres.conf and cgroup.conf.
Slurm is successfully deployed on some of the world’s largest supercomputers including systems listed in the TOP500. It has the performance capability to manage tens of thousands of nodes and millions of jobs simultaneously.
Slurm is a powerful open-source and flexible platform used to manage job execution and resource planning in high performance computing systems. It is widely preferred in academic and institutional research environments and serves as the default system in the majority of supercomputing centers worldwide.
National Energy Research Scientific Computing Center (NERSC). *NERSC Slurm User Guide.* DOE Office of Science. Accessed May 5, 2025. https://docs.nersc.gov/jobs/slurm/.
SchedMD LLC. *Slurm Workload Manager Documentation.* Accessed May 5, 2025. https://slurm.schedmd.com/documentation.html.
TOP500.org. "TOP500 List – Supercomputers Using Slurm." Accessed May 5, 2025. https://www.top500.org/statistics/list/.
Texas Advanced Computing Center (TACC). *Using Slurm on HPC Clusters.* Accessed May 5, 2025. https://portal.tacc.utexas.edu/user-guides/slurm.
Yoo, Andy B., Marty Jette, and Mark Grondona. “Slurm: Simple Linux Utility for Resource Management.” *Job Scheduling Strategies for Parallel Processing*, edited by Dror G. Feitelson, Larry Rudolph, and Uwe Schwiegelshohn, 44–60. Berlin: Springer, 2003. .
What is Slurm?
Core Components
Use Cases
Job Scheduling and Queuing
Customization and Extensibility
Performance and Scalability
Comparison of Slurm with Other Systems