Learn how to effectively schedule and manage your system workload using Slurm; the free, open source and highly scalable cluster management and job scheduling system for Linux clusters. Slurm is in use today on roughly half of the largest systems in the world servicing a broad spectrum of applications. Slurm developers have been working closely with NVIDIA to provide capabilities specifically focused on the needs of GPU management. This includes a multitude of new options to specify GPU requirements for a job in various ways (GPU count per job, node, socket and/or task), additional resource requirements for allocated GPUs (CPUs and/or memory per GPU), how spawned tasks should be bound to allocated GPUs, and control over GPU frequency and voltage. An introduction to Slurm's design and capabilities will be presented with a focus on managing workloads for GPUs.