Job Step Launch under Slurm and Cluster System Management (CSM)
For a list of the current Sandia platforms, with links to documentation, please see https://computing.sandia.gov/platforms.
This page provides a quick-start guide to Slurm, providing examples of how to perform common tasks.
This page provides a brief overview of batch system concepts. It is provided as a foundation to understanding the system for running jobs on the Sierra compute clusters.
Slurm is a combined batch scheduler and resource manager that allows users to run their jobs on Livermore Computing’s (LC) high performance computing (HPC) clusters. This document describes the process for submitting and running jobs under the Slurm Workload Manager.
Livermore Computing (LC) provides a large variety of High Performance Computing (HPC) clusters. However, there are only two batch schedulers that run user jobs on those clusters. Slurm is the batch scheduler and resource manager that schedules almost all LC clusters. The exception is the IBM Sierra clusters (aka CORAL systems) which run the Spectrum LSF scheduler.
We are scheduling large dedicated application runs on M&IC machines, usually during a weekend. Some of these runs may require that the entire machine be dedicated to them.
Retrieve files on the SCF by using SFTP/SCP, FTP or Hopper to make a copy of a file from the transfer server to your local machine. Initiate the FTP client software, connect to fastfis.llnl.gov, and complete the authentication process (by specifying a user name and SCF password). After you have connected to to fastfis.llnl.gov, change to your FROM directory and list the file(s) in the directory.
This page lists available online tutorials related to parallel programming and using LC's HPC systems.