Exercise 1
Preparation:
- Login to the workshop machine
- Workshops differ in how this is done. The instructor will go over this beforehand.
- Copy the exercise files into your home directory then cd into it:
- cp -R /usr/global/docs/training/blaise/slurmmoab ~
- cd slurmmoab
- List the contents of your subdirectory. You should notice the following files:
Description | Files | |
---|---|---|
Simple shell script used for first exercise |
Document
|
|
Parallel program source code - C version |
Document
|
|
Parallel program source code - Fortran version |
Document
|
|
Hybrid parallel (MPI + threads) program source code - C version |
Document
|
|
Multiple jobs from single batch script example - C version |
Document
|
|
Solutions to exercises | Slurm Document
Document
Document
Document
Document
Document
Document
Document
|
Moab Document
Document
Document
Document
Document
Document
Document
Document
|
Review your cluster's batch configuration
- Try the commands below:
- news job.lim.machine - where machine is the name of the cluster
- sinfo -s
- mjstat | head
- Questions:
- Which queues are configured?
- How many nodes are there in each queue?
- What are the batch queue node and time limits?
- What states are the nodes in (alloc, idle, etc.)?
Find out which banks are available
- To see which banks (accounts) are available to you on this cluster, simply issue the mshare command. Note that it displays your bank allocation and usage information also.
- The mdiag -u classXX command can be used to list your banks (accounts) and also show your valid QOS options.
- To view the entire bank hierarchy, use the mshare -t root command.
Create and run a job script
- Using your favorite text editor (vi/vim, emacs, nedit, gedit, nano...), create a job script that does the following:
- Runs under the classXX login shell (/bin/tcsh)
- Sets a time limit of 5 minutes
- Requests 1 node
- Runs in the pReserved queue
- Writes the batch output to a file name of your choosing
- Gives your job a unique name
- Changes to your slurmmoab subdirectory
- Issues the name of the host it is running on
- Issues the jobid for this job
- Shows your path
- Runs the exercise1 executable provided to you
- Sleeps for a few minutes (so you can have time to check on it)
- For reference, you can review the appropriate solution file:
Documentor
Document
- Submit your job using either the sbatch (Slurm) command or the msub (Moab) command.
- Was your job script accepted?
- What was its jobid number?
- Problems? Check your script against the
Documentor
Documentsolution file for errors.
Monitor your job
- The tutorial described several ways to monitor your job, including:
- squeue
- mjstat
- showq
- checkjob
- mdiag -j
- Try any/all of these commands, noting their similarities and differences.
- Hint: you may want to pipe the output of the more verbose commands into grep with the jobid or your workshop username. For example:
- showq | grep class04
- If you run out of time, you can submit another job with a longer "sleep".
- If you have questions about the output of these commands, check the tutorial and/or man pages.
- After your job completes, examine its status using the checkjob and showq -c | grep jobid commands.
Check your job's output
- Review the output file from your job.
- Where did you find it?
- Is it named what you specified?
- Is the output what should be expected? Compare your output file to the
Documentor
Documentoutput file.
- You may also want to look at the
Documentexecutable.
This completes Exercise 1
Exercise 2
Still logged into the workshop cluster?
- If so, then continue to the next step. If not, then login as you did previously for Exercise 1.
Holding and releasing a job
- Using your same job script, submit the job so that it is held. This can be done on the msub or sbatch command line, or from within the script itself. Try both ways. If you have any questions see the tutorial.
- Verify that your job is actually in a holding state
- Release your job(s) so that they run to completion.
- Verify that the job release actually took effect
Canceling a job
- Once again, submit your job script.
- Try to cancel it before it completes. You can do this when its queued or when it's running. If you have any questions see the tutorial.
- Confirm that the job is actually cancelled. Also, check its post-execution status with the checkjob command.
Running in standby mode
- Modify your job script so that it will run in standby mode. The
Documentand
Documentfiles are provided for reference.
- Submit your job script.
- When your job starts to run, verify that it is running in standby mode. One way to do this is use the checkjob command and look for qos:standby near the top of the output.
- Submit your job script again but be sure to have the job HELD.
- Confirm that the job is held.
- Now change the qos from standby to normal for this job. If you have any questions see the Standby section of the tutorial.
- Confirm that the qos was changed.
- Cancel the job (or release it and let it run) when you're sure that it was changed.
Run a parallel job
- Using the slurm1 or moab1 example file, copy it to a new file - call it whatever you'd like.
- Modify your new file so that:
- Four nodes are requested
- A new output file name is used
- It compiles either mpi_array.c (use "mpicc") or mpi_array.f (use "mpif77").
- Lists the names of the nodes used to run the job
- Runs a 48-task MPI job using the mpi_array executable you created in the previous step.
- The
Documentand
Documentexample files are provided for reference.
- Submit your job and monitor it, making sure it is using the number of nodes/tasks specified.
- Check your output file to verify that things worked. See
Documentas a comparison.
Run a hybrid (MPI + threads) parallel job
- The example file mpithreads.c combines MPI with pthreads. The basic idea is to run one MPI task per node, and then spawn one thread for each core on that node. The threads do the actual work and MPI is used to collect the results across all nodes. Feel free to examine the source code if you'd like.
- Using the slurm3 or moab3 example file, copy it to a new file - call it whatever you'd like.
- Modify your new file so that:
- A new output file name is used
- Compiles the mpithreads.c file (use "mpicc -pthread")
- Runs a 4-task MPI job using the executable you created in the previous step. However, this time run with only one task per node. This will permit the threads spawned by each MPI task to use the available cores on a node without competition from the threads of other MPI tasks.
- The
Documentand
Documentexample files are provided for reference.
- Submit your job and monitor it, making sure it is using the number of nodes/tasks specified.
- Check your output file to verify that things worked. See
Documentas a comparison.
Run multiple jobs from a single batch script
- The
Documentand
Documentexample files demonstrate how to run multiple jobs from a single batch script.
- Review either example file and note what is being done:
- Four nodes are requested
- A simple executable is compiled
- Four 1-node jobs are launched to run simultaneously
- Submit either example file and then review its output when it completes.
When will my job start?
- Some of the most frequently asked questions by users include;
- "When will my job start?"
- "Why won't my job start?"
- "Why are jobs sitting idle when there are ample unused nodes?"
- There are several common answers to these questions. Assuming that there are no system problems or errors in the user's job submission script, one of the most common reasons has to do with a job's calculated priority and the scheduler's fair-share algorithms.
- Use one of the commands below to generate a list of eligible jobs and their priorities:
sprio -l | more mdiag -p -v | more
- As you scroll through the list, note that it is sorted by jobid.
- To make the list more meaningful, sort it by priority (highest to lowest):
sprio -l | sort -r -k 3,3 mdiag -p -v | sort -r -k 3,3
- You can now find where any job is relative to other jobs in the queue.
- Columns 4-9 show the factors used to compute priority values.
- You also use the checkjob jobid command to view the scheduler's current estimate on when your job will start. Look for the line that shows "StartTime:" (if it exists). For example:
% checkjob 87889 ... WallTime: 00:00:00 of 1-00:00:00 SubmitTime: Wed Jun 7 10:40:27 (Time Queued Total: 00:01:48 Eligible: 00:01:48) StartTime: Thu Jun 8 10:41:57 Total Requested Tasks: 1 Total Requested Nodes: 1 Partition: pbatch Dedicated Resources Per Task: lscratchf Node Access: SINGLEJOB ...
- Sometimes the squeue --start command can be used to get an estimate for job start times. And sometimes it can't...
- Note that start times can change dynamically if new jobs with a higher priority are submitted.
Try sview
- The sview utility provides a graphical view of all user jobs running on a cluster. Give it a try if you haven't already.
Documentation - if you still have time
- The best launching spot for LC Slurm / Moab documentation is: https://hpc.llnl.gov/banks-jobs/running-jobs
This completes the exercise.
Please complete the online evaluation form if you have not already done so for this tutorial.
Where would you like to go now?
- Agenda
- Back to the tutorial