Slurm is the batch system available. Any script can be submitted as a job with no changes, but you might want to see Writing SLURM jobs to customise it.
To submit a script as a serial job with default options enter the command:
sbatch yourscript.sh |
You may query the queues to see the jobs currently running or pending with:
squeue |
And cancel a job with
scancel <jobid> |
The "scancel" command should be executed on a login node on the same cluster as the job. |
See the Slurm documentation for more details on the different commands available to submit, query or cancel jobs.
These are the different QoS (or queues) available for standard users on the four complexes:
| QoS name | Type | Suitable for... | Shared nodes | Maximum jobs per user | Maximum nodes per user | Default / Max Wall Clock Limit | Default / Max CPUs | Default memory per cpu |
|---|---|---|---|---|---|---|---|---|
| ng | GPU | serial and small parallel jobs with GPU. It is de default | Yes | - | 4 | average runtime + standard deviation / 2 days | 1 / - | 2900 MB |
| dg | GPU | short debug jobs requiring GPU | Yes | 1 | 2 | average runtime + standard deviation / 30 min | 1 / - | 2900 MB |
Different limits on the different QoSs may be introduced or changed as the system evolves. |
If you want to get all the details of a particular QoS on the system, you may run, for example:
|
If you are submitting jobs from a different platform via ssh, please use the ag-batch dedicated node instead of the *-login equivalent
ssh ag-batch "sbatch myjob.sh" |
<style>
div#content h2 a::after {
content: " - [read more]";
}
</style> |