Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Slurm is the batch system available. Any script can be submitted as a job with no changes, but you might want to see Writing SLURM jobs to customise it.

To submit a script as a serial job with default options enter the command:

...

And cancel a job with

No Format
scancel <jobid>


Note

Currently the "scancel" command shall be executed on the login node of the same cluster where the job is running.

See the Slurm documentation for more details on the different commands available to submit, query or cancel jobs.

...

These are the different QoS (or queues) available for standard users on the four complexes:

QoS nameTypeSuitable for...Shared nodes Maximum jobs per userDefault / Max Wall Clock LimitDefault / Max CPUsDefault / Max Memory
nffractionalserial and small parallel jobs. It is the defaultYes-
2 day
average runtime + standard deviation / 2 days1 / 648 GB / 128 GB
niinteractiveserial and small parallel interactive jobsYes1
1 day  
12 hours / 7 days1 / 328 GB / 32 GB
npparallelparallel jobs requiring more than half a nodeNo-average runtime  + standard deviation / 2 days-240GB / 240 GB per node (all usable memory in a node)


Show If
groupecmwf

GPU special Partition

On the AC complex there is also the ng queue that gives access to the special partition with GPU-enabled nodes. See HPC2020: GPU usage for AI and Machine Learning for all the details on how to make use of those special resources.

Excerpt


QoS nameTypeSuitable for...Shared nodes Maximum jobs per userDefault / Max Wall Clock LimitDefault / Max CPUsDefault / Max Memory per node
ngGPUserial and small parallel jobs. It is the defaultYes-average runtime + standard deviation / 2 days1 / -

...

8 GB /  500 GB



ECS

For those using ECS, these are the different QoS (or queues) available for standard users of this service:

QoS nameTypeSuitable for...Shared nodes Maximum jobs per userDefault / Max Wall Clock LimitDefault / Max CPUsDefault / Max Memory
effractionalserial and small parallel jobs - ECGATE serviceYes-2 day average job runtime + standard deviation / 2 days1 / 88 GB / 16 GB
eiinteractiveserial and small parallel interactive jobs - ECGATE serviceYes11 day 12 hours  / 7 days1 / 48 GB / 8 GB
ellongserial and small parallel interactive jobs - ECGATE serviceYes-7 day average job runtime + standard deviation  / 7 days1 / 88 GB / 16 GB
etTime-critical Option 1

serial and small parallel Time-Critical jobs.

Only usable through ECACCESS Time Critical Option-1

Yes-12 hours average job runtime + standard deviation  / 12 hours1 / 88 GB / 16 GB


Info
titleTime limit management

See HPC2020: Job Runtime Management for more information on how the default Wall Clock Time limit is calculated.


Note
titleLimits are not set in stone

Different limits on the different QoSs may be introduced or changed as the system evolves.


Tip
titleChecking QoS setup

If you want to get all the details of a particular QoS on the system, you may run, for example:

No Format
sacctmgr list qos names=nf


Submitting jobs remotely

If you are submitting jobs from a different platform via ssh, please use the *-batch dedicated nodes instead of the *-login equivalents:

  • For generic remote job submission on HPCF: hpc-batch or hpc2020-batch
  • For remote job submission on a specific HPCF complex: <complex_name>-batch
  • For remote job submission to the ECS virtual complex: ecs-batch

For example, to submit a job from a remote platform onto the Atos HCPF: 

No Format
ssh hpc-batch "sbatch myjob.sh"
Note
titleWork in progress
Different limits on the different QoSs may be introduced or changed as the system evolves to its final configuration.


HTML
<style>
div#content h2 a::after {
 content: " - [read more]";
}
</style>

...