Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

These are the different QoS (or queues) available for standard users on the four complexes:

QoS nameTypeSuitable for...Shared nodes Maximum jobs per user
Default / Max Wall Clock LimitDefault / Max CPUsDefault / Max Memorynffractionalserial and small parallel jobs. It is the defaultYes-average runtime + standard deviation / 2 days1 / 648 GB / 128 GBnpparallelparallel jobs requiring more than half a nodeNo-average runtime  + standard deviation / 2 days-240GB / 240 GB per node (all usable memory in a node)

GPU special Partition

On the AC complex you will find dedicated queues for special partition with GPU-enabled nodes. See AG: GPU usage for AI and Machine Learning for all the details on how to make use of those special resources.

...

Maximum

...

nodes per userDefault / Max Wall Clock LimitDefault / Max CPUsDefault / Max Memory per node
ngGPUserial and small parallel jobs with GPU. It is de defaultYes-4average runtime + standard deviation / 2 days1 / -8 GB /  500 GB
dgGPUshort debug jobs requiring GPUYes12average runtime + standard deviation / 30 min1 / -8 GB /  500

...

ECS

For those using ECS, these are the different QoS (or queues) available for standard users of this service:

QoS nameTypeSuitable for...Shared nodes Maximum jobs per userDefault / Max Wall Clock LimitDefault / Max CPUsDefault / Max Memoryeffractionalserial and small parallel jobs - ECGATE serviceYes-average job runtime + standard deviation / 2 days1 / 88
GB
/ 16 GBellongserial and small parallel interactive jobs - ECGATE serviceYes-
average job runtime + standard deviation  / 7 days1 / 88 GB / 16 GBetTime-critical Option 1

serial and small parallel Time-Critical jobs.

Only usable through ECACCESS Time Critical Option-1

Yes-average job runtime + standard deviation  / 12 hours1 / 88 GB / 16 GB

Interactive sessions - ecinteractive

Using the "ecinteractive" command, jobs will be submitted in either of these queues depending on if the user has access to the full HPC service (ni queue, for Member State users) or the ECS service (ei queue, for co-operating States, as service previously known as ECGATE)


Multiexcerpt
MultiExcerptNameecinteractive
QoS nameTypeSuitable for...Shared nodes Maximum jobs per userDefault / Max Wall Clock LimitDefault / Max CPUsDefault / Max Memory
niinteractiveserial and small parallel interactive jobsYes112 hours / 7 days1 / 328 GB / 32 GB
eiinteractiveserial and small parallel interactive jobs - ECGATE serviceYes112 hours  / 7 days1 / 48 GB / 8 GB
Info
titleTime limit management

See AG: Job Runtime Management for more information on how the default Wall Clock Time limit is calculated.

...

Tip
titleChecking QoS setup

If you want to get all the details of a particular QoS on the system, you may run, for example:

No Format
sacctmgr list qos names=nfng


Submitting jobs remotely

If you are submitting jobs from a different platform via ssh, please use the *ag-batch dedicated nodes node instead of the *-login equivalents:

  • For generic remote job submission on HPCF: hpc-batch or hpc2020-batch
  • For remote job submission on a specific HPCF complex: <complex_name>-batch
  • For remote job submission to the ECS virtual complex: ecs-batch

For example, to submit a job from a remote platform onto the Atos HCPF: equivalent

No Format
ssh hpcag-batch "sbatch myjob.sh"


HTML
<style>
div#content h2 a::after {
 content: " - [read more]";
}
</style>

...