For timing purposes you will need to create a batch job.
SLURM batch job example which you can download here.
Launching MPI-programs on our cluster with OpenMPI : do NOT use mpirun – instead srun (which is SLURM-based launcher). Otherwise with mpirun you end up creating detached tasks that do not communicate with each other.
To initialize device under OpenACC use the following coding in Fortran
To select MPI-task specific GPU-device (either 0 or 1 in our system), use the following coding in MPI-codes (with help of nodeinfo subroutine):
Go to the training home page