Versionen im Vergleich

Schlüssel

  • Diese Zeile wurde hinzugefügt.
  • Diese Zeile wurde entfernt.
  • Formatierung wurde geändert.
Kommentar: updated partition chart to reflect new test nodes

...

Slurm partitionNode numberCPUMain memory (GB)GPUs per nodeGPU hardwareWalltime (hh:mm:ss)Description
gpu-a1003634Ice Lake 8360Y10004NVIDIA Tesla A100 80GB 24:00:00full node exclusive
gpu-a100:shared54NVIDIA Tesla A100 80GB shared node access, exclusive use of the requested GPUs
gpu-a100:shared:mig128 (4 x 7)1 to 28 1g.10gb A100 MIG slices

shared node access, shared GPU devices via Multi Instance GPU. Each of the four GPUs is logically split into usable seven slices with 10 GB of GPU memory associated to each slice

gpu-a100:test24NVIDIA Tesla A100 80GB 01:00:00nodes reserved for short job tests before scheduling longer jobs with more resources

See Slurm usage how to pass a 24h walltime limit with job dependencies.

...