NHR center NHR@ZIB follows NHR-wide regulations.
...
NHR@ZIB operates system Lise which hold different Compute cluster containing different types of compute nodes each. The charge rates for the partitions are given in the table(slurm-)partitions representing the specific hardware.
Cluster | Partition | Charge (core hour) per 1 node per 1 h occupancy time | Increased charge rate due to |
---|---|---|---|
CPU cluster | standard96 standard96:test | 96 | |
large96 large96:test large96:shared | 144 | high memory layout | |
huge96 | 192 | high memory layout | |
GPU A100 cluster | gpu-a100 | 600 | four NVidia A100 (80 GB) per compute node |
gpu-a100:shared | 150 per GPU | 600 for four NVidia A100 (80 GB) per node | |
gpu-a100:shared:mig | 21.43 per MiG slice | four NVidia A100 (80 GB) splitted each into two 2g.10gb slices (8 per node and currently 24 in total) and one 3g.20gb slice (4 per node and currently 12 in total) | |
GPU PVC cluster | gpu-pvc | free of charge | test phase |
...
one node in partition | charged "core hours" per 1h occupancy time | increased charge rate due to |
---|---|---|
standard96 standard96:test | 96 | |
large96 large96:test large96:shared | 144 | high memory layout |
huge96 | 192 | high memory layout |
medium40 medium40:test | 40 | |
large40 large40:test | 80 | high memory layout |
gpu | 375 | four NVidia V100 (32 GB) GPUs per node |
grete | 600 | four NVidia A100 (40 GB) |
grete:shared | 150 per GPU | 600: four NVidia A100 (40 GB) per node |
grete:interactive grete:preemptible | 47 per MiG slice | four NVidia A100 (40 GB) splitted each into two 2g.10gb slices (8 per node and currently 24 in total) and one 3g.20gb slice (4 per node and currently 12 in total) |
Job Charge
The charge of core hours for a batch job depends on the number of nodes, the wallclock time used by the job, and the charge rate for the partition used. For a batch job with
...