Atlassian uses cookies to improve your browsing experience, perform analytics and research, and conduct advertising. Accept all cookies to indicate that you agree to our use of cookies on your device. Atlassian cookies and tracking notice, (opens new window)
User Manual
Teams
, (opens new window)

User Manual
Results will update as you type.
  • Application Guide
  • Status of System
  • Usage Guide
  • Compute partitions
    • CPU CLX partition
      • Workflow CPU CLX
      • Slurm partition CPU CLX
      • Examples and Recipes
      • Fat Tree OPA network of CLX partition
      • Operating system migration from CentOS to Rocky Linux
    • CPU Genoa partition
    • GPU A100 partition
    • GPU PVC partition
    • Next-Gen Technology Pool
  • Software
  • FAQ
  • NHR Community
  • Contact
    You‘re viewing this with anonymous access, so some content might be blocked.
    /
    Slurm partition CPU CLX

    Slurm partition CPU CLX

    Okt. 02, 2024

    Content

    Code execution

    After creation of

    • a binary (executable, model code) like in Workflow CPU CLX,
    • a slurm job script like in Workflow CPU CLX with a slurm partition in table Partition for CPU CLX,

    submit the slurm job script to execute the binary on compute nodes.

    > sbatch myjobscipt.slurm
    Submitted batch job 8028673
    > ls slurm-8028673.out
    slurm-8028673.out

    Partition for CPU CLX

    The compute nodes of the CPU cluster of system Lise are organised via the following Slurm partitions.

    Partition name

    Node count

    CPU

    Main memory (GB)

    Max. nodes
    per job

    Max. jobs per user (running/ queued)

    Wall time limit (hh:mm:ss)Remark
    cpu-clx948Cascade 9242362512

    128 / 500

    12:00:00default
    cpu-clx:test16 dedicated
    +128 on demand
    362 161 / 50001:00:00test nodes with higher priority but less wall time
    cpu-clx:ssd50362
    128/50012:00:00local 2TB SSD for IO
    cpu-clx:large287478128 / 50012:00:00fat memory nodes
    blogin1-2.nhr.zib.de
    cpu-clx:huge215221128 / 50024:00:00

    very fat memory nodes for data pre- and post-processing

    See Slurm usage how to pass the 12h wall time limit with job dependencies.

    Which partition to choose?

    The default partition cpu-clx is suitable for most calculations. The :test partitions are, as the name suggests, intended for shorter and smaller test runs. These have a higher priority and a few dedicated nodes, but provide only limited resources. Shared nodes are suitable for pre- and post-processing. A job running on a shared node is only accounted for its core fraction (cores of job / all cores per node). All non-shared nodes are exclusive to one job only at a time.

    The available home/local-ssd/work/perm file systems are discussed under File Systems.

    For an overview of all Slurm partitions and status of nodes: sinfo -r
    For detailed information about a particular nodes: scontrol show node <nodename>

    Charge rates for accounting

    Charge rates for the Slurm partitions can be found under Accounting.

    Fat-Tree Communication Network of Lise

    See Fat Tree OPA network of CLX partition

    List of CPUs


    Short nameLink to manufacturer specificationsWhere to findUnits per node

    Cores per unit

    Clock speed
    [GHz]

    Cascade 9242Intel Cascade Lake Platinum 9242 (CLX-AP)CPU partition "Lise"2482.3
    Cascade 4210Intel Cascade Lake Silver 4210 (CLX)blogin[1-8]2102.2
    , multiple selections available,
    {"serverDuration": 10, "requestCorrelationId": "cc3f2ac7f0a4468f9508686e8fbffc33"}