From ed11b609c0dedda786a69fe1d1e0d1eee0fe7b9f Mon Sep 17 00:00:00 2001 From: Silvan Sievers Date: Wed, 18 Oct 2023 15:29:19 +0200 Subject: [PATCH] Fix typos. Co-authored-by: Jendrik Seipp --- lab/environments.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/lab/environments.py b/lab/environments.py index fbe895707..3f74ee19a 100644 --- a/lab/environments.py +++ b/lab/environments.py @@ -140,7 +140,7 @@ class SlurmEnvironment(Environment): * "infai_1": 24 nodes with 16 cores, 64GB memory, 500GB Sata (default) * "infai_2": 24 nodes with 20 cores, 128GB memory, 240GB SSD - * "infai_2": 12 nodes with 128 cores, 512GB memory, 240GB SSD + * "infai_3": 12 nodes with 128 cores, 512GB memory, 240GB SSD *qos* must be a valid Slurm QOS name. In Basel this must be "normal". @@ -157,7 +157,7 @@ class SlurmEnvironment(Environment): letters K, M or G. The default is "3872M". The value for *memory_per_cpu* should not surpass the amount of memory that is available per core, which is "3872M" for infai_1, "6354M" for - infai_2, and "4028M" for infai_3 Processes that surpass the + infai_2, and "4028M" for infai_3. Processes that surpass the *memory_per_cpu* limit are terminated with SIGKILL. To impose a soft limit that can be caught from within your programs, you can use the ``memory_limit`` kwarg of @@ -179,7 +179,7 @@ class SlurmEnvironment(Environment): >>> env1 = BaselSlurmEnvironment(partition="infai_1", memory_per_cpu="3872M") >>> env2 = BaselSlurmEnvironment(partition="infai_2", memory_per_cpu="6354M") - >>> env2 = BaselSlurmEnvironment(partition="infai_3", memory_per_cpu="4028M") + >>> env3 = BaselSlurmEnvironment(partition="infai_3", memory_per_cpu="4028M") Example that reserves 12 GiB of memory on infai_1: @@ -206,7 +206,7 @@ class SlurmEnvironment(Environment): >>> # 12 * 1024 / 4028 = 3.05 -> round to next int -> 4 cores per task >>> # 12G / 4 = 3G per core >>> env = BaselSlurmEnvironment( - ... partition="infai_2", + ... partition="infai_3", ... memory_per_cpu="3G", ... cpus_per_task=4, ... )