From d89ef9157edf5a8c3f14628399376b5ae29e30db Mon Sep 17 00:00:00 2001 From: HazelGrant Date: Tue, 21 May 2024 10:17:31 -0400 Subject: [PATCH] Fixes typos in submit/slurm.yml.erb files --- .../apps/bc_desktop/submit/slurm.yml.erb | 54 +++++++++---------- .../apps/bc_desktop/submit/slurm.yml.erb | 12 ++--- 2 files changed, 33 insertions(+), 33 deletions(-) diff --git a/apps.awesim.org/apps/bc_desktop/submit/slurm.yml.erb b/apps.awesim.org/apps/bc_desktop/submit/slurm.yml.erb index 228b53b..e8b719c 100644 --- a/apps.awesim.org/apps/bc_desktop/submit/slurm.yml.erb +++ b/apps.awesim.org/apps/bc_desktop/submit/slurm.yml.erb @@ -32,35 +32,35 @@ end slurm_args = case node_type - # 'any' case handled by scheduler, this is just a quick short circuit - when "any" - plus_gpus(base_slug_arms + any_node, ["--gpus-per-node", "#{gpu_count}"]) - when "any-40core" - base_slurm_args + p18_node - when "any-48core" - base_slurm_args + p20_node + # 'any' case handled by scheduler, this is just a quick short circuit + when "any" + plus_gpus(base_slurm_args + any_node, ["--gpus-per-node", "#{gpu_count}"]) + when "any-40core" + base_slurm_args + p18_node + when "any-48core" + base_slurm_args + p20_node - when "gpu-any" - plus_gpus(base_slug_arms + any_node, ["--gpus-per-node", "#{gpu_count}"]) - when "gpu-40core" - plus_gpus(base_slug_arms + p18_node, ["--gpus-per-node", "#{gpu_count}"]) - when "gpu-48core" - plus_gpus(base_slug_arms + p20_node, ["--gpus-per-node", "#{gpu_count}"]) - when "vis" - plus_gpus(base_slug_arms + any_node, ["--gpus-per-node", "#{gpu_count}", "--gres", "vis"]) - when "densegpu" - plus_gpus(base_slug_arms + p20_node, ["--gpus-per-node", "4"]) + when "gpu-any" + plus_gpus(base_slurm_args + any_node, ["--gpus-per-node", "#{gpu_count}"]) + when "gpu-40core" + plus_gpus(base_slurm_args + p18_node, ["--gpus-per-node", "#{gpu_count}"]) + when "gpu-48core" + plus_gpus(base_slurm_args + p20_node, ["--gpus-per-node", "#{gpu_count}"]) + when "vis" + plus_gpus(base_slurm_args + any_node, ["--gpus-per-node", "#{gpu_count}", "--gres", "vis"]) + when "densegpu" + plus_gpus(base_slurm_args + p20_node, ["--gpus-per-node", "4"]) - # using partitions here is easier than specifying memory requests - when "largemem" - partition = bc_num_slots.to_i > 1 ? "largemem-parallel" : "largemem" - base_slurm_args + tasks_per_node + ["--partition", partition ] - when "hugemem" - partition = bc_num_slots.to_i > 1 ? "hugemem-parallel" : "hugemem" - base_slurm_args + tasks_per_node + ["--partition", partition ] - else - base_slurm_args - end + # using partitions here is easier than specifying memory requests + when "largemem" + partition = bc_num_slots.to_i > 1 ? "largemem-parallel" : "largemem" + base_slurm_args + tasks_per_node + ["--partition", partition ] + when "hugemem" + partition = bc_num_slots.to_i > 1 ? "hugemem-parallel" : "hugemem" + base_slurm_args + tasks_per_node + ["--partition", partition ] + else + base_slurm_args + end image = '/apps/project/ondemand/singularity/mate-rhel8/mate-rhel8.sif' -%> diff --git a/ondemand.osc.edu/apps/bc_desktop/submit/slurm.yml.erb b/ondemand.osc.edu/apps/bc_desktop/submit/slurm.yml.erb index 228b53b..ef2bca0 100644 --- a/ondemand.osc.edu/apps/bc_desktop/submit/slurm.yml.erb +++ b/ondemand.osc.edu/apps/bc_desktop/submit/slurm.yml.erb @@ -34,22 +34,22 @@ slurm_args = case node_type # 'any' case handled by scheduler, this is just a quick short circuit when "any" - plus_gpus(base_slug_arms + any_node, ["--gpus-per-node", "#{gpu_count}"]) + plus_gpus(base_slurm_args + any_node, ["--gpus-per-node", "#{gpu_count}"]) when "any-40core" base_slurm_args + p18_node when "any-48core" base_slurm_args + p20_node when "gpu-any" - plus_gpus(base_slug_arms + any_node, ["--gpus-per-node", "#{gpu_count}"]) + plus_gpus(base_slurm_args + any_node, ["--gpus-per-node", "#{gpu_count}"]) when "gpu-40core" - plus_gpus(base_slug_arms + p18_node, ["--gpus-per-node", "#{gpu_count}"]) + plus_gpus(base_slurm_args + p18_node, ["--gpus-per-node", "#{gpu_count}"]) when "gpu-48core" - plus_gpus(base_slug_arms + p20_node, ["--gpus-per-node", "#{gpu_count}"]) + plus_gpus(base_slurm_args + p20_node, ["--gpus-per-node", "#{gpu_count}"]) when "vis" - plus_gpus(base_slug_arms + any_node, ["--gpus-per-node", "#{gpu_count}", "--gres", "vis"]) + plus_gpus(base_slurm_args + any_node, ["--gpus-per-node", "#{gpu_count}", "--gres", "vis"]) when "densegpu" - plus_gpus(base_slug_arms + p20_node, ["--gpus-per-node", "4"]) + plus_gpus(base_slurm_args + p20_node, ["--gpus-per-node", "4"]) # using partitions here is easier than specifying memory requests when "largemem"