Bjørn-Helge Mevik
2014-09-18 10:46:11 UTC
We are running slurm 2.6.9, with the following configs:
SelectType=select/cons_res
SelectTypeParameters=CR_CPU_Memory
TaskPlugin=task/cgroup
ProctrackType=proctrack/cgroup
i.e., we hand out CPUs and memory to jobs, and use cgroups to enforce
the memory limit. Jobs are required to use --mem-per-cpu.
In some cases, it would be very nice to be able to «override» the memory
limit. Specifically, when a job specifies --exclusive, it would be very
useful if the job would be allowed to use more memory than there is RAM
on the node, because some programs use a lot of memory for a short
while. (It would of course be the user's responsibility if the job/node
crashed.)
Does anyone have any idea about how this could be achieved? Some
job_start plugin that disabled cgroups if a job specifies --exclusive
(or some other, plugin-implemented switch), perhaps? Has anyone tried
something like this?
SelectType=select/cons_res
SelectTypeParameters=CR_CPU_Memory
TaskPlugin=task/cgroup
ProctrackType=proctrack/cgroup
i.e., we hand out CPUs and memory to jobs, and use cgroups to enforce
the memory limit. Jobs are required to use --mem-per-cpu.
In some cases, it would be very nice to be able to «override» the memory
limit. Specifically, when a job specifies --exclusive, it would be very
useful if the job would be allowed to use more memory than there is RAM
on the node, because some programs use a lot of memory for a short
while. (It would of course be the user's responsibility if the job/node
crashed.)
Does anyone have any idea about how this could be achieved? Some
job_start plugin that disabled cgroups if a job specifies --exclusive
(or some other, plugin-implemented switch), perhaps? Has anyone tried
something like this?
--
Regards,
Bjørn-Helge Mevik, dr. scient,
Department for Research Computing, University of Oslo
Regards,
Bjørn-Helge Mevik, dr. scient,
Department for Research Computing, University of Oslo