1slurmd(8)                        Slurm Daemon                        slurmd(8)
2
3
4

NAME

6       slurmd - The compute node daemon for Slurm.
7
8

SYNOPSIS

10       slurmd [OPTIONS...]
11
12

DESCRIPTION

14       slurmd  is the compute node daemon of Slurm. It monitors all tasks run‐
15       ning on the compute node , accepts work (tasks),  launches  tasks,  and
16       kills running tasks upon request.
17
18

OPTIONS

20       -b     Report  node  rebooted  when  daemon restarted. Used for testing
21              purposes.
22
23       -c     Clear system locks as needed. This may  be  required  if  slurmd
24              terminated abnormally.
25
26       -C     Print  the  actual hardware configuration (not the configuration
27              from the slurm.conf file) and exit.  The format of output is the
28              same  as  used  in slurm.conf to describe a node's configuration
29              plus its uptime.
30
31       --conf <node parameters>
32              Used in conjunction with the -Z option. Used to override or  de‐
33              fine additional parameters of a dynamic node using the same syn‐
34              tax and parameters used to define nodes in the slurm.conf. Spec‐
35              ifying  any  of CPUs, Boards, SocketsPerBoard, CoresPerSocket or
36              ThreadsPerCore will override the defaults defined by the -C  op‐
37              tion. NodeName and Port are not supported.
38
39              For example if slurmd -C reports
40              NodeName=node1 CPUs=16 Boards=1 SocketsPerBoard=1 CoresPerSocket=8 ThreadsPerCore=2 RealMemory=31848
41
42              the  following  --conf  specifications  will generate the corre‐
43              sponding node definitions:
44              --conf "Gres=gpu:2"
45              NodeName=node1 CPUs=16 Boards=1 SocketsPerBoard=1 CoresPerSocket=8 ThreadsPerCore=2 RealMemory=31848 Gres=gpu:2
46
47              --conf "RealMemory=30000"
48              NodeName=node1 CPUs=16 Boards=1 SocketsPerBoard=1 CoresPerSocket=8 ThreadsPerCore=2 RealMemory=30000
49
50              --conf "CPUs=16"
51              NodeName=node1 CPUs=16 RealMemory=331848
52
53              --conf "CPUs=16 RealMemory=30000 Gres=gpu:2"
54              NodeName=node1 CPUs=16 RealMemory=30000 Gres=gpu:2"
55
56       --conf-server <host>[:<port>]
57              Comma-separated list of controllers, the first being the primary
58              slurmctld.  A  port  can (optionally) be specified for each con‐
59              troller. These hosts are where the slurmd will fetch the config‐
60              uration from when running in "configless" mode.
61
62       -d <file>
63              Specify  the  fully qualified pathname to the slurmstepd program
64              to be used for shepherding user job steps. This  can  be  useful
65              for testing purposes.
66
67       -D     Run  slurmd  in the foreground. Error and debug messages will be
68              copied to stderr.
69
70       -f <file>
71              Read configuration from the specified file. See NOTES below.
72
73       -F[feature]
74              Start this node as a Dynamic Future node. It will try to match a
75              node  definition  with  a  state of FUTURE, optionally using the
76              specified feature to match the node definition.
77
78       -G     Print  Generic  RESource  (GRES)   configuration   (based   upon
79              slurm.conf  GRES  merged  with gres.conf contents for this node)
80              and exit.
81
82       -h     Help; print a brief summary of command options.
83
84       -L <file>
85              Write log messages to the specified file.
86
87       -M     Lock slurmd pages into system memory using mlockall (2) to  dis‐
88              able  paging of the slurmd process. This may help in cases where
89              nodes are marked DOWN during periods of heavy swap activity.  If
90              the  mlockall (2) system call is not available, an error will be
91              printed to the log and slurmd will continue as normal.
92
93              It is suggested to  set  LaunchParameters=slurmstepd_memlock  in
94              slurm.conf(5) when setting -M.
95
96       -n <value>
97              Set  the daemon's nice value to the specified value, typically a
98              negative number.  Also note the PropagatePrioProcess  configura‐
99              tion parameter.
100
101       -N <nodename>
102              Run the daemon with the given nodename. Used to emulate a larger
103              system with more than one slurmd daemon per node. Requires  that
104              Slurm  be built using the --enable-multiple-slurmd configure op‐
105              tion.
106
107       -s     Change working directory of slurmd to SlurmdLogFile path if pos‐
108              sible,  or  to SlurmdSpoolDir otherwise. If both of them fail it
109              will fallback to /var/tmp.
110
111       -v     Verbose operation. Multiple -v's increase verbosity.
112
113       -V, --version
114              Print version information and exit.
115
116       -Z     Start this node as a Dynamic Normal node. If no --conf is speci‐
117              fied,  then the slurmd will register with the same hardware con‐
118              figuration as defined by the -C option.
119

ENVIRONMENT VARIABLES

121       The following environment variables can be used  to  override  settings
122       compiled into slurmd.
123
124
125       SLURM_CONF          The location of the Slurm configuration file.  This
126                           is overridden by explicitly naming a  configuration
127                           file on the command line.
128

SIGNALS

130       SIGTERM SIGINT
131              slurmd will shutdown cleanly, waiting for in-progress rollups to
132              finish.
133
134       SIGHUP Reloads the slurm configuration files, similar to 'scontrol  re‐
135              configure'.
136
137       SIGUSR2
138              Reread  the  log level from the configs, and then reopen the log
139              file.  This should be used when setting up logrotate(8).
140
141       SIGPIPE
142              This signal is explicitly ignored.
143

CORE FILE LOCATION

145       If slurmd is started with the -D option then  the  core  file  will  be
146       written  to  the current working directory.  Otherwise if SlurmdLogFile
147       is a fully qualified path name (starting with a slash), the  core  file
148       will  be  written to the same directory as the log file.  Otherwise the
149       core  file  will  be  written  to  the  SlurmSpoolDir   directory,   or
150       "/var/tmp/"  as  a last resort. If none of the above directories can be
151       written, no core file will be produced.
152
153

NOTES

155       It may be useful to experiment with different slurmd specific  configu‐
156       ration  parameters using a distinct configuration file (e.g. timeouts).
157       However, this special configuration file will not be used by the slurm‐
158       ctld daemon or the Slurm programs, unless you specifically tell each of
159       them to use it. If you desire changing communication ports,  the  loca‐
160       tion  of  the  temporary file system, or other parameters used by other
161       Slurm components, change the common configuration file, slurm.conf.
162
163       If you are using configless mode with a login node that runs a  lot  of
164       client  commands, you may consider running slurmd on that machine so it
165       can manage a cached version of the configuration files. Otherwise, each
166       client  command  will  use the DNS record to contact the controller and
167       get the configuration information, which could place additional load on
168       the controller.
169
170

COPYING

172       Copyright  (C)  2002-2007  The Regents of the University of California.
173       Copyright (C) 2008-2010 Lawrence Livermore  National  Security.   Copy‐
174       right  (C)  2010-2022  SchedMD LLC.  Produced at Lawrence Livermore Na‐
175       tional Laboratory (cf, DISCLAIMER).
176
177       This file is part of Slurm, a resource  management  program.   For  de‐
178       tails, see <https://slurm.schedmd.com/>.
179
180       Slurm  is free software; you can redistribute it and/or modify it under
181       the terms of the GNU General Public License as published  by  the  Free
182       Software  Foundation;  either version 2 of the License, or (at your op‐
183       tion) any later version.
184
185       Slurm is distributed in the hope that it will be  useful,  but  WITHOUT
186       ANY  WARRANTY;  without even the implied warranty of MERCHANTABILITY or
187       FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General  Public  License
188       for more details.
189
190

FILES

192       /etc/slurm.conf
193
194

SEE ALSO

196       slurm.conf(5), slurmctld(8)
197
198
199
200August 2022                      Slurm Daemon                        slurmd(8)
Impressum