1slurmd(8) Slurm Daemon slurmd(8)
2
3
4
6 slurmd - The compute node daemon for Slurm.
7
8
10 slurmd [OPTIONS...]
11
12
14 slurmd is the compute node daemon of Slurm. It monitors all tasks run‐
15 ning on the compute node , accepts work (tasks), launches tasks, and
16 kills running tasks upon request.
17
18
20 -b Report node rebooted when daemon restarted. Used for testing
21 purposes.
22
23 -c Clear system locks as needed. This may be required if slurmd
24 terminated abnormally.
25
26 -C Print the actual hardware configuration (not the configuration
27 from the slurm.conf file) and exit. The format of output is the
28 same as used in slurm.conf to describe a node's configuration
29 plus its uptime.
30
31 --conf <node parameters>
32 Used in conjunction with the -Z option. Used to override or de‐
33 fine additional parameters of a dynamic node using the same syn‐
34 tax and parameters used to define nodes in the slurm.conf. Spec‐
35 ifying any of CPUs, Boards, SocketsPerBoard, CoresPerSocket or
36 ThreadsPerCore will override the defaults defined by the -C op‐
37 tion. NodeName and Port are not supported.
38
39 For example if slurmd -C reports
40 NodeName=node1 CPUs=16 Boards=1 SocketsPerBoard=1 CoresPerSocket=8 ThreadsPerCore=2 RealMemory=31848
41
42 the following --conf specifications will generate the corre‐
43 sponding node definitions:
44 --conf "Gres=gpu:2"
45 NodeName=node1 CPUs=16 Boards=1 SocketsPerBoard=1 CoresPerSocket=8 ThreadsPerCore=2 RealMemory=31848 Gres=gpu:2
46
47 --conf "RealMemory=30000"
48 NodeName=node1 CPUs=16 Boards=1 SocketsPerBoard=1 CoresPerSocket=8 ThreadsPerCore=2 RealMemory=30000
49
50 --conf "CPUs=16"
51 NodeName=node1 CPUs=16 RealMemory=331848
52
53 --conf "CPUs=16 RealMemory=30000 Gres=gpu:2"
54 NodeName=node1 CPUs=16 RealMemory=30000 Gres=gpu:2"
55
56 --conf-server <host>[:<port>]
57 Comma-separated list of controllers, the first being the primary
58 slurmctld. A port can (optionally) be specified for each con‐
59 troller. These hosts are where the slurmd will fetch the config‐
60 uration from when running in "configless" mode.
61
62 -d <file>
63 Specify the fully qualified pathname to the slurmstepd program
64 to be used for shepherding user job steps. This can be useful
65 for testing purposes.
66
67 -D Run slurmd in the foreground. Error and debug messages will be
68 copied to stderr.
69
70 -f <file>
71 Read configuration from the specified file. See NOTES below.
72
73 -F[feature]
74 Start this node as a Dynamic Future node. It will try to match a
75 node definition with a state of FUTURE, optionally using the
76 specified feature to match the node definition.
77
78 -G Print Generic RESource (GRES) configuration (based upon
79 slurm.conf GRES merged with gres.conf contents for this node)
80 and exit.
81
82 -h Help; print a brief summary of command options.
83
84 -L <file>
85 Write log messages to the specified file.
86
87 -M Lock slurmd pages into system memory using mlockall (2) to dis‐
88 able paging of the slurmd process. This may help in cases where
89 nodes are marked DOWN during periods of heavy swap activity. If
90 the mlockall (2) system call is not available, an error will be
91 printed to the log and slurmd will continue as normal.
92
93 It is suggested to set LaunchParameters=slurmstepd_memlock in
94 slurm.conf(5) when setting -M.
95
96 -n <value>
97 Set the daemon's nice value to the specified value, typically a
98 negative number. Also note the PropagatePrioProcess configura‐
99 tion parameter.
100
101 -N <nodename>
102 Run the daemon with the given nodename. Used to emulate a larger
103 system with more than one slurmd daemon per node. Requires that
104 Slurm be built using the --enable-multiple-slurmd configure op‐
105 tion.
106
107 -s Change working directory of slurmd to SlurmdLogFile path if pos‐
108 sible, or to SlurmdSpoolDir otherwise. If both of them fail it
109 will fallback to /var/tmp.
110
111 -v Verbose operation. Multiple -v's increase verbosity.
112
113 -V, --version
114 Print version information and exit.
115
116 -Z Start this node as a Dynamic Normal node. If no --conf is speci‐
117 fied, then the slurmd will register with the same hardware con‐
118 figuration as defined by the -C option.
119
121 The following environment variables can be used to override settings
122 compiled into slurmd.
123
124
125 SLURM_CONF The location of the Slurm configuration file. This
126 is overridden by explicitly naming a configuration
127 file on the command line.
128
130 SIGTERM SIGINT
131 slurmd will shutdown cleanly, waiting for in-progress rollups to
132 finish.
133
134 SIGHUP Reloads the slurm configuration files, similar to 'scontrol re‐
135 configure'.
136
137 SIGUSR2
138 Reread the log level from the configs, and then reopen the log
139 file. This should be used when setting up logrotate(8).
140
141 SIGPIPE
142 This signal is explicitly ignored.
143
145 If slurmd is started with the -D option then the core file will be
146 written to the current working directory. Otherwise if SlurmdLogFile
147 is a fully qualified path name (starting with a slash), the core file
148 will be written to the same directory as the log file. Otherwise the
149 core file will be written to the SlurmSpoolDir directory, or
150 "/var/tmp/" as a last resort. If none of the above directories can be
151 written, no core file will be produced.
152
153
155 It may be useful to experiment with different slurmd specific configu‐
156 ration parameters using a distinct configuration file (e.g. timeouts).
157 However, this special configuration file will not be used by the slurm‐
158 ctld daemon or the Slurm programs, unless you specifically tell each of
159 them to use it. If you desire changing communication ports, the loca‐
160 tion of the temporary file system, or other parameters used by other
161 Slurm components, change the common configuration file, slurm.conf.
162
163 If you are using configless mode with a login node that runs a lot of
164 client commands, you may consider running slurmd on that machine so it
165 can manage a cached version of the configuration files. Otherwise, each
166 client command will use the DNS record to contact the controller and
167 get the configuration information, which could place additional load on
168 the controller.
169
170
172 Copyright (C) 2002-2007 The Regents of the University of California.
173 Copyright (C) 2008-2010 Lawrence Livermore National Security. Copy‐
174 right (C) 2010-2022 SchedMD LLC. Produced at Lawrence Livermore Na‐
175 tional Laboratory (cf, DISCLAIMER).
176
177 This file is part of Slurm, a resource management program. For de‐
178 tails, see <https://slurm.schedmd.com/>.
179
180 Slurm is free software; you can redistribute it and/or modify it under
181 the terms of the GNU General Public License as published by the Free
182 Software Foundation; either version 2 of the License, or (at your op‐
183 tion) any later version.
184
185 Slurm is distributed in the hope that it will be useful, but WITHOUT
186 ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
187 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
188 for more details.
189
190
192 /etc/slurm.conf
193
194
196 slurm.conf(5), slurmctld(8)
197
198
199
200August 2022 Slurm Daemon slurmd(8)