1scontrol(1) Slurm Commands scontrol(1)
2
3
4
6 scontrol - view or modify Slurm configuration and state.
7
8
10 scontrol [OPTIONS...] [COMMAND...]
11
12
14 scontrol is used to view or modify Slurm configuration including: job,
15 job step, node, partition, reservation, and overall system configura‐
16 tion. Most of the commands can only be executed by user root or an Ad‐
17 ministrator. If an attempt to view or modify configuration information
18 is made by an unauthorized user, an error message will be printed and
19 the requested action will not occur. If no command is entered on the
20 execute line, scontrol will operate in an interactive mode and prompt
21 for input. It will continue prompting for input and executing commands
22 until explicitly terminated. If a command is entered on the execute
23 line, scontrol will execute that command and terminate. All commands
24 and options are case-insensitive, although node names, partition names,
25 and reservation names are case-sensitive (node names "LX" and "lx" are
26 distinct). All commands and options can be abbreviated to the extent
27 that the specification is unique. A modified Slurm configuration can be
28 written to a file using the scontrol write config command. The result‐
29 ing file will be named using the convention "slurm.conf.<datetime>" and
30 located in the same directory as the original "slurm.conf" file. The
31 directory containing the original slurm.conf must be writable for this
32 to occur.
33
34
36 -a, --all
37 When the show command is used, then display all partitions,
38 their jobs and jobs steps. This causes information to be dis‐
39 played about partitions that are configured as hidden and parti‐
40 tions that are unavailable to user's group.
41
42 -M, --clusters=<string>
43 The cluster to issue commands to. Only one cluster name may be
44 specified. Note that the SlurmDBD must be up for this option to
45 work properly. This option implicitly sets the --local option.
46
47 -d, --details
48 Causes the show command to provide additional details where
49 available.
50
51 --federation
52 Report jobs from federation if a member of one.
53
54 -F, --future
55 Report nodes in FUTURE state.
56
57 -h, --help
58 Print a help message describing the usage of scontrol.
59
60 --hide Do not display information about hidden partitions, their jobs
61 and job steps. By default, neither partitions that are config‐
62 ured as hidden nor those partitions unavailable to user's group
63 will be displayed (i.e. this is the default behavior).
64
65 --local
66 Show only information local to this cluster. Ignore other clus‐
67 ters in the federated if a member of one. Overrides --federa‐
68 tion.
69
70 -o, --oneliner
71 Print information one line per record.
72
73 -Q, --quiet
74 Print no warning or informational messages, only fatal error
75 messages.
76
77 --sibling
78 Show all sibling jobs on a federated cluster. Implies --federa‐
79 tion.
80
81 -u, --uid=<uid>
82 Attempt to update a job as user <uid> instead of the invoking
83 user id.
84
85 -v, --verbose
86 Print detailed event logging. Multiple -v's will further in‐
87 crease the verbosity of logging. By default only errors will be
88 displayed.
89
90 -V , --version
91 Print version information and exit.
92
94 abort Instruct the Slurm controller to terminate immediately and gen‐
95 erate a core file. See "man slurmctld" for information about
96 where the core file will be written.
97
98 cancel_reboot <NodeList>
99 Cancel pending reboots on nodes. The node will be undrain'ed and
100 the reason cleared if the node was drained by an ASAP reboot.
101
102 cluster <CLUSTER_NAME>
103 The cluster to issue commands to. Only one cluster name may be
104 specified.
105
106 create <SPECIFICATION>
107 Create a new partition or reservation. See the full list of pa‐
108 rameters below. Include the tag "res" to create a reservation
109 without specifying a reservation name.
110
111 completing
112 Display all jobs in a COMPLETING state along with associated
113 nodes in either a COMPLETING or DOWN state.
114
115 delete <SPECIFICATION>
116 Delete the entry with the specified SPECIFICATION. The two
117 SPECIFICATION choices are PartitionName=<name> and Reserva‐
118 tion=<name>. Reservations and partitions should have no associ‐
119 ated jobs at the time of their deletion (modify the jobs first).
120 If the specified partition is in use, the request is denied.
121
122 errnumstr <ERRNO>
123 Given a Slurm error number, return a descriptive string.
124
125 fsdampeningfactor <FACTOR>
126 Set the FairShareDampeningFactor in slurmctld.
127
128 help Display a description of scontrol options and commands.
129
130 hold <job_list>
131 Prevent a pending job from being started (sets its priority to
132 0). Use the release command to permit the job to be scheduled.
133 The job_list argument is a comma separated list of job IDs OR
134 "jobname=" with the job's name, which will attempt to hold all
135 jobs having that name. Note that when a job is held by a system
136 administrator using the hold command, only a system administra‐
137 tor may release the job for execution (also see the uhold com‐
138 mand). When the job is held by its owner, it may also be re‐
139 leased by the job's owner. Additionally, attempting to hold a
140 running job will have not suspend or cancel it. But, it will set
141 the job priority to 0 and update the job reason field, which
142 would hold the job if it was requeued at a later time.
143
144 notify <job_id> <message>
145 Send a message to standard error of the salloc or srun command
146 or batch job associated with the specified job_id.
147
148 pidinfo <proc_id>
149 Print the Slurm job id and scheduled termination time corre‐
150 sponding to the supplied process id, proc_id, on the current
151 node. This will work only with processes on node on which scon‐
152 trol is run, and only for those processes spawned by Slurm and
153 their descendants.
154
155 listpids [<job_id>[.<step_id>]] [<NodeName>]
156 Print a listing of the process IDs in a job step (if JO‐
157 BID.STEPID is provided), or all of the job steps in a job (if
158 job_id is provided), or all of the job steps in all of the jobs
159 on the local node (if job_id is not provided or job_id is "*").
160 This will work only with processes on the node on which scontrol
161 is run, and only for those processes spawned by Slurm and their
162 descendants. Note that some Slurm configurations (ProctrackType
163 value of pgid) are unable to identify all processes associated
164 with a job or job step.
165
166 Note that the NodeName option is only really useful when you
167 have multiple slurmd daemons running on the same host machine.
168 Multiple slurmd daemons on one host are, in general, only used
169 by Slurm developers.
170
171 ping Ping the primary and secondary slurmctld daemon and report if
172 they are responding.
173
174 reboot [ASAP] [nextstate={RESUME|DOWN}] [reason=<reason>] {ALL|<Node‐
175 List>}
176 Reboot the nodes in the system when they become idle using the
177 RebootProgram as configured in Slurm's slurm.conf file. Each
178 node will have the "REBOOT" flag added to its node state. After
179 a node reboots and the slurmd daemon starts up again, the
180 HealthCheckProgram will run once. Then, the slurmd daemon will
181 register itself with the slurmctld daemon and the "REBOOT" flag
182 will be cleared. The node's "DRAIN" state flag will be cleared
183 if the reboot was "ASAP", nextstate=resume or down. The "ASAP"
184 option adds the "DRAIN" flag to each node's state, preventing
185 additional jobs from running on the node so it can be rebooted
186 and returned to service "As Soon As Possible" (i.e. ASAP).
187 "ASAP" will also set the node reason to "Reboot ASAP" if the
188 "reason" option isn't specified. If the "nextstate" option is
189 specified as "DOWN", then the node will remain in a down state
190 after rebooting. If "nextstate" is specified as "RESUME", then
191 the nodes will resume as normal and the node's reason and
192 "DRAIN" state will be cleared. Resuming nodes will be consid‐
193 ered as available in backfill future scheduling and won't be re‐
194 placed by idle nodes in a reservation. The "reason" option sets
195 each node's reason to a user-defined message. A default reason
196 of "reboot requested" is set if no other reason is set on the
197 node. The reason will be appended with: "reboot issued" when
198 the reboot is issued; "reboot complete" when the node registers
199 and has a "nextstate" of "DOWN"; or "reboot timed out" when the
200 node fails to register within ResumeTimeout. You must specify
201 either a list of nodes or that ALL nodes are to be rebooted.
202 NOTE: By default, this command does not prevent additional jobs
203 from being scheduled on any nodes before reboot. To do this,
204 you can either use the "ASAP" option or explicitly drain the
205 nodes beforehand. You can alternately create an advanced reser‐
206 vation to prevent additional jobs from being initiated on nodes
207 to be rebooted. Pending reboots can be cancelled by using
208 "scontrol cancel_reboot <node>" or setting the node state to
209 "CANCEL_REBOOT". A node will be marked "DOWN" if it doesn't re‐
210 boot within ResumeTimeout.
211
212 reconfigure
213 Instruct all Slurm daemons to re-read the configuration file.
214 This command does not restart the daemons. This mechanism can
215 be used to modify configuration parameters set in slurm.conf.
216 The Slurm controller (slurmctld) forwards the request to all
217 other daemons (slurmd daemon on each compute node). Running jobs
218 continue execution. Most configuration parameters can be
219 changed by just running this command; however, there are parame‐
220 ters that require a restart of the relevant Slurm daemons. Pa‐
221 rameters requiring a restart will be noted in the slurm.conf(5)
222 man page. The slurmctld daemon and all slurmd daemons must also
223 be restarted if nodes are added to or removed from the cluster.
224
225 release <job_list>
226 Release a previously held job to begin execution. The job_list
227 argument is a comma separated list of job IDs OR "jobname=" with
228 the job's name, which will attempt to hold all jobs having that
229 name. Also see hold.
230
231 requeue [<option>] <job_list>
232 Requeue a running, suspended or finished Slurm batch job into
233 pending state. The job_list argument is a comma separated list
234 of job IDs. The command accepts the following option:
235
236 Incomplete
237 Operate only on jobs (or tasks of a job array) which have
238 not completed. Specifically only jobs in the following
239 states will be requeued: CONFIGURING, RUNNING, STOPPED or
240 SUSPENDED.
241
242 requeuehold [<option>] <job_list>
243 Requeue a running, suspended or finished Slurm batch job into
244 pending state, moreover the job is put in held state (priority
245 zero). The job_list argument is a comma separated list of job
246 IDs. A held job can be released using scontrol to reset its
247 priority (e.g. "scontrol release <job_id>"). The command ac‐
248 cepts the following options:
249
250 Incomplete
251 Operate only on jobs (or tasks of a job array) which have
252 not completed. Specifically only jobs in the following
253 states will be requeued: CONFIGURING, RUNNING, STOPPED or
254 SUSPENDED.
255
256 State=SpecialExit
257 The "SpecialExit" keyword specifies that the job has to
258 be put in a special state JOB_SPECIAL_EXIT. The "scon‐
259 trol show job" command will display the JobState as SPE‐
260 CIAL_EXIT, while the "squeue" command as SE.
261
262 resume <job_list>
263 Resume a previously suspended job. The job_list argument is a
264 comma separated list of job IDs. Also see suspend.
265
266 NOTE: A suspended job releases its CPUs for allocation to other
267 jobs. Resuming a previously suspended job may result in multi‐
268 ple jobs being allocated the same CPUs, which could trigger gang
269 scheduling with some configurations or severe degradation in
270 performance with other configurations. Use of the scancel com‐
271 mand to send SIGSTOP and SIGCONT signals would stop a job with‐
272 out releasing its CPUs for allocation to other jobs and would be
273 a preferable mechanism in many cases. If performing system
274 maintenance you may want to use suspend/resume in the following
275 way. Before suspending set all nodes to draining or set all par‐
276 titions to down so that no new jobs can be scheduled. Then sus‐
277 pend jobs. Once maintenance is done resume jobs then resume
278 nodes and/or set all partitions back to up. Use with caution.
279
280 schedloglevel <LEVEL>
281 Enable or disable scheduler logging. LEVEL may be "0", "1",
282 "disable" or "enable". "0" has the same effect as "disable". "1"
283 has the same effect as "enable". This value is temporary and
284 will be overwritten when the slurmctld daemon reads the
285 slurm.conf configuration file (e.g. when the daemon is restarted
286 or scontrol reconfigure is executed) if the SlurmSchedLogLevel
287 parameter is present.
288
289 setdebug <LEVEL>
290 Change the debug level of the slurmctld daemon for all active
291 logging channels not originally configured off (quiet). LEVEL
292 may be an integer value between zero and nine (using the same
293 values as SlurmctldDebug in the slurm.conf file) or the name of
294 the most detailed message type to be printed: "quiet", "fatal",
295 "error", "info", "verbose", "debug", "debug2", "debug3", "de‐
296 bug4", or "debug5". This value is temporary and will be over‐
297 written whenever the slurmctld daemon reads the slurm.conf con‐
298 figuration file (e.g. when the daemon is restarted or scontrol
299 reconfigure is executed).
300
301 setdebugflags [+|-]<FLAG>
302 Add or remove DebugFlags of the slurmctld daemon. See "man
303 slurm.conf" for a list of supported DebugFlags. NOTE: Changing
304 the value of some DebugFlags will have no effect without
305 restarting the slurmctld daemon, which would set DebugFlags
306 based upon the contents of the slurm.conf configuration file.
307
308 show <ENTITY>[=<ID>] or <ENTITY> [<ID>]
309 Display the state of the specified entity with the specified
310 identification.
311
312 aliases
313 Returns all NodeName values associated with a given Node‐
314 Hostname (useful to get the list of virtual nodes associ‐
315 ated with a real node in a configuration where multiple
316 slurmd daemons execute on a single compute node).
317
318 assoc_mgr
319 Displays the current contents of the slurmctld's internal
320 cache for users, associations and/or qos. The output can
321 be filtered by different record types:
322
323 users=<user1>[...,<userN>]
324 Limit the User Records displayed to those with the
325 specified user name(s).
326
327 accounts=<acct1>[...,<acctN>]
328 Limit the Association Records displayed to those
329 with the specified account name(s).
330
331 qos=<qos1>[...,<qosN>]
332 Limit the QOS Records displayed to those with the
333 specified QOS name(s).
334
335 flags={users|assoc|qos}
336 Specify the desired record type to be displayed.
337 If no flags are specified, all record types are
338 displayed.
339
340 bbstat Displays output from Cray's burst buffer status tool. Op‐
341 tions following bbstat are passed directly to the dwstat
342 command by the slurmctld daemon and the response returned
343 to the user. Equivalent to dwstat.
344
345 burstbuffer
346 Displays the current status of the BurstBuffer plugin.
347
348 config Displays parameter names from the configuration files in
349 mixed case (e.g. SlurmdPort=7003) while derived parame‐
350 ters names are in upper case only (e.g. SLURM_VERSION).
351
352 daemons
353 Reports which daemons should be running on this node.
354
355 dwstat Displays output from Cray's burst buffer status tool. Op‐
356 tions following dwstat are passed directly to the dwstat
357 command by the slurmctld daemon and the response returned
358 to the user. Equivalent to bbstat.
359
360 federation
361 The federation name that the controller is part of and
362 the sibling clusters part of the federation will be
363 listed.
364
365 frontend
366 Shows configured frontend nodes.
367
368 hostlist
369 Takes a list of host names and prints the hostlist ex‐
370 pression for them (the inverse of hostnames). hostlist
371 can also take the absolute pathname of a file (beginning
372 with the character '/') containing a list of hostnames.
373 Multiple node names may be specified using simple node
374 range expressions (e.g. "lx[10-20]"). All other ID values
375 must identify a single element. The job step ID is of the
376 form "job_id.step_id", (e.g. "1234.1"). slurmd reports
377 the current status of the slurmd daemon executing on the
378 same node from which the scontrol command is executed
379 (the local host). It can be useful to diagnose problems.
380 By default hostlist does not sort the node list or make
381 it unique (e.g. tux2,tux1,tux2 = tux[2,1-2]). If you
382 wanted a sorted list use hostlistsorted (e.g.
383 tux2,tux1,tux2 = tux[1-2,2]). By default, all elements
384 of the entity type specified are printed.
385
386 hostlistsorted
387 Takes a list of host names and prints a sorted, unique
388 hostlist expression for them. See hostlist.
389
390 hostnames
391 Takes an optional hostlist expression as input and writes
392 a list of individual host names to standard output (one
393 per line). If no hostlist expression is supplied, the
394 contents of the SLURM_JOB_NODELIST environment variable
395 is used. For example "tux[1-3]" is mapped to
396 "tux1","tux2" and "tux3" (one hostname per line).
397
398 job Displays statistics about all jobs by default. If an op‐
399 tional jobid is specified, details for just that job will
400 be displayed. If the job does not specify
401 socket-per-node, cores-per-socket or threads-per-core
402 then it will display '*' in the ReqS:C:T=*:*:* field.
403
404 licenses
405 Displays statistics about all configured licenses (local
406 and remote) by default. If an optional license name is
407 specified, details for just that license will be dis‐
408 played.
409
410
411 node Displays statistics about all nodes by default. If an op‐
412 tional nodename is specified, details for just that node
413 will be displayed.
414
415 partition
416 Displays statistics about all partitions by default. If
417 an optional partition name is specified, details for just
418 that partition will be displayed.
419
420 reservation
421 Displays statistics about all reservations by default. If
422 an optional reservation name is specified, details for
423 just that reservation will be displayed.
424
425 slurmd Displays statistics for the slurmd running on the current
426 node.
427
428 step Displays statistics about all job steps by default. If an
429 optional jobid is specified, details about steps for just
430 that job will be displayed. If a jobid.stepid is speci‐
431 fied, details for just that step will be displayed.
432
433 topology
434 Displays information about the defined topology layout.
435 If a switch is specified, information about that switch
436 will be shown. If one node name is specified, all
437 switches connected to that node (and their parent
438 switches) will be shown. If more than one node name is
439 specified, only switches that connect to all named nodes
440 will be shown.
441
442 shutdown <OPTION>
443 Instruct Slurm daemons to save current state and terminate. By
444 default, the Slurm controller (slurmctld) forwards the request
445 all other daemons (slurmd daemon on each compute node). An OP‐
446 TION of slurmctld or controller results in only the slurmctld
447 daemon being shutdown and the slurmd daemons remaining active.
448
449 suspend <job_list>
450 Suspend a running job. The job_list argument is a comma sepa‐
451 rated list of job IDs. Use the resume command to resume its ex‐
452 ecution. User processes must stop on receipt of SIGSTOP signal
453 and resume upon receipt of SIGCONT for this operation to be ef‐
454 fective. Not all architectures and configurations support job
455 suspension. If a suspended job is requeued, it will be placed
456 in a held state. The time a job is suspended will not count
457 against a job's time limit. Only an operator, administrator,
458 SlurmUser, or root can suspend jobs.
459
460 takeover [<INDEX>]
461 Instruct one of Slurm's backup controllers (slurmctld) to take
462 over system control. By default the first backup controller (IN‐
463 DEX=1) requests control from the primary and waits for its ter‐
464 mination. After that, it switches from backup mode to controller
465 mode. If primary controller can not be contacted, it directly
466 switches to controller mode. This can be used to speed up the
467 Slurm controller fail-over mechanism when the primary node is
468 down. This can be used to minimize disruption if the computer
469 executing the primary Slurm controller is scheduled down.
470 (Note: Slurm's primary controller will take the control back at
471 startup.)
472
473 top <job_list>
474 Move the specified job IDs to the top of the queue of jobs be‐
475 longing to the identical user ID, partition name, account, and
476 QOS. The job_list argument is a comma separated ordered list of
477 job IDs. Any job not matching all of those fields will not be
478 effected. Only jobs submitted to a single partition will be ef‐
479 fected. This operation changes the order of jobs by adjusting
480 job nice values. The net effect on that user's throughput will
481 be negligible to slightly negative. This operation is disabled
482 by default for non-privileged (non-operator, admin, SlurmUser,
483 or root) users. This operation may be enabled for non-privileged
484 users by the system administrator by including the option "en‐
485 able_user_top" in the SchedulerParameters configuration parame‐
486 ter.
487
488 token [lifespan=<lifespan>] [username=<username>]
489 Return an auth token which can be used to support JWT authenti‐
490 cation if AuthAltTypes=auth/jwt has been enabled on the system.
491 Supports two optional arguments. lifespan= may be used to spec‐
492 ify the token's lifespan in seconds. username (only available to
493 SlurmUser/root) may be used to request a token for a different
494 username.
495
496 uhold <job_list>
497 Prevent a pending job from being started (sets its priority to
498 0). The job_list argument is a space separated list of job IDs
499 or job names. Use the release command to permit the job to be
500 scheduled. This command is designed for a system administrator
501 to hold a job so that the job owner may release it rather than
502 requiring the intervention of a system administrator (also see
503 the hold command).
504
505 update <SPECIFICATION>
506 Update job, step, node, partition, or reservation configuration
507 per the supplied specification. SPECIFICATION is in the same
508 format as the Slurm configuration file and the output of the
509 show command described above. It may be desirable to execute the
510 show command (described above) on the specific entity you want
511 to update, then use cut-and-paste tools to enter updated config‐
512 uration values to the update. Note that while most configuration
513 values can be changed using this command, not all can be changed
514 using this mechanism. In particular, the hardware configuration
515 of a node or the physical addition or removal of nodes from the
516 cluster may only be accomplished through editing the Slurm con‐
517 figuration file and executing the reconfigure command (described
518 above).
519
520 version
521 Display the version number of scontrol being executed.
522
523 wait_job <job_id>
524 Wait until a job and all of its nodes are ready for use or the
525 job has entered some termination state. This option is particu‐
526 larly useful in the Slurm Prolog or in the batch script itself
527 if nodes are powered down and restarted automatically as needed.
528
529 NOTE: Don't use scontrol wait_job in PrologSlurmctld or Prolog
530 with PrologFlags=Alloc as this will result in a deadlock.
531
532 NOTE: When using wait_job for an array job, use the SLURM_JOB_ID
533 environment variable to reference the job rather than the
534 SLURM_ARRAY_JOB_ID variable.
535
536
537 write batch_script <job_id> [<optional_filename>]
538 Write the batch script for a given job_id to a file or to std‐
539 out. The file will default to slurm-<job_id>.sh if the optional
540 filename argument is not given. The script will be written to
541 stdout if - is given instead of a filename. The batch script
542 can only be retrieved by an admin or operator, or by the owner
543 of the job.
544
545 write config <optional_filename>
546 Write the current configuration to a file with the naming con‐
547 vention of "slurm.conf.<datetime>" in the same directory as the
548 original slurm.conf file. If a filename is given that file lo‐
549 cation with a .<datetime> suffix is created.
550
552 NOTE: All commands listed below can be used in the interactive mode,
553 but NOT on the initial command line.
554
555
556 all Show all partitions, their jobs and jobs steps. This causes in‐
557 formation to be displayed about partitions that are configured
558 as hidden and partitions that are unavailable to user's group.
559
560 details
561 Causes the show command to provide additional details where
562 available. Job information will include CPUs and NUMA memory
563 allocated on each node. Note that on computers with hyper‐
564 threading enabled and Slurm configured to allocate cores, each
565 listed CPU represents one physical core. Each hyperthread on
566 that core can be allocated a separate task, so a job's CPU count
567 and task count may differ. See the --cpu-bind and --mem-bind
568 option descriptions in srun man pages for more information. The
569 details option is currently only supported for the show job com‐
570 mand.
571
572 exit Terminate scontrol interactive session.
573
574 hide Do not display partition, job or jobs step information for par‐
575 titions that are configured as hidden or partitions that are un‐
576 available to the user's group. This is the default behavior.
577
578 oneliner
579 Print information one line per record.
580
581 quiet Print no warning or informational messages, only fatal error
582 messages.
583
584 quit Terminate the execution of scontrol.
585
586 verbose
587 Print detailed event logging. This includes time-stamps on data
588 structures, record counts, etc.
589
590 !! Repeat the last command executed.
591
593 Note that update requests done by either root, SlurmUser or Administra‐
594 tors are not subject to certain restrictions. For instance, if an Ad‐
595 ministrator changes the QOS on a pending job, certain limits such as
596 the TimeLimit will not be changed automatically as changes made by the
597 Administrators are allowed to violate these restrictions.
598
599
600 Account=<account>
601 Account name to be changed for this job's resource use. Value
602 may be cleared with blank data value, "Account=".
603
604 AdminComment=<spec>
605 Arbitrary descriptive string. Can only be set by a Slurm admin‐
606 istrator.
607
608 ArrayTaskThrottle=<count>
609 Specify the maximum number of tasks in a job array that can exe‐
610 cute at the same time. Set the count to zero in order to elimi‐
611 nate any limit. The task throttle count for a job array is re‐
612 ported as part of its ArrayTaskId field, preceded with a percent
613 sign. For example "ArrayTaskId=1-10%2" indicates the maximum
614 number of running tasks is limited to 2.
615
616 BurstBuffer=<spec>
617 Burst buffer specification to be changed for this job's resource
618 use. Value may be cleared with blank data value, "Burst‐
619 Buffer=". Format is burst buffer plugin specific.
620
621 Clusters=<spec>
622 Specifies the clusters that the federated job can run on.
623
624 ClusterFeatures=<spec>
625 Specifies features that a federated cluster must have to have a
626 sibling job submitted to it. Slurm will attempt to submit a sib‐
627 ling job to a cluster if it has at least one of the specified
628 features.
629
630 Comment=<spec>
631 Arbitrary descriptive string.
632
633 Contiguous={yes|no}
634 Set the job's requirement for contiguous (consecutive) nodes to
635 be allocated. Possible values are "YES" and "NO". Only the
636 Slurm administrator or root can change this parameter.
637
638 CoreSpec=<count>
639 Number of cores to reserve per node for system use. The job
640 will be charged for these cores, but be unable to use them.
641 Will be reported as "*" if not constrained.
642
643 CPUsPerTask=<count>
644 Change the CPUsPerTask job's value.
645
646 Deadline=<time_spec>
647 It accepts times of the form HH:MM:SS to specify a deadline to a
648 job at a specific time of day (seconds are optional). You may
649 also specify midnight, noon, fika (3 PM) or teatime (4 PM) and
650 you can have a time-of-day suffixed with AM or PM for a deadline
651 in the morning or the evening. You can specify a deadline for
652 the job with a date of the form MMDDYY or MM/DD/YY or MM.DD.YY,
653 or a date and time as YYYY-MM-DD[THH:MM[:SS]]. You can also
654 give times like now + count time-units, where the time-units can
655 be seconds (default), minutes, hours, days, or weeks and you can
656 tell Slurm to put a deadline for tomorrow with the keyword to‐
657 morrow. The specified deadline must be later than the current
658 time. Only pending jobs can have the deadline updated. Only
659 the Slurm administrator or root can change this parameter.
660
661 DelayBoot=<time_spec>
662 Change the time to decide whether to reboot nodes in order to
663 satisfy job's feature specification if the job has been eligible
664 to run for less than this time period. See salloc/sbatch man
665 pages option --delay-boot.
666
667 Dependency=<dependency_list>
668 Defer job's initiation until specified job dependency specifica‐
669 tion is satisfied. Cancel dependency with an empty depen‐
670 dency_list (e.g. "Dependency="). <dependency_list> is of the
671 form <type:job_id[:job_id][,type:job_id[:job_id]]>. Many jobs
672 can share the same dependency and these jobs may even belong to
673 different users.
674
675 after:job_id[:jobid...]
676 This job can begin execution after the specified jobs
677 have begun execution or been canceled.
678
679 afterany:job_id[:jobid...]
680 This job can begin execution after the specified jobs
681 have terminated.
682
683 afternotok:job_id[:jobid...]
684 This job can begin execution after the specified jobs
685 have terminated in some failed state (non-zero exit code,
686 node failure, timed out, etc).
687
688 afterok:job_id[:jobid...]
689 This job can begin execution after the specified jobs
690 have successfully executed (ran to completion with an
691 exit code of zero).
692
693 singleton
694 This job can begin execution after any previously
695 launched jobs sharing the same job name and user have
696 terminated. In other words, only one job by that name
697 and owned by that user can be running or suspended at any
698 point in time.
699
700 EligibleTime=<time_spec>
701 See StartTime.
702
703 EndTime
704 The time the job is expected to terminate based on the job's
705 time limit. When the job ends sooner, this field will be up‐
706 dated with the actual end time.
707
708 ExcNodeList=<nodes>
709 Set the job's list of excluded node. Multiple node names may be
710 specified using simple node range expressions (e.g.
711 "lx[10-20]"). Value may be cleared with blank data value, "Exc‐
712 NodeList=".
713
714 Features=<features>
715 Set the job's required node features. The list of features may
716 include multiple feature names separated by ampersand (AND)
717 and/or vertical bar (OR) operators. For example: Fea‐
718 tures="opteron&video" or Features="fast|faster". In the first
719 example, only nodes having both the feature "opteron" AND the
720 feature "video" will be used. There is no mechanism to specify
721 that you want one node with feature "opteron" and another node
722 with feature "video" in case no node has both features. If only
723 one of a set of possible options should be used for all allo‐
724 cated nodes, then use the OR operator and enclose the options
725 within square brackets. For example: "Fea‐
726 tures=[rack1|rack2|rack3|rack4]" might be used to specify that
727 all nodes must be allocated on a single rack of the cluster, but
728 any of those four racks can be used. A request can also specify
729 the number of nodes needed with some feature by appending an as‐
730 terisk and count after the feature name. For example "Fea‐
731 tures=graphics*4" indicates that at least four allocated nodes
732 must have the feature "graphics." Parenthesis are also sup‐
733 ported for features to be ANDed together. For example "Fea‐
734 tures=[(knl&a2a&flat)*4&haswell*2]" indicates the resource allo‐
735 cation should include 4 nodes with ALL of the features "knl",
736 "a2a", and "flat" plus 2 nodes with the feature "haswell". Con‐
737 straints with node counts may only be combined with AND opera‐
738 tors. Value may be cleared with blank data value, for example
739 "Features=".
740
741 Gres=<list>
742 Specifies a comma-delimited list of generic consumable re‐
743 sources. The format of each entry on the list is
744 "name[:count[*cpu]]". The name is that of the consumable re‐
745 source. The count is the number of those resources with a de‐
746 fault value of 1. The specified resources will be allocated to
747 the job on each node allocated unless "*cpu" is appended, in
748 which case the resources will be allocated on a per cpu basis.
749 The available generic consumable resources is configurable by
750 the system administrator. A list of available generic consum‐
751 able resources will be printed and the command will exit if the
752 option argument is "help". Examples of use include
753 "Gres=gpus:2*cpu,disk=40G" and "Gres=help".
754
755 JobId=<job_list>
756 Identify the job(s) to be updated. The job_list may be a comma
757 separated list of job IDs. Either JobId or JobName is required.
758
759 Licenses=<name>
760 Specification of licenses (or other resources available on all
761 nodes of the cluster) as described in salloc/sbatch/srun man
762 pages.
763
764 MailType=<types>
765 Set the mail event types. Valid type values are NONE, BEGIN,
766 END, FAIL, REQUEUE, ALL (equivalent to BEGIN, END, FAIL, RE‐
767 QUEUE, and STAGE_OUT), STAGE_OUT (burst buffer stage out and
768 teardown completed), TIME_LIMIT, TIME_LIMIT_90 (reached 90 per‐
769 cent of time limit), TIME_LIMIT_80 (reached 80 percent of time
770 limit), TIME_LIMIT_50 (reached 50 percent of time limit) and AR‐
771 RAY_TASKS (send emails for each array task). Multiple type val‐
772 ues may be specified in a comma separated list. Unless the AR‐
773 RAY_TASKS option is specified, mail notifications on job BEGIN,
774 END and FAIL apply to a job array as a whole rather than gener‐
775 ating individual email messages for each task in the job array.
776
777 MailUser=<name>
778 Set the user to receive email notification of state changes. A
779 blank string will set the mail user to the default which is the
780 submitting user.
781
782 MinCPUsNode=<count>
783 Set the job's minimum number of CPUs per node to the specified
784 value.
785
786 MinMemoryCPU=<megabytes>
787 Set the job's minimum real memory required per allocated CPU to
788 the specified value. Either MinMemoryCPU or MinMemoryNode may be
789 set, but not both.
790
791 MinMemoryNode=<megabytes>
792 Set the job's minimum real memory required per node to the spec‐
793 ified value. Either MinMemoryCPU or MinMemoryNode may be set,
794 but not both.
795
796 MinTmpDiskNode=<megabytes>
797 Set the job's minimum temporary disk space required per node to
798 the specified value. Only the Slurm administrator or root can
799 change this parameter.
800
801 TimeMin=<timespec>
802 Change TimeMin value which specifies the minimum time limit min‐
803 utes of the job.
804
805 JobName=<name>
806 Identify the name of jobs to be modified or set the job's name
807 to the specified value. When used to identify jobs to be modi‐
808 fied, all jobs belonging to all users are modified unless the
809 UserID option is used to identify a specific user. Either JobId
810 or JobName is required.
811
812 Name[=<name>]
813 See JobName.
814
815 Nice[=<adjustment>]
816 Update the job with an adjusted scheduling priority within
817 Slurm. With no adjustment value the scheduling priority is de‐
818 creased by 100. A negative nice value increases the priority,
819 otherwise decreases it. The adjustment range is +/- 2147483645.
820 Only privileged users can specify a negative adjustment.
821
822 NodeList=<nodes>
823 Change the nodes allocated to a running job to shrink its size.
824 The specified list of nodes must be a subset of the nodes cur‐
825 rently allocated to the job. Multiple node names may be speci‐
826 fied using simple node range expressions (e.g. "lx[10-20]"). Af‐
827 ter a job's allocation is reduced, subsequent srun commands must
828 explicitly specify node and task counts which are valid for the
829 new allocation.
830
831 NumCPUs=<min_count>[-<max_count>]
832 Set the job's minimum and optionally maximum count of CPUs to be
833 allocated.
834
835 NumNodes=<min_count>[-<max_count>]
836 Set the job's minimum and optionally maximum count of nodes to
837 be allocated. If the job is already running, use this to spec‐
838 ify a node count less than currently allocated and resources
839 previously allocated to the job will be relinquished. After a
840 job's allocation is reduced, subsequent srun commands must ex‐
841 plicitly specify node and task counts which are valid for the
842 new allocation. Also see the NodeList parameter above. This is
843 the same as ReqNodes.
844
845 NumTasks=<count>
846 Set the job's count of required tasks to the specified value.
847 This is the same as ReqProcs.
848
849 OverSubscribe={yes|no}
850 Set the job's ability to share compute resources (i.e. individ‐
851 ual CPUs) with other jobs. Possible values are "YES" and "NO".
852 This option can only be changed for pending jobs.
853
854 Partition=<name>
855 Set the job's partition to the specified value.
856
857 Priority=<number>
858 Set the job's priority to the specified value. Note that a job
859 priority of zero prevents the job from ever being scheduled. By
860 setting a job's priority to zero it is held. Set the priority
861 to a non-zero value to permit it to run. Explicitly setting a
862 job's priority clears any previously set nice value and removes
863 the priority/multifactor plugin's ability to manage a job's pri‐
864 ority. In order to restore the priority/multifactor plugin's
865 ability to manage a job's priority, hold and then release the
866 job. Only the Slurm administrator or root can increase job's
867 priority.
868
869 QOS=<name>
870 Set the job's QOS (Quality Of Service) to the specified value.
871 Value may be cleared with blank data value, "QOS=".
872
873 Reboot={yes|no}
874 Set the job's flag that specifies whether to force the allocated
875 nodes to reboot before starting the job. This is only supported
876 with some system configurations and therefore it could be
877 silently ignored.
878
879 ReqCores=<count>
880 Change the job's requested Cores count.
881
882 ReqNodeList=<nodes>
883 Set the job's list of required node. Multiple node names may be
884 specified using simple node range expressions (e.g.
885 "lx[10-20]"). Value may be cleared with blank data value, "Re‐
886 qNodeList=".
887
888 ReqNodes=<min_count>[-<max_count>]
889 See NumNodes.
890
891 ReqProcs=<count>
892 See NumTasks.
893
894 ReqSockets=<count>
895 Change the job's requested socket count.
896
897 ReqThreads=<count>
898 Change the job's requested threads count.
899
900 IRequeue={0|1}
901 Stipulates whether a job should be requeued after a node fail‐
902 ure: 0 for no, 1 for yes.
903
904 ReservationName=<name>
905 Set the job's reservation to the specified value. Value may be
906 cleared with blank data value, "ReservationName=".
907
908 ResetAccrueTime
909 Set the job's accrue time value to 'now' meaning it will lose
910 any time previously accrued for priority. Helpful if you have a
911 large queue of jobs already in the queue and want to start lim‐
912 iting how many jobs can accrue time without waiting for the
913 queue to flush out.
914
915 SiteFactor=<account>
916 Specify the job's admin priority factor in the range of
917 +/-2147483645. Only privileged users can modify the value.
918
919 StdOut=<filepath>
920 Set the batch job's stdout file path.
921
922 Shared={yes|no}
923 See OverSubscribe option above.
924
925 StartTime=<time_spec>
926 Set the job's earliest initiation time. It accepts times of the
927 form HH:MM:SS to run a job at a specific time of day (seconds
928 are optional). (If that time is already past, the next day is
929 assumed.) You may also specify midnight, noon, fika (3 PM) or
930 teatime (4 PM) and you can have a time-of-day suffixed with AM
931 or PM for running in the morning or the evening. You can also
932 say what day the job will be run, by specifying a date of the
933 form MMDDYY or MM/DD/YY or MM.DD.YY, or a date and time as
934 YYYY-MM-DD[THH:MM[:SS]]. You can also give times like now +
935 count time-units, where the time-units can be seconds (default),
936 minutes, hours, days, or weeks and you can tell Slurm to run the
937 job today with the keyword today and to run the job tomorrow
938 with the keyword tomorrow.
939
940 Notes on date/time specifications:
941 - although the 'seconds' field of the HH:MM:SS time specifica‐
942 tion is allowed by the code, note that the poll time of the
943 Slurm scheduler is not precise enough to guarantee dispatch of
944 the job on the exact second. The job will be eligible to start
945 on the next poll following the specified time. The exact poll
946 interval depends on the Slurm scheduler (e.g., 60 seconds with
947 the default sched/builtin).
948 - if no time (HH:MM:SS) is specified, the default is
949 (00:00:00).
950 - if a date is specified without a year (e.g., MM/DD) then the
951 current year is assumed, unless the combination of MM/DD and
952 HH:MM:SS has already passed for that year, in which case the
953 next year is used.
954
955 Switches=<count>[@<max-time-to-wait>]
956 When a tree topology is used, this defines the maximum count of
957 switches desired for the job allocation. If Slurm finds an allo‐
958 cation containing more switches than the count specified, the
959 job remain pending until it either finds an allocation with de‐
960 sired switch count or the time limit expires. By default there
961 is no switch count limit and no time limit delay. Set the count
962 to zero in order to clean any previously set count (disabling
963 the limit). The job's maximum time delay may be limited by the
964 system administrator using the SchedulerParameters configuration
965 parameter with the max_switch_wait parameter option. Also see
966 wait-for-switch.
967
968 wait-for-switch=<seconds>
969 Change max time to wait for a switch <seconds> secs.
970
971 TasksPerNode=<count>
972 Change the job's requested TasksPerNode.
973
974 ThreadSpec=<count>
975 Number of threads to reserve per node for system use. The job
976 will be charged for these threads, but be unable to use them.
977 Will be reported as "*" if not constrained.
978
979 TimeLimit=<time>
980 The job's time limit. Output format is [days-]hours:min‐
981 utes:seconds or "UNLIMITED". Input format (for update command)
982 set is minutes, minutes:seconds, hours:minutes:seconds,
983 days-hours, days-hours:minutes or days-hours:minutes:seconds.
984 Time resolution is one minute and second values are rounded up
985 to the next minute. If changing the time limit of a job, either
986 specify a new time limit value or precede the time and equal
987 sign with a "+" or "-" to increment or decrement the current
988 time limit (e.g. "TimeLimit+=30"). In order to increment or
989 decrement the current time limit, the JobId specification must
990 precede the TimeLimit specification. Note that incrementing or
991 decrementing the time limit for a job array is only allowed be‐
992 fore the job array has been split into more than one job record.
993 Only the Slurm administrator or root can increase job's Time‐
994 Limit.
995
996 UserID=<UID or name>
997 Used with the JobName option to identify jobs to be modified.
998 Either a user name or numeric ID (UID), may be specified.
999
1000 WCKey=<key>
1001 Set the job's workload characterization key to the specified
1002 value.
1003
1004 WorkDir=<directory_name>
1005 Set the job's working directory to the specified value. Note
1006 that this may only be set for jobs in the PENDING state, and
1007 that jobs may fail to launch if they rely on relative paths to
1008 the originally submitted WorkDir.
1009
1011 The "show" command, when used with the "job" or "job <jobid>" entity
1012 displays detailed information about a job or jobs. Much of this infor‐
1013 mation may be modified using the "update job" command as described
1014 above. However, the following fields displayed by the show job command
1015 are read-only and cannot be modified:
1016
1017
1018 AllocNode:Sid
1019 Local node and system id making the resource allocation.
1020
1021 BatchFlag
1022 Jobs submitted using the sbatch command have BatchFlag set to 1.
1023 Jobs submitted using other commands have BatchFlag set to 0.
1024
1025 ExitCode=<exit>:<sig>
1026 Exit status reported for the job by the wait() function. The
1027 first number is the exit code, typically as set by the exit()
1028 function. The second number of the signal that caused the
1029 process to terminate if it was terminated by a signal.
1030
1031 GroupId
1032 The group under which the job was submitted.
1033
1034 JobState
1035 The current state of the job.
1036
1037 NodeListIndices
1038 The NodeIndices expose the internal indices into the node table
1039 associated with the node(s) allocated to the job.
1040
1041 NtasksPerN:B:S:C=<tasks_per_node>:<tasks_per_base‐
1042 board>:<tasks_per_socket>:<tasks_per_core>
1043 Specifies the number of tasks to be started per hardware compo‐
1044 nent (node, baseboard, socket and core). Unconstrained values
1045 may be shown as "0" or "*".
1046
1047 PreemptEligibleTime
1048 Time the job becomes eligible for preemption. Modified by Pre‐
1049 emptExemptTime, either from the global option in slurm.conf or
1050 the job QOS. This is hidden if the job has not started or if
1051 PreemptMode=OFF.
1052
1053 PreemptTime
1054 Time at which job was signaled that it was selected for preemp‐
1055 tion. This value is only meaningful for PreemptMode=CANCEL and
1056 PreemptMode=REQUEUE and for jobs in a partition or QOS that has
1057 a GraceTime value designated. This is hidden if the job has not
1058 started or if PreemptMode=OFF.
1059
1060 PreSusTime
1061 Time the job ran prior to last suspend.
1062
1063 Reason The reason job is not running: e.g., waiting "Resources".
1064
1065 ReqB:S:C:T=<baseboard_count>:<socket_per_base‐
1066 board_count>:<core_per_socket_count>:<thread_per_core_count>
1067 Specifies the count of various hardware components requested by
1068 the job. Unconstrained values may be shown as "0" or "*".
1069
1070 SecsPreSuspend=<seconds>
1071 If the job is suspended, this is the run time accumulated by the
1072 job (in seconds) prior to being suspended.
1073
1074 Socks/Node=<count>
1075 Count of desired sockets per node
1076
1077 SubmitTime
1078 The time and date stamp (in localtime) the job was submitted.
1079 The format of the output is identical to that of the EndTime
1080 field.
1081
1082 NOTE: If a job is requeued, the submit time is reset. To obtain
1083 the original submit time it is necessary to use the "sacct -j
1084 <job_id[.<step_id>]" command also designating the -D or --dupli‐
1085 cate option to display all duplicate entries for a job.
1086
1087 SuspendTime
1088 Time the job was last suspended or resumed.
1089
1090 NOTE on information displayed for various job states: When you
1091 submit a request for the "show job" function the scontrol
1092 process makes an RPC request call to slurmctld with a RE‐
1093 QUEST_JOB_INFO message type. If the state of the job is PEND‐
1094 ING, then it returns some detail information such as: min_nodes,
1095 min_procs, cpus_per_task, etc. If the state is other than PEND‐
1096 ING the code assumes that it is in a further state such as RUN‐
1097 NING, COMPLETE, etc. In these cases the code explicitly returns
1098 zero for these values. These values are meaningless once the job
1099 resources have been allocated and the job has started.
1100
1102 StepId=<job_id>[.<step_id>]
1103 Identify the step to be updated. If the job_id is given, but no
1104 step_id is specified then all steps of the identified job will
1105 be modified. This specification is required.
1106
1107 TimeLimit=<time>
1108 The job's time limit. Output format is [days-]hours:min‐
1109 utes:seconds or "UNLIMITED". Input format (for update command)
1110 set is minutes, minutes:seconds, hours:minutes:seconds,
1111 days-hours, days-hours:minutes or days-hours:minutes:seconds.
1112 Time resolution is one minute and second values are rounded up
1113 to the next minute. If changing the time limit of a step, ei‐
1114 ther specify a new time limit value or precede the time with a
1115 "+" or "-" to increment or decrement the current time limit
1116 (e.g. "TimeLimit=+30"). In order to increment or decrement the
1117 current time limit, the StepId specification must precede the
1118 TimeLimit specification.
1119
1121 NodeName=<name>
1122 Identify the node(s) to be updated. Multiple node names may be
1123 specified using simple node range expressions (e.g.
1124 "lx[10-20]"). This specification is required.
1125
1126 ActiveFeatures=<features>
1127 Identify the feature(s) currently active on the specified node.
1128 Any previously active feature specification will be overwritten
1129 with the new value. Also see AvailableFeatures. Typically Ac‐
1130 tiveFeatures will be identical to AvailableFeatures; however Ac‐
1131 tiveFeatures may be configured as a subset of the AvailableFea‐
1132 tures. For example, a node may be booted in multiple configura‐
1133 tions. In that case, all possible configurations may be identi‐
1134 fied as AvailableFeatures, while ActiveFeatures would identify
1135 the current node configuration. When updating the ActiveFea‐
1136 tures with scontrol, the change is only made in slurmctld. When
1137 using a node_features plugin the state/features of the node must
1138 be updated on the node such that a new node start will report
1139 the updated state/features.
1140
1141 AvailableFeatures=<features>
1142 Identify the feature(s) available on the specified node. Any
1143 previously defined available feature specification will be over‐
1144 written with the new value. AvailableFeatures assigned via
1145 scontrol will only persist across the restart of the slurmctld
1146 daemon with the -R option and state files preserved or slurm‐
1147 ctld's receipt of a SIGHUP. Update slurm.conf with any changes
1148 meant to be persistent across normal restarts of slurmctld or
1149 the execution of scontrol reconfig.
1150
1151 Note: Available features being removed via scontrol must not be
1152 active (i.e. remove them from ActiveFeatures first).
1153
1154 Comment=<comment>
1155 Arbitrary descriptive string. Use quotes to enclose a comment
1156 having more than one word
1157
1158 CpuBind=<node>
1159 Specify the task binding mode to be used by default for this
1160 node. Supported options include: "none", "socket", "ldom"
1161 (NUMA), "core", "thread" and "off" (remove previous binding
1162 mode).
1163
1164 Extra=<comment>
1165 Arbitrary string on the node. Use quotes to enclose a string
1166 having more than one word.
1167
1168 Gres=<gres>
1169 Identify generic resources to be associated with the specified
1170 node. Any previously defined generic resources will be over‐
1171 written with the new value. Specifications for multiple generic
1172 resources should be comma separated. Each resource specifica‐
1173 tion consists of a name followed by an optional colon with a nu‐
1174 meric value (default value is one) (e.g. "Gres=band‐
1175 width:10000"). Modification of GRES count associated with spe‐
1176 cific files (e.g. GPUs) is not allowed other than to set their
1177 count on a node to zero. In order to change the GRES count to
1178 another value, modify your slurm.conf and gres.conf files and
1179 restart daemons. If GRES are associated with specific sockets,
1180 that information will be reported For example if all 4 GPUs on a
1181 node are all associated with socket zero, then
1182 "Gres=gpu:4(S:0)". If associated with sockets 0 and 1 then
1183 "Gres=gpu:4(S:0-1)". The information of which specific GPUs are
1184 associated with specific GPUs is not reported, but only avail‐
1185 able by parsing the gres.conf file. Generic resources assigned
1186 via scontrol will only persist across the restart of the slurm‐
1187 ctld daemon with the -R option and state files preserved or
1188 slurmctld's receipt of a SIGHUP. Update slurm.conf with any
1189 changes meant to be persistent across normal restarts of slurm‐
1190 ctld or the execution of scontrol reconfig.
1191
1192 NodeAddr=<node address>
1193 Name that a node should be referred to in establishing a commu‐
1194 nications path. This name will be used as an argument to the
1195 getaddrinfo() function for identification. If a node range ex‐
1196 pression is used to designate multiple nodes, they must exactly
1197 match the entries in the NodeName (e.g. "NodeName=lx[0-7]
1198 NodeAddr=elx[0-7]"). NodeAddr may also contain IP addresses.
1199
1200 NodeHostname=<node hostname>
1201 Typically this would be the string that "/bin/hostname -s" re‐
1202 turns. It may also be the fully qualified domain name as re‐
1203 turned by "/bin/hostname -f" (e.g. "foo1.bar.com"), or any valid
1204 domain name associated with the host through the host database
1205 (/etc/hosts) or DNS, depending on the resolver settings. Note
1206 that if the short form of the hostname is not used, it may pre‐
1207 vent use of hostlist expressions (the numeric portion in brack‐
1208 ets must be at the end of the string). A node range expression
1209 can be used to specify a set of nodes. If an expression is used,
1210 the number of nodes identified by NodeHostname must be identical
1211 to the number of nodes identified by NodeName.
1212
1213 Reason=<reason>
1214 Identify the reason the node is in a "DOWN", "DRAINED", "DRAIN‐
1215 ING", "FAILING" or "FAIL" state. Use quotes to enclose a reason
1216 having more than one word.
1217
1218 State=<state>
1219 Assign one of the following states/actions to the node(s) speci‐
1220 fied by the update command.
1221
1222 CANCEL_REBOOT
1223 Cancels a pending reboot on the node (same as scontrol
1224 cancel_reboot <node>).
1225
1226 DOWN Stop all running and suspended jobs and make the node un‐
1227 available for new jobs.
1228
1229 DRAIN Indicates that no new jobs may be started on this node.
1230 Existing jobs are allowed to run to completion, leaving
1231 the node in a DRAINED state once all the jobs have com‐
1232 pleted.
1233
1234 FAIL Similar to DRAIN except that some applications will seek
1235 to relinquish those nodes before the job completes.
1236
1237 FUTURE Indicates the node is not fully configured, but is ex‐
1238 pected to be available at some point in the future.
1239
1240 NoResp This will set the "Not Responding" flag for a node with‐
1241 out changing its underlying state.
1242
1243 POWER_DOWN
1244 Will use the configured SuspendProgram program to explic‐
1245 itly place a node in power saving mode. If a node is al‐
1246 ready in the process of being powered down, the command
1247 will only change the state of the node but won't have any
1248 effect until the configured SuspendTimeout is reached.
1249 Use of this command can be useful in situations where a
1250 ResumeProgram, like capmc in Cray machines, is stalled
1251 and one wants to restore the node to "IDLE" manually. In
1252 this case rebooting the node and setting the state to
1253 "POWER_DOWN" will cancel the previous "POWER_UP" state
1254 and the node will become "IDLE".
1255
1256 POWER_DOWN_ASAP
1257 Will drain the node and mark it for power down. Currently
1258 running jobs will complete first and no additional jobs
1259 will be allocated to the node.
1260
1261 POWER_DOWN_FORCE
1262 Will cancel all jobs on the node, power it down, and re‐
1263 set its state to "IDLE".
1264
1265 POWER_UP
1266 Will use the configured ResumeProgram program to explic‐
1267 itly move a node out of power saving mode. If a node is
1268 already in the process of being powered up, the command
1269 will only change the state of the node but won't have any
1270 effect until the configured ResumeTimeout is reached.
1271
1272 RESUME Not an actual node state, but will change a node state
1273 from DRAIN, DRAINING, DOWN or REBOOT to IDLE and NoResp.
1274 slurmctld will then attempt to contact slurmd to request
1275 that the node register itself. Once registered, the node
1276 state will then remove the NoResp flag and will resume
1277 normal operations. It will also clear the POWERING_DOWN
1278 state of a node and make it eligible to be allocted.
1279
1280 UNDRAIN
1281 Clears the node from being drained (like RESUME), but
1282 will not change the node's base state (e.g. DOWN).
1283 UNDRAIN requires a valid node registration before new
1284 jobs can be scheduled on the node. Setting a node DOWN
1285 will cause all running and suspended jobs on that node to
1286 be terminated.
1287
1288 While all of the above states are valid, some of them are not
1289 valid new node states given their prior state.
1290
1291 NOTE: The scontrol command should not be used to change node
1292 state on Cray systems. Use Cray tools such as xtprocadmin in‐
1293 stead.
1294
1295 Weight=<weight>
1296 Identify weight to be associated with specified nodes. This al‐
1297 lows dynamic changes to weight associated with nodes, which will
1298 be used for the subsequent node allocation decisions. Weight
1299 assigned via scontrol will only persist across the restart of
1300 the slurmctld daemon with the -R option and state files pre‐
1301 served or slurmctld's receipt of a SIGHUP. Update slurm.conf
1302 with any changes meant to be persistent across normal restarts
1303 of slurmctld or the execution of scontrol reconfig.
1304
1306 AllocMem
1307 The total memory, in MB, currently allocated by jobs on the
1308 node.
1309
1310 CPUSpecList
1311 The list of Slurm abstract CPU IDs on this node reserved for ex‐
1312 clusive use by the Slurm compute node daemons (slurmd, slurm‐
1313 stepd).
1314
1315 FreeMem
1316 The total memory, in MB, currently free on the node as reported
1317 by the OS.
1318
1319 LastBusyTime
1320 The last time the node was busy (i.e. last time the node had
1321 jobs on it). This time is used in PowerSave to determine when to
1322 suspend nodes (e.g. now - LastBusy > SuspendTime).
1323
1324 MemSpecLimit
1325 The combined memory limit, in megabytes, on this node for the
1326 Slurm compute node daemons (slurmd, slurmstepd).
1327
1328 RealMemory
1329 The total memory, in MB, on the node.
1330
1331 State Identify the state(s) assigned to the node with '+' delimited
1332 state flags.
1333
1334 States:
1335
1336 ALLOCATED
1337 Indicates that the node has all CPUs allocated to job(s)
1338 running on the node.
1339
1340 DOWN The node does not have any running jobs and is unavail‐
1341 able for new work.
1342
1343 ERROR The node is in an error state. Consult the logs for more
1344 information about what caused this state.
1345
1346 FUTURE The node is currently not fully configured, but expected
1347 to be available at some point in the indefinite future
1348 for use.
1349
1350 IDLE Indicates that the node is available for work but does
1351 not currently have any jobs assigned to it.
1352
1353 MIXED Indicates that the node is in multiple states. For in‐
1354 stance if only part of the node is ALLOCATED and the rest
1355 of the node is IDLE the state will be MIXED.
1356
1357 UNKNOWN
1358 The node has not yet registered with the controller and
1359 its state is not known.
1360
1361 Flags:
1362
1363 CLOUD Indicates that the node is configured as a cloud node, to
1364 be brought up on demand, but not currently running.
1365
1366 COMPLETING
1367 Indicates that the only job on the node or that all jobs
1368 on the node are in the process of completing.
1369
1370 DRAIN The node is not accepting any new jobs and any currently
1371 running jobs will complete.
1372
1373 DYNAMIC
1374 Slurm allows you to define multiple types of nodes in a
1375 FUTURE state. When starting slurmd on a node you can
1376 specify the -F flag to have the node match and use an ex‐
1377 isting definition in your slurm.conf file. The DYNAMIC
1378 state indicates that the node was started as a Dynamic
1379 Future node.
1380
1381 INVALID_REG
1382 The node did not register correctly with the controller.
1383 This happens when a node registers with less resources
1384 than configured in the slurm.conf file. The node will
1385 clear from this state with a valid registration (i.e. a
1386 slurmd restart is required).
1387
1388 MAINTENANCE
1389 The node is currently in a reservation that includes the
1390 maintenance flag.
1391
1392 NOT_RESPONDING
1393 Node is not responding.
1394
1395 PERFCTRS
1396 Indicates that Network Performance Counters associated
1397 with this node are in use, rendering this node as not us‐
1398 able for any other jobs.
1399
1400 POWER_DOWN
1401 Node is pending power down.
1402
1403 POWERED_DOWN
1404 Node is currently powered down and not capable of running
1405 any jobs.
1406
1407 POWERING_DOWN
1408 Node is in the process of powering down.
1409
1410 POWERING_UP
1411 Node is in the process of powering up.
1412
1413 PLANNED
1414 The node is earmarked for a job that will start in the
1415 future.
1416
1417 REBOOT_ISSUED
1418 A reboot request has been sent to the agent configured to
1419 handle this request.
1420
1421 REBOOT_REQUESTED
1422 A request to reboot this node has been made, but hasn't
1423 been handled yet.
1424
1425 RESERVED
1426 Indicates the node is in an advanced reservation and not
1427 generally available.
1428
1429 The meaning of the energy information is as follows:
1430
1431 CurrentWatts
1432 The instantaneous power consumption of the node at the
1433 time of the last node energy accounting sample, in watts.
1434
1435 LowestJoules
1436 The energy consumed by the node between the last time it
1437 was powered on and the last time it was registered by
1438 slurmd, in joules.
1439
1440 ConsumedJoules
1441 The energy consumed by the node between the last time it
1442 was registered by the slurmd daemon and the last node en‐
1443 ergy accounting sample, in joules.
1444
1445 If the reported value is "n/s" (not supported), the node does
1446 not support the configured AcctGatherEnergyType plugin. If the
1447 reported value is zero, energy accounting for nodes is disabled.
1448
1449
1450 The meaning of the external sensors information is as follows:
1451
1452 ExtSensorsJoules
1453 The energy consumed by the node between the last time it
1454 was powered on and the last external sensors plugin node
1455 sample, in joules.
1456
1457 ExtSensorsWatts
1458 The instantaneous power consumption of the node at the
1459 time of the last external sensors plugin node sample, in
1460 watts.
1461
1462 ExtSensorsTemp
1463 The temperature of the node at the time of the last ex‐
1464 ternal sensors plugin node sample, in celsius.
1465
1466 If the reported value is "n/s" (not supported), the node does
1467 not support the configured ExtSensorsType plugin.
1468
1470 FrontendName=<name>
1471 Identify the front end node to be updated. This specification is
1472 required.
1473
1474 Reason=<reason>
1475 Identify the reason the node is in a "DOWN" or "DRAIN" state.
1476 Use quotes to enclose a reason having more than one word.
1477
1478 State=<state>
1479 Identify the state to be assigned to the front end node. Possi‐
1480 ble values are "DOWN", "DRAIN" or "RESUME". If you want to re‐
1481 move a front end node from service, you typically want to set
1482 its state to "DRAIN". "RESUME" is not an actual node state, but
1483 will return a "DRAINED", "DRAINING", or "DOWN" front end node to
1484 service, either "IDLE" or "ALLOCATED" state as appropriate.
1485 Setting a front end node "DOWN" will cause all running and sus‐
1486 pended jobs on that node to be terminated.
1487
1489 AllowGroups=<name>
1490 Identify the user groups which may use this partition. Multiple
1491 groups may be specified in a comma separated list. To permit
1492 all groups to use the partition specify "AllowGroups=ALL".
1493
1494 AllocNodes=<name>
1495 Comma separated list of nodes from which users can execute jobs
1496 in the partition. Node names may be specified using the node
1497 range expression syntax described above. The default value is
1498 "ALL".
1499
1500 Alternate=<partition name>
1501 Alternate partition to be used if the state of this partition is
1502 "DRAIN" or "INACTIVE." The value "NONE" will clear a previously
1503 set alternate partition.
1504
1505 CpuBind=<node>
1506 Specify the task binding mode to be used by default for this
1507 partition. Supported options include: "none", "socket", "ldom"
1508 (NUMA), "core", "thread" and "off" (remove previous binding
1509 mode).
1510
1511 Default={yes|no}
1512 Specify if this partition is to be used by jobs which do not ex‐
1513 plicitly identify a partition to use. Possible output values
1514 are "YES" and "NO". In order to change the default partition of
1515 a running system, use the scontrol update command and set De‐
1516 fault=yes for the partition that you want to become the new de‐
1517 fault.
1518
1519 DefaultTime=<time>
1520 Run time limit used for jobs that don't specify a value. If not
1521 set then MaxTime will be used. Format is the same as for Max‐
1522 Time.
1523
1524 DefMemPerCPU=<MB>
1525 Set the default memory to be allocated per CPU for jobs in this
1526 partition. The memory size is specified in megabytes.
1527
1528 DefMemPerNode=<MB>
1529 Set the default memory to be allocated per node for jobs in this
1530 partition. The memory size is specified in megabytes.
1531
1532 DisableRootJobs={yes|no}
1533 Specify if jobs can be executed as user root. Possible values
1534 are "YES" and "NO".
1535
1536 GraceTime=<seconds>
1537 Specifies, in units of seconds, the preemption grace time to be
1538 extended to a job which has been selected for preemption. The
1539 default value is zero, no preemption grace time is allowed on
1540 this partition or qos. (Meaningful only for PreemptMode=CANCEL)
1541
1542 Hidden={yes|no}
1543 Specify if the partition and its jobs should be hidden from
1544 view. Hidden partitions will by default not be reported by
1545 Slurm APIs or commands. Possible values are "YES" and "NO".
1546
1547 JobDefaults=<specs>
1548 Specify job default values using a comma-delimited list of
1549 "key=value" pairs. Supported keys include
1550
1551 DefCpuPerGPU Default number of CPUs per allocated GPU.
1552
1553 DefMemPerGPU Default memory limit (in megabytes) per allocated
1554 GPU.
1555
1556 MaxMemPerCPU=<MB>
1557 Set the maximum memory to be allocated per CPU for jobs in this
1558 partition. The memory size is specified in megabytes.
1559
1560 MaxMemPerCNode=<MB>
1561 Set the maximum memory to be allocated per node for jobs in this
1562 partition. The memory size is specified in megabytes.
1563
1564 MaxNodes=<count>
1565 Set the maximum number of nodes which will be allocated to any
1566 single job in the partition. Specify a number, "INFINITE" or
1567 "UNLIMITED". Changing the MaxNodes of a partition has no effect
1568 upon jobs that have already begun execution.
1569
1570 MaxTime=<time>
1571 The maximum run time for jobs. Output format is
1572 [days-]hours:minutes:seconds or "UNLIMITED". Input format (for
1573 update command) is minutes, minutes:seconds, hours:minutes:sec‐
1574 onds, days-hours, days-hours:minutes or days-hours:minutes:sec‐
1575 onds. Time resolution is one minute and second values are
1576 rounded up to the next minute. Changing the MaxTime of a parti‐
1577 tion has no effect upon jobs that have already begun execution.
1578
1579 MinNodes=<count>
1580 Set the minimum number of nodes which will be allocated to any
1581 single job in the partition. Changing the MinNodes of a parti‐
1582 tion has no effect upon jobs that have already begun execution.
1583 Increasing this value may prevent pending jobs from starting,
1584 even if they were submitted without -N/--nodes specification.
1585 If you do get in that situation, updating the MinNodes value of
1586 a pending job using the scontrol command will allow that job to
1587 be scheduled.
1588
1589 Nodes=<name>
1590 Identify the node(s) to be associated with this partition. Mul‐
1591 tiple node names may be specified using simple node range ex‐
1592 pressions (e.g. "lx[10-20]"). Note that jobs may only be asso‐
1593 ciated with one partition at any time. Specify a blank data
1594 value to remove all nodes from a partition: "Nodes=". Changing
1595 the Nodes in a partition has no effect upon jobs that have al‐
1596 ready begun execution.
1597
1598 OverTimeLimit=<count>
1599 Number of minutes by which a job can exceed its time limit be‐
1600 fore being canceled. The configured job time limit is treated
1601 as a soft limit. Adding OverTimeLimit to the soft limit pro‐
1602 vides a hard limit, at which point the job is canceled. This is
1603 particularly useful for backfill scheduling, which bases upon
1604 each job's soft time limit. A partition-specific OverTimeLimit
1605 will override any global OverTimeLimit value. If not specified,
1606 the global OverTimeLimit value will take precedence. May not
1607 exceed 65533 minutes. An input value of "UNLIMITED" will clear
1608 any previously configured partition-specific OverTimeLimit
1609 value.
1610
1611 OverSubscribe={yes|no|exclusive|force}[:<job_count>]
1612 Specify if compute resources (i.e. individual CPUs) in this par‐
1613 tition can be shared by multiple jobs. Possible values are
1614 "YES", "NO", "EXCLUSIVE" and "FORCE". An optional job count
1615 specifies how many jobs can be allocated to use each resource.
1616
1617 PartitionName=<name>
1618 Identify the partition to be updated. This specification is re‐
1619 quired.
1620
1621 PreemptMode=<mode>
1622 Reset the mechanism used to preempt jobs in this partition if
1623 PreemptType is configured to preempt/partition_prio. The default
1624 preemption mechanism is specified by the cluster-wide Preempt‐
1625 Mode configuration parameter. Possible values are "OFF", "CAN‐
1626 CEL", "REQUEUE" and "SUSPEND".
1627
1628 Priority=<count>
1629 Jobs submitted to a higher priority partition will be dispatched
1630 before pending jobs in lower priority partitions and if possible
1631 they will preempt running jobs from lower priority partitions.
1632 Note that a partition's priority takes precedence over a job's
1633 priority. The value may not exceed 65533.
1634
1635 PriorityJobFactor=<count>
1636 Partition factor used by priority/multifactor plugin in calcu‐
1637 lating job priority. The value may not exceed 65533. Also see
1638 PriorityTier.
1639
1640 PriorityTier=<count>
1641 Jobs submitted to a partition with a higher priority tier value
1642 will be dispatched before pending jobs in partition with lower
1643 priority tier value and, if possible, they will preempt
1644 running jobs from partitions with lower priority tier values.
1645 Note that a partition's priority tier takes precedence over a
1646 job's priority. The value may not exceed 65533. Also see Pri‐
1647 orityJobFactor.
1648
1649 QOS=<QOSname|blank to remove>
1650 Set the partition QOS with a QOS name or to remove the Partition
1651 QOS leave the option blank.
1652
1653 RootOnly={yes|no}
1654 Specify if only allocation requests initiated by user root will
1655 be satisfied. This can be used to restrict control of the par‐
1656 tition to some meta-scheduler. Possible values are "YES" and
1657 "NO".
1658
1659 ReqResv={yes|no}
1660 Specify if only allocation requests designating a reservation
1661 will be satisfied. This is used to restrict partition usage to
1662 be allowed only within a reservation. Possible values are "YES"
1663 and "NO".
1664
1665 Shared={yes|no|exclusive|force}[:<job_count>]
1666 Renamed to OverSubscribe, see option descriptions above.
1667
1668 State={up|down|drain|inactive}
1669 Specify if jobs can be allocated nodes or queued in this parti‐
1670 tion. Possible values are "UP", "DOWN", "DRAIN" and "INACTIVE".
1671
1672 UP Designates that new jobs may queued on the partition,
1673 and that jobs may be allocated nodes and run from the
1674 partition.
1675
1676 DOWN Designates that new jobs may be queued on the parti‐
1677 tion, but queued jobs may not be allocated nodes and
1678 run from the partition. Jobs already running on the
1679 partition continue to run. The jobs must be explicitly
1680 canceled to force their termination.
1681
1682 DRAIN Designates that no new jobs may be queued on the par‐
1683 tition (job submission requests will be denied with an
1684 error message), but jobs already queued on the parti‐
1685 tion may be allocated nodes and run. See also the
1686 "Alternate" partition specification.
1687
1688 INACTIVE Designates that no new jobs may be queued on the par‐
1689 tition, and jobs already queued may not be allocated
1690 nodes and run. See also the "Alternate" partition
1691 specification.
1692
1693 TRESBillingWeights=<TRES Billing Weights>
1694 TRESBillingWeights is used to define the billing weights of each
1695 TRES type that will be used in calculating the usage of a job.
1696 The calculated usage is used when calculating fairshare and when
1697 enforcing the TRES billing limit on jobs. Updates affect new
1698 jobs and not existing jobs. See the slurm.conf man page for
1699 more information.
1700
1702 Reservation=<name>
1703 Identify the name of the reservation to be created, updated, or
1704 deleted. This parameter is required for update and is the only
1705 parameter for delete. For create, if you do not want to give a
1706 reservation name, use "scontrol create res ..." and a name will
1707 be created automatically.
1708
1709 Accounts=<account list>
1710 List of accounts permitted to use the reserved nodes, for exam‐
1711 ple "Accounts=physcode1,physcode2". A user in any of the ac‐
1712 counts may use the reserved nodes. A new reservation must spec‐
1713 ify Users or Groups and/or Accounts. If both Users/Groups and
1714 Accounts are specified, a job must match both in order to use
1715 the reservation. Accounts can also be denied access to reserva‐
1716 tions by preceding all of the account names with '-'. Alter‐
1717 nately precede the equal sign with '-'. For example, "Ac‐
1718 counts=-physcode1,-physcode2" or "Accounts-=physcode1,physcode2"
1719 will permit any account except physcode1 and physcode2 to use
1720 the reservation. You can add or remove individual accounts from
1721 an existing reservation by using the update command and adding a
1722 '+' or '-' sign before the '=' sign. If accounts are denied ac‐
1723 cess to a reservation (account name preceded by a '-'), then all
1724 other accounts are implicitly allowed to use the reservation and
1725 it is not possible to also explicitly specify allowed accounts.
1726
1727 BurstBuffer=<buffer_spec>[,<buffer_spec>,...]
1728 Specification of burst buffer resources which are to be re‐
1729 served. "buffer_spec" consists of four elements:
1730 [plugin:][type:]#[units] "plugin" is the burst buffer plugin
1731 name, currently either "datawarp" or "generic". If no plugin is
1732 specified, the reservation applies to all configured burst buf‐
1733 fer plugins. "type" specifies a Cray generic burst buffer re‐
1734 source, for example "nodes". if "type" is not specified, the
1735 number is a measure of storage space. The "units" may be "N"
1736 (nodes), "K|KiB", "M|MiB", "G|GiB", "T|TiB", "P|PiB" (for powers
1737 of 1024) and "KB", "MB", "GB", "TB", "PB" (for powers of 1000).
1738 The default units are bytes for reservations of storage space.
1739 For example "BurstBuffer=datawarp:2TB" (reserve 2TB of storage
1740 plus 3 nodes from the Cray plugin) or "BurstBuffer=100GB" (re‐
1741 serve 100 GB of storage from all configured burst buffer plug‐
1742 ins). Jobs using this reservation are not restricted to these
1743 burst buffer resources, but may use these reserved resources
1744 plus any which are generally available. NOTE: Usually Slurm in‐
1745 terprets KB, MB, GB, TB, PB, TB units as powers of 1024, but for
1746 Burst Buffers size specifications Slurm supports both IEC/SI
1747 formats. This is because the CRAY API for managing DataWarps
1748 supports both formats.
1749
1750 CoreCnt=<num>
1751 This option is only supported when SelectType=select/cons_res or
1752 select/cons_tres. Identify number of cores to be reserved. If
1753 NodeCnt is used without the FIRST_CORES flag, this is the total
1754 number of cores to reserve where cores per node is
1755 CoreCnt/NodeCnt. If a nodelist is used, or if NodeCnt is used
1756 with the FIRST_CORES flag, this should be an array of core num‐
1757 bers by node: Nodes=node[1-5] CoreCnt=2,2,3,3,4 or
1758 flags=FIRST_CORES NodeCnt=5 CoreCnt=1,2,1,3,2.
1759
1760 Licenses=<license>
1761 Specification of licenses (or other resources available on all
1762 nodes of the cluster) which are to be reserved. License names
1763 can be followed by a colon and count (the default count is one).
1764 Multiple license names should be comma separated (e.g. "Li‐
1765 censes=foo:4,bar"). A new reservation must specify one or more
1766 resource to be included: NodeCnt, Nodes and/or Licenses. If a
1767 reservation includes Licenses, but no NodeCnt or Nodes, then the
1768 option Flags=LICENSE_ONLY must also be specified. Jobs using
1769 this reservation are not restricted to these licenses, but may
1770 use these reserved licenses plus any which are generally avail‐
1771 able.
1772
1773 MaxStartDelay[=<timespec>]
1774 Change MaxStartDelay value which specifies the maximum time an
1775 eligible job not requesting this reservation can delay a job re‐
1776 questing it. Default is none. Valid formats are minutes, min‐
1777 utes:seconds, hours:minutes:seconds, days-hours, days-hours:min‐
1778 utes, days-hours:minutes:seconds. Time resolution is one minute
1779 and second values are rounded up to the next minute. Output for‐
1780 mat is always [days-]hours:minutes:seconds.
1781
1782 NodeCnt=<num>[,<num>,...]
1783 Identify number of nodes to be reserved. The number can include
1784 a suffix of "k" or "K", in which case the number specified is
1785 multiplied by 1024. A new reservation must specify one or more
1786 resource to be included: NodeCnt, Nodes and/or Licenses.
1787
1788 Nodes=<name>
1789 Identify the node(s) to be reserved. Multiple node names may be
1790 specified using simple node range expressions (e.g.
1791 "Nodes=lx[10-20]"). Specify a blank data value to remove all
1792 nodes from a reservation: "Nodes=". A new reservation must
1793 specify one or more resource to be included: NodeCnt, Nodes
1794 and/or Licenses. A specification of "ALL" will reserve all
1795 nodes. Set Flags=PART_NODES and PartitionName= in order for
1796 changes in the nodes associated with a partition to also be re‐
1797 flected in the nodes associated with a reservation.
1798
1799 StartTime=<time_spec>
1800 The start time for the reservation. A new reservation must
1801 specify a start time. It accepts times of the form HH:MM:SS for
1802 a specific time of day (seconds are optional). (If that time is
1803 already past, the next day is assumed.) You may also specify
1804 midnight, noon, fika (3 PM) or teatime (4 PM) and you can have a
1805 time-of-day suffixed with AM or PM for running in the morning or
1806 the evening. You can also say what day the job will be run, by
1807 specifying a date of the form MMDDYY or MM/DD/YY or MM.DD.YY, or
1808 a date and time as YYYY-MM-DD[THH:MM[:SS]]. You can also give
1809 times like now + count time-units, where the time-units can be
1810 seconds (default), minutes, hours, days, or weeks and you can
1811 tell Slurm to run the job today with the keyword today and to
1812 run the job tomorrow with the keyword tomorrow. You cannot up‐
1813 date the StartTime of a reservation in ACTIVE state.
1814
1815 EndTime=<time_spec>
1816 The end time for the reservation. A new reservation must spec‐
1817 ify an end time or a duration. Valid formats are the same as
1818 for StartTime.
1819
1820 Duration=<time>
1821 The length of a reservation. A new reservation must specify an
1822 end time or a duration. Valid formats are minutes, minutes:sec‐
1823 onds, hours:minutes:seconds, days-hours, days-hours:minutes,
1824 days-hours:minutes:seconds, or UNLIMITED. Time resolution is
1825 one minute and second values are rounded up to the next minute.
1826 Output format is always [days-]hours:minutes:seconds.
1827
1828 PartitionName=<name>
1829 Partition used to reserve nodes from. This will attempt to allo‐
1830 cate all nodes in the specified partition unless you request
1831 fewer resources than are available with CoreCnt, NodeCnt or
1832 TRES. Jobs will be allowed to use this reservation even if run‐
1833 ning in a different partition. There only needs to be overlap‐
1834 ping nodes from that different partition and the nodes used in
1835 the reservation.
1836
1837 Flags=<flags>
1838 Flags associated with the reservation. You can add or remove
1839 individual flags from an existing reservation by adding a '+' or
1840 '-' sign before the '=' sign. For example: Flags-=DAILY (NOTE:
1841 this shortcut is not supported for all flags). Currently sup‐
1842 ported flags include:
1843
1844 ANY_NODES This is a reservation for burst buffers and/or li‐
1845 censes only and not compute nodes. If this flag
1846 is set, a job using this reservation may use the
1847 associated burst buffers and/or licenses plus any
1848 compute nodes. If this flag is not set, a job us‐
1849 ing this reservation may use only the nodes and
1850 licenses associated with the reservation.
1851
1852 DAILY Repeat the reservation at the same time every day.
1853
1854 FLEX Permit jobs requesting the reservation to begin
1855 prior to the reservation's start time, end after
1856 the reservation's end time, and use any resources
1857 inside and/or outside of the reservation regard‐
1858 less of any constraints possibly set in the reser‐
1859 vation. A typical use case is to prevent jobs not
1860 explicitly requesting the reservation from using
1861 those reserved resources rather than forcing jobs
1862 requesting the reservation to use those resources
1863 in the time frame reserved. Another use case could
1864 be to always have a particular number of nodes
1865 with a specific feature reserved for a specific
1866 account so users in this account may use this
1867 nodes plus possibly other nodes without this fea‐
1868 ture.
1869
1870 FIRST_CORES Use the lowest numbered cores on a node only. Flag
1871 removal with '-=' is not supported.
1872
1873 IGNORE_JOBS Ignore currently running jobs when creating the
1874 reservation. This can be especially useful when
1875 reserving all nodes in the system for maintenance.
1876
1877 LICENSE_ONLY See ANY_NODES.
1878
1879 MAINT Maintenance mode, receives special accounting
1880 treatment. This reservation is permitted to use
1881 resources that are already in another reservation.
1882
1883 MAGNETIC This flag allows jobs to be considered for this
1884 reservation even if they didn't request it.
1885
1886 NO_HOLD_JOBS_AFTER
1887 By default, when a reservation ends the reserva‐
1888 tion request will be removed from any pending jobs
1889 submitted to the reservation and will be put into
1890 a held state. Use this flag to let jobs run out‐
1891 side of the reservation after the reservation is
1892 gone. Flag removal with '-=' is not supported.
1893
1894 OVERLAP This reservation can be allocated resources that
1895 are already in another reservation. Flag removal
1896 with '-=' is not supported.
1897
1898 PART_NODES This flag can be used to reserve all nodes within
1899 the specified partition. PartitionName and
1900 Nodes=ALL must be specified with this flag.
1901
1902 PURGE_COMP[=<timespec>]
1903 Purge the reservation if it is ever idle for time‐
1904 spec (no jobs associated with it). If timespec
1905 isn't given then 5 minutes is the default. Valid
1906 timespec formats are minutes, minutes:seconds,
1907 hours:minutes:seconds, days-hours, days-hours:min‐
1908 utes, days-hours:minutes:seconds. Time resolution
1909 is one minute and second values are rounded up to
1910 the next minute. Output format is always
1911 [days-]hours:minutes:seconds.
1912
1913 REPLACE Nodes which are DOWN, DRAINED, or allocated to
1914 jobs are automatically replenished using idle re‐
1915 sources. This option can be used to maintain a
1916 constant number of idle resources available for
1917 pending jobs (subject to availability of idle re‐
1918 sources). This should be used with the NodeCnt
1919 reservation option; do not identify specific nodes
1920 to be included in the reservation. Flag removal
1921 with '-=' is not supported.
1922
1923 NOTE: Removing a node from the cluster while in a
1924 reservation with the REPLACE flag will not cause
1925 it to be replaced.
1926
1927 REPLACE_DOWN Nodes which are DOWN or DRAINED are automatically
1928 replenished using idle resources. This option can
1929 be used to maintain a constant sized pool of re‐
1930 sources available for pending jobs (subject to
1931 availability of idle resources). This should be
1932 used with the NodeCnt reservation option; do not
1933 identify specific nodes to be included in the
1934 reservation. Flag removal with '-=' is not sup‐
1935 ported.
1936
1937 NOTE: Removing a node from the cluster while in a
1938 reservation with the REPLACE_DOWN flag will not
1939 cause it to be replaced.
1940
1941 SPEC_NODES Reservation is for specific nodes (output only).
1942
1943 STATIC_ALLOC Make it so after the nodes are selected for a
1944 reservation they don't change. Without this op‐
1945 tion when nodes are selected for a reservation and
1946 one goes down the reservation will select a new
1947 node to fill the spot.
1948
1949 TIME_FLOAT The reservation start time is relative to the cur‐
1950 rent time and moves forward through time (e.g. a
1951 StartTime=now+10minutes will always be 10 minutes
1952 in the future). Repeating (e.g. DAILY) floating
1953 reservations are not supported. Flag cannot be
1954 added to or removed from an existing reservation.
1955
1956 WEEKDAY Repeat the reservation at the same time on every
1957 weekday (Monday, Tuesday, Wednesday, Thursday and
1958 Friday).
1959
1960 WEEKEND Repeat the reservation at the same time on every
1961 weekend day (Saturday and Sunday).
1962
1963 WEEKLY Repeat the reservation at the same time every
1964 week.
1965
1966 Features=<features>
1967 Set the reservation's required node features. Multiple values
1968 may be "&" separated if all features are required (AND opera‐
1969 tion) or separated by "|" if any of the specified features are
1970 required (OR operation). Parenthesis are also supported for
1971 features to be ANDed together with counts of nodes having the
1972 specified features. For example "Fea‐
1973 tures=[(knl&a2a&flat)*4&haswell*2]" indicates the advanced
1974 reservation should include 4 nodes with ALL of the features
1975 "knl", "a2a", and "flat" plus 2 nodes with the feature
1976 "haswell".
1977
1978 Value may be cleared with blank data value, "Features=".
1979
1980 Groups=<group list>
1981 List of groups permitted to use the reserved nodes, for example
1982 "Group=bio,chem". A new reservation must specify Users or
1983 Groups and/or Accounts. If both Users/Groups and Accounts are
1984 specified, a job must match both in order to use the reserva‐
1985 tion. Unlike users groups do not allow denied access to reser‐
1986 vations. You can add or remove individual groups from an exist‐
1987 ing reservation by using the update command and adding a '+' or
1988 '-' sign before the '=' sign. NOTE: Groups and Users are mutu‐
1989 ally exclusive in reservations, if you want to switch between
1990 the 2 you must update the reservation with a group='' or user=''
1991 and fill in the opposite with the appropriate setting.
1992
1993 Skip Used on a reoccurring reservation, skip to the next reservation
1994 iteration. NOTE: Only available for update.
1995
1996 Users=<user list>
1997 List of users permitted to use the reserved nodes, for example
1998 "User=jones1,smith2". A new reservation must specify Users or
1999 Groups and/or Accounts. If both Users/Groups and Accounts are
2000 specified, a job must match both in order to use the reserva‐
2001 tion. Users can also be denied access to reservations by pre‐
2002 ceding all of the user names with '-'. Alternately precede the
2003 equal sign with '-'. For example, "User=-jones1,-smith2" or
2004 "User-=jones1,smith2" will permit any user except jones1 and
2005 smith2 to use the reservation. You can add or remove individual
2006 users from an existing reservation by using the update command
2007 and adding a '+' or '-' sign before the '=' sign. If users are
2008 denied access to a reservation (user name preceded by a '-'),
2009 then all other users are implicitly allowed to use the reserva‐
2010 tion and it is not possible to also explicitly specify allowed
2011 users. NOTE: Groups and Users are mutually exclusive in reser‐
2012 vations, if you want to switch between the 2 you must update the
2013 reservation with a group='' or user='' and fill in the opposite
2014 with the appropriate setting.
2015
2016 TRES=<tres_spec>
2017 Comma-separated list of TRES required for the reservation. Cur‐
2018 rent supported TRES types with reservations are: CPU, Node, Li‐
2019 cense and BB. CPU and Node follow the same format as CoreCnt and
2020 NodeCnt parameters respectively. License names can be followed
2021 by an equal '=' and a count:
2022
2023 License/<name1>=<count1>[,License/<name2>=<count2>,...]
2024
2025 BurstBuffer can be specified in a similar way as BurstBuffer pa‐
2026 rameter. The only difference is that colon symbol ':' should be
2027 replaced by an equal '=' in order to follow the TRES format.
2028
2029 Some examples of TRES valid specifications:
2030
2031 TRES=cpu=5,bb/cray=4,license/iop1=1,license/iop2=3
2032
2033 TRES=node=5k,license/iop1=2
2034
2035 As specified in CoreCnt, if a nodelist is specified, cpu can be
2036 an array of core numbers by node: nodes=compute[1-3]
2037 TRES=cpu=2,2,1,bb/cray=4,license/iop1=2
2038
2039 Please note that CPU, Node, License and BB can override CoreCnt,
2040 NodeCnt, Licenses and BurstBuffer parameters respectively. Also
2041 CPU represents CoreCnt, in a reservation and will be adjusted if
2042 you have threads per core on your nodes.
2043
2044 Note that a reservation that contains nodes or cores is associ‐
2045 ated with one partition, and can't span resources over multiple
2046 partitions. The only exception from this is when the reserva‐
2047 tion is created with explicitly requested nodes.
2048
2050 Executing scontrol sends a remote procedure call to slurmctld. If
2051 enough calls from scontrol or other Slurm client commands that send re‐
2052 mote procedure calls to the slurmctld daemon come in at once, it can
2053 result in a degradation of performance of the slurmctld daemon, possi‐
2054 bly resulting in a denial of service.
2055
2056 Do not run scontrol or other Slurm client commands that send remote
2057 procedure calls to slurmctld from loops in shell scripts or other pro‐
2058 grams. Ensure that programs limit calls to scontrol to the minimum nec‐
2059 essary for the information you are trying to gather.
2060
2061
2063 Some scontrol options may be set via environment variables. These envi‐
2064 ronment variables, along with their corresponding options, are listed
2065 below. (Note: Command line options will always override these set‐
2066 tings.)
2067
2068
2069 SCONTROL_ALL -a, --all
2070
2071 SCONTROL_FEDERATION --federation
2072
2073 SCONTROL_FUTURE -F, --future
2074
2075 SCONTROL_LOCAL --local
2076
2077 SCONTROL_SIBLING --sibling
2078
2079 SLURM_BITSTR_LEN Specifies the string length to be used for holding
2080 a job array's task ID expression. The default
2081 value is 64 bytes. A value of 0 will print the
2082 full expression with any length required. Larger
2083 values may adversely impact the application perfor‐
2084 mance.
2085
2086 SLURM_CLUSTERS Same as --clusters
2087
2088 SLURM_CONF The location of the Slurm configuration file.
2089
2090 SLURM_CONF_OUT When running 'write config', the location of the
2091 Slurm configuration file to be written.
2092
2093 SLURM_TIME_FORMAT Specify the format used to report time stamps. A
2094 value of standard, the default value, generates
2095 output in the form
2096 "year-month-dateThour:minute:second". A value of
2097 relative returns only "hour:minute:second" if the
2098 current day. For other dates in the current year
2099 it prints the "hour:minute" preceded by "Tomorr"
2100 (tomorrow), "Ystday" (yesterday), the name of the
2101 day for the coming week (e.g. "Mon", "Tue", etc.),
2102 otherwise the date (e.g. "25 Apr"). For other
2103 years it returns a date month and year without a
2104 time (e.g. "6 Jun 2012"). All of the time stamps
2105 use a 24 hour format.
2106
2107 A valid strftime() format can also be specified.
2108 For example, a value of "%a %T" will report the day
2109 of the week and a time stamp (e.g. "Mon 12:34:56").
2110
2111 SLURM_TOPO_LEN Specify the maximum size of the line when printing
2112 Topology. If not set, the default value is unlim‐
2113 ited.
2114
2116 When using SlurmDBD, users who have an AdminLevel defined (Operator or
2117 Admin) and users who are account coordinators are given the authority
2118 to view and modify jobs, reservations, nodes, etc., as defined in the
2119 following table - regardless of whether a PrivateData restriction has
2120 been defined in the slurm.conf file.
2121
2122 scontrol show job(s): Admin, Operator, Coordinator
2123 scontrol update job: Admin, Operator, Coordinator
2124 scontrol requeue: Admin, Operator, Coordinator
2125 scontrol show step(s): Admin, Operator, Coordinator
2126 scontrol update step: Admin, Operator, Coordinator
2127
2128 scontrol show node: Admin, Operator
2129 scontrol update node: Admin
2130
2131 scontrol create partition: Admin
2132 scontrol show partition: Admin, Operator
2133 scontrol update partition: Admin
2134 scontrol delete partition: Admin
2135
2136 scontrol create reservation: Admin, Operator
2137 scontrol show reservation: Admin, Operator
2138 scontrol update reservation: Admin, Operator
2139 scontrol delete reservation: Admin, Operator
2140
2141 scontrol reconfig: Admin
2142 scontrol shutdown: Admin
2143 scontrol takeover: Admin
2144
2145
2147 $ scontrol
2148 scontrol: show part debug
2149 PartitionName=debug
2150 AllocNodes=ALL AllowGroups=ALL Default=YES
2151 DefaultTime=NONE DisableRootJobs=NO Hidden=NO
2152 MaxNodes=UNLIMITED MaxTime=UNLIMITED MinNodes=1
2153 Nodes=snowflake[0-48]
2154 Priority=1 RootOnly=NO OverSubscribe=YES:4
2155 State=UP TotalCPUs=694 TotalNodes=49
2156 scontrol: update PartitionName=debug MaxTime=60:00 MaxNodes=4
2157 scontrol: show job 71701
2158 JobId=71701 Name=hostname
2159 UserId=da(1000) GroupId=da(1000)
2160 Priority=66264 Account=none QOS=normal WCKey=*123
2161 JobState=COMPLETED Reason=None Dependency=(null)
2162 TimeLimit=UNLIMITED Requeue=1 Restarts=0 BatchFlag=0 ExitCode=0:0
2163 SubmitTime=2010-01-05T10:58:40 EligibleTime=2010-01-05T10:58:40
2164 StartTime=2010-01-05T10:58:40 EndTime=2010-01-05T10:58:40
2165 SuspendTime=None SecsPreSuspend=0
2166 Partition=debug AllocNode:Sid=snowflake:4702
2167 ReqNodeList=(null) ExcNodeList=(null)
2168 NodeList=snowflake0
2169 NumNodes=1 NumCPUs=10 CPUs/Task=2 ReqS:C:T=1:1:1
2170 MinCPUsNode=2 MinMemoryNode=0 MinTmpDiskNode=0
2171 Features=(null) Reservation=(null)
2172 OverSubscribe=OK Contiguous=0 Licenses=(null) Network=(null)
2173 scontrol: update JobId=71701 TimeLimit=30:00 Priority=500
2174 scontrol: show hostnames tux[1-3]
2175 tux1
2176 tux2
2177 tux3
2178 scontrol: create res StartTime=2009-04-01T08:00:00 Duration=5:00:00 Users=dbremer NodeCnt=10
2179 Reservation created: dbremer_1
2180 scontrol: update Reservation=dbremer_1 Flags=Maint NodeCnt=20
2181 scontrol: delete Reservation=dbremer_1
2182 scontrol: quit
2183
2184
2186 Copyright (C) 2002-2007 The Regents of the University of California.
2187 Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER).
2188 Copyright (C) 2008-2010 Lawrence Livermore National Security.
2189 Copyright (C) 2010-2022 SchedMD LLC.
2190
2191 This file is part of Slurm, a resource management program. For de‐
2192 tails, see <https://slurm.schedmd.com/>.
2193
2194 Slurm is free software; you can redistribute it and/or modify it under
2195 the terms of the GNU General Public License as published by the Free
2196 Software Foundation; either version 2 of the License, or (at your op‐
2197 tion) any later version.
2198
2199 Slurm is distributed in the hope that it will be useful, but WITHOUT
2200 ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
2201 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
2202 for more details.
2203
2205 /etc/slurm.conf
2206
2208 scancel(1), sinfo(1), squeue(1), slurm_create_partition [22m(3),
2209 slurm_delete_partition [22m(3), slurm_load_ctl_conf [22m(3), slurm_load_jobs
2210 (3), slurm_load_node (3), slurm_load_partitions (3), slurm_reconfigure
2211 (3), slurm_requeue [22m(3), slurm_resume [22m(3), slurm_shutdown [22m(3),
2212 slurm_suspend (3), slurm_takeover (3), slurm_update_job (3), slurm_up‐
2213 date_node (3), slurm_update_partition (3), slurm.conf(5), slurmctld(8)
2214
2215
2216
2217March 2022 Slurm Commands scontrol(1)