1scontrol(1)                     Slurm Commands                     scontrol(1)
2
3
4

NAME

6       scontrol - view or modify Slurm configuration and state.
7
8

SYNOPSIS

10       scontrol [OPTIONS...] [COMMAND...]
11
12

DESCRIPTION

14       scontrol  is used to view or modify Slurm configuration including: job,
15       job step, node, partition, reservation, and overall  system  configura‐
16       tion.  Most  of  the  commands  can only be executed by user root or an
17       Administrator. If an attempt to view or modify  configuration  informa‐
18       tion  is made by an unauthorized user, an error message will be printed
19       and the requested action will not occur.  If no command is  entered  on
20       the  execute  line,  scontrol  will  operate in an interactive mode and
21       prompt for input. It will continue prompting for  input  and  executing
22       commands  until  explicitly  terminated. If a command is entered on the
23       execute line, scontrol will execute that  command  and  terminate.  All
24       commands  and options are case-insensitive, although node names, parti‐
25       tion names, and reservation names are case-sensitive (node  names  "LX"
26       and  "lx" are distinct). All commands and options can be abbreviated to
27       the extent that the specification is unique. A modified Slurm  configu‐
28       ration  can  be  written to a file using the scontrol write config com‐
29       mand.  The  resulting  file  will  be  named   using   the   convention
30       "slurm.conf.<datetime>" and located in the same directory as the origi‐
31       nal "slurm.conf" file. The directory containing the original slurm.conf
32       must be writable for this to occur.
33
34

OPTIONS

36       -a, --all
37              When  the  show  command  is  used, then display all partitions,
38              their jobs and jobs steps. This causes information  to  be  dis‐
39              played about partitions that are configured as hidden and parti‐
40              tions that are unavailable to user's group.
41
42       -d, --details
43              Causes the show command  to  provide  additional  details  where
44              available.
45
46       --federation
47              Report jobs from from federation if a member of one.
48
49       -F, --future
50              Report nodes in FUTURE state.
51
52       -h, --help
53              Print a help message describing the usage of scontrol.
54
55       --hide Do  not  display information about hidden partitions, their jobs
56              and job steps.  By default, neither partitions that are  config‐
57              ured  as hidden nor those partitions unavailable to user's group
58              will be displayed (i.e. this is the default behavior).
59
60       --local
61              Show only information local to this cluster. Ignore other  clus‐
62              ters  in  the  federated if a member of one. Overrides --federa‐
63              tion.
64
65       -M, --clusters=<string>
66              The cluster to issue commands to. Only one cluster name  may  be
67              specified.  Note that the SlurmDBD must be up for this option to
68              work properly.  This option implicitly sets the --local option.
69
70
71       -o, --oneliner
72              Print information one line per record.
73
74       -Q, --quiet
75              Print no warning or informational  messages,  only  fatal  error
76              messages.
77
78       --sibling
79              Show  all sibling jobs on a federated cluster. Implies --federa‐
80              tion.
81
82       -u, --uid=<uid>
83              Attempt to update a job as user <uid> instead  of  the  invoking
84              user id.
85
86       -v, --verbose
87              Print   detailed  event  logging.  Multiple  -v's  will  further
88              increase the verbosity of logging. By default only  errors  will
89              be displayed.
90
91
92       -V , --version
93              Print version information and exit.
94
95       COMMANDS
96
97
98       abort  Instruct  the Slurm controller to terminate immediately and gen‐
99              erate a core file.  See "man slurmctld"  for  information  about
100              where the core file will be written.
101
102
103       cancel_reboot <NodeList>
104              Cancel pending reboots on nodes.
105
106
107       checkpoint CKPT_OP ID
108              Perform a checkpoint activity on the job step(s) with the speci‐
109              fied identification.  ID can be used to identify a specific  job
110              (e.g. "<job_id>", which applies to all of its existing steps) or
111              a specific job  step  (e.g.  "<job_id>.<step_id>").   Acceptable
112              values for CKPT_OP include:
113
114              able        Test if presently not disabled, report start time if
115                          checkpoint in progress
116
117              create      Create a checkpoint and continue the job or job step
118
119              disable     Disable future checkpoints
120
121              enable      Enable future checkpoints
122
123              error       Report the result for the last  checkpoint  request,
124                          error code and message
125
126              restart     Restart execution of the previously checkpointed job
127                          or job step
128
129              requeue     Create a checkpoint and requeue the batch job,  com‐
130                          bines vacate and restart operations
131
132              vacate      Create  a  checkpoint  and  terminate the job or job
133                          step
134       Acceptable values for CKPT_OP include:
135
136              MaxWait=<seconds>   Maximum time for checkpoint to  be  written.
137                                  Default  value  is  10  seconds.  Valid with
138                                  create and vacate options only.
139
140              ImageDir=<directory_name>
141                                  Location of  checkpoint  file.   Valid  with
142                                  create,  vacate  and  restart  options only.
143                                  This value takes precedent over any --check‐
144                                  point-dir  value specified at job submission
145                                  time.
146
147              StickToNodes        If set, resume job on  the  same  nodes  are
148                                  previously  used.   Valid  with  the restart
149                                  option only.
150
151
152       cluster CLUSTER_NAME
153              The cluster to issue commands to. Only one cluster name  may  be
154              specified.
155
156
157       create SPECIFICATION
158              Create  a  new  partition  or reservation.  See the full list of
159              parameters below.  Include the tag "res" to create a reservation
160              without specifying a reservation name.
161
162
163       completing
164              Display  all  jobs  in  a COMPLETING state along with associated
165              nodes in either a COMPLETING or DOWN state.
166
167
168       delete SPECIFICATION
169              Delete the entry with  the  specified  SPECIFICATION.   The  two
170              SPECIFICATION  choices  are  PartitionName=<name>  and  Reserva‐
171              tion=<name>. Reservations and partitions should have no  associ‐
172              ated jobs at the time of their deletion (modify the jobs first).
173              If the specified partition is in use, the request is denied.
174
175
176       errnumstr ERRNO
177              Given a Slurm error number, return a descriptive string.
178
179
180       fsdampeningfactor FACTOR
181              Set the FairShareDampeningFactor in slurmctld.
182
183
184       help   Display a description of scontrol options and commands.
185
186
187       hold job_list
188              Prevent a pending job from being started (sets it's priority  to
189              0).   Use the release command to permit the job to be scheduled.
190              The job_list argument is a comma separated list of  job  IDs  OR
191              "jobname="  with  the job's name, which will attempt to hold all
192              jobs having that name.  Note that when a job is held by a system
193              administrator  using the hold command, only a system administra‐
194              tor may release the job for execution (also see the  uhold  com‐
195              mand).  When  the  job  is  held  by  its  owner, it may also be
196              released by the job's owner.  Additionally, attempting to hold a
197              running job will have not suspend or cancel it. But, it will set
198              the job priority to 0 and update the  job  reason  field,  which
199              would hold the job if it was requeued at a later time.
200
201
202       notify job_id message
203              Send  a  message to standard error of the salloc or srun command
204              or batch job associated with the specified job_id.
205
206
207       pidinfo proc_id
208              Print the Slurm job id and  scheduled  termination  time  corre‐
209              sponding  to  the  supplied  process id, proc_id, on the current
210              node.  This will work only with processes on node on which scon‐
211              trol  is  run, and only for those processes spawned by Slurm and
212              their descendants.
213
214
215       listpids [job_id[.step_id]] [NodeName]
216              Print  a  listing  of  the  process  IDs  in  a  job  step   (if
217              JOBID.STEPID  is provided), or all of the job steps in a job (if
218              job_id is provided), or all of the job steps in all of the  jobs
219              on  the local node (if job_id is not provided or job_id is "*").
220              This will work only with processes on the node on which scontrol
221              is  run, and only for those processes spawned by Slurm and their
222              descendants. Note that some Slurm configurations  (ProctrackType
223              value  of  pgid) are unable to identify all processes associated
224              with a job or job step.
225
226              Note that the NodeName option is only  really  useful  when  you
227              have  multiple  slurmd daemons running on the same host machine.
228              Multiple slurmd daemons on one host are, in general,  only  used
229              by Slurm developers.
230
231
232       ping   Ping  the  primary  and secondary slurmctld daemon and report if
233              they are responding.
234
235
236       reboot     [ASAP]      [nextstate=<RESUME|DOWN>]      [reason=<reason>]
237       <ALL|NodeList>
238              Reboot  the  nodes in the system when they become idle using the
239              RebootProgram as configured in Slurm's  slurm.conf  file.   Each
240              node will have the "REBOOT" flag added to its node state.  After
241              a node reboots and  the  slurmd  daemon  starts  up  again,  the
242              HealthCheckProgram  will  run once. Then, the slurmd daemon will
243              register itself with the slurmctld daemon and the "REBOOT"  flag
244              will  be  cleared.   If  the node reason is "Reboot ASAP", Slurm
245              will clear the node's "DRAIN" state flag as  well.   The  "ASAP"
246              option  adds  the  "DRAIN" flag to each node's state, preventing
247              additional jobs from running on the node so it can  be  rebooted
248              and  returned  to  service  "As  Soon  As Possible" (i.e. ASAP).
249              "ASAP" will also set the node reason to  "Reboot  ASAP"  if  the
250              "reason"  option  isn't specified.  If the "nextstate" option is
251              specified as "DOWN", then the node will remain in a  down  state
252              after  rebooting.  If "nextstate" is specified as "RESUME", then
253              the nodes will resume as normal when the node registers and  the
254              node  reason will be cleared.  Resuming nodes will be considered
255              as available in backfill future scheduling and won't be replaced
256              by  idle nodes in a reservation.  When using the "nextstate" and
257              "reason" options together  the  reason  will  be  appended  with
258              "reboot  issued" when the reboot is issued and "reboot complete"
259              when the node registers with a "nextstate" of "DOWN".  The "rea‐
260              son"  option  sets each node's reason to a user-defined message.
261              You must specify either a list of nodes or that ALL nodes are to
262              be  rebooted.   NOTE:  By default, this command does not prevent
263              additional jobs from being scheduled on any nodes before reboot.
264              To  do  this, you can either use the "ASAP" option or explicitly
265              drain the nodes  beforehand.   You  can  alternately  create  an
266              advanced  reservation to prevent additional jobs from being ini‐
267              tiated on nodes to be rebooted.  Pending  reboots  can  be  can‐
268              celled  by  using "scontrol cancel_reboot <node>" or setting the
269              node state to "CANCEL_REBOOT".  A node will be marked "DOWN"  if
270              it doesn't reboot within ResumeTimeout.
271
272
273       reconfigure
274              Instruct  all  Slurm  daemons to re-read the configuration file.
275              This command does not restart the daemons.  This mechanism would
276              be  used  to  modify  configuration  parameters (Epilog, Prolog,
277              SlurmctldLogFile, SlurmdLogFile, etc.).   The  Slurm  controller
278              (slurmctld)  forwards the request all other daemons (slurmd dae‐
279              mon on each compute  node).  Running  jobs  continue  execution.
280              Most  configuration  parameters  can  be changed by just running
281              this command, however, Slurm  daemons  should  be  shutdown  and
282              restarted  if  any  of  these  parameters  are  to  be  changed:
283              AuthType, BackupAddr,  BackupController,  ControlAddr,  Control‐
284              Mach, PluginDir, StateSaveLocation, SlurmctldPort or SlurmdPort.
285              The slurmctld daemon and all slurmd daemons must be restarted if
286              nodes are added to or removed from the cluster.
287
288
289       release job_list
290              Release  a previously held job to begin execution.  The job_list
291              argument is a comma separated list of job IDs OR "jobname=" with
292              the  job's name, which will attempt to hold all jobs having that
293              name.  Also see hold.
294
295
296       requeue [option] job_list
297              Requeue a running, suspended or finished Slurm  batch  job  into
298              pending  state.  The job_list argument is a comma separated list
299              of job IDs.  The command accepts the following option:
300
301              Incomplete
302                     Operate only on jobs (or tasks of a job array) which have
303                     not  completed.   Specifically only jobs in the following
304                     states will be requeued: CONFIGURING, RUNNING, STOPPED or
305                     SUSPENDED.
306
307
308       requeuehold [option] job_list
309              Requeue  a  running,  suspended or finished Slurm batch job into
310              pending state, moreover the job is put in held  state  (priority
311              zero).   The  job_list argument is a comma separated list of job
312              IDs.  A held job can be released using  scontrol  to  reset  its
313              priority   (e.g.   "scontrol  release  <job_id>").  The  command
314              accepts the following options:
315
316              Incomplete
317                     Operate only on jobs (or tasks of a job array) which have
318                     not  completed.   Specifically only jobs in the following
319                     states will be requeued: CONFIGURING, RUNNING, STOPPED or
320                     SUSPENDED.
321
322              State=SpecialExit
323                     The  "SpecialExit"  keyword specifies that the job has to
324                     be put in a special state JOB_SPECIAL_EXIT.   The  "scon‐
325                     trol  show job" command will display the JobState as SPE‐
326                     CIAL_EXIT, while the "squeue" command as SE.
327
328
329       resume job_list
330              Resume a previously suspended job.  The job_list argument  is  a
331              comma separated list of job IDs.  Also see suspend.
332
333              NOTE:  A suspended job releases its CPUs for allocation to other
334              jobs.  Resuming a previously suspended job may result in  multi‐
335              ple jobs being allocated the same CPUs, which could trigger gang
336              scheduling with some configurations  or  severe  degradation  in
337              performance  with other configurations.  Use of the scancel com‐
338              mand to send SIGSTOP and SIGCONT signals would stop a job  with‐
339              out releasing its CPUs for allocation to other jobs and would be
340              a preferable mechanism in  many  cases.   If  performing  system
341              maintenance  you may want to use suspend/resume in the following
342              way. Before suspending set all nodes to draining or set all par‐
343              titions  to down so that no new jobs can be scheduled. Then sus‐
344              pend jobs. Once maintenance is  done  resume  jobs  then  resume
345              nodes and/or set all partitions back to up.  Use with caution.
346
347
348       schedloglevel LEVEL
349              Enable  or  disable  scheduler  logging.  LEVEL may be "0", "1",
350              "disable" or "enable". "0" has the same effect as "disable". "1"
351              has  the  same  effect as "enable".  This value is temporary and
352              will  be  overwritten  when  the  slurmctld  daemon  reads   the
353              slurm.conf configuration file (e.g. when the daemon is restarted
354              or scontrol reconfigure is executed) if  the  SlurmSchedLogLevel
355              parameter is present.
356
357
358       setdebug LEVEL
359              Change the debug level of the slurmctld daemon.  LEVEL may be an
360              integer value between zero and nine (using the  same  values  as
361              SlurmctldDebug  in  the slurm.conf file) or the name of the most
362              detailed message type to be printed: "quiet", "fatal",  "error",
363              "info",  "verbose",  "debug",  "debug2",  "debug3", "debug4", or
364              "debug5".  This value is temporary and will be overwritten when‐
365              ever  the  slurmctld  daemon  reads the slurm.conf configuration
366              file (e.g. when the daemon is restarted or scontrol  reconfigure
367              is executed).
368
369
370       setdebugflags [+|-]FLAG
371              Add  or  remove  DebugFlags  of  the slurmctld daemon.  See "man
372              slurm.conf" for a list of supported DebugFlags.  NOTE:  Changing
373              the  value  of  some  DebugFlags  will  have  no  effect without
374              restarting the slurmctld  daemon,  which  would  set  DebugFlags
375              based upon the contents of the slurm.conf configuration file.
376
377
378       show ENTITY ID
379              or
380
381       show ENTITY=ID
382              Display  the  state  of  the specified entity with the specified
383              identification.   ENTITY  may  be  aliases,  assoc_mgr,  bbstat,
384              burstbuffer, config, daemons, dwstat, federation, frontend, job,
385              node, partition, powercap, reservation, slurmd, step,  topology,
386              hostlist, hostlistsorted or hostnames ID can be used to identify
387              a specific element of the identified entity: job ID, node  name,
388              partition  name, reservation name, or job step ID for job, node,
389              partition, or step respectively.  For an  ENTITY  of  bbstat  or
390              dwstat  (they are equivalent) optional arguments are the options
391              of the local status command.  The status commands will  be  exe‐
392              cuted  by  the slurmctld daemon and its response returned to the
393              user.  For an ENTITY of topology, the ID may be a node or switch
394              name.   If one node name is specified, all switches connected to
395              that node (and their parent switches) will be  shown.   If  more
396              than  one  node name is specified, only switches that connect to
397              all named nodes will be shown.  aliases will return all NodeName
398              values  associated  to  a  given NodeHostname (useful to get the
399              list of virtual nodes associated with a real node in a  configu‐
400              ration where multiple slurmd daemons execute on a single compute
401              node).  assoc_mgr displays the current contents  of  the  slurm‐
402              ctld's internal cache for users, associations and/or qos. The ID
403              may               be                users=<user1>,[...,<userN>],
404              accounts=<acct1>,[...,<acctN>],  qos=<qos1>,[...,<qosN>]  and/or
405              flags=<users,assoc,qos>, used to filter the desired  section  to
406              be  displayed.  If no flags are specified, all sections are dis‐
407              played.  burstbuffer displays the current status of  the  Burst‐
408              Buffer plugin.  config displays parameter names from the config‐
409              uration files in mixed case (e.g. SlurmdPort=7003) while derived
410              parameters  names  are  in upper case only (e.g. SLURM_VERSION).
411              hostnames takes an optional hostlist  expression  as  input  and
412              writes  a  list of individual host names to standard output (one
413              per line). If no hostlist expression is supplied,  the  contents
414              of  the  SLURM_JOB_NODELIST  environment  variable  is used. For
415              example "tux[1-3]" is mapped to "tux1","tux2"  and  "tux3"  (one
416              hostname  per  line).   hostlist  takes a list of host names and
417              prints the hostlist expression for them (the  inverse  of  host‐
418              names).   hostlist can also take the absolute pathname of a file
419              (beginning with the character '/') containing a  list  of  host‐
420              names.   Multiple  node names may be specified using simple node
421              range expressions (e.g. "lx[10-20]"). All other ID  values  must
422              identify  a  single  element.  The  job  step  ID is of the form
423              "job_id.step_id", (e.g. "1234.1").  slurmd reports  the  current
424              status  of  the  slurmd  daemon  executing on the same node from
425              which the scontrol command is executed (the local host). It  can
426              be  useful  to  diagnose problems.  By default hostlist does not
427              sort the node list or make  it  unique  (e.g.  tux2,tux1,tux2  =
428              tux[2,1-2]).   If  you  wanted  a sorted list use hostlistsorted
429              (e.g. tux2,tux1,tux2 = tux[1-2,2]).  By default, all elements of
430              the entity type specified are printed.  For an ENTITY of job, if
431              the job does not specify  socket-per-node,  cores-per-socket  or
432              threads-per-core  then  it  will  display  '*' in ReqS:C:T=*:*:*
433              field. For an ENTITY of federation, the federation name that the
434              controller  is part of and the sibling clusters part of the fed‐
435              eration will be listed.
436
437
438       shutdown OPTION
439              Instruct Slurm daemons to save current state and terminate.   By
440              default,  the  Slurm controller (slurmctld) forwards the request
441              all other daemons (slurmd daemon  on  each  compute  node).   An
442              OPTION  of slurmctld or controller results in only the slurmctld
443              daemon being shutdown and the slurmd daemons remaining active.
444
445
446       suspend job_list
447              Suspend a running job.  The job_list argument is a  comma  sepa‐
448              rated  list  of  job  IDs.  Use the resume command to resume its
449              execution.  User processes must stop on receipt of SIGSTOP  sig‐
450              nal  and resume upon receipt of SIGCONT for this operation to be
451              effective.  Not all architectures and configurations support job
452              suspension.   If  a suspended job is requeued, it will be placed
453              in a held state.  The time a job is  suspended  will  not  count
454              against  a  job's  time limit.  Only an operator, administrator,
455              SlurmUser, or root can suspend jobs.
456
457
458       takeover
459              Instruct Slurm's backup controller (slurmctld) to take over sys‐
460              tem  control.   Slurm's  backup controller requests control from
461              the primary and  waits  for  its  termination.  After  that,  it
462              switches  from  backup  mode to controller mode. If primary con‐
463              troller can not be contacted, it directly switches to controller
464              mode.  This  can  be  used  to  speed  up  the  Slurm controller
465              fail-over mechanism when the primary node is down.  This can  be
466              used  to  minimize disruption if the computer executing the pri‐
467              mary Slurm controller is scheduled down.  (Note: Slurm's primary
468              controller will take the control back at startup.)
469
470
471       top job_list
472              Move  the  specified  job  IDs  to  the top of the queue of jobs
473              belonging to the identical user ID, partition name, account, and
474              QOS.  The job_list argument is a comma separated ordered list of
475              job IDs.  Any job not matching all of those fields will  not  be
476              effected.   Only  jobs  submitted  to a single partition will be
477              effected.  This operation changes the order of jobs by adjusting
478              job  nice values.  The net effect on that user's throughput will
479              be negligible to slightly negative.  This operation is  disabled
480              by  default  for non-privileged (non-operator, admin, SlurmUser,
481              or root) users. This operation may be enabled for non-privileged
482              users  by  the  system  administrator  by  including  the option
483              "enable_user_top"  in  the   SchedulerParameters   configuration
484              parameter.
485
486
487       uhold job_list
488              Prevent  a pending job from being started (sets it's priority to
489              0).  The job_list argument is a space separated list of job  IDs
490              or  job  names.  Use the release command to permit the job to be
491              scheduled.  This command is designed for a system  administrator
492              to  hold  a job so that the job owner may release it rather than
493              requiring the intervention of a system administrator  (also  see
494              the hold command).
495
496
497       update SPECIFICATION
498              Update  job,  step, node, partition, powercapping or reservation
499              configuration per the supplied specification.  SPECIFICATION  is
500              in  the same format as the Slurm configuration file and the out‐
501              put of the show command described above. It may be desirable  to
502              execute  the  show  command  (described  above)  on the specific
503              entity you want to update, then use cut-and-paste tools to enter
504              updated configuration values to the update. Note that while most
505              configuration values can be changed using this command, not  all
506              can be changed using this mechanism. In particular, the hardware
507              configuration of a node or the physical addition or  removal  of
508              nodes  from the cluster may only be accomplished through editing
509              the Slurm configuration file and executing the reconfigure  com‐
510              mand (described above).
511
512
513       version
514              Display the version number of scontrol being executed.
515
516
517       wait_job job_id
518              Wait  until  a job and all of its nodes are ready for use or the
519              job has entered some termination state. This option is  particu‐
520              larly  useful  in the Slurm Prolog or in the batch script itself
521              if nodes are powered down and restarted automatically as needed.
522
523              NOTE: Don't use scontrol wait_job in PrologCtld or  Prolog  with
524              PrologFlags=Alloc as this will result in deadlock.
525
526
527       write batch_script job_id optional_filename
528              Write  the  batch script for a given job_id to a file or to std‐
529              out. The file will default to slurm-<job_id>.sh if the  optional
530              filename  argument  is  not given. The script will be written to
531              stdout if - is given instead of a filename.   The  batch  script
532              can  only  be retrieved by an admin or operator, or by the owner
533              of the job.
534
535
536       write config
537              Write the current configuration to a file with the  naming  con‐
538              vention  of "slurm.conf.<datetime>" in the same directory as the
539              original slurm.conf file.
540
541
542       INTERACTIVE COMMANDS
543              NOTE: All commands listed below can be used in  the  interactive
544              mode, but NOT on the initial command line.
545
546
547       all    Show  all  partitions,  their  jobs  and jobs steps. This causes
548              information to be displayed about partitions that are configured
549              as hidden and partitions that are unavailable to user's group.
550
551
552       details
553              Causes  the  show  command  to  provide additional details where
554              available.  Job information will include CPUs  and  NUMA  memory
555              allocated  on  each  node.   Note  that on computers with hyper‐
556              threading enabled and Slurm configured to allocate  cores,  each
557              listed  CPU  represents  one physical core.  Each hyperthread on
558              that core can be allocated a separate task, so a job's CPU count
559              and  task  count  may differ.  See the --cpu-bind and --mem-bind
560              option descriptions in srun man pages for more information.  The
561              details option is currently only supported for the show job com‐
562              mand.
563
564
565       exit   Terminate scontrol interactive session.
566
567
568       hide   Do not display partition, job or jobs step information for  par‐
569              titions  that  are  configured  as hidden or partitions that are
570              unavailable to the user's group.  This is the default behavior.
571
572
573       oneliner
574              Print information one line per record.
575
576
577       quiet  Print no warning or informational  messages,  only  fatal  error
578              messages.
579
580
581       quit   Terminate the execution of scontrol.
582
583
584       verbose
585              Print detailed event logging.  This includes time-stamps on data
586              structures, record counts, etc.
587
588
589       !!     Repeat the last command executed.
590
591
592       SPECIFICATIONS FOR UPDATE COMMAND, JOBS
593
594       Note that update requests done by either root, SlurmUser or Administra‐
595       tors  are  not  subject  to  certain  restrictions. For instance, if an
596       Administrator changes the QOS on a pending job, certain limits such  as
597       the  TimeLimit will not be changed automatically as changes made by the
598       Administrators are allowed to violate these restrictions.
599
600
601       Account=<account>
602              Account name to be changed for this job's resource  use.   Value
603              may be cleared with blank data value, "Account=".
604
605       AdminComment=<spec>
606              Arbitrary  descriptive string. Can only be set by a Slurm admin‐
607              istrator.
608
609       ArrayTaskThrottle=<count>
610              Specify the maximum number of tasks in a job array that can exe‐
611              cute at the same time.  Set the count to zero in order to elimi‐
612              nate any limit.  The task throttle count  for  a  job  array  is
613              reported  as part of its ArrayTaskId field, preceded with a per‐
614              cent sign.  For example "ArrayTaskId=1-10%2" indicates the maxi‐
615              mum number of running tasks is limited to 2.
616
617       BurstBuffer=<spec>
618              Burst buffer specification to be changed for this job's resource
619              use.  Value may  be  cleared  with  blank  data  value,  "Burst‐
620              Buffer=".  Format is burst buffer plugin specific.
621
622       Clusters=<spec>
623              Specifies the clusters that the federated job can run on.
624
625       ClusterFeatures=<spec>
626              Specifies  features that a federated cluster must have to have a
627              sibling job submitted to it. Slurm will attempt to submit a sib‐
628              ling  job  to  a cluster if it has at least one of the specified
629              features.
630
631       Comment=<spec>
632              Arbitrary descriptive string.
633
634       Contiguous=<yes|no>
635              Set the job's requirement for contiguous (consecutive) nodes  to
636              be  allocated.   Possible  values  are "YES" and "NO".  Only the
637              Slurm administrator or root can change this parameter.
638
639       CoreSpec=<count>
640              Number of cores to reserve per node for  system  use.   The  job
641              will  be  charged  for  these  cores, but be unable to use them.
642              Will be reported as "*" if not constrained.
643
644       CPUsPerTask=<count>
645              Change the CPUsPerTask job's value.
646
647       Deadline=<time_spec>
648              It accepts times of the form HH:MM:SS to specify a deadline to a
649              job  at  a specific time of day (seconds are optional).  You may
650              also specify midnight, noon, fika (3 PM) or teatime (4  PM)  and
651              you can have a time-of-day suffixed with AM or PM for a deadline
652              in the morning or the evening.  You can specify a  deadline  for
653              the  job with a date of the form MMDDYY or MM/DD/YY or MM.DD.YY,
654              or a date and time as  YYYY-MM-DD[THH:MM[:SS]].   You  can  also
655              give times like now + count time-units, where the time-units can
656              be minutes, hours, days, or weeks and you can tell Slurm to  put
657              a  deadline  for tomorrow with the keyword tomorrow.  The speci‐
658              fied deadline must be later than the current time.  Only pending
659              jobs  can have the deadline updated.  Only the Slurm administra‐
660              tor or root can change this parameter.
661
662       DelayBoot=<time_spec>
663              Change the time to decide whether to reboot nodes  in  order  to
664              satisfy job's feature specification if the job has been eligible
665              to run for less than this time  period.  See  salloc/sbatch  man
666              pages option --delay-boot.
667
668       Dependency=<dependency_list>
669              Defer job's initiation until specified job dependency specifica‐
670              tion is satisfied.   Cancel  dependency  with  an  empty  depen‐
671              dency_list  (e.g.  "Dependency=").   <dependency_list> is of the
672              form <type:job_id[:job_id][,type:job_id[:job_id]]>.   Many  jobs
673              can  share the same dependency and these jobs may even belong to
674              different  users.
675
676              after:job_id[:jobid...]
677                     This job can begin execution  after  the  specified  jobs
678                     have begun execution.
679
680              afterany:job_id[:jobid...]
681                     This  job  can  begin  execution after the specified jobs
682                     have terminated.
683
684              afternotok:job_id[:jobid...]
685                     This job can begin execution  after  the  specified  jobs
686                     have terminated in some failed state (non-zero exit code,
687                     node failure, timed out, etc).
688
689              afterok:job_id[:jobid...]
690                     This job can begin execution  after  the  specified  jobs
691                     have  successfully  executed  (ran  to completion with an
692                     exit code of zero).
693
694              singleton
695                     This  job  can  begin  execution  after  any   previously
696                     launched  jobs  sharing  the  same job name and user have
697                     terminated.  In other words, only one job  by  that  name
698                     and owned by that user can be running or suspended at any
699                     point in time.
700
701       EligibleTime=<time_spec>
702              See StartTime.
703
704       EndTime
705              The time the job is expected to terminate  based  on  the  job's
706              time  limit.   When  the  job  ends  sooner,  this field will be
707              updated with the actual end time.
708
709       ExcNodeList=<nodes>
710              Set the job's list of excluded node. Multiple node names may  be
711              specified    using   simple   node   range   expressions   (e.g.
712              "lx[10-20]").  Value may be cleared with blank data value, "Exc‐
713              NodeList=".
714
715       Features=<features>
716              Set  the job's required node features.  The list of features may
717              include multiple feature  names  separated  by  ampersand  (AND)
718              and/or   vertical   bar   (OR)  operators.   For  example:  Fea‐
719              tures="opteron&video" or Features="fast|faster".  In  the  first
720              example,  only  nodes  having both the feature "opteron" AND the
721              feature "video" will be used.  There is no mechanism to  specify
722              that  you  want one node with feature "opteron" and another node
723              with feature "video" in case no node has both features.  If only
724              one  of  a  set of possible options should be used for all allo‐
725              cated nodes, then use the OR operator and  enclose  the  options
726              within      square     brackets.      For     example:     "Fea‐
727              tures=[rack1|rack2|rack3|rack4]" might be used to  specify  that
728              all nodes must be allocated on a single rack of the cluster, but
729              any of those four racks can be used.  A request can also specify
730              the  number  of  nodes  needed with some feature by appending an
731              asterisk and count after the feature name.   For  example  "Fea‐
732              tures=graphics*4"  indicates  that at least four allocated nodes
733              must have the feature "graphics."   Parenthesis  are  also  sup‐
734              ported  for  features  to  be ANDed together.  For example "Fea‐
735              tures=[(knl&a2a&flat)*4&haswell*2]" indicates the resource allo‐
736              cation  should  include  4 nodes with ALL of the features "knl",
737              "a2a", and "flat" plus 2 nodes with the feature "haswell".  Con‐
738              straints  with  node counts may only be combined with AND opera‐
739              tors.  Value may be cleared with blank data value,  for  example
740              "Features=".
741
742
743       Gres=<list>
744              Specifies   a   comma   delimited  list  of  generic  consumable
745              resources.   The  format  of  each  entry   on   the   list   is
746              "name[:count[*cpu]]".   The  name  is  that  of  the  consumable
747              resource.  The count is the number of  those  resources  with  a
748              default  value  of 1.  The specified resources will be allocated
749              to the job on each node allocated unless "*cpu" is appended,  in
750              which  case  the resources will be allocated on a per cpu basis.
751              The available generic consumable resources  is  configurable  by
752              the  system  administrator.  A list of available generic consum‐
753              able resources will be printed and the command will exit if  the
754              option   argument   is   "help".    Examples   of   use  include
755              "Gres=gpus:2*cpu,disk=40G" and "Gres=help".
756
757
758       JobId=<job_list>
759              Identify the job(s) to be updated.  The job_list may be a  comma
760              separated list of job IDs.  Either JobId or JobName is required.
761
762       Licenses=<name>
763              Specification  of  licenses (or other resources available on all
764              nodes of the cluster) as  described  in  salloc/sbatch/srun  man
765              pages.
766
767       MinCPUsNode=<count>
768              Set  the  job's minimum number of CPUs per node to the specified
769              value.
770
771       MinMemoryCPU=<megabytes>
772              Set the job's minimum real memory required per allocated CPU  to
773              the specified value. Either MinMemoryCPU or MinMemoryNode may be
774              set, but not both.
775
776       MinMemoryNode=<megabytes>
777              Set the job's minimum real memory required per node to the spec‐
778              ified  value.   Either MinMemoryCPU or MinMemoryNode may be set,
779              but not both.
780
781       MinTmpDiskNode=<megabytes>
782              Set the job's minimum temporary disk space required per node  to
783              the  specified  value.  Only the Slurm administrator or root can
784              change this parameter.
785
786       imeMin=<timespec>
787              Change TimeMin value which specifies the minimum time limit min‐
788              utes of the job.
789
790       JobName=<name>
791              Identify  the  name of jobs to be modified or set the job's name
792              to the specified value.  When used to identify jobs to be  modi‐
793              fied,  all  jobs  belonging to all users are modified unless the
794              UserID option is used to identify a specific user.  Either JobId
795              or JobName is required.
796
797       Name[=<name>]
798              See JobName.
799
800       Nice[=<adjustment>]
801              Update  the  job  with  an  adjusted  scheduling priority within
802              Slurm. With no  adjustment  value  the  scheduling  priority  is
803              decreased  by 100. A negative nice value increases the priority,
804              otherwise decreases it. The adjustment range is +/-  2147483645.
805              Only privileged users can specify a negative adjustment.
806
807       NodeList=<nodes>
808              Change the nodes allocated to a running job to shrink it's size.
809              The specified list of nodes must be a subset of the  nodes  cur‐
810              rently  allocated  to the job. Multiple node names may be speci‐
811              fied using simple node  range  expressions  (e.g.  "lx[10-20]").
812              After  a  job's  allocation is reduced, subsequent srun commands
813              must explicitly specify node and task counts which are valid for
814              the new allocation.
815
816       NumCPUs=<min_count>[-<max_count>]
817              Set the job's minimum and optionally maximum count of CPUs to be
818              allocated.
819
820       NumNodes=<min_count>[-<max_count>]
821              Set the job's minimum and optionally maximum count of  nodes  to
822              be  allocated.  If the job is already running, use this to spec‐
823              ify a node count less than  currently  allocated  and  resources
824              previously  allocated  to  the job will be relinquished. After a
825              job's allocation  is  reduced,  subsequent  srun  commands  must
826              explicitly  specify node and task counts which are valid for the
827              new allocation. Also see the NodeList parameter above.  This  is
828              the same than ReqNodes.
829
830       NumTasks=<count>
831              Set  the  job's  count of required tasks to the specified value.
832              This is the same than ReqProcs.
833
834       OverSubscribe=<yes|no>
835              Set the job's ability to share compute resources (i.e.  individ‐
836              ual  CPUs)  with other jobs. Possible values are "YES" and "NO".
837              This option can only be changed for pending jobs.
838
839       Partition=<name>
840              Set the job's partition to the specified value.
841
842       Priority=<number>
843              Set the job's priority to the specified value.  Note that a  job
844              priority of zero prevents the job from ever being scheduled.  By
845              setting a job's priority to zero it is held.  Set  the  priority
846              to  a  non-zero value to permit it to run.  Explicitly setting a
847              job's priority clears any previously set nice value and  removes
848              the priority/multifactor plugin's ability to manage a job's pri‐
849              ority.  In order to restore  the  priority/multifactor  plugin's
850              ability  to  manage  a job's priority, hold and then release the
851              job.  Only the Slurm administrator or root  can  increase  job's
852              priority.
853
854       QOS=<name>
855              Set  the  job's QOS (Quality Of Service) to the specified value.
856              Value may be cleared with blank data value, "QOS=".
857
858       Reboot=<yes|no>
859              Set the job's flag that specifies whether to force the allocated
860              nodes  to reboot before starting the job. This is only supported
861              with some  system  configurations  and  therefore  it  could  be
862              silently ignored.
863
864       ReqCores=<count>
865              Change the job's requested Cores count.
866
867       ReqNodeList=<nodes>
868              Set  the job's list of required node. Multiple node names may be
869              specified   using   simple   node   range   expressions    (e.g.
870              "lx[10-20]").   Value  may  be  cleared  with  blank data value,
871              "ReqNodeList=".
872
873       ReqNodes=<min_count>[-<max_count>]
874              See NumNodes.
875
876       ReqProcs=<count>
877              See NumTasks.
878
879       ReqSockets=<count>
880              Change the job's requested socket count.
881
882       ReqThreads=<count>
883              Change the job's requested threads count.
884
885       Requeue=<0|1>
886              Stipulates whether a job should be requeued after a  node  fail‐
887              ure: 0 for no, 1 for yes.
888
889       ReservationName=<name>
890              Set  the job's reservation to the specified value.  Value may be
891              cleared with blank data value, "ReservationName=".
892
893       ResetAccrueTime
894              Reset the job's accrue time value to 0 meaning it will loose any
895              time  previously  accrued  for  priority.  Helpful if you have a
896              large queue of jobs already in the queue and want to start  lim‐
897              iting  how  many  jobs  can  accrue time without waiting for the
898              queue to flush out.
899
900       SiteFactor=<account>
901              Specify  the  job's  admin  priority  factor  in  the  range  of
902              +/-2147483645.  Only privileged users can modify the value.
903
904       StdOut=<filepath>
905              Set the batch job's stdout file path.
906
907       Shared=<yes|no>
908              See OverSubscribe option above.
909
910       StartTime=<time_spec>
911              Set the job's earliest initiation time.  It accepts times of the
912              form HH:MM:SS to run a job at a specific time  of  day  (seconds
913              are  optional).   (If that time is already past, the next day is
914              assumed.)  You may also specify midnight, noon, fika (3  PM)  or
915              teatime  (4  PM) and you can have a time-of-day suffixed with AM
916              or PM for running in the morning or the evening.  You  can  also
917              say  what  day  the job will be run, by specifying a date of the
918              form MMDDYY or MM/DD/YY or MM.DD.YY,  or  a  date  and  time  as
919              YYYY-MM-DD[THH:MM[:SS]].   You  can  also  give times like now +
920              count time-units, where the time-units can  be  minutes,  hours,
921              days,  or weeks and you can tell Slurm to run the job today with
922              the keyword today and to run the job tomorrow with  the  keyword
923              tomorrow.
924
925              Notes on date/time specifications:
926               -  although the 'seconds' field of the HH:MM:SS time specifica‐
927              tion is allowed by the code, note that  the  poll  time  of  the
928              Slurm  scheduler  is not precise enough to guarantee dispatch of
929              the job on the exact second.  The job will be eligible to  start
930              on  the  next  poll following the specified time. The exact poll
931              interval depends on the Slurm scheduler (e.g., 60  seconds  with
932              the default sched/builtin).
933               -   if   no  time  (HH:MM:SS)  is  specified,  the  default  is
934              (00:00:00).
935               - if a date is specified without a year (e.g., MM/DD) then  the
936              current  year  is  assumed,  unless the combination of MM/DD and
937              HH:MM:SS has already passed for that year,  in  which  case  the
938              next year is used.
939
940       Switches=<count>[@<max-time-to-wait>]
941              When  a tree topology is used, this defines the maximum count of
942              switches desired for the job allocation. If Slurm finds an allo‐
943              cation  containing  more  switches than the count specified, the
944              job remain pending until it  either  finds  an  allocation  with
945              desired switch count or the time limit expires. By default there
946              is no switch count limit and no time limit delay. Set the  count
947              to  zero  in  order to clean any previously set count (disabling
948              the limit).  The job's maximum time delay may be limited by  the
949              system administrator using the SchedulerParameters configuration
950              parameter with the max_switch_wait parameter option.   Also  see
951              wait-for-switch.
952
953
954       wait-for-switch=<seconds>
955              Change max time to wait for a switch <seconds> secs.
956
957
958       TasksPerNode=<count>
959              Change the job's requested TasksPerNode.
960
961
962       ThreadSpec=<count>
963              Number  of  threads to reserve per node for system use.  The job
964              will be charged for these threads, but be unable  to  use  them.
965              Will be reported as "*" if not constrained.
966
967
968       TimeLimit=<time>
969              The  job's  time  limit.   Output  format  is  [days-]hours:min‐
970              utes:seconds or "UNLIMITED".  Input format (for update  command)
971              set    is   minutes,   minutes:seconds,   hours:minutes:seconds,
972              days-hours,  days-hours:minutes  or  days-hours:minutes:seconds.
973              Time  resolution  is one minute and second values are rounded up
974              to the next minute.  If changing the time limit of a job, either
975              specify  a  new  time  limit value or precede the time and equal
976              sign with a "+" or "-" to increment  or  decrement  the  current
977              time  limit  (e.g.  "TimeLimit+=30").  In  order to increment or
978              decrement the current time limit, the JobId  specification  must
979              precede  the TimeLimit specification.  Note that incrementing or
980              decrementing the time limit for a  job  array  is  only  allowed
981              before  the  job  array  has  been  split into more than one job
982              record.  Only the Slurm administrator or root can increase job's
983              TimeLimit.
984
985
986       UserID=<UID or name>
987              Used  with  the  JobName option to identify jobs to be modified.
988              Either a user name or numeric ID (UID), may be specified.
989
990
991       WCKey=<key>
992              Set the job's workload characterization  key  to  the  specified
993              value.
994
995
996       WorkDir=<directory_name>
997              Set  the  job's  working  directory to the specified value. Note
998              that this may only be set for jobs in  the  PENDING  state,  and
999              that  jobs  may fail to launch if they rely on relative paths to
1000              the originally submitted WorkDir.
1001
1002
1003       SPECIFICATIONS FOR SHOW COMMAND, JOBS
1004
1005       The "show" command, when used with the "job" or  "job  <jobid>"  entity
1006       displays detailed information about a job or jobs.  Much of this infor‐
1007       mation may be modified using the  "update  job"  command  as  described
1008       above.  However, the following fields displayed by the show job command
1009       are read-only and cannot be modified:
1010
1011
1012       AllocNode:Sid
1013              Local node and system id making the resource allocation.
1014
1015       BatchFlag
1016              Jobs submitted using the sbatch command have BatchFlag set to 1.
1017              Jobs submitted using other commands have BatchFlag set to 0.
1018
1019       ExitCode=<exit>:<sig>
1020              Exit  status  reported  for the job by the wait() function.  The
1021              first number is the exit code, typically as set  by  the  exit()
1022              function.   The  second  number  of  the  signal that caused the
1023              process to terminate if it was terminated by a signal.
1024
1025       GroupId
1026              The group under which the job was submitted.
1027
1028       JobState
1029              The current state of the job.
1030
1031       NodeListIndices
1032              The NodeIndices expose the internal indices into the node  table
1033              associated with the node(s) allocated to the job.
1034
1035       NtasksPerN:B:S:C=
1036              <tasks_per_node>:<tasks_per_base‐
1037              board>:<tasks_per_socket>:<tasks_per_core> Specifies the number
1038              of tasks to be started per hardware component (node, baseboard,
1039              socket and core).  Unconstrained values may be shown as "0" or
1040              "*".
1041
1042       PreemptEligibleTime
1043              Time  the  job becomes eligible for preemption. Modified by Pre‐
1044              emptExemptTime, either from the global option in  slurm.conf  or
1045              the  job  QOS.  This  is hidden if the job has not started or if
1046              PreemptMode=OFF.
1047
1048       PreemptTime
1049              Time at which job was signaled that it was selected for  preemp‐
1050              tion.  (Meaningful only for PreemptMode=CANCEL and the partition
1051              or QOS with which the job is associated has  a  GraceTime  value
1052              designated.)  This  is  hidden  if the job has not started or if
1053              PreemptMode=OFF.
1054
1055       PreSusTime
1056              Time the job ran prior to last suspend.
1057
1058       Reason The reason job is not running: e.g., waiting "Resources".
1059
1060       ReqB:S:C:T=
1061              <baseboard_count>:<socket_per_base‐
1062              board_count>:<core_per_socket_count>:<thread_per_core_count>
1063              Specifies the count of various hardware components requested by
1064              the job.  Unconstrained values may be shown as "0" or "*".
1065
1066       SecsPreSuspend=<seconds>
1067              If the job is suspended, this is the run time accumulated by the
1068              job (in seconds) prior to being suspended.
1069
1070       Socks/Node=<count>
1071              Count of desired sockets per node
1072
1073       SubmitTime
1074              The time  and  date stamp (in localtime) the job was  submitted.
1075              The  format  of  the  output is identical to that of the EndTime
1076              field.
1077
1078              NOTE: If a job is requeued, the submit time is reset.  To obtain
1079              the  original  submit  time it is necessary to use the "sacct -j
1080              <job_id[.<step_id>]" command also designating the -D or --dupli‐
1081              cate option to display all duplicate entries for a job.
1082
1083       SuspendTime
1084              Time the job was last suspended or resumed.
1085
1086       NOTE on information displayed for various job states:
1087              When  you submit a request for the "show job" function the scon‐
1088              trol process makes an RPC  request  call  to  slurmctld  with  a
1089              REQUEST_JOB_INFO message type.  If the state of the job is PEND‐
1090              ING, then it returns some detail information such as: min_nodes,
1091              min_procs,  cpus_per_task, etc. If the state is other than PEND‐
1092              ING the code assumes that it is in a further state such as  RUN‐
1093              NING,  COMPLETE, etc. In these cases the code explicitly returns
1094              zero for these values. These values are meaningless once the job
1095              resources have been allocated and the job has started.
1096
1097
1098       SPECIFICATIONS FOR UPDATE COMMAND, STEPS
1099
1100       StepId=<job_id>[.<step_id>]
1101              Identify the step to be updated.  If the job_id is given, but no
1102              step_id is specified then all steps of the identified  job  will
1103              be modified.  This specification is required.
1104
1105       CompFile=<completion file>
1106              Update a step with information about a steps completion.  Can be
1107              useful if step statistics aren't directly  available  through  a
1108              jobacct_gather  plugin.  The file is a space-delimited file with
1109              format for Version 1 is as follows
1110
1111              1 34461 0 2 0 3 1361906011 1361906015 1 1 3368 13357 /bin/sleep
1112              A B     C D E F G          H          I J K    L     M
1113
1114              Field Descriptions:
1115
1116              A file version
1117              B ALPS apid
1118              C inblocks
1119              D outblocks
1120              E exit status
1121              F number of allocated CPUs
1122              G start time
1123              H end time
1124              I utime
1125              J stime
1126              K maxrss
1127              L uid
1128              M command name
1129
1130       TimeLimit=<time>
1131              The  job's  time  limit.   Output  format  is  [days-]hours:min‐
1132              utes:seconds  or "UNLIMITED".  Input format (for update command)
1133              set   is   minutes,   minutes:seconds,    hours:minutes:seconds,
1134              days-hours,  days-hours:minutes  or  days-hours:minutes:seconds.
1135              Time resolution is one minute and second values are  rounded  up
1136              to  the  next  minute.   If  changing  the time limit of a step,
1137              either specify a new time limit value or precede the time with a
1138              "+"  or  "-"  to  increment  or decrement the current time limit
1139              (e.g. "TimeLimit=+30"). In order to increment or  decrement  the
1140              current  time  limit,  the StepId specification must precede the
1141              TimeLimit specification.
1142
1143
1144       SPECIFICATIONS FOR UPDATE COMMAND, NODES
1145
1146       NodeName=<name>
1147              Identify the node(s) to be updated. Multiple node names  may  be
1148              specified    using   simple   node   range   expressions   (e.g.
1149              "lx[10-20]"). This specification is required.
1150
1151
1152       ActiveFeatures=<features>
1153              Identify the feature(s) currently active on the specified  node.
1154              Any  previously active feature specification will be overwritten
1155              with the new  value.   Also  see  AvailableFeatures.   Typically
1156              ActiveFeatures  will  be identical to AvailableFeatures; however
1157              ActiveFeatures may be configured as a subset of  the  Available‐
1158              Features. For example, a node may be booted in multiple configu‐
1159              rations. In that case, all possible configurations may be  iden‐
1160              tified as AvailableFeatures, while ActiveFeatures would identify
1161              the current node configuration.
1162
1163
1164       AvailableFeatures=<features>
1165              Identify the feature(s) available on the  specified  node.   Any
1166              previously defined available feature specification will be over‐
1167              written with the  new  value.   AvailableFeatures  assigned  via
1168              scontrol  will  only persist across the restart of the slurmctld
1169              daemon with the -R option and state files  preserved  or  slurm‐
1170              ctld's  receipt of a SIGHUP.  Update slurm.conf with any changes
1171              meant to be persistent across normal restarts  of  slurmctld  or
1172              the execution of scontrol reconfig.  Also see ActiveFeatures.
1173
1174
1175       CpuBind=<node>
1176              Specify  the  task  binding  mode to be used by default for this
1177              node.  Supported options  include:  "none",  "board",  "socket",
1178              "ldom" (NUMA), "core", "thread" and "off" (remove previous bind‐
1179              ing mode).
1180
1181
1182       Gres=<gres>
1183              Identify generic resources to be associated with  the  specified
1184              node.   Any  previously  defined generic resources will be over‐
1185              written with the new value.  Specifications for multiple generic
1186              resources  should  be comma separated.  Each resource specifica‐
1187              tion consists of a name followed by an  optional  colon  with  a
1188              numeric   value   (default   value  is  one)  (e.g.  "Gres=band‐
1189              width:10000").  Modification of GRES count associated with  spe‐
1190              cific  files  (e.g. GPUs) is not allowed other than to set their
1191              count on a node to zero.  In order to change the GRES  count  to
1192              another  value,  modify  your slurm.conf and gres.conf files and
1193              restart daemons.  If GRES as associated with  specific  sockets,
1194              that information will be reported For example if all 4 GPUs on a
1195              node   are   all   associated    with    socket    zero,    then
1196              "Gres=gpu:4(S:0)".  If  associated  with  sockets  0  and 1 then
1197              "Gres=gpu:4(S:0-1)".  The information of which specific GPUs are
1198              associated  with  specific GPUs is not reported, but only avail‐
1199              able by parsing the gres.conf file.  Generic resources  assigned
1200              via  scontrol will only persist across the restart of the slurm‐
1201              ctld daemon with the -R option  and  state  files  preserved  or
1202              slurmctld's  receipt  of  a  SIGHUP.  Update slurm.conf with any
1203              changes meant to be persistent across normal restarts of  slurm‐
1204              ctld or the execution of scontrol reconfig.
1205
1206
1207       Reason=<reason>
1208              Identify  the reason the node is in a "DOWN", "DRAINED", "DRAIN‐
1209              ING", "FAILING" or "FAIL" state.  Use quotes to enclose a reason
1210              having more than one word.
1211
1212
1213       State=<state>
1214              Identify  the  state  to  be assigned to the node. Possible node
1215              states are "NoResp", "ALLOC", "ALLOCATED", "COMPLETING", "DOWN",
1216              "DRAIN",  "FAIL",  "FAILING", "FUTURE" "IDLE", "MAINT", "MIXED",
1217              "PERFCTRS/NPC", "RESERVED", "POWER_DOWN",  "POWER_UP",  "RESUME"
1218              or "UNDRAIN". Not all of those states can be set using the scon‐
1219              trol command only the following  can:  "CANCEL_REBOOT",  "DOWN",
1220              "DRAIN",  "FAIL",  "FUTURE",  "RESUME",  "NoResp", "POWER_DOWN",
1221              "POWER_UP" and "UNDRAIN".  If a node is in a  "MIXED"  state  it
1222              usually  means  the node is in multiple states.  For instance if
1223              only part of the node is "ALLOCATED" and the rest of the node is
1224              "IDLE"  the state will be "MIXED".  If you want to remove a node
1225              from service, you typically want to set it's state  to  "DRAIN".
1226              "CANCEL_REBOOT"  cancels  a  pending reboot on the node (same as
1227              "scontrol  cancel_reboot  <node>").   "FAILING"  is  similar  to
1228              "DRAIN"  except  that  some applications will seek to relinquish
1229              those nodes before the job completes.  "PERFCTRS/NPC"  indicates
1230              that  Network Performance Counters associated with this node are
1231              in use, rendering this node as not usable for  any  other  jobs.
1232              "RESERVED"  indicates the node is in an advanced reservation and
1233              not generally available.  "RESUME" is not an actual node  state,
1234              but  will change a node state from "DRAINED", "DRAINING", "DOWN"
1235              or "REBOOT" to either "IDLE" or "ALLOCATED" state  as  appropri‐
1236              ate.   "RESUME"  will  also clear the "POWERING_DOWN" state of a
1237              node and make it eligible to be allocted.  "UNDRAIN" clears  the
1238              node from being drained (like "RESUME"), but will not change the
1239              node's base state (e.g. "DOWN").  Setting  a  node  "DOWN"  will
1240              cause  all  running and suspended jobs on that node to be termi‐
1241              nated.  "POWER_DOWN" and "POWER_UP" will use the configured Sus‐
1242              pendProg  and  ResumeProg programs to explicitly place a node in
1243              or out of a power saving mode. If  a  node  is  already  in  the
1244              process  of  being  powered  up  or  down, the command will only
1245              change the state of the node but won't have any effect until the
1246              configured  ResumeTimeout  or SuspendTimeout is reached.  Use of
1247              this command can be useful in situations where a ResumeProg like
1248              capmc  in  Cray machines is stalled and one wants to restore the
1249              node to "IDLE" manually, in this case  rebooting  the  node  and
1250              setting  the  state  to  "POWER_DOWN"  will  cancel the previous
1251              "POWER_UP" state and the node will become "IDLE".  The  "NoResp"
1252              state  will only set the "NoResp" flag for a node without chang‐
1253              ing its underlying state.  While all of  the  above  states  are
1254              valid,  some  of  them are not valid new node states given their
1255              prior state.  If the node state code printed is followed by "~",
1256              this  indicates  the  node  is  presently in a power saving mode
1257              (typically running at reduced frequency).   If  the  node  state
1258              code  is  followed  by "#", this indicates the node is presently
1259              being powered up or configured.  If the node state code is  fol‐
1260              lowed by "$", this indicates the node is currently in a reserva‐
1261              tion with a flag value of "maintenance".  If the node state code
1262              is  followed by "@", this indicates the node is currently sched‐
1263              uled  to  be  rebooted.   Generally  only  "DRAIN",  "FAIL"  and
1264              "RESUME"  should be used.  NOTE: The scontrol command should not
1265              be used to change node state on Cray  systems.  Use  Cray  tools
1266              such as xtprocadmin instead.
1267
1268
1269       Weight=<weight>
1270              Identify  weight  to  be  associated  with specified nodes. This
1271              allows dynamic changes to weight associated  with  nodes,  which
1272              will  be  used  for  the  subsequent  node allocation decisions.
1273              Weight assigned  via  scontrol  will  only  persist  across  the
1274              restart  of  the  slurmctld  daemon with the -R option and state
1275              files preserved or slurmctld's  receipt  of  a  SIGHUP.   Update
1276              slurm.conf with any changes meant to be persistent across normal
1277              restarts of slurmctld or the execution of scontrol reconfig.
1278
1279
1280       SPECIFICATIONS FOR SHOW COMMAND, NODES
1281
1282       The meaning of the energy information is as follows:
1283
1284
1285       CurrentWatts
1286              The instantaneous power consumption of the node at the  time  of
1287              the last node energy accounting sample, in watts.
1288
1289
1290       LowestJoules
1291              The  energy  consumed  by  the node between the last time it was
1292              powered on and the last time it was  registered  by  slurmd,  in
1293              joules.
1294
1295
1296       ConsumedJoules
1297              The  energy  consumed  by  the node between the last time it was
1298              registered by  the  slurmd  daemon  and  the  last  node  energy
1299              accounting sample, in joules.
1300
1301
1302       If  the reported value is "n/s" (not supported), the node does not sup‐
1303       port the configured AcctGatherEnergyType plugin. If the reported  value
1304       is zero, energy accounting for nodes is disabled.
1305
1306
1307       The meaning of the external sensors information is as follows:
1308
1309
1310       ExtSensorsJoules
1311              The  energy  consumed  by  the node between the last time it was
1312              powered on and the last external sensors plugin node sample,  in
1313              joules.
1314
1315
1316
1317       ExtSensorsWatts
1318              The  instantaneous  power consumption of the node at the time of
1319              the last external sensors plugin node sample, in watts.
1320
1321
1322       ExtSensorsTemp
1323              The temperature of the node at the time  of  the  last  external
1324              sensors plugin node sample, in celsius.
1325
1326
1327       If  the reported value is "n/s" (not supported), the node does not sup‐
1328       port the configured ExtSensorsType plugin.
1329
1330
1331       The meaning of the resource specialization information is as follows:
1332
1333
1334       CPUSpecList
1335              The list of Slurm abstract CPU IDs on  this  node  reserved  for
1336              exclusive  use by the Slurm compute node daemons (slurmd, slurm‐
1337              stepd).
1338
1339
1340       MemSpecLimit
1341              The combined memory limit, in megabytes, on this  node  for  the
1342              Slurm compute node daemons (slurmd, slurmstepd).
1343
1344
1345       The meaning of the memory information is as follows:
1346
1347
1348       RealMemory
1349              The total memory, in MB, on the node.
1350
1351
1352       AllocMem
1353              The  total  memory,  in  MB,  currently allocated by jobs on the
1354              node.
1355
1356
1357       FreeMem
1358              The total memory, in MB, currently free on the node as  reported
1359              by the OS.
1360
1361
1362       SPECIFICATIONS FOR UPDATE COMMAND, FRONTEND
1363
1364
1365       FrontendName=<name>
1366              Identify the front end node to be updated. This specification is
1367              required.
1368
1369
1370       Reason=<reason>
1371              Identify the reason the node is in a "DOWN"  or  "DRAIN"  state.
1372              Use quotes to enclose a reason having more than one word.
1373
1374
1375       State=<state>
1376              Identify  the state to be assigned to the front end node. Possi‐
1377              ble values are "DOWN", "DRAIN" or  "RESUME".   If  you  want  to
1378              remove  a front end node from service, you typically want to set
1379              it's state to "DRAIN".  "RESUME" is not an  actual  node  state,
1380              but  will  return  a  "DRAINED", "DRAINING", or "DOWN" front end
1381              node to service, either "IDLE" or "ALLOCATED" state as appropri‐
1382              ate.  Setting a front end node "DOWN" will cause all running and
1383              suspended jobs on that node to be terminated.
1384
1385
1386       SPECIFICATIONS FOR CREATE, UPDATE, AND DELETE COMMANDS, PARTITIONS
1387
1388       AllowGroups=<name>
1389              Identify the user groups which may use this partition.  Multiple
1390              groups  may  be  specified in a comma separated list.  To permit
1391              all groups to use the partition specify "AllowGroups=ALL".
1392
1393
1394       AllocNodes=<name>
1395              Comma separated list of nodes from which users can execute  jobs
1396              in  the  partition.   Node names may be specified using the node
1397              range expression syntax described above.  The default  value  is
1398              "ALL".
1399
1400
1401       Alternate=<partition name>
1402              Alternate partition to be used if the state of this partition is
1403              "DRAIN" or "INACTIVE."  The value "NONE" will clear a previously
1404              set alternate partition.
1405
1406
1407       CpuBind=<node>
1408              Specify  the  task  binding  mode to be used by default for this
1409              partition.   Supported   options   include:   "none",   "board",
1410              "socket", "ldom" (NUMA), "core", "thread" and "off" (remove pre‐
1411              vious binding mode).
1412
1413
1414       Default=<yes|no>
1415              Specify if this partition is to be used by  jobs  which  do  not
1416              explicitly  identify a partition to use.  Possible output values
1417              are "YES" and "NO".  In order to change the default partition of
1418              a  running  system,  use  the  scontrol  update  command and set
1419              Default=yes for the partition that you want to  become  the  new
1420              default.
1421
1422
1423       DefaultTime=<time>
1424              Run  time limit used for jobs that don't specify a value. If not
1425              set then MaxTime will be used.  Format is the same as  for  Max‐
1426              Time.
1427
1428
1429       DefMemPerCPU=<MB>
1430              Set  the default memory to be allocated per CPU for jobs in this
1431              partition.  The memory size is specified in megabytes.
1432
1433       DefMemPerNode=<MB>
1434              Set the default memory to be allocated per node for jobs in this
1435              partition.  The memory size is specified in megabytes.
1436
1437
1438       DisableRootJobs=<yes|no>
1439              Specify  if  jobs can be executed as user root.  Possible values
1440              are "YES" and "NO".
1441
1442
1443       GraceTime=<seconds>
1444              Specifies, in units of seconds, the preemption grace time to  be
1445              extended  to  a job which has been selected for preemption.  The
1446              default value is zero, no preemption grace time  is  allowed  on
1447              this partition or qos.  (Meaningful only for PreemptMode=CANCEL)
1448
1449
1450       Hidden=<yes|no>
1451              Specify  if  the  partition  and  its jobs should be hidden from
1452              view.  Hidden partitions will by  default  not  be  reported  by
1453              Slurm APIs or commands.  Possible values are "YES" and "NO".
1454
1455
1456       JobDefaults=<specs>
1457              Specify  job  default  values  using  a  comma delimited list of
1458              "key=value" pairs.  Supported keys include
1459
1460              DefCpuPerGPU  Default number of CPUs per allocated GPU.
1461
1462              DefMemPerGPU  Default memory limit (in megabytes) per  allocated
1463                            GPU.
1464
1465
1466       MaxMemPerCPU=<MB>
1467              Set  the maximum memory to be allocated per CPU for jobs in this
1468              partition.  The memory size is specified in megabytes.
1469
1470       MaxMemPerCNode=<MB>
1471              Set the maximum memory to be allocated per node for jobs in this
1472              partition.  The memory size is specified in megabytes.
1473
1474
1475       MaxNodes=<count>
1476              Set  the  maximum number of nodes which will be allocated to any
1477              single job in the partition. Specify  a  number,  "INFINITE"  or
1478              "UNLIMITED".  Changing the MaxNodes of a partition has no effect
1479              upon jobs that have already begun execution.
1480
1481
1482       MaxTime=<time>
1483              The   maximum   run   time   for   jobs.    Output   format   is
1484              [days-]hours:minutes:seconds  or "UNLIMITED".  Input format (for
1485              update command) is minutes, minutes:seconds,  hours:minutes:sec‐
1486              onds,  days-hours, days-hours:minutes or days-hours:minutes:sec‐
1487              onds.  Time resolution is  one  minute  and  second  values  are
1488              rounded up to the next minute.  Changing the MaxTime of a parti‐
1489              tion has no effect upon jobs that have already begun execution.
1490
1491
1492       MinNodes=<count>
1493              Set the minimum number of nodes which will be allocated  to  any
1494              single  job in the partition.  Changing the MinNodes of a parti‐
1495              tion has no effect upon jobs that have already begun execution.
1496
1497
1498       Nodes=<name>
1499              Identify the node(s) to be associated with this partition.  Mul‐
1500              tiple  node  names  may  be  specified  using  simple node range
1501              expressions (e.g. "lx[10-20]").  Note  that  jobs  may  only  be
1502              associated with one partition at any time.  Specify a blank data
1503              value to remove all nodes from a partition: "Nodes=".   Changing
1504              the  Nodes  in  a  partition  has  no effect upon jobs that have
1505              already begun execution.
1506
1507
1508       OverTimeLimit=<count>
1509              Number of minutes by which a  job  can  exceed  its  time  limit
1510              before being canceled.  The configured job time limit is treated
1511              as a soft limit.  Adding OverTimeLimit to the  soft  limit  pro‐
1512              vides a hard limit, at which point the job is canceled.  This is
1513              particularly useful for backfill scheduling,  which  bases  upon
1514              each  job's soft time limit.  A partition-specific OverTimeLimit
1515              will override any global OverTimeLimit value.  If not specified,
1516              the  global  OverTimeLimit  value will take precedence.  May not
1517              exceed exceed 65533 minutes.  An input value of "UNLIMITED" will
1518              clear any previously configured partition-specific OverTimeLimit
1519              value.
1520
1521
1522       OverSubscribe=<yes|no|exclusive|force>[:<job_count>]
1523              Specify if compute resources (i.e. individual CPUs) in this par‐
1524              tition  can  be  shared  by  multiple jobs.  Possible values are
1525              "YES", "NO", "EXCLUSIVE" and "FORCE".   An  optional  job  count
1526              specifies how many jobs can be allocated to use each resource.
1527
1528
1529       PartitionName=<name>
1530              Identify  the  partition  to  be  updated. This specification is
1531              required.
1532
1533
1534       PreemptMode=<mode>
1535              Reset the mechanism used to preempt jobs in  this  partition  if
1536              PreemptType is configured to preempt/partition_prio. The default
1537              preemption mechanism is specified by the  cluster-wide  Preempt‐
1538              Mode  configuration parameter.  Possible values are "OFF", "CAN‐
1539              CEL", "CHECKPOINT", "REQUEUE" and "SUSPEND".
1540
1541
1542       Priority=<count>
1543              Jobs submitted to a higher priority partition will be dispatched
1544              before pending jobs in lower priority partitions and if possible
1545              they will preempt running jobs from lower  priority  partitions.
1546              Note  that  a partition's priority takes precedence over a job's
1547              priority.  The value may not exceed 65533.
1548
1549
1550       PriorityJobFactor=<count>
1551              Partition factor used by priority/multifactor plugin  in  calcu‐
1552              lating  job priority.  The value may not exceed 65533.  Also see
1553              PriorityTier.
1554
1555
1556       PriorityTier=<count>
1557              Jobs submitted to a partition with a higher priority tier  value
1558              will  be  dispatched before pending jobs in partition with lower
1559              priority tier value and,  if   possible,   they   will   preempt
1560              running  jobs  from  partitions with lower priority tier values.
1561              Note that a partition's priority tier takes  precedence  over  a
1562              job's  priority.  The value may not exceed 65533.  Also see Pri‐
1563              orityJobFactor.
1564
1565
1566       QOS=<QOSname|blank to remove>
1567              Set the partition QOS with a QOS name or to remove the Partition
1568              QOS leave the option blank.
1569
1570
1571       RootOnly=<yes|no>
1572              Specify  if only allocation requests initiated by user root will
1573              be satisfied.  This can be used to restrict control of the  par‐
1574              tition  to  some  meta-scheduler.  Possible values are "YES" and
1575              "NO".
1576
1577
1578       ReqResv=<yes|no>
1579              Specify if only allocation requests  designating  a  reservation
1580              will  be satisfied.  This is used to restrict partition usage to
1581              be allowed only within a reservation.  Possible values are "YES"
1582              and "NO".
1583
1584
1585       Shared=<yes|no|exclusive|force>[:<job_count>]
1586              Renamed to OverSubscribe, see option descriptions above.
1587
1588
1589       State=<up|down|drain|inactive>
1590              Specify  if jobs can be allocated nodes or queued in this parti‐
1591              tion.  Possible values are "UP", "DOWN", "DRAIN" and "INACTIVE".
1592
1593              UP        Designates that new jobs may queued on the  partition,
1594                        and  that jobs may be allocated nodes and run from the
1595                        partition.
1596
1597              DOWN      Designates that new jobs may be queued on  the  parti‐
1598                        tion,  but  queued jobs may not be allocated nodes and
1599                        run from the partition. Jobs already  running  on  the
1600                        partition continue to run. The jobs must be explicitly
1601                        canceled to force their termination.
1602
1603              DRAIN     Designates that no new jobs may be queued on the  par‐
1604                        tition (job submission requests will be denied with an
1605                        error message), but jobs already queued on the  parti‐
1606                        tion  may  be  allocated  nodes and run.  See also the
1607                        "Alternate" partition specification.
1608
1609              INACTIVE  Designates that no new jobs may be queued on the  par‐
1610                        tition,  and  jobs already queued may not be allocated
1611                        nodes and run.  See  also  the  "Alternate"  partition
1612                        specification.
1613
1614
1615       TRESBillingWeights=<TRES Billing Weights>
1616              TRESBillingWeights is used to define the billing weights of each
1617              TRES type that will be used in calculating the usage of  a  job.
1618              The calculated usage is used when calculating fairshare and when
1619              enforcing the TRES billing limit on jobs.   Updates  affect  new
1620              jobs  and  not  existing  jobs.  See the slurm.conf man page for
1621              more information.
1622
1623
1624
1625       SPECIFICATIONS FOR UPDATE COMMAND, POWERCAP
1626
1627
1628       PowerCap=<count>
1629              Set the amount of watts the cluster is limited  to.   Specify  a
1630              number,  "INFINITE"  to  enable  the power capping logic without
1631              power restriction or "0" to disable  the  power  capping  logic.
1632              Update slurm.conf with any changes meant to be persistent across
1633              normal restarts of slurmctld or the execution of scontrol recon‐
1634              fig.
1635
1636
1637       SPECIFICATIONS FOR CREATE, UPDATE, AND DELETE COMMANDS, RESERVATIONS
1638
1639
1640
1641       Reservation=<name>
1642              Identify  the  name  of  the  reservation  to be created,
1643              updated, or deleted.   This  parameter  is  required  for
1644              update and is the only parameter for delete.  For create,
1645              if you do not want to give a reservation name, use "scon‐
1646              trol create res ..." and a name will be created automati‐
1647              cally.
1648
1649
1650       Accounts=<account list>
1651              List of accounts permitted to use the reserved nodes, for
1652              example "Accounts=physcode1,physcode2".  A user in any of
1653              the accounts may use the reserved nodes.  A new  reserva‐
1654              tion  must  specify Users and/or Accounts.  If both Users
1655              and Accounts are specified, a  job  must  match  both  in
1656              order  to  use  the  reservation.   Accounts  can also be
1657              denied access to reservations by  preceding  all  of  the
1658              account  names  with  '-'.  Alternately precede the equal
1659              sign        with        '-'.         For         example,
1660              "Accounts=-physcode1,-physcode2"                       or
1661              "Accounts-=physcode1,physcode2" will permit  any  account
1662              except  physcode1  and  physcode2 to use the reservation.
1663              You can add or remove individual accounts from an  exist‐
1664              ing  reservation by using the update command and adding a
1665              '+' or '-' sign before the '='  sign.   If  accounts  are
1666              denied  access to a reservation (account name preceded by
1667              a '-'), then all other accounts are implicitly allowed to
1668              use  the  reservation  and  it  is  not  possible to also
1669              explicitly specify allowed accounts.
1670
1671
1672       BurstBuffer=<buffer_spec>[,<buffer_spec>,...]
1673              Specification of burst buffer resources which are  to  be
1674              reserved.    "buffer_spec"  consists  of  four  elements:
1675              [plugin:][type:]#[units] "plugin"  is  the  burst  buffer
1676              plugin  name,  currently  either "datawarp" or "generic".
1677              If no plugin is specified, the reservation applies to all
1678              configured burst buffer plugins.  "type" specifies a Cray
1679              generic burst buffer resource, for example  "nodes".   if
1680              "type" is not specified, the number is a measure of stor‐
1681              age space.  The "units"  may  be  "N"  (nodes),  "K|KiB",
1682              "M|MiB",  "G|GiB",  "T|TiB", "P|PiB" (for powers of 1024)
1683              and "KB", "MB", "GB", "TB", "PB" (for  powers  of  1000).
1684              The  default  units are bytes for reservations of storage
1685              space.  For example  "BurstBuffer=datawarp:2TB"  (reserve
1686              2TB  of  storage  plus  3  nodes from the Cray plugin) or
1687              "BurstBuffer=100GB" (reserve 100 GB of storage  from  all
1688              configured burst buffer plugins).  Jobs using this reser‐
1689              vation  are  not  restricted  to   these   burst   buffer
1690              resources,  but may use these reserved resources plus any
1691              which  are  generally  available.   NOTE:  Usually  Slurm
1692              interprets  KB,  MB,  GB,  TB,  PB, TB units as powers of
1693              1024, but for Burst  Buffers  size  specifications  Slurm
1694              supports  both  IEC/SI formats.  This is because the CRAY
1695              API for managing DataWarps supports both formats.
1696
1697
1698       CoreCnt=<num>
1699              This   option   is   only    supported    when    Select‐
1700              Type=select/cons_res.  Identify  number  of  cores  to be
1701              reserved. If NodeCnt  is  used  without  the  FIRST_CORES
1702              flag,  this is the total number of cores to reserve where
1703              cores per node is  CoreCnt/NodeCnt.   If  a  nodelist  is
1704              used,  or  if  NodeCnt is used with the FIRST_CORES flag,
1705              this  should  be  an  array  of  core  numbers  by  node:
1706              Nodes=node[1-5]  CoreCnt=2,2,3,3,4  or  flags=FIRST_CORES
1707              NodeCnt=5 CoreCnt=1,2,1,3,2.
1708
1709
1710       Licenses=<license>
1711              Specification of licenses (or other  resources  available
1712              on  all  nodes  of the cluster) which are to be reserved.
1713              License names can be followed by a colon and  count  (the
1714              default  count is one).  Multiple license names should be
1715              comma  separated  (e.g.  "Licenses=foo:4,bar").   A   new
1716              reservation  must  specify  one  or  more  resource to be
1717              included: NodeCnt, Nodes and/or Licenses.  If a  reserva‐
1718              tion includes Licenses, but no NodeCnt or Nodes, then the
1719              option Flags=LICENSE_ONLY must also be  specified.   Jobs
1720              using  this  reservation  are  not  restricted  to  these
1721              licenses, but may use these reserved  licenses  plus  any
1722              which are generally available.
1723
1724
1725       NodeCnt=<num>[,num,...]
1726              Identify  number  of nodes to be reserved. The number can
1727              include a suffix of "k" or "K", in which case the  number
1728              specified  is multiplied by 1024.  A new reservation must
1729              specify one or more resource  to  be  included:  NodeCnt,
1730              Nodes and/or Licenses.
1731
1732
1733       Nodes=<name>
1734              Identify  the node(s) to be reserved. Multiple node names
1735              may be specified  using  simple  node  range  expressions
1736              (e.g.  "Nodes=lx[10-20]").  Specify a blank data value to
1737              remove all nodes from a  reservation:  "Nodes=".   A  new
1738              reservation  must  specify  one  or  more  resource to be
1739              included: NodeCnt, Nodes and/or Licenses. A specification
1740              of "ALL" will reserve all nodes. Set Flags=PART_NODES and
1741              PartitionName= in order for changes in the nodes  associ‐
1742              ated  with  a partition to also be reflected in the nodes
1743              associated with a reservation.
1744
1745
1746       StartTime=<time_spec>
1747              The start time for the reservation.   A  new  reservation
1748              must  specify a start time.  It accepts times of the form
1749              HH:MM:SS  for  a  specific  time  of  day  (seconds   are
1750              optional).   (If  that time is already past, the next day
1751              is assumed.)  You may also specify midnight,  noon,  fika
1752              (3  PM)  or teatime (4 PM) and you can have a time-of-day
1753              suffixed with AM or PM for running in the morning or  the
1754              evening.   You can also say what day the job will be run,
1755              by specifying a date of the form MMDDYY  or  MM/DD/YY  or
1756              MM.DD.YY,  or a date and time as YYYY-MM-DD[THH:MM[:SS]].
1757              You can also give times  like  now  +  count  time-units,
1758              where  the  time-units  can  be  minutes, hours, days, or
1759              weeks and you can tell Slurm to run the  job  today  with
1760              the  keyword  today  and to run the job tomorrow with the
1761              keyword tomorrow. You cannot update the  StartTime  of  a
1762              reservation in ACTIVE state.
1763
1764
1765       EndTime=<time_spec>
1766              The end time for the reservation.  A new reservation must
1767              specify an end time or a duration.  Valid formats are the
1768              same as for StartTime.
1769
1770
1771       Duration=<time>
1772              The  length  of  a  reservation.   A new reservation must
1773              specify an end time or a  duration.   Valid  formats  are
1774              minutes,      minutes:seconds,     hours:minutes:seconds,
1775              days-hours,  days-hours:minutes,  days-hours:minutes:sec‐
1776              onds,  or  UNLIMITED.   Time resolution is one minute and
1777              second values are rounded up to the next  minute.  Output
1778              format is always [days-]hours:minutes:seconds.
1779
1780
1781       PartitionName=<name>
1782              Identify the partition to be reserved.
1783
1784
1785       Flags=<flags>
1786              Flags  associated  with  the reservation.  You can add or
1787              remove individual flags from an existing  reservation  by
1788              adding  a '+' or '-' sign before the '=' sign.  For exam‐
1789              ple: Flags-=DAILY (NOTE: this shortcut is  not  supported
1790              for all flags).  Currently supported flags include:
1791
1792              ANY_NODES     This  is  a  reservation  for burst buffers
1793                            and/or licenses only and not compute nodes.
1794                            If  this  flag  is  set,  a  job using this
1795                            reservation may use  the  associated  burst
1796                            buffers  and/or  licenses  plus any compute
1797                            nodes.  If this flag  is  not  set,  a  job
1798                            using  this  reservation  may  use only the
1799                            nodes  and  licenses  associated  with  the
1800                            reservation.
1801
1802              DAILY         Repeat  the  reservation  at  the same time
1803                            every day.
1804
1805              FLEX          Permit jobs requesting the  reservation  to
1806                            begin  prior  to  the  reservation's  start
1807                            time, end after the reservation's end time,
1808                            and use any resources inside and/or outside
1809                            of the reservation regardless of  any  con‐
1810                            straints possibly set in the reservation. A
1811                            typical use case is  to  prevent  jobs  not
1812                            explicitly  requesting the reservation from
1813                            using those reserved resources rather  than
1814                            forcing  jobs requesting the reservation to
1815                            use  those  resources  in  the  time  frame
1816                            reserved.  Another  use  case  could  be to
1817                            always have a particular  number  of  nodes
1818                            with a specific feature reserved for a spe‐
1819                            cific account so users in this account  may
1820                            use  this  nodes  plus possibly other nodes
1821                            without this feature.
1822
1823              FIRST_CORES   Use the lowest numbered  cores  on  a  node
1824                            only.
1825
1826              IGNORE_JOBS   Ignore currently running jobs when creating
1827                            the reservation.  This  can  be  especially
1828                            useful when reserving all nodes in the sys‐
1829                            tem for maintenance.
1830
1831              LICENSE_ONLY  See ANY_NODES.
1832
1833              MAINT         Maintenance mode, receives special account‐
1834                            ing treatment.  This partition is permitted
1835                            to  use  resources  that  are  already   in
1836                            another reservation.
1837
1838              NO_HOLD_JOBS_AFTER
1839                            By  default,  when  a  reservation ends the
1840                            reservation request will  be  removed  from
1841                            any  pending jobs submitted to the reserva‐
1842                            tion and will be put  into  a  held  state.
1843                            Use  this  flag  to let jobs run outside of
1844                            the reservation after  the  reservation  is
1845                            gone.
1846
1847              OVERLAP       This reservation can be allocated resources
1848                            that are already in another reservation.
1849
1850              PART_NODES    This flag can be used to reserve all  nodes
1851                            within the specified partition.  Partition‐
1852                            Name and Nodes=ALL  must  be  specified  or
1853                            this option is ignored.
1854
1855              PURGE_COMP    Purge the reservation once the last associ‐
1856                            ated job has completed.  Once the  reserva‐
1857                            tion has been created, it must be populated
1858                            within 5 minutes of its start  time  or  it
1859                            will  be  purged  before any jobs have been
1860                            run.
1861
1862              REPLACE       Nodes which are DOWN, DRAINED, or allocated
1863                            to jobs are automatically replenished using
1864                            idle resources.  This option can be used to
1865                            maintain   a   constant   number   of  idle
1866                            resources available for pending jobs  (sub‐
1867                            ject  to  availability  of idle resources).
1868                            This should be used with the NodeCnt reser‐
1869                            vation  option;  do  not  identify specific
1870                            nodes to be included in the reservation.
1871
1872              REPLACE_DOWN  Nodes which are DOWN or DRAINED  are  auto‐
1873                            matically replenished using idle resources.
1874                            This option can be used to maintain a  con‐
1875                            stant sized pool of resources available for
1876                            pending jobs (subject  to  availability  of
1877                            idle  resources).  This should be used with
1878                            the  NodeCnt  reservation  option;  do  not
1879                            identify  specific  nodes to be included in
1880                            the reservation.
1881
1882              SPEC_NODES    Reservation is for specific  nodes  (output
1883                            only)
1884
1885              STATIC_ALLOC  Make it so after the nodes are selected for
1886                            a reservation they don't  change.   Without
1887                            this  option  when nodes are selected for a
1888                            reservation and one goes down the  reserva‐
1889                            tion  will  select  a  new node to fill the
1890                            spot.
1891
1892              TIME_FLOAT    The reservation start time is  relative  to
1893                            the  current time and moves forward through
1894                            time (e.g. a  StartTime=now+10minutes  will
1895                            always be 10 minutes in the future).
1896
1897              WEEKDAY       Repeat  the reservation at the same time on
1898                            every weekday (Monday, Tuesday,  Wednesday,
1899                            Thursday and Friday).
1900
1901              WEEKEND       Repeat  the reservation at the same time on
1902                            every weekend day (Saturday and Sunday).
1903
1904              WEEKLY        Repeat the reservation  at  the  same  time
1905                            every week.
1906
1907
1908       Features=<features>
1909              Set  the  reservation's  required node features. Multiple
1910              values may be "&" separated if all features are  required
1911              (AND  operation) or separated by "|" if any of the speci‐
1912              fied features are required (OR  operation).   Parenthesis
1913              are also supported for features to be ANDed together with
1914              counts of nodes having the specified features.  For exam‐
1915              ple "Features=[(knl&a2a&flat)*4&haswell*2]" indicates the
1916              advanced reservation should include 4 nodes with  ALL  of
1917              the  features  "knl", "a2a", and "flat" plus 2 nodes with
1918              the feature "haswell".
1919
1920              Value may be cleared with blank data value, "Features=".
1921
1922
1923       Users=<user list>
1924              List of users permitted to use the  reserved  nodes,  for
1925              example  "User=jones1,smith2".   A  new  reservation must
1926              specify  Users  and/or  Accounts.   If  both  Users   and
1927              Accounts are specified, a job must match both in order to
1928              use the reservation.  Users can also be denied access  to
1929              reservations by preceding all of the user names with '-'.
1930              Alternately precede the equal sign with '-'.   For  exam‐
1931              ple, "User=-jones1,-smith2" or "User-=jones1,smith2" will
1932              permit any user except  jones1  and  smith2  to  use  the
1933              reservation.  You can add or remove individual users from
1934              an existing reservation by using the update  command  and
1935              adding  a  '+' or '-' sign before the '=' sign.  If users
1936              are denied access to a reservation (user name preceded by
1937              a  '-'),  then  all other users are implicitly allowed to
1938              use the reservation  and  it  is  not  possible  to  also
1939              explicitly specify allowed users.
1940
1941
1942       TRES=<tres_spec>
1943              Comma-separated  list  of  TRES required for the reserva‐
1944              tion. Current supported TRES types with reservations are:
1945              CPU,  Node,  License and BB. CPU and Node follow the same
1946              format as CoreCnt and  NodeCnt  parameters  respectively.
1947              License  names  can  be  followed  by  an equal '=' and a
1948              count:
1949
1950              License/<name1>=<count1>[,License/<name2>=<count2>,...]
1951
1952              BurstBuffer can be specified in a similar way  as  Burst‐
1953              Buffer  parameter. The only difference is that colon sym‐
1954              bol ':' should be replaced by an equal '='  in  order  to
1955              follow the TRES format.
1956
1957              Some examples of TRES valid specifications:
1958
1959              TRES=cpu=5,bb/cray=4,license/iop1=1,license/iop2=3
1960
1961              TRES=node=5k,license/iop1=2
1962
1963              As  specified in CoreCnt, if a nodelist is specified, cpu
1964              can be an array  of  core  numbers  by  node:  nodes=com‐
1965              pute[1-3] TRES=cpu=2,2,1,bb/cray=4,license/iop1=2
1966
1967              Please  note  that CPU, Node, License and BB can override
1968              CoreCnt, NodeCnt,  Licenses  and  BurstBuffer  parameters
1969              respectively.  Also CPU represents CoreCnt, in a reserva‐
1970              tion and will be adjusted if you have threads per core on
1971              your nodes.
1972
1973
1974       SPECIFICATIONS FOR SHOW COMMAND, LAYOUTS
1975
1976       Without  options,  lists  all  configured layouts. With a layout
1977       specified, shows entities with following options:
1978
1979
1980       Key=<value>
1981              Keys/Values to update for the entities. The  format  must
1982              respect the layout.d configuration files. Key=Type cannot
1983              be updated. One Key/Value is  required,  several  can  be
1984              set.
1985
1986       Entity=<value>
1987              Entities to show, default is not used. Can be set to "*".
1988
1989       Type=<value>
1990              Type of entities to show, default is not used.
1991
1992       nolayout
1993              If  not  used,  only  entities with defining the tree are
1994              shown.  With the option, only leaves are shown.
1995
1996

ENVIRONMENT VARIABLES

1998       Some scontrol options may  be  set  via  environment  variables.
1999       These  environment  variables,  along  with  their corresponding
2000       options, are  listed  below.  (Note:  Commandline  options  will
2001       always override these settings.)
2002
2003       SCONTROL_ALL        -a, --all
2004
2005       SCONTROL_FEDERATION --federation
2006
2007       SCONTROL_FUTURE     -F, --future
2008
2009       SCONTROL_LOCAL      --local
2010
2011       SCONTROL_SIBLING    --sibling
2012
2013       SLURM_BITSTR_LEN    Specifies  the  string length to be used for
2014                           holding a job array's  task  ID  expression.
2015                           The default value is 64 bytes.  A value of 0
2016                           will print  the  full  expression  with  any
2017                           length    required.    Larger   values   may
2018                           adversely  impact  the  application  perfor‐
2019                           mance.
2020
2021       SLURM_CLUSTERS      Same as --clusters
2022
2023       SLURM_CONF          The  location  of  the  Slurm  configuration
2024                           file.
2025
2026       SLURM_TIME_FORMAT   Specify  the  format  used  to  report  time
2027                           stamps.  A  value  of  standard, the default
2028                           value,  generates   output   in   the   form
2029                           "year-month-dateThour:minute:second".      A
2030                           value    of    relative     returns     only
2031                           "hour:minute:second"  if  the  current  day.
2032                           For other  dates  in  the  current  year  it
2033                           prints   the   "hour:minute"   preceded   by
2034                           "Tomorr" (tomorrow),  "Ystday"  (yesterday),
2035                           the  name  of  the  day  for the coming week
2036                           (e.g. "Mon",  "Tue",  etc.),  otherwise  the
2037                           date  (e.g.  "25  Apr").  For other years it
2038                           returns a date month and year without a time
2039                           (e.g.  "6 Jun 2012"). All of the time stamps
2040                           use a 24 hour format.
2041
2042                           A valid strftime() format can also be speci‐
2043                           fied.  For  example, a value of "%a %T" will
2044                           report the day of the week and a time  stamp
2045                           (e.g. "Mon 12:34:56").
2046
2047
2048       SLURM_TOPO_LEN      Specify  the  maximum  size of the line when
2049                           printing Topology. If not set,  the  default
2050                           value is unlimited.
2051
2052

AUTHORIZATION

2054       When  using  the  Slurm  db, users who have AdminLevel's defined
2055       (Operator or Admin) and users who are account  coordinators  are
2056       given  the  authority  to  view  and  modify jobs, reservations,
2057       nodes, etc., as defined in the following table -  regardless  of
2058       whether  a  PrivateData  restriction  has  been  defined  in the
2059       slurm.conf file.
2060
2061       scontrol show job(s):        Admin, Operator, Coordinator
2062       scontrol update job:         Admin, Operator, Coordinator
2063       scontrol requeue:            Admin, Operator, Coordinator
2064       scontrol show step(s):       Admin, Operator, Coordinator
2065       scontrol update step:        Admin, Operator, Coordinator
2066
2067       scontrol show node:          Admin, Operator
2068       scontrol update node:        Admin
2069
2070       scontrol create partition:   Admin
2071       scontrol show partition:     Admin, Operator
2072       scontrol update partition:   Admin
2073       scontrol delete partition:   Admin
2074
2075       scontrol create reservation: Admin, Operator
2076       scontrol show reservation:   Admin, Operator
2077       scontrol update reservation: Admin, Operator
2078       scontrol delete reservation: Admin, Operator
2079
2080       scontrol reconfig:           Admin
2081       scontrol shutdown:           Admin
2082       scontrol takeover:           Admin
2083
2084

EXAMPLES

2086       # scontrol
2087       scontrol: show part debug
2088       PartitionName=debug
2089          AllocNodes=ALL AllowGroups=ALL Default=YES
2090          DefaultTime=NONE DisableRootJobs=NO Hidden=NO
2091          MaxNodes=UNLIMITED MaxTime=UNLIMITED MinNodes=1
2092          Nodes=snowflake[0-48]
2093          Priority=1 RootOnly=NO OverSubscribe=YES:4
2094          State=UP TotalCPUs=694 TotalNodes=49
2095       scontrol: update PartitionName=debug MaxTime=60:00 MaxNodes=4
2096       scontrol: show job 71701
2097       JobId=71701 Name=hostname
2098          UserId=da(1000) GroupId=da(1000)
2099          Priority=66264 Account=none QOS=normal WCKey=*123
2100          JobState=COMPLETED Reason=None Dependency=(null)
2101          TimeLimit=UNLIMITED Requeue=1  Restarts=0  BatchFlag=0  Exit‐
2102       Code=0:0
2103          SubmitTime=2010-01-05T10:58:40                      Eligible‐
2104       Time=2010-01-05T10:58:40
2105          StartTime=2010-01-05T10:58:40 EndTime=2010-01-05T10:58:40
2106          SuspendTime=None SecsPreSuspend=0
2107          Partition=debug AllocNode:Sid=snowflake:4702
2108          ReqNodeList=(null) ExcNodeList=(null)
2109          NodeList=snowflake0
2110          NumNodes=1 NumCPUs=10 CPUs/Task=2 ReqS:C:T=1:1:1
2111          MinCPUsNode=2 MinMemoryNode=0 MinTmpDiskNode=0
2112          Features=(null) Reservation=(null)
2113          OverSubscribe=OK Contiguous=0 Licenses=(null) Network=(null)
2114       scontrol: update JobId=71701 TimeLimit=30:00 Priority=500
2115       scontrol: show hostnames tux[1-3]
2116       tux1
2117       tux2
2118       tux3
2119       scontrol:   create   res   StartTime=2009-04-01T08:00:00   Dura‐
2120       tion=5:00:00 Users=dbremer NodeCnt=10
2121       Reservation created: dbremer_1
2122       scontrol: update Reservation=dbremer_1 Flags=Maint NodeCnt=20
2123       scontrol: delete Reservation=dbremer_1
2124       scontrol: quit
2125
2126

COPYING

2128       Copyright  (C)  2002-2007 The Regents of the University of Cali‐
2129       fornia.  Produced at Lawrence Livermore National Laboratory (cf,
2130       DISCLAIMER).
2131       Copyright (C) 2008-2010 Lawrence Livermore National Security.
2132       Copyright (C) 2010-2018 SchedMD LLC.
2133
2134       This  file is part of Slurm, a resource management program.  For
2135       details, see <https://slurm.schedmd.com/>.
2136
2137       Slurm is free software; you can redistribute it and/or modify it
2138       under  the  terms of the GNU General Public License as published
2139       by the  Free  Software  Foundation;  either  version  2  of  the
2140       License, or (at your option) any later version.
2141
2142       Slurm  is  distributed  in  the hope that it will be useful, but
2143       WITHOUT ANY WARRANTY; without even the implied warranty of  MER‐
2144       CHANTABILITY  or  FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
2145       General Public License for more details.
2146

FILES

2148       /etc/slurm.conf
2149

SEE ALSO

2151       scancel(1),   sinfo(1),   squeue(1),    slurm_checkpoint    (3),
2152       slurm_create_partition    (3),    slurm_delete_partition    (3),
2153       slurm_load_ctl_conf (3),  slurm_load_jobs  (3),  slurm_load_node
2154       (3),    slurm_load_partitions    (3),   slurm_reconfigure   (3),
2155       slurm_requeue  (3),  slurm_resume   (3),   slurm_shutdown   (3),
2156       slurm_suspend  (3),  slurm_takeover  (3),  slurm_update_job (3),
2157       slurm_update_node     (3),      slurm_update_partition      (3),
2158       slurm.conf(5), slurmctld(8)
2159
2160
2161
2162August 2019                     Slurm Commands                     scontrol(1)
Impressum