1CONDOR_Q(1)                     HTCondor Manual                    CONDOR_Q(1)
2
3
4

NAME

6       condor_q - HTCondor Manual
7
8       Display information about jobs in queue
9
10
11

SYNOPSIS

13       condor_q [-help [Universe | State] ]
14
15       condor_q  [-debug  ] [general options ] [restriction list ] [output op‐
16       tions ] [analyze options ]
17

DESCRIPTION

19       condor_q displays information about jobs in the HTCondor job queue.  By
20       default, condor_q queries the local job queue, but this behavior may be
21       modified by specifying one of the general options.
22
23       As of version 8.5.2, condor_q defaults to  querying  only  the  current
24       user's  jobs.  This default is overridden when the restriction list has
25       usernames and/or job ids, when the -submitter  or  -allusers  arguments
26       are  specified,  or  when the current user is a queue superuser. It can
27       also be overridden by setting the  CONDOR_Q_ONLY_MY_JOBS  configuration
28       macro to False.
29
30       As of version 8.5.6, condor_q defaults to batch-mode output (see -batch
31       in the Options section below). The old  behavior  can  be  obtained  by
32       specifying  -nobatch on the command line. To change the default back to
33       its  pre-8.5.6  value,  set  the  new   configuration   variable   CON‐
34       DOR_Q_DASH_BATCH_IS_DEFAULT to False.
35

BATCHES OF JOBS

37       As  of version 8.5.6, condor_q defaults to displaying information about
38       batches of jobs, rather than individual jobs.  The  intention  is  that
39       this  will  be  a more useful, and user-friendly, format for users with
40       large numbers of jobs in the queue. Ideally, users will  specify  mean‐
41       ingful  batch  names for their jobs, to make it easier to keep track of
42       related jobs.
43
44       (For information about specifying batch names for your  jobs,  see  the
45       condor_submit and condor_submit_dag manual pages.)
46
47       A batch of jobs is defined as follows:
48
49       • An  entire  workflow  (a  DAG or hierarchy of nested DAGs) (note that
50         condor_dagman now specifies a default batch name for all  jobs  in  a
51         given workflow)
52
53       • All jobs in a single cluster
54
55       • All  jobs  submitted  by  a single user that have the same executable
56         specified in their submit file (unless submitted with different batch
57         names)
58
59       • All  jobs  submitted  by  a single user that have the same batch name
60         specified in their submit file or on the condor_submit or condor_sub‐
61         mit_dag command line.
62

OUTPUT

64       There  are many output options that modify the output generated by con‐
65       dor_q. The effects of these options, and the meanings  of  the  various
66       output data, are described below.
67
68   Output options
69       If  the -long option is specified, condor_q displays a long description
70       of the queried jobs by printing the entire job  ClassAd  for  all  jobs
71       matching  the  restrictions,  if  any. Individual attributes of the job
72       ClassAd can be displayed by means of the -format option, which displays
73       attributes  with  a  printf(3)  format, or with the -autoformat option.
74       Multiple -format options may be specified in the option list to display
75       several attributes of the job.
76
77       For  most  output  options (except as specified), the last line of con‐
78       dor_q output contains a summary of the queue: the total number of jobs,
79       and  the  number of jobs in the completed, removed, idle, running, held
80       and suspended states.
81
82       If no output options are specified,  condor_q  now  defaults  to  batch
83       mode,  and displays the following columns of information, with one line
84       of output per batch of jobs:
85
86          OWNER, BATCH_NAME, SUBMITTED, DONE, RUN, IDLE, [HOLD,] TOTAL, JOB_IDS
87
88       Note that the HOLD column is only shown if there are held jobs  in  the
89       output or if there are no jobs in the output.
90
91       If  the  -nobatch  option is specified, condor_q displays the following
92       columns of information, with one line of output per job:
93
94          ID, OWNER, SUBMITTED, RUN_TIME, ST, PRI, SIZE, CMD
95
96       If the -dag option is specified (in conjunction  with  -nobatch),  con‐
97       dor_q  displays  the following columns of information, with one line of
98       output per job; the owner is shown only for top-level jobs, and for all
99       other jobs (including sub-DAGs) the node name is shown:
100
101          ID, OWNER/NODENAME, SUBMITTED, RUN_TIME, ST, PRI, SIZE, CMD
102
103       If  the  -run  option is specified (in conjunction with -nobatch), con‐
104       dor_q displays the following columns of information, with one  line  of
105       output per running job:
106
107          ID, OWNER, SUBMITTED, RUN_TIME, HOST(S)
108
109       Also note that the -run option disables output of the totals line.
110
111       If  the -grid option is specified, condor_q displays the following col‐
112       umns of information, with one line of output per job:
113
114          ID, OWNER, STATUS, GRID->MANAGER, HOST, GRID_JOB_ID
115
116       If the -grid:ec2 option is specified, condor_q displays  the  following
117       columns of information, with one line of output per job:
118
119          ID, OWNER, STATUS, INSTANCE ID, CMD
120
121       If  the  -goodput  option is specified, condor_q displays the following
122       columns of information, with one line of output per job:
123
124          ID, OWNER, SUBMITTED, RUN_TIME, GOODPUT, CPU_UTIL, Mb/s
125
126       If the -io option is specified, condor_q displays the following columns
127       of information, with one line of output per job:
128
129          ID, OWNER, RUNS, ST, INPUT, OUTPUT, RATE, MISC
130
131       If  the  -cputime  option  is specified (in conjunction with -nobatch),
132       condor_q displays the following columns of information, with  one  line
133       of output per job:
134
135          ID, OWNER, SUBMITTED, CPU_TIME, ST, PRI, SIZE, CMD
136
137       If  the -hold option is specified, condor_q displays the following col‐
138       umns of information, with one line of output per job:
139
140          ID, OWNER, HELD_SINCE, HOLD_REASON
141
142       If the -totals option is specified, condor_q displays only one line  of
143       output  no  matter  how many jobs and batches of jobs are in the queue.
144       That line of output contains the total number of jobs, and  the  number
145       of  jobs  in  the completed, removed, idle, running, held and suspended
146       states.
147
148   Output data
149       The available output data are as follows:
150
151          ID     (Non-batch mode only) The cluster/process id of the  HTCondor
152                 job.
153
154          OWNER  The owner of the job or batch of jobs.
155
156          OWNER/NODENAME
157                 (-dag  only)  The  owner of a job or the DAG node name of the
158                 job.
159
160          BATCH_NAME
161                 (Batch mode only) The batch name of the job or batch of jobs.
162
163          SUBMITTED
164                 The month, day, hour, and minute the job was submitted to the
165                 queue.
166
167          DONE   (Batch  mode only) The number of job procs that are done, but
168                 still in the queue.
169
170          RUN    (Batch mode only) The number of job procs that are running.
171
172          IDLE   (Batch mode only) The number of job procs  that  are  in  the
173                 queue but idle.
174
175          HOLD   (Batch  mode  only)  The  number of job procs that are in the
176                 queue but held.
177
178          TOTAL  (Batch mode only) The total number of job procs in the queue,
179                 unless  the  batch  is a DAG, in which case this is the total
180                 number of clusters in the queue. Note: for  non-DAG  batches,
181                 the  TOTAL  column  contains  correct  values only in version
182                 8.5.7 and later.
183
184          JOB_IDS
185                 (Batch mode only) The range  of  job  IDs  belonging  to  the
186                 batch.
187
188          RUN_TIME
189                 (Non-batch  mode only) Wall-clock time accumulated by the job
190                 to date in days, hours, minutes, and seconds.
191
192          ST     (Non-batch mode only) Current status of the job, which varies
193                 somewhat  according to the job universe and the timing of up‐
194                 dates. H = on hold, R = running, I = idle (waiting for a  ma‐
195                 chine  to  execute  on), C = completed, X = removed, S = sus‐
196                 pended (execution of a running job temporarily  suspended  on
197                 execute  node),  < = transferring input (or queued to do so),
198                 and > = transferring output (or queued to do so).
199
200          PRI    (Non-batch mode only) User specified  priority  of  the  job,
201                 displayed as an integer, with higher numbers corresponding to
202                 better priority.
203
204          SIZE   (Non-batch mode only) The peak amount  of  memory  in  Mbytes
205                 consumed  by the job; note this value is only refreshed peri‐
206                 odically. The actual value reported is  taken  from  the  job
207                 ClassAd  attribute  MemoryUsage if this attribute is defined,
208                 and from job attribute ImageSize otherwise.
209
210          CMD    (Non-batch mode only) The name of  the  executable.  For  EC2
211                 jobs, this field is arbitrary.
212
213          HOST(S)
214                 (-run only) The host where the job is running.
215
216          STATUS (-grid  only) The state that HTCondor believes the job is in.
217                 Possible values are grid-type specific, but include:
218
219                     PENDING
220                            The job is waiting for resources to become  avail‐
221                            able in order to run.
222
223                     ACTIVE The  job  has received resources, and the applica‐
224                            tion is executing.
225
226                     FAILED The job terminated before completion because of an
227                            error,  user-triggered cancel, or system-triggered
228                            cancel.
229
230                     DONE   The job completed successfully.
231
232                     SUSPENDED
233                            The job has been suspended. Resources  which  were
234                            allocated  for this job may have been released due
235                            to a scheduler-specific reason.
236
237                     UNSUBMITTED
238                            The job has not been submitted  to  the  scheduler
239                            yet, pending the reception of the GLOBUS_GRAM_PRO‐
240                            TOCOL_JOB_SIGNAL_COMMIT_REQUEST  signal   from   a
241                            client.
242
243                     STAGE_IN
244                            The  job  manager is staging in files, in order to
245                            run the job.
246
247                     STAGE_OUT
248                            The job manager is staging out files generated  by
249                            the job.
250
251                     UNKNOWN
252                            Unknown
253
254          GRID->MANAGER
255                 (-grid  only)  A guess at what remote batch system is running
256                 the job. It is a guess, because HTCondor looks at the  Globus
257                 jobmanager  contact  string to attempt identification. If the
258                 value is fork, the job is running on the remote host  without
259                 a jobmanager. Values may also be condor, lsf, or pbs.
260
261          HOST   (-grid only) The host to which the job was submitted.
262
263          GRID_JOB_ID
264                 (-grid only) (More information needed here.)
265
266          INSTANCE ID
267                 (-grid:ec2 only) Usually EC2 instance ID; may be blank or the
268                 client token, depending on job progress.
269
270          GOODPUT
271                 (-goodput only) The percentage of RUN_TIME for this job which
272                 has been saved in a checkpoint. A low GOODPUT value indicates
273                 that the job is failing to checkpoint. If a job has  not  yet
274                 attempted a checkpoint, this column contains [?????].
275
276          CPU_UTIL
277                 (-goodput  only) The ratio of CPU_TIME to RUN_TIME for check‐
278                 pointed work. A low CPU_UTIL indicates that the  job  is  not
279                 running  efficiently,  perhaps because it is I/O bound or be‐
280                 cause the job requires more memory than available on the  re‐
281                 mote  workstations.   If  the job has not (yet) checkpointed,
282                 this column contains [??????].
283
284          Mb/s   (-goodput only) The network usage of this  job,  in  Megabits
285                 per  second  of run-time.  READ The total number of bytes the
286                 application has read from files and sockets.  WRITE The total
287                 number  of  bytes  the  application  has written to files and
288                 sockets.  SEEK The total number of seek operations the appli‐
289                 cation has performed on files.  XPUT The effective throughput
290                 (average bytes read and written per second) from the applica‐
291                 tion's point of view.  BUFSIZE The maximum number of bytes to
292                 be buffered per file.  BLOCKSIZE The desired block  size  for
293                 large  data  transfers.  These  fields are updated when a job
294                 produces a checkpoint or completes. If a job has not yet pro‐
295                 duced a checkpoint, this information is not available.
296
297          INPUT  (-io  only)  For standard universe, FileReadBytes; otherwise,
298                 BytesRecvd.
299
300          OUTPUT (-io only) For standard universe, FileWriteBytes;  otherwise,
301                 BytesSent.
302
303          RATE   (-io  only)  For  standard universe, FileReadBytes+FileWrite‐
304                 Bytes; otherwise, BytesRecvd+BytesSent.
305
306          MISC   (-io only) JobUniverse.
307
308          CPU_TIME
309                 (-cputime only) The remote CPU time accumulated by the job to
310                 date  (which has been stored in a checkpoint) in days, hours,
311                 minutes, and seconds. (If the job is currently running,  time
312                 accumulated  during  the current run is not shown. If the job
313                 has  not  produced  a  checkpoint,   this   column   contains
314                 0+00:00:00.)
315
316          HELD_SINCE
317                 (-hold only) Month, day, hour and minute at which the job was
318                 held.
319
320          HOLD_REASON
321                 (-hold only) The hold reason for the job.
322
323   Analyze
324       The -analyze or -better-analyze options can be used  to  determine  why
325       certain jobs are not running by performing an analysis on a per machine
326       basis for each machine in the pool. The reasons can vary  among  failed
327       constraints, insufficient priority, resource owner preferences and pre‐
328       vention of preemption by the PREEMPTION_REQUIREMENTS
329         expression. If the analyze option -verbose is  specified  along  with
330       the  -analyze  option, the reason for failure is displayed on a per ma‐
331       chine basis. -better-analyze differs from -analyze in that it  will  do
332       matchmaking  analysis on jobs even if they are currently running, or if
333       the reason they are not running is not due to matchmaking. -better-ana‐
334       lyze  also  produces more thorough analysis of complex Requirements and
335       shows the values of relevant job ClassAd attributes. When only a single
336       machine  is  being analyzed via -machine or -mconstraint, the values of
337       relevant attributes of the machine ClassAd are also displayed.
338

RESTRICTIONS

340       To restrict the display to jobs of interest, a list of zero or more re‐
341       striction options may be supplied. Each restriction may be one of:
342
343cluster.process,  which  matches  jobs  which belong to the specified
344         cluster and have the specified process number;
345
346cluster (without a process), which matches all jobs belonging to  the
347         specified cluster;
348
349owner, which matches all jobs owned by the specified owner;
350
351-constraint expression, which matches all jobs that satisfy the spec‐
352         ified ClassAd expression;
353
354-unmatchable expression, which matches all jobs that do not match any
355         slot that would be considered by -better-analyze ;
356
357-allusers,  which  overrides the default restriction of only matching
358         jobs submitted by the current user.
359
360       If cluster or cluster.process is specified, and the job  matching  that
361       restriction  is  a  condor_dagman job, information for all jobs of that
362       DAG is displayed in batch mode (in non-batch mode, only the condor_dag‐
363       man job itself is displayed).
364
365       If  no  owner restrictions are present, the job matches the restriction
366       list if it matches at least one restriction in the list. If  owner  re‐
367       strictions  are  present, the job matches the list if it matches one of
368       the owner restrictions and at least one non-owner restriction.
369

OPTIONS

371          -debug Causes debugging information to be sent to stderr,  based  on
372                 the value of the configuration variable TOOL_DEBUG.
373
374          -batch (output  option)  Show  a single line of progress information
375                 for a batch of jobs, where a batch is defined as follows:
376
377                 • An entire workflow (a DAG or hierarchy of nested DAGs)
378
379                 • All jobs in a single cluster
380
381                 • All jobs submitted by a single user that have the same exe‐
382                   cutable specified in their submit file
383
384                 • All  jobs  submitted  by  a  single user that have the same
385                   batch name specified in their submit file or  on  the  con‐
386                   dor_submit or condor_submit_dag command line.
387
388                 Also change the output columns as noted above.
389
390                 Note that, as of version 8.5.6, -batch is the default, unless
391                 the CONDOR_Q_DASH_BATCH_IS_DEFAULT configuration variable  is
392                 set to False.
393
394          -nobatch
395                 (output option) Show a line for each job (turn off the -batch
396                 option).
397
398          -global
399                 (general option) Queries all job queues in the pool.
400
401          -submitter submitter
402                 (general option) List jobs of a specific submitter in the en‐
403                 tire pool, not just for a single condor_schedd.
404
405          -name name
406                 (general  option)  Query only the job queue of the named con‐
407                 dor_schedd daemon.
408
409          -pool centralmanagerhostname[:portnumber]
410                 (general option) Use the centralmanagerhostname as  the  cen‐
411                 tral  manager to locate condor_schedd daemons. The default is
412                 the COLLECTOR_HOST, as specified in the configuration.
413
414          -jobads file
415                 (general option) Display jobs from a list of ClassAds from  a
416                 file,  instead  of  the  real ClassAds from the condor_schedd
417                 daemon. This is most useful for debugging purposes. The Clas‐
418                 sAds  appear  as  if  condor_q  -long is used with the header
419                 stripped out.
420
421          -userlog file
422                 (general option) Display jobs, with  job  information  coming
423                 from  a job event log, instead of from the real ClassAds from
424                 the condor_schedd daemon. This is most useful  for  automated
425                 testing  of  the  status of jobs known to be in the given job
426                 event log, because it reduces the load on the  condor_schedd.
427                 A  job event log does not contain all of the job information,
428                 so some fields in the  normal  output  of  condor_q  will  be
429                 blank.
430
431          -autocluster
432                 (output  option) Output condor_schedd daemon auto cluster in‐
433                 formation. For each auto cluster, output the unique ID of the
434                 auto cluster along with the number of jobs in that auto clus‐
435                 ter. This option is intended to be  used  together  with  the
436                 -long  option  to output the ClassAds representing auto clus‐
437                 ters. The ClassAds can then be used to identify  or  classify
438                 the  demand for sets of machine resources, which will be use‐
439                 ful in the on-demand creation of execute  nodes  for  glidein
440                 services.
441
442          -cputime
443                 (output   option)   Instead  of  wall-clock  allocation  time
444                 (RUN_TIME), display remote CPU time accumulated by the job to
445                 date in days, hours, minutes, and seconds. If the job is cur‐
446                 rently running, time accumulated during the  current  run  is
447                 not shown. Note that this option has no effect unless used in
448                 conjunction with -nobatch.
449
450          -currentrun
451                 (output option) Normally, RUN_TIME contains all the time  ac‐
452                 cumulated  during  the current run plus all previous runs. If
453                 this option is specified, RUN_TIME only displays the time ac‐
454                 cumulated so far on this current run.
455
456          -dag   (output  option) Display DAG node jobs under their DAGMan in‐
457                 stance.  Child nodes are listed using indentation to show the
458                 structure of the DAG. Note that this option has no effect un‐
459                 less used in conjunction with -nobatch.
460
461          -expert
462                 (output option) Display shorter error messages.
463
464          -grid  (output option) Get information only about jobs submitted  to
465                 grid resources.
466
467          -grid:ec2
468                 (output  option) Get information only about jobs submitted to
469                 grid resources and display it in a format  better-suited  for
470                 EC2 than the default.
471
472          -goodput
473                 (output option) Display job goodput statistics.
474
475          -help [Universe | State]
476                 (output  option) Print usage info, and, optionally, addition‐
477                 ally print job universes or job states.
478
479          -hold  (output option) Get information about jobs in the hold state.
480                 Also displays the time the job was placed into the hold state
481                 and the reason why the job was placed in the hold state.
482
483          -limit Number
484                 (output option) Limit the number of items output to Number.
485
486          -io    (output option) Display job input/output summaries.
487
488          -long  (output option) Display entire job ClassAds  in  long  format
489                 (one attribute per line).
490
491          -run   (output option) Get information about running jobs. Note that
492                 this option has no effect unless  used  in  conjunction  with
493                 -nobatch.
494
495          -stream-results
496                 (output  option) Display results as jobs are fetched from the
497                 job queue rather than storing results  in  memory  until  all
498                 jobs  have  been  fetched. This can reduce memory consumption
499                 when fetching large numbers  of  jobs,  but  if  condor_q  is
500                 paused while displaying results, this could result in a time‐
501                 out in communication with condor_schedd.
502
503          -totals
504                 (output option) Display only the totals.
505
506          -version
507                 (output option) Print the HTCondor version and exit.
508
509          -wide  (output option) If this option is specified, and the  command
510                 portion of the output would cause the output to extend beyond
511                 80 columns, display beyond the 80 columns.
512
513          -xml   (output option) Display entire job ClassAds  in  XML  format.
514                 The  XML format is fully defined in the reference manual, ob‐
515                 tained  from  the  ClassAds  web  page,  with   a   link   at
516                 http://htcondor.org/classad/classad.html.
517
518          -json  (output option) Display entire job ClassAds in JSON format.
519
520          -attributes Attr1[,Attr2 ...]
521                 (output  option) Explicitly list the attributes, by name in a
522                 comma separated list, which should be  displayed  when  using
523                 the  -xml, -json or -long options. Limiting the number of at‐
524                 tributes increases the efficiency of the query.
525
526          -format fmt attr
527                 (output option) Display attribute or expression attr in  for‐
528                 mat  fmt.  To  display the attribute or expression the format
529                 must contain a single printf(3)-style  conversion  specifier.
530                 Attributes  must  be  from  the  job ClassAd. Expressions are
531                 ClassAd expressions and may refer to attributes  in  the  job
532                 ClassAd.  If  the attribute is not present in a given ClassAd
533                 and cannot be parsed as an expression, then the format option
534                 will  be  silently skipped. %r prints the unevaluated, or raw
535                 values. The conversion specifier must match the type  of  the
536                 attribute  or  expression. %s is suitable for strings such as
537                 Owner, %d for integers such as ClusterId, and %f for floating
538                 point numbers such as RemoteWallClockTime.  %v identifies the
539                 type of the attribute, and then prints the value in an appro‐
540                 priate  format.  %V identifies the type of the attribute, and
541                 then prints the value in an appropriate format  as  it  would
542                 appear  in the -long format. As an example, strings used with
543                 %V will have quote marks. An incorrect format will result  in
544                 undefined behavior. Do not use more than one conversion spec‐
545                 ifier in a given format. More than one  conversion  specifier
546                 will  result  in  undefined  behavior. To output multiple at‐
547                 tributes repeat the -format option once for each desired  at‐
548                 tribute.  Like printf(3) style formats, one may include other
549                 text that will be reproduced directly. A format  without  any
550                 conversion  specifiers  may be specified, but an attribute is
551                 still required. Include a backslash followed  by  an  'n'  to
552                 specify a line break.
553
554          -autoformat[:jlhVr,tng]  attr1  [attr2 ...] or -af[:jlhVr,tng] attr1
555          [attr2 ...]
556                 (output option) Display attribute(s) or expression(s) format‐
557                 ted  in  a default way according to attribute types. This op‐
558                 tion takes an arbitrary number of attribute  names  as  argu‐
559                 ments, and prints out their values, with a space between each
560                 value and a newline character after the  last  value.  It  is
561                 like  the  -format option without format strings. This output
562                 option does not work in conjunction with any of  the  options
563                 -run, -currentrun, -hold, -grid, -goodput, or -io.
564
565                 It is assumed that no attribute names begin with a dash char‐
566                 acter, so that the next word that begins  with  dash  is  the
567                 start  of  the next option. The autoformat option may be fol‐
568                 lowed by a colon character and formatting qualifiers to devi‐
569                 ate the output formatting from the default:
570
571                 j print the job ID as the first field,
572
573                 l label each field,
574
575                 h print column headings before the first line of output,
576
577                 V  use  %V  rather  than %v for formatting (string values are
578                 quoted),
579
580                 r print "raw", or unevaluated values,
581
582                 , add a comma character after each field,
583
584                 t add a tab character before each field instead  of  the  de‐
585                 fault space character,
586
587                 n add a newline character after each field,
588
589                 g add a newline character between ClassAds, and suppress spa‐
590                 ces before each field.
591
592                 Use -af:h to get tabular values with headings.
593
594                 Use -af:lrng to get -long equivalent format.
595
596                 The newline and comma characters may not  be  used  together.
597                 The l and h characters may not be used together.
598
599          -analyze[:<qual>]
600                 (analyze  option)  Perform  a matchmaking analysis on why the
601                 requested jobs are not running. First a simple  analysis  de‐
602                 termines  if  the  job  is  not running due to not being in a
603                 runnable state. If the job is in a runnable state, then  this
604                 option  is  equivalent  to -better-analyze. <qual> is a comma
605                 separated list containing one or more of
606
607                 priority to consider user priority during the analysis
608
609                 summary to show a one line summary for each job or machine
610                 reverse to analyze machines, rather than jobs
611
612
613          -better-analyze[:<qual>]
614                 (analyze option) Perform a more detailed matchmaking analysis
615                 to  determine how many resources are available to run the re‐
616                 quested jobs. This option is never meaningful  for  Scheduler
617                 universe  jobs and only meaningful for grid universe jobs do‐
618                 ing matchmaking. When this option is used in conjunction with
619                 the -unmatchable option, The output will be a list of job ids
620                 that don't match any of the  available  slots.  <qual>  is  a
621                 comma separated list containing one or more of
622
623                 priority to consider user priority during the analysis
624
625                 summary to show a one line summary for each job or machine
626                 reverse to analyze machines, rather than jobs
627
628
629          -machine name
630                 (analyze  option)  When  doing  matchmaking analysis, analyze
631                 only machine ClassAds that have slot or  machine  names  that
632                 match the given name.
633
634          -mconstraint expression
635                 (analyze  option) When doing matchmaking analysis, match only
636                 machine ClassAds which  match  the  ClassAd  expression  con‐
637                 straint.
638
639          -slotads file
640                 (analyze option) When doing matchmaking analysis, use the ma‐
641                 chine ClassAds from the file instead of  the  ones  from  the
642                 condor_collector  daemon.  This  is most useful for debugging
643                 purposes. The ClassAds appear as if  condor_status  -long  is
644                 used.
645
646          -userprios file
647                 (analyze  option) When doing matchmaking analysis with prior‐
648                 ity, read user priorities from the file rather than the  ones
649                 from  the  condor_negotiator  daemon. This is most useful for
650                 debugging purposes or to  speed  up  analysis  in  situations
651                 where the condor_negotiator daemon is slow to respond to con‐
652                 dor_userprio requests. The file should be in the format  pro‐
653                 duced by condor_userprio -long.
654
655          -nouserprios
656                 (analyze  option)  Do  not  consider user priority during the
657                 analysis.
658
659          -reverse-analyze
660                 (analyze option) Analyze machine requirements against jobs.
661
662          -verbose
663                 (analyze option) When doing analysis, show progress  and  in‐
664                 clude the names of specific machines in the output.
665

GENERAL REMARKS

667       The default output from condor_q is formatted to be human readable, not
668       script readable. In an effort to make the output fit within 80  charac‐
669       ters, values in some fields might be truncated. Furthermore, the HTCon‐
670       dor Project can (and does) change the formatting of this default output
671       as  we  see fit. Therefore, any script that is attempting to parse data
672       from condor_q is strongly encouraged to use  the  -format  option  (de‐
673       scribed above, examples given below).
674
675       Although  -analyze  provides  a very good first approximation, the ana‐
676       lyzer cannot diagnose all possible situations, because the analysis  is
677       based on instantaneous and local information. Therefore, there are some
678       situations such as when  several  submitters  are  contending  for  re‐
679       sources, or if the pool is rapidly changing state which cannot be accu‐
680       rately diagnosed.
681
682       Options -goodput, -cputime, and -io are most useful for  standard  uni‐
683       verse  jobs,  since  they rely on values computed when a job produces a
684       checkpoint.
685
686       It is possible to to hold jobs that are in the X state. To  avoid  this
687       it  is  best to construct a -constraint expression that option contains
688       JobStatus != 3 if the user wishes to avoid this condition.
689

EXAMPLES

691       The -format option provides a way to specify both  the  job  attributes
692       and  formatting  of those attributes. There must be only one conversion
693       specification per -format option. As an  example,  to  list  only  Jane
694       Doe's jobs in the queue, choosing to print and format only the owner of
695       the job, the command line arguments for the job, and the process ID  of
696       the job:
697
698          $ condor_q -submitter jdoe -format "%s" Owner -format " %s " Args -format " ProcId = %d\n" ProcId
699          jdoe 16386 2800 ProcId = 0
700          jdoe 16386 3000 ProcId = 1
701          jdoe 16386 3200 ProcId = 2
702          jdoe 16386 3400 ProcId = 3
703          jdoe 16386 3600 ProcId = 4
704          jdoe 16386 4200 ProcId = 7
705
706       To  display only the JobID's of Jane Doe's jobs you can use the follow‐
707       ing.
708
709          $ condor_q -submitter jdoe -format "%d." ClusterId -format "%d\n" ProcId
710          27.0
711          27.1
712          27.2
713          27.3
714          27.4
715          27.7
716
717       An example that shows the analysis in summary format:
718
719          $ condor_q -analyze:summary
720
721          -- Submitter: submit-1.chtc.wisc.edu : <192.168.100.43:9618?sock=11794_95bb_3> :
722           submit-1.chtc.wisc.edu
723          Analyzing matches for 5979 slots
724                      Autocluster  Matches    Machine     Running  Serving
725           JobId     Members/Idle  Reqmnts  Rejects Job  Users Job Other User Avail Owner
726          ---------- ------------ -------- ------------ ---------- ---------- ----- -----
727          25764522.0  7/0             5910        820   7/10       5046        34   smith
728          25764682.0  9/0             2172        603   9/9        1531        29   smith
729          25765082.0  18/0            2172        603   18/9       1531        29   smith
730          25765900.0  1/0             2172        603   1/9        1531        29   smith
731
732       An example that shows summary information by machine:
733
734          $ condor_q -ana:sum,rev
735
736          -- Submitter: s-1.chtc.wisc.edu : <192.168.100.43:9618?sock=11794_95bb_3> : s-1.chtc.wisc.edu
737          Analyzing matches for 2885 jobs
738                                          Slot  Slot's Req    Job's Req     Both
739          Name                            Type  Matches Job  Matches Slot    Match %
740          ------------------------        ---- ------------  ------------ ----------
741          slot1@INFO.wisc.edu             Stat         2729  0                  0.00
742          slot2@INFO.wisc.edu             Stat         2729  0                  0.00
743          slot1@aci-001.chtc.wisc.edu     Part            0  2793               0.00
744          slot1_1@a-001.chtc.wisc.edu     Dyn          2644  2792              91.37
745          slot1_2@a-001.chtc.wisc.edu     Dyn          2623  2601              85.10
746          slot1_3@a-001.chtc.wisc.edu     Dyn          2644  2632              85.82
747          slot1_4@a-001.chtc.wisc.edu     Dyn          2644  2792              91.37
748          slot1@a-002.chtc.wisc.edu       Part            0  2633               0.00
749          slot1_10@a-002.chtc.wisc.edu    Den          2623  2601              85.10
750
751       An example with two independent DAGs in the queue:
752
753          $ condor_q
754
755          -- Schedd: wenger@manta.cs.wisc.edu : <128.105.14.228:35169?...
756          OWNER  BATCH_NAME    SUBMITTED   DONE   RUN    IDLE  TOTAL JOB_IDS
757          wenger DAG: 3696    2/12 11:55      _     10      _     10 3698.0 ... 3707.0
758          wenger DAG: 3697    2/12 11:55      1      1      1     10 3709.0 ... 3710.0
759
760          14 jobs; 0 completed, 0 removed, 1 idle, 13 running, 0 held, 0 suspended
761
762       Note that the "13 running" in the last line is two more than the  total
763       of  the  RUN  column, because the two condor_dagman jobs themselves are
764       counted in the last line but not the RUN column.
765
766       Also note that the "completed" value in the last line does  not  corre‐
767       spond to the total of the DONE column, because the "completed" value in
768       the last line only counts jobs that are  completed  but  still  in  the
769       queue,  whereas  the  DONE column counts jobs that are no longer in the
770       queue.
771
772       Here's an example with a held job, illustrating  the  addition  of  the
773       HOLD column to the output:
774
775          $ condor_q
776
777          -- Schedd: wenger@manta.cs.wisc.edu : <128.105.14.228:9619?...
778          OWNER  BATCH_NAME        SUBMITTED   DONE   RUN    IDLE   HOLD  TOTAL JOB_IDS
779          wenger CMD: /bin/slee   9/13 16:25      _      3      _      1      4 599.0 ...
780
781          4 jobs; 0 completed, 0 removed, 0 idle, 3 running, 1 held, 0 suspended
782
783       Here  are  some examples with a nested-DAG workflow in the queue, which
784       is one of the most  complicated  cases.  The  workflow  consists  of  a
785       top-level  DAG with nodes NodeA and NodeB, each with two two-proc clus‐
786       ters; and a sub-DAG SubZ with nodes NodeSA and NodeSB,  each  with  two
787       two-proc clusters.
788
789       First of all, non-batch mode with all of the node jobs in the queue:
790
791          $ condor_q -nobatch
792
793          -- Schedd: wenger@manta.cs.wisc.edu : <128.105.14.228:9619?...
794           ID      OWNER            SUBMITTED     RUN_TIME ST PRI SIZE CMD
795           591.0   wenger          9/13 16:05   0+00:00:13 R  0    2.4 condor_dagman -p 0
796           592.0   wenger          9/13 16:05   0+00:00:07 R  0    0.0 sleep 60
797           592.1   wenger          9/13 16:05   0+00:00:07 R  0    0.0 sleep 300
798           593.0   wenger          9/13 16:05   0+00:00:07 R  0    0.0 sleep 60
799           593.1   wenger          9/13 16:05   0+00:00:07 R  0    0.0 sleep 300
800           594.0   wenger          9/13 16:05   0+00:00:07 R  0    2.4 condor_dagman -p 0
801           595.0   wenger          9/13 16:05   0+00:00:01 R  0    0.0 sleep 60
802           595.1   wenger          9/13 16:05   0+00:00:01 R  0    0.0 sleep 300
803           596.0   wenger          9/13 16:05   0+00:00:01 R  0    0.0 sleep 60
804           596.1   wenger          9/13 16:05   0+00:00:01 R  0    0.0 sleep 300
805
806          10 jobs; 0 completed, 0 removed, 0 idle, 10 running, 0 held, 0 suspended
807
808       Now  non-batch  mode  with  the  -dag  option  (unfortunately, condor_q
809       doesn't do a good job of grouping procs in the same cluster together):
810
811          $ condor_q -nobatch -dag
812
813          -- Schedd: wenger@manta.cs.wisc.edu : <128.105.14.228:9619?...
814           ID      OWNER/NODENAME      SUBMITTED     RUN_TIME ST PRI SIZE CMD
815           591.0   wenger             9/13 16:05   0+00:00:27 R  0    2.4 condor_dagman -
816           592.0    |-NodeA           9/13 16:05   0+00:00:21 R  0    0.0 sleep 60
817           593.0    |-NodeB           9/13 16:05   0+00:00:21 R  0    0.0 sleep 60
818           594.0    |-SubZ            9/13 16:05   0+00:00:21 R  0    2.4 condor_dagman -
819           595.0     |-NodeSA         9/13 16:05   0+00:00:15 R  0    0.0 sleep 60
820           596.0     |-NodeSB         9/13 16:05   0+00:00:15 R  0    0.0 sleep 60
821           592.1    |-NodeA           9/13 16:05   0+00:00:21 R  0    0.0 sleep 300
822           593.1    |-NodeB           9/13 16:05   0+00:00:21 R  0    0.0 sleep 300
823           595.1     |-NodeSA         9/13 16:05   0+00:00:15 R  0    0.0 sleep 300
824           596.1     |-NodeSB         9/13 16:05   0+00:00:15 R  0    0.0 sleep 300
825
826          10 jobs; 0 completed, 0 removed, 0 idle, 10 running, 0 held, 0 suspended
827
828       Now, finally, the non-batch (default) mode:
829
830          $ condor_q
831
832          -- Schedd: wenger@manta.cs.wisc.edu : <128.105.14.228:9619?...
833          OWNER  BATCH_NAME     SUBMITTED   DONE   RUN    IDLE  TOTAL JOB_IDS
834          wenger ex1.dag+591   9/13 16:05      _      8      _      5 592.0 ... 596.1
835
836          10 jobs; 0 completed, 0 removed, 0 idle, 10 running, 0 held, 0 suspended
837
838       There are several things about this output that may be slightly confus‐
839       ing:
840
841       • The  TOTAL  column  is less than the RUN column. This is because, for
842         DAG node jobs, their contribution to the TOTAL column is  the  number
843         of  clusters,  not the number of procs (but their contribution to the
844         RUN column is the number of procs). So the four DAG nodes  (8  procs)
845         contribute  4,  and  the  sub-DAG contributes 1, to the TOTAL column.
846         (But, somewhat confusingly, the sub-DAG job is not counted in the RUN
847         column.)
848
849       • The  sum  of  the  RUN  and IDLE columns (8) is less than the 10 jobs
850         listed in the  totals  line  at  the  bottom.  This  is  because  the
851         top-level DAG and sub-DAG jobs are not counted in the RUN column, but
852         they are counted in the totals line.
853
854       Now here is non-batch mode after proc 0 of each node job has finished:
855
856          $ condor_q -nobatch
857
858          -- Schedd: wenger@manta.cs.wisc.edu : <128.105.14.228:9619?...
859           ID      OWNER            SUBMITTED     RUN_TIME ST PRI SIZE CMD
860           591.0   wenger          9/13 16:05   0+00:01:19 R  0    2.4 condor_dagman -p 0
861           592.1   wenger          9/13 16:05   0+00:01:13 R  0    0.0 sleep 300
862           593.1   wenger          9/13 16:05   0+00:01:13 R  0    0.0 sleep 300
863           594.0   wenger          9/13 16:05   0+00:01:13 R  0    2.4 condor_dagman -p 0
864           595.1   wenger          9/13 16:05   0+00:01:07 R  0    0.0 sleep 300
865           596.1   wenger          9/13 16:05   0+00:01:07 R  0    0.0 sleep 300
866
867          6 jobs; 0 completed, 0 removed, 0 idle, 6 running, 0 held, 0 suspended
868
869       The same state also with the -dag option:
870
871          $ condor_q -nobatch -dag
872
873          -- Schedd: wenger@manta.cs.wisc.edu : <128.105.14.228:9619?...
874           ID      OWNER/NODENAME      SUBMITTED     RUN_TIME ST PRI SIZE CMD
875           591.0   wenger             9/13 16:05   0+00:01:30 R  0    2.4 condor_dagman -
876           592.1    |-NodeA           9/13 16:05   0+00:01:24 R  0    0.0 sleep 300
877           593.1    |-NodeB           9/13 16:05   0+00:01:24 R  0    0.0 sleep 300
878           594.0    |-SubZ            9/13 16:05   0+00:01:24 R  0    2.4 condor_dagman -
879           595.1     |-NodeSA         9/13 16:05   0+00:01:18 R  0    0.0 sleep 300
880           596.1     |-NodeSB         9/13 16:05   0+00:01:18 R  0    0.0 sleep 300
881
882          6 jobs; 0 completed, 0 removed, 0 idle, 6 running, 0 held, 0 suspended
883
884       And, finally, that state in batch (default) mode:
885
886          $ condor_q
887
888          -- Schedd: wenger@manta.cs.wisc.edu : <128.105.14.228:9619?...
889          OWNER  BATCH_NAME     SUBMITTED   DONE   RUN    IDLE  TOTAL JOB_IDS
890          wenger ex1.dag+591   9/13 16:05      _      4      _      5 592.1 ... 596.1
891
892          6 jobs; 0 completed, 0 removed, 0 idle, 6 running, 0 held, 0 suspended
893

EXIT STATUS

895       condor_q will exit with a status value of 0 (zero) upon success, and it
896       will exit with the value 1 (one) upon failure.
897

AUTHOR

899       HTCondor Team
900
902       1990-2022,  Center for High Throughput Computing, Computer Sciences De‐
903       partment, University of Wisconsin-Madison, Madison,  WI,  US.  Licensed
904       under the Apache License, Version 2.0.
905
906
907
908
9098.8                              Jun 13, 2022                      CONDOR_Q(1)
Impressum