1PCS(8) System Administration Utilities PCS(8)
2
3
4
6 pcs - pacemaker/corosync configuration system
7
9 pcs [-f file] [-h] [commands]...
10
12 Control and configure pacemaker and corosync.
13
15 -h, --help
16 Display usage and exit.
17
18 -f file
19 Perform actions on file instead of active CIB.
20 Commands supporting the option use the initial state of the
21 specified file as their input and then overwrite the file with
22 the state reflecting the requested operation(s).
23 A few commands only use the specified file in read-only mode
24 since their effect is not a CIB modification.
25
26 --debug
27 Print all network traffic and external commands run.
28
29 --version
30 Print pcs version information. List pcs capabilities if --full
31 is specified.
32
33 --request-timeout=<timeout>
34 Timeout for each outgoing request to another node in seconds.
35 Default is 60s.
36
37 Commands:
38 cluster
39 Configure cluster options and nodes.
40
41 resource
42 Manage cluster resources.
43
44 stonith
45 Manage fence devices.
46
47 constraint
48 Manage resource constraints.
49
50 property
51 Manage pacemaker properties.
52
53 acl
54 Manage pacemaker access control lists.
55
56 qdevice
57 Manage quorum device provider on the local host.
58
59 quorum
60 Manage cluster quorum settings.
61
62 booth
63 Manage booth (cluster ticket manager).
64
65 status
66 View cluster status.
67
68 config
69 View and manage cluster configuration.
70
71 pcsd
72 Manage pcs daemon.
73
74 host
75 Manage hosts known to pcs/pcsd.
76
77 node
78 Manage cluster nodes.
79
80 alert
81 Manage pacemaker alerts.
82
83 client
84 Manage pcsd client configuration.
85
86 dr
87 Manage disaster recovery configuration.
88
89 tag
90 Manage pacemaker tags.
91
92 resource
93 [status [<resource id | tag id>] [node=<node>] [--hide-inactive]]
94 Show status of all currently configured resources. If --hide-in‐
95 active is specified, only show active resources. If a resource
96 or tag id is specified, only show status of the specified re‐
97 source or resources in the specified tag. If node is specified,
98 only show status of resources configured for the specified node.
99
100 config [--output-format text|cmd|json] [<resource id>]...
101 Show options of all currently configured resources or if re‐
102 source ids are specified show the options for the specified re‐
103 source ids. There are 3 formats of output available: 'cmd',
104 'json' and 'text', default is 'text'. Format 'text' is a human
105 friendly output. Format 'cmd' prints pcs commands which can be
106 used to recreate the same configuration. Format 'json' is a ma‐
107 chine oriented output of the configuration.
108
109 list [filter] [--nodesc]
110 Show list of all available resource agents (if filter is pro‐
111 vided then only resource agents matching the filter will be
112 shown). If --nodesc is used then descriptions of resource agents
113 are not printed.
114
115 describe [<standard>:[<provider>:]]<type> [--full]
116 Show options for the specified resource. If --full is specified,
117 all options including advanced and deprecated ones are shown.
118
119 create <resource id> [<standard>:[<provider>:]]<type> [resource op‐
120 tions] [op <operation action> <operation options> [<operation action>
121 <operation options>]...] [meta <meta options>...] [clone [<clone id>]
122 [<clone options>] | promotable [<clone id>] [<promotable options>] |
123 --group <group id> [--before <resource id> | --after <resource id>] |
124 bundle <bundle id>] [--disabled] [--no-default-ops] [--wait[=n]]
125 Create specified resource. If clone is used a clone resource is
126 created. If promotable is used a promotable clone resource is
127 created. If --group is specified the resource is added to the
128 group named. You can use --before or --after to specify the po‐
129 sition of the added resource relatively to some resource already
130 existing in the group. If bundle is specified, resource will be
131 created inside of the specified bundle. If --disabled is speci‐
132 fied the resource is not started automatically. If --no-de‐
133 fault-ops is specified, only monitor operations are created for
134 the resource and all other operations use default settings. If
135 --wait is specified, pcs will wait up to 'n' seconds for the re‐
136 source to start and then return 0 if the resource is started, or
137 1 if the resource has not yet started. If 'n' is not specified
138 it defaults to 60 minutes.
139
140 Example: Create a new resource called 'VirtualIP' with IP ad‐
141 dress 192.168.0.99, netmask of 32, monitored everything 30 sec‐
142 onds, on eth2: pcs resource create VirtualIP ocf:heart‐
143 beat:IPaddr2 ip=192.168.0.99 cidr_netmask=32 nic=eth2 op monitor
144 interval=30s
145
146 delete <resource id|group id|bundle id|clone id>
147 Deletes the resource, group, bundle or clone (and all resources
148 within the group/bundle/clone).
149
150 remove <resource id|group id|bundle id|clone id>
151 Deletes the resource, group, bundle or clone (and all resources
152 within the group/bundle/clone).
153
154 enable <resource id | tag id>... [--wait[=n]]
155 Allow the cluster to start the resources. Depending on the rest
156 of the configuration (constraints, options, failures, etc), the
157 resources may remain stopped. If --wait is specified, pcs will
158 wait up to 'n' seconds for the resources to start and then re‐
159 turn 0 if the resources are started, or 1 if the resources have
160 not yet started. If 'n' is not specified it defaults to 60 min‐
161 utes.
162
163 disable <resource id | tag id>... [--safe [--brief] [--no-strict]]
164 [--simulate [--brief]] [--wait[=n]]
165 Attempt to stop the resources if they are running and forbid the
166 cluster from starting them again. Depending on the rest of the
167 configuration (constraints, options, failures, etc), the re‐
168 sources may remain started.
169 If --safe is specified, no changes to the cluster configuration
170 will be made if other than specified resources would be affected
171 in any way. If --brief is also specified, only errors are
172 printed.
173 If --no-strict is specified, no changes to the cluster configu‐
174 ration will be made if other than specified resources would get
175 stopped or demoted. Moving resources between nodes is allowed.
176 If --simulate is specified, no changes to the cluster configura‐
177 tion will be made and the effect of the changes will be printed
178 instead. If --brief is also specified, only a list of affected
179 resources will be printed.
180 If --wait is specified, pcs will wait up to 'n' seconds for the
181 resources to stop and then return 0 if the resources are stopped
182 or 1 if the resources have not stopped. If 'n' is not specified
183 it defaults to 60 minutes.
184
185 safe-disable <resource id | tag id>... [--brief] [--no-strict] [--simu‐
186 late [--brief]] [--wait[=n]] [--force]
187 Attempt to stop the resources if they are running and forbid the
188 cluster from starting them again. Depending on the rest of the
189 configuration (constraints, options, failures, etc), the re‐
190 sources may remain started. No changes to the cluster configura‐
191 tion will be made if other than specified resources would be af‐
192 fected in any way.
193 If --brief is specified, only errors are printed.
194 If --no-strict is specified, no changes to the cluster configu‐
195 ration will be made if other than specified resources would get
196 stopped or demoted. Moving resources between nodes is allowed.
197 If --simulate is specified, no changes to the cluster configura‐
198 tion will be made and the effect of the changes will be printed
199 instead. If --brief is also specified, only a list of affected
200 resources will be printed.
201 If --wait is specified, pcs will wait up to 'n' seconds for the
202 resources to stop and then return 0 if the resources are stopped
203 or 1 if the resources have not stopped. If 'n' is not specified
204 it defaults to 60 minutes.
205 If --force is specified, checks for safe disable will be
206 skipped.
207
208 restart <resource id> [node] [--wait=n]
209 Restart the resource specified. If a node is specified and if
210 the resource is a clone or bundle it will be restarted only on
211 the node specified. If --wait is specified, then we will wait up
212 to 'n' seconds for the resource to be restarted and return 0 if
213 the restart was successful or 1 if it was not.
214
215 debug-start <resource id> [--full]
216 This command will force the specified resource to start on this
217 node ignoring the cluster recommendations and print the output
218 from starting the resource. Using --full will give more de‐
219 tailed output. This is mainly used for debugging resources that
220 fail to start.
221
222 debug-stop <resource id> [--full]
223 This command will force the specified resource to stop on this
224 node ignoring the cluster recommendations and print the output
225 from stopping the resource. Using --full will give more de‐
226 tailed output. This is mainly used for debugging resources that
227 fail to stop.
228
229 debug-promote <resource id> [--full]
230 This command will force the specified resource to be promoted on
231 this node ignoring the cluster recommendations and print the
232 output from promoting the resource. Using --full will give more
233 detailed output. This is mainly used for debugging resources
234 that fail to promote.
235
236 debug-demote <resource id> [--full]
237 This command will force the specified resource to be demoted on
238 this node ignoring the cluster recommendations and print the
239 output from demoting the resource. Using --full will give more
240 detailed output. This is mainly used for debugging resources
241 that fail to demote.
242
243 debug-monitor <resource id> [--full]
244 This command will force the specified resource to be monitored
245 on this node ignoring the cluster recommendations and print the
246 output from monitoring the resource. Using --full will give
247 more detailed output. This is mainly used for debugging re‐
248 sources that fail to be monitored.
249
250 move <resource id> [destination node] [--promoted] [--strict]
251 [--wait[=n]]
252 Move the resource off the node it is currently running on. This
253 is achieved by creating a -INFINITY location constraint to ban
254 the node. If destination node is specified the resource will be
255 moved to that node by creating an INFINITY location constraint
256 to prefer the destination node. The constraint needed for moving
257 the resource will be automatically removed once the resource is
258 running on it's new location. The command will fail in case it
259 is not possible to verify that the resource will not be moved
260 back after deleting the constraint.
261
262 If --strict is specified, the command will also fail if other
263 resources would be affected.
264
265 If --promoted is used the scope of the command is limited to the
266 Promoted role and promotable clone id must be used (instead of
267 the resource id).
268
269 If --wait is specified, pcs will wait up to 'n' seconds for the
270 resource to move and then return 0 on success or 1 on error. If
271 'n' is not specified it defaults to 60 minutes.
272
273 NOTE: This command has been changed in pcs-0.11. It is equiva‐
274 lent to command 'resource move <resource id> --autodelete' from
275 pcs-0.10.9. Legacy functionality of the 'resource move' command
276 is still available as 'resource move-with-constraint <resource
277 id>'.
278
279 If you want the resource to preferably avoid running on some
280 nodes but be able to failover to them use 'pcs constraint loca‐
281 tion avoids'.
282
283 move-with-constraint <resource id> [destination node] [lifetime=<life‐
284 time>] [--promoted] [--wait[=n]]
285 Move the resource off the node it is currently running on by
286 creating a -INFINITY location constraint to ban the node. If
287 destination node is specified the resource will be moved to that
288 node by creating an INFINITY location constraint to prefer the
289 destination node.
290
291 If lifetime is specified then the constraint will expire after
292 that time, otherwise it defaults to infinity and the constraint
293 can be cleared manually with 'pcs resource clear' or 'pcs con‐
294 straint delete'. Lifetime is expected to be specified as ISO
295 8601 duration (see https://en.wikipedia.org/wiki/ISO_8601#Dura‐
296 tions).
297
298 If --promoted is used the scope of the command is limited to the
299 Promoted role and promotable clone id must be used (instead of
300 the resource id).
301
302 If --wait is specified, pcs will wait up to 'n' seconds for the
303 resource to move and then return 0 on success or 1 on error. If
304 'n' is not specified it defaults to 60 minutes.
305
306 If you want the resource to preferably avoid running on some
307 nodes but be able to failover to them use 'pcs constraint loca‐
308 tion avoids'.
309
310 ban <resource id> [node] [--promoted] [lifetime=<lifetime>]
311 [--wait[=n]]
312 Prevent the resource id specified from running on the node (or
313 on the current node it is running on if no node is specified) by
314 creating a -INFINITY location constraint.
315
316 If --promoted is used the scope of the command is limited to the
317 Promoted role and promotable clone id must be used (instead of
318 the resource id).
319
320 If lifetime is specified then the constraint will expire after
321 that time, otherwise it defaults to infinity and the constraint
322 can be cleared manually with 'pcs resource clear' or 'pcs con‐
323 straint delete'. Lifetime is expected to be specified as ISO
324 8601 duration (see https://en.wikipedia.org/wiki/ISO_8601#Dura‐
325 tions).
326
327 If --wait is specified, pcs will wait up to 'n' seconds for the
328 resource to move and then return 0 on success or 1 on error. If
329 'n' is not specified it defaults to 60 minutes.
330
331 If you want the resource to preferably avoid running on some
332 nodes but be able to failover to them use 'pcs constraint loca‐
333 tion avoids'.
334
335 clear <resource id> [node] [--promoted] [--expired] [--wait[=n]]
336 Remove constraints created by move and/or ban on the specified
337 resource (and node if specified).
338
339 If --promoted is used the scope of the command is limited to the
340 Promoted role and promotable clone id must be used (instead of
341 the resource id).
342
343 If --expired is specified, only constraints with expired life‐
344 times will be removed.
345
346 If --wait is specified, pcs will wait up to 'n' seconds for the
347 operation to finish (including starting and/or moving resources
348 if appropriate) and then return 0 on success or 1 on error. If
349 'n' is not specified it defaults to 60 minutes.
350
351 standards
352 List available resource agent standards supported by this in‐
353 stallation (OCF, LSB, etc.).
354
355 providers
356 List available OCF resource agent providers.
357
358 agents [standard[:provider]]
359 List available agents optionally filtered by standard and
360 provider.
361
362 update <resource id> [resource options] [op [<operation action> <opera‐
363 tion options>]...] [meta <meta operations>...] [--wait[=n]]
364 Add, remove or change options of specified resource, clone or
365 multi-state resource. Unspecified options will be kept un‐
366 changed. If you wish to remove an option, set it to empty value,
367 i.e. 'option_name='.
368
369 If an operation (op) is specified it will update the first found
370 operation with the same action on the specified resource. If no
371 operation with that action exists then a new operation will be
372 created. (WARNING: all existing options on the updated operation
373 will be reset if not specified.) If you want to create multiple
374 monitor operations you should use the 'op add' & 'op remove'
375 commands.
376
377 If --wait is specified, pcs will wait up to 'n' seconds for the
378 changes to take effect and then return 0 if the changes have
379 been processed or 1 otherwise. If 'n' is not specified it de‐
380 faults to 60 minutes.
381
382 op add <resource id> <operation action> [operation properties]
383 Add operation for specified resource.
384
385 op delete <resource id> <operation action> [<operation properties>...]
386 Remove specified operation (note: you must specify the exact op‐
387 eration properties to properly remove an existing operation).
388
389 op delete <operation id>
390 Remove the specified operation id.
391
392 op remove <resource id> <operation action> [<operation properties>...]
393 Remove specified operation (note: you must specify the exact op‐
394 eration properties to properly remove an existing operation).
395
396 op remove <operation id>
397 Remove the specified operation id.
398
399 op defaults [config] [--all] [--full] [--no-check-expired]
400 List currently configured default values for operations. If
401 --all is specified, also list expired sets of values. If --full
402 is specified, also list ids. If --no-expire-check is specified,
403 do not evaluate whether sets of values are expired.
404
405 op defaults <name>=<value>...
406 Set default values for operations.
407 NOTE: Defaults do not apply to resources / stonith devices which
408 override them with their own defined values.
409
410 op defaults set create [<set options>] [meta [<name>=<value>]...] [rule
411 [<expression>]]
412 Create a new set of default values for resource / stonith device
413 operations. You may specify a rule describing resources /
414 stonith devices and / or operations to which the set applies.
415
416 Set options are: id, score
417
418 Expression looks like one of the following:
419 op <operation name> [interval=<interval>]
420 resource [<standard>]:[<provider>]:[<type>]
421 defined|not_defined <node attribute>
422 <node attribute> lt|gt|lte|gte|eq|ne [string|integer|num‐
423 ber|version] <value>
424 date gt|lt <date>
425 date in_range [<date>] to <date>
426 date in_range <date> to duration <duration options>
427 date-spec <date-spec options>
428 <expression> and|or <expression>
429 (<expression>)
430
431 You may specify all or any of 'standard', 'provider' and 'type'
432 in a resource expression. For example: 'resource ocf::' matches
433 all resources of 'ocf' standard, while 'resource ::Dummy'
434 matches all resources of 'Dummy' type regardless of their stan‐
435 dard and provider.
436
437 Dates are expected to conform to ISO 8601 format.
438
439 Duration options are: hours, monthdays, weekdays, yearsdays,
440 months, weeks, years, weekyears, moon. Value for these options
441 is an integer.
442
443 Date-spec options are: hours, monthdays, weekdays, yearsdays,
444 months, weeks, years, weekyears, moon. Value for these options
445 is an integer or a range written as integer-integer.
446
447 NOTE: Defaults do not apply to resources / stonith devices which
448 override them with their own defined values.
449
450 op defaults set delete [<set id>]...
451 Delete specified options sets.
452
453 op defaults set remove [<set id>]...
454 Delete specified options sets.
455
456 op defaults set update <set id> [meta [<name>=<value>]...]
457 Add, remove or change values in specified set of default values
458 for resource / stonith device operations. Unspecified options
459 will be kept unchanged. If you wish to remove an option, set it
460 to empty value, i.e. 'option_name='.
461
462 NOTE: Defaults do not apply to resources / stonith devices which
463 override them with their own defined values.
464
465 op defaults update <name>=<value>...
466 Add, remove or change default values for operations. This is a
467 simplified command useful for cases when you only manage one set
468 of default values. Unspecified options will be kept unchanged.
469 If you wish to remove an option, set it to empty value, i.e.
470 'option_name='.
471
472 NOTE: Defaults do not apply to resources / stonith devices which
473 override them with their own defined values.
474
475 meta <resource id | group id | clone id> <meta options> [--wait[=n]]
476 Add specified options to the specified resource, group or clone.
477 Meta options should be in the format of name=value, options may
478 be removed by setting an option without a value. If --wait is
479 specified, pcs will wait up to 'n' seconds for the changes to
480 take effect and then return 0 if the changes have been processed
481 or 1 otherwise. If 'n' is not specified it defaults to 60 min‐
482 utes.
483 Example: pcs resource meta TestResource failure-timeout=50 re‐
484 source-stickiness=
485
486 group list
487 Show all currently configured resource groups and their re‐
488 sources.
489
490 group add <group id> <resource id> [resource id] ... [resource id]
491 [--before <resource id> | --after <resource id>] [--wait[=n]]
492 Add the specified resource to the group, creating the group if
493 it does not exist. If the resource is present in another group
494 it is moved to the new group. If the group remains empty after
495 move, it is deleted (for cloned groups, the clone is deleted as
496 well). The delete operation may fail in case the group is refer‐
497 enced within the configuration, e.g. by constraints. In that
498 case, use 'pcs resource ungroup' command prior to moving all re‐
499 sources out of the group.
500
501 You can use --before or --after to specify the position of the
502 added resources relatively to some resource already existing in
503 the group. By adding resources to a group they are already in
504 and specifying --after or --before you can move the resources in
505 the group.
506
507 If --wait is specified, pcs will wait up to 'n' seconds for the
508 operation to finish (including moving resources if appropriate)
509 and then return 0 on success or 1 on error. If 'n' is not speci‐
510 fied it defaults to 60 minutes.
511
512 group delete <group id> [resource id]... [--wait[=n]]
513 Remove the group (note: this does not remove any resources from
514 the cluster) or if resources are specified, remove the specified
515 resources from the group. If --wait is specified, pcs will wait
516 up to 'n' seconds for the operation to finish (including moving
517 resources if appropriate) and the return 0 on success or 1 on
518 error. If 'n' is not specified it defaults to 60 minutes.
519
520 group remove <group id> [resource id]... [--wait[=n]]
521 Remove the group (note: this does not remove any resources from
522 the cluster) or if resources are specified, remove the specified
523 resources from the group. If --wait is specified, pcs will wait
524 up to 'n' seconds for the operation to finish (including moving
525 resources if appropriate) and the return 0 on success or 1 on
526 error. If 'n' is not specified it defaults to 60 minutes.
527
528 ungroup <group id> [resource id]... [--wait[=n]]
529 Remove the group (note: this does not remove any resources from
530 the cluster) or if resources are specified, remove the specified
531 resources from the group. If --wait is specified, pcs will wait
532 up to 'n' seconds for the operation to finish (including moving
533 resources if appropriate) and the return 0 on success or 1 on
534 error. If 'n' is not specified it defaults to 60 minutes.
535
536 clone <resource id | group id> [<clone id>] [clone options]...
537 [--wait[=n]]
538 Set up the specified resource or group as a clone. If --wait is
539 specified, pcs will wait up to 'n' seconds for the operation to
540 finish (including starting clone instances if appropriate) and
541 then return 0 on success or 1 on error. If 'n' is not specified
542 it defaults to 60 minutes.
543
544 promotable <resource id | group id> [<clone id>] [clone options]...
545 [--wait[=n]]
546 Set up the specified resource or group as a promotable clone.
547 This is an alias for 'pcs resource clone <resource id> pro‐
548 motable=true'.
549
550 unclone <clone id | resource id | group id> [--wait[=n]]
551 Remove the specified clone or the clone which contains the spec‐
552 ified group or resource (the resource or group will not be re‐
553 moved). If --wait is specified, pcs will wait up to 'n' seconds
554 for the operation to finish (including stopping clone instances
555 if appropriate) and then return 0 on success or 1 on error. If
556 'n' is not specified it defaults to 60 minutes.
557
558 bundle create <bundle id> container <container type> [<container op‐
559 tions>] [network <network options>] [port-map <port options>]... [stor‐
560 age-map <storage options>]... [meta <meta options>] [--disabled]
561 [--wait[=n]]
562 Create a new bundle encapsulating no resources. The bundle can
563 be used either as it is or a resource may be put into it at any
564 time. If --disabled is specified, the bundle is not started au‐
565 tomatically. If --wait is specified, pcs will wait up to 'n'
566 seconds for the bundle to start and then return 0 on success or
567 1 on error. If 'n' is not specified it defaults to 60 minutes.
568
569 bundle reset <bundle id> [container <container options>] [network <net‐
570 work options>] [port-map <port options>]... [storage-map <storage op‐
571 tions>]... [meta <meta options>] [--disabled] [--wait[=n]]
572 Configure specified bundle with given options. Unlike bundle up‐
573 date, this command resets the bundle according given options -
574 no previous options are kept. Resources inside the bundle are
575 kept as they are. If --disabled is specified, the bundle is not
576 started automatically. If --wait is specified, pcs will wait up
577 to 'n' seconds for the bundle to start and then return 0 on suc‐
578 cess or 1 on error. If 'n' is not specified it defaults to 60
579 minutes.
580
581 bundle update <bundle id> [container <container options>] [network
582 <network options>] [port-map (add <port options>) | (delete | remove
583 <id>...)]... [storage-map (add <storage options>) | (delete | remove
584 <id>...)]... [meta <meta options>] [--wait[=n]]
585 Add, remove or change options of specified bundle. Unspecified
586 options will be kept unchanged. If you wish to remove an option,
587 set it to empty value, i.e. 'option_name='.
588
589 If you wish to update a resource encapsulated in the bundle, use
590 the 'pcs resource update' command instead and specify the re‐
591 source id.
592
593 If --wait is specified, pcs will wait up to 'n' seconds for the
594 operation to finish (including moving resources if appropriate)
595 and then return 0 on success or 1 on error. If 'n' is not speci‐
596 fied it defaults to 60 minutes.
597
598 manage <resource id | tag id>... [--monitor]
599 Set resources listed to managed mode (default). If --monitor is
600 specified, enable all monitor operations of the resources.
601
602 unmanage <resource id | tag id>... [--monitor]
603 Set resources listed to unmanaged mode. When a resource is in
604 unmanaged mode, the cluster is not allowed to start nor stop the
605 resource. If --monitor is specified, disable all monitor opera‐
606 tions of the resources.
607
608 defaults [config] [--all] [--full] [--no-check-expired]
609 List currently configured default values for resources / stonith
610 devices. If --all is specified, also list expired sets of val‐
611 ues. If --full is specified, also list ids. If --no-expire-check
612 is specified, do not evaluate whether sets of values are ex‐
613 pired.
614
615 defaults <name>=<value>...
616 Set default values for resources / stonith devices.
617 NOTE: Defaults do not apply to resources / stonith devices which
618 override them with their own defined values.
619
620 defaults set create [<set options>] [meta [<name>=<value>]...] [rule
621 [<expression>]]
622 Create a new set of default values for resources / stonith de‐
623 vices. You may specify a rule describing resources / stonith de‐
624 vices to which the set applies.
625
626 Set options are: id, score
627
628 Expression looks like one of the following:
629 resource [<standard>]:[<provider>]:[<type>]
630 date gt|lt <date>
631 date in_range [<date>] to <date>
632 date in_range <date> to duration <duration options>
633 date-spec <date-spec options>
634 <expression> and|or <expression>
635 (<expression>)
636
637 You may specify all or any of 'standard', 'provider' and 'type'
638 in a resource expression. For example: 'resource ocf::' matches
639 all resources of 'ocf' standard, while 'resource ::Dummy'
640 matches all resources of 'Dummy' type regardless of their stan‐
641 dard and provider.
642
643 Dates are expected to conform to ISO 8601 format.
644
645 Duration options are: hours, monthdays, weekdays, yearsdays,
646 months, weeks, years, weekyears, moon. Value for these options
647 is an integer.
648
649 Date-spec options are: hours, monthdays, weekdays, yearsdays,
650 months, weeks, years, weekyears, moon. Value for these options
651 is an integer or a range written as integer-integer.
652
653 NOTE: Defaults do not apply to resources / stonith devices which
654 override them with their own defined values.
655
656 defaults set delete [<set id>]...
657 Delete specified options sets.
658
659 defaults set remove [<set id>]...
660 Delete specified options sets.
661
662 defaults set update <set id> [meta [<name>=<value>]...]
663 Add, remove or change values in specified set of default values
664 for resources / stonith devices. Unspecified options will be
665 kept unchanged. If you wish to remove an option, set it to empty
666 value, i.e. 'option_name='.
667
668 NOTE: Defaults do not apply to resources / stonith devices which
669 override them with their own defined values.
670
671 defaults update <name>=<value>...
672 Add, remove or change default values for resources / stonith de‐
673 vices. This is a simplified command useful for cases when you
674 only manage one set of default values. Unspecified options will
675 be kept unchanged. If you wish to remove an option, set it to
676 empty value, i.e. 'option_name='.
677
678 NOTE: Defaults do not apply to resources / stonith devices which
679 override them with their own defined values.
680
681 cleanup [<resource id | stonith id>] [node=<node>] [operation=<opera‐
682 tion> [interval=<interval>]] [--strict]
683 Make the cluster forget failed operations from history of the
684 resource / stonith device and re-detect its current state. This
685 can be useful to purge knowledge of past failures that have
686 since been resolved.
687
688 If the named resource is part of a group, or one numbered in‐
689 stance of a clone or bundled resource, the clean-up applies to
690 the whole collective resource unless --strict is given.
691
692 If a resource id / stonith id is not specified then all re‐
693 sources / stonith devices will be cleaned up.
694
695 If a node is not specified then resources / stonith devices on
696 all nodes will be cleaned up.
697
698 refresh [<resource id | stonith id>] [node=<node>] [--strict]
699 Make the cluster forget the complete operation history (includ‐
700 ing failures) of the resource / stonith device and re-detect its
701 current state. If you are interested in forgetting failed opera‐
702 tions only, use the 'pcs resource cleanup' command.
703
704 If the named resource is part of a group, or one numbered in‐
705 stance of a clone or bundled resource, the refresh applies to
706 the whole collective resource unless --strict is given.
707
708 If a resource id / stonith id is not specified then all re‐
709 sources / stonith devices will be refreshed.
710
711 If a node is not specified then resources / stonith devices on
712 all nodes will be refreshed.
713
714 failcount [show [<resource id | stonith id>] [node=<node>] [opera‐
715 tion=<operation> [interval=<interval>]]] [--full]
716 Show current failcount for resources and stonith devices, op‐
717 tionally filtered by a resource / stonith device, node, opera‐
718 tion and its interval. If --full is specified do not sum fail‐
719 counts per resource / stonith device and node. Use 'pcs resource
720 cleanup' or 'pcs resource refresh' to reset failcounts.
721
722 relocate dry-run [resource1] [resource2] ...
723 The same as 'relocate run' but has no effect on the cluster.
724
725 relocate run [resource1] [resource2] ...
726 Relocate specified resources to their preferred nodes. If no
727 resources are specified, relocate all resources. This command
728 calculates the preferred node for each resource while ignoring
729 resource stickiness. Then it creates location constraints which
730 will cause the resources to move to their preferred nodes. Once
731 the resources have been moved the constraints are deleted auto‐
732 matically. Note that the preferred node is calculated based on
733 current cluster status, constraints, location of resources and
734 other settings and thus it might change over time.
735
736 relocate show
737 Display current status of resources and their optimal node ig‐
738 noring resource stickiness.
739
740 relocate clear
741 Remove all constraints created by the 'relocate run' command.
742
743 utilization [<resource id> [<name>=<value> ...]]
744 Add specified utilization options to specified resource. If re‐
745 source is not specified, shows utilization of all resources. If
746 utilization options are not specified, shows utilization of
747 specified resource. Utilization option should be in format
748 name=value, value has to be integer. Options may be removed by
749 setting an option without a value. Example: pcs resource uti‐
750 lization TestResource cpu= ram=20
751
752 relations <resource id> [--full]
753 Display relations of a resource specified by its id with other
754 resources in a tree structure. Supported types of resource rela‐
755 tions are: ordering constraints, ordering set constraints, rela‐
756 tions defined by resource hierarchy (clones, groups, bundles).
757 If --full is used, more verbose output will be printed.
758
759 cluster
760 setup <cluster name> (<node name> [addr=<node address>]...)... [trans‐
761 port knet|udp|udpu [<transport options>] [link <link options>]... [com‐
762 pression <compression options>] [crypto <crypto options>]] [totem
763 <totem options>] [quorum <quorum options>] [--no-cluster-uuid] ([--en‐
764 able] [--start [--wait[=<n>]]] [--no-keys-sync]) | [--corosync_conf
765 <path>]
766 Create a cluster from the listed nodes and synchronize cluster
767 configuration files to them. If --corosync_conf is specified, do
768 not connect to other nodes and save corosync.conf to the speci‐
769 fied path; see 'Local only mode' below for details.
770
771 Nodes are specified by their names and optionally their ad‐
772 dresses. If no addresses are specified for a node, pcs will con‐
773 figure corosync to communicate with that node using an address
774 provided in 'pcs host auth' command. Otherwise, pcs will config‐
775 ure corosync to communicate with the node using the specified
776 addresses.
777
778 Transport knet:
779 This is the default transport. It allows configuring traffic en‐
780 cryption and compression as well as using multiple addresses
781 (links) for nodes.
782 Transport options are: ip_version, knet_pmtud_interval,
783 link_mode
784 Link options are: link_priority, linknumber, mcastport, ping_in‐
785 terval, ping_precision, ping_timeout, pong_count, transport (udp
786 or sctp)
787 Each 'link' followed by options sets options for one link in the
788 order the links are defined by nodes' addresses. You can set
789 link options for a subset of links using a linknumber. See exam‐
790 ples below.
791 Compression options are: level, model, threshold
792 Crypto options are: cipher, hash, model
793 By default, encryption is enabled with cipher=aes256 and
794 hash=sha256. To disable encryption, set cipher=none and
795 hash=none.
796
797 Transports udp and udpu:
798 These transports are limited to one address per node. They do
799 not support traffic encryption nor compression.
800 Transport options are: ip_version, netmtu
801 Link options are: bindnetaddr, broadcast, mcastaddr, mcastport,
802 ttl
803
804 Totem and quorum can be configured regardless of used transport.
805 Totem options are: block_unlisted_ips, consensus, downcheck,
806 fail_recv_const, heartbeat_failures_allowed, hold, join,
807 max_messages, max_network_delay, merge, miss_count_const,
808 send_join, seqno_unchanged_const, token, token_coefficient, to‐
809 ken_retransmit, token_retransmits_before_loss_const, window_size
810 Quorum options are: auto_tie_breaker, last_man_standing,
811 last_man_standing_window, wait_for_all
812
813 Transports and their options, link, compression, crypto and
814 totem options are all documented in corosync.conf(5) man page;
815 knet link options are prefixed 'knet_' there, compression op‐
816 tions are prefixed 'knet_compression_' and crypto options are
817 prefixed 'crypto_'. Quorum options are documented in votequo‐
818 rum(5) man page.
819
820 --no-cluster-uuid will not generate a unique ID for the cluster.
821 --enable will configure the cluster to start on nodes boot.
822 --start will start the cluster right after creating it. --wait
823 will wait up to 'n' seconds for the cluster to start.
824 --no-keys-sync will skip creating and distributing pcsd SSL cer‐
825 tificate and key and corosync and pacemaker authkey files. Use
826 this if you provide your own certificates and keys.
827
828 Local only mode:
829 By default, pcs connects to all specified nodes to verify they
830 can be used in the new cluster and to send cluster configuration
831 files to them. If this is not what you want, specify
832 --corosync_conf option followed by a file path. Pcs will save
833 corosync.conf to the specified file and will not connect to
834 cluster nodes. These are the tasks that pcs skips in that case:
835 * make sure the nodes are not running or configured to run a
836 cluster already
837 * make sure cluster packages are installed on all nodes and
838 their versions are compatible
839 * make sure there are no cluster configuration files on any node
840 (run 'pcs cluster destroy' and remove pcs_settings.conf file on
841 all nodes)
842 * synchronize corosync and pacemaker authkeys, /etc/corosync/au‐
843 thkey and /etc/pacemaker/authkey respectively, and the
844 corosync.conf file
845 * authenticate the cluster nodes against each other ('pcs clus‐
846 ter auth' or 'pcs host auth' command)
847 * synchronize pcsd certificates (so that pcs web UI can be used
848 in an HA mode)
849
850 Examples:
851 Create a cluster with default settings:
852 pcs cluster setup newcluster node1 node2
853 Create a cluster using two links:
854 pcs cluster setup newcluster node1 addr=10.0.1.11
855 addr=10.0.2.11 node2 addr=10.0.1.12 addr=10.0.2.12
856 Set link options for all links. Link options are matched to the
857 links in order. The first link (link 0) has sctp transport, the
858 second link (link 1) has mcastport 55405:
859 pcs cluster setup newcluster node1 addr=10.0.1.11
860 addr=10.0.2.11 node2 addr=10.0.1.12 addr=10.0.2.12 transport
861 knet link transport=sctp link mcastport=55405
862 Set link options for the second and fourth links only. Link op‐
863 tions are matched to the links based on the linknumber option
864 (the first link is link 0):
865 pcs cluster setup newcluster node1 addr=10.0.1.11
866 addr=10.0.2.11 addr=10.0.3.11 addr=10.0.4.11 node2
867 addr=10.0.1.12 addr=10.0.2.12 addr=10.0.3.12 addr=10.0.4.12
868 transport knet link linknumber=3 mcastport=55405 link linknum‐
869 ber=1 transport=sctp
870 Create a cluster using udp transport with a non-default port:
871 pcs cluster setup newcluster node1 node2 transport udp link
872 mcastport=55405
873
874 config [show] [--output-format cmd|json|text] [--corosync_conf <path>]
875 Show cluster configuration. There are 3 formats of output avail‐
876 able: 'cmd', 'json' and 'text', default is 'text'. Format 'text'
877 is a human friendly output. Format 'cmd' prints pcs commands
878 which can be used to recreate the same configuration. Format
879 'json' is a machine oriented output of the configuration. If
880 --corosync_conf is specified, configuration file specified by
881 <path> is used instead of the current cluster configuration.
882
883 config update [transport <transport options>] [compression <compression
884 options>] [crypto <crypto options>] [totem <totem options>]
885 [--corosync_conf <path>]
886 Update cluster configuration. Unspecified options will be kept
887 unchanged. If you wish to remove an option, set it to empty
888 value, i.e. 'option_name='.
889
890 If --corosync_conf is specified, update cluster configuration in
891 a file specified by <path>.
892
893 All options are documented in corosync.conf(5) man page. There
894 are different transport options for transport types. Compression
895 and crypto options are only available for knet transport. Totem
896 options can be set regardless of the transport type.
897 Transport options for knet transport are: ip_version, knet_pm‐
898 tud_interval, link_mode
899 Transport options for udp and updu transports are: ip_version,
900 netmtu
901 Compression options are: level, model, threshold
902 Crypto options are: cipher, hash, model
903 Totem options are: block_unlisted_ips, consensus, downcheck,
904 fail_recv_const, heartbeat_failures_allowed, hold, join,
905 max_messages, max_network_delay, merge, miss_count_const,
906 send_join, seqno_unchanged_const, token, token_coefficient, to‐
907 ken_retransmit, token_retransmits_before_loss_const, window_size
908
909 config uuid generate [--corosync_conf <path>] [--force]
910 Generate a new cluster UUID and distribute it to all cluster
911 nodes. Cluster UUID is not used by the cluster stack in any way,
912 it is provided to easily distinguish between multiple clusters
913 in a multi-cluster environment since the cluster name does not
914 have to be unique.
915
916 If --corosync_conf is specified, update cluster configuration in
917 file specified by <path>.
918
919 If --force is specified, existing UUID will be overwritten.
920
921 authkey corosync [<path>]
922 Generate a new corosync authkey and distribute it to all cluster
923 nodes. If <path> is specified, do not generate a key and use key
924 from the file.
925
926 start [--all | <node>... ] [--wait[=<n>]] [--request-timeout=<seconds>]
927 Start a cluster on specified node(s). If no nodes are specified
928 then start a cluster on the local node. If --all is specified
929 then start a cluster on all nodes. If the cluster has many nodes
930 then the start request may time out. In that case you should
931 consider setting --request-timeout to a suitable value. If
932 --wait is specified, pcs waits up to 'n' seconds for the cluster
933 to get ready to provide services after the cluster has success‐
934 fully started.
935
936 stop [--all | <node>... ] [--request-timeout=<seconds>]
937 Stop a cluster on specified node(s). If no nodes are specified
938 then stop a cluster on the local node. If --all is specified
939 then stop a cluster on all nodes. If the cluster is running re‐
940 sources which take long time to stop then the stop request may
941 time out before the cluster actually stops. In that case you
942 should consider setting --request-timeout to a suitable value.
943
944 kill Force corosync and pacemaker daemons to stop on the local node
945 (performs kill -9). Note that init system (e.g. systemd) can de‐
946 tect that cluster is not running and start it again. If you want
947 to stop cluster on a node, run pcs cluster stop on that node.
948
949 enable [--all | <node>... ]
950 Configure cluster to run on node boot on specified node(s). If
951 node is not specified then cluster is enabled on the local node.
952 If --all is specified then cluster is enabled on all nodes.
953
954 disable [--all | <node>... ]
955 Configure cluster to not run on node boot on specified node(s).
956 If node is not specified then cluster is disabled on the local
957 node. If --all is specified then cluster is disabled on all
958 nodes.
959
960 auth [-u <username>] [-p <password>]
961 Authenticate pcs/pcsd to pcsd on nodes configured in the local
962 cluster.
963
964 status View current cluster status (an alias of 'pcs status cluster').
965
966 sync Sync cluster configuration (files which are supported by all
967 subcommands of this command) to all cluster nodes.
968
969 sync corosync
970 Sync corosync configuration to all nodes found from current
971 corosync.conf file.
972
973 cib [filename] [scope=<scope> | --config]
974 Get the raw xml from the CIB (Cluster Information Base). If a
975 filename is provided, we save the CIB to that file, otherwise
976 the CIB is printed. Specify scope to get a specific section of
977 the CIB. Valid values of the scope are: acls, alerts, configura‐
978 tion, constraints, crm_config, fencing-topology, nodes, op_de‐
979 faults, resources, rsc_defaults, tags. --config is the same as
980 scope=configuration. Do not specify a scope if you want to edit
981 the saved CIB using pcs (pcs -f <command>).
982
983 cib-push <filename> [--wait[=<n>]] [diff-against=<filename_original> |
984 scope=<scope> | --config]
985 Push the raw xml from <filename> to the CIB (Cluster Information
986 Base). You can obtain the CIB by running the 'pcs cluster cib'
987 command, which is recommended first step when you want to per‐
988 form desired modifications (pcs -f <command>) for the one-off
989 push.
990 If diff-against is specified, pcs diffs contents of filename
991 against contents of filename_original and pushes the result to
992 the CIB.
993 Specify scope to push a specific section of the CIB. Valid val‐
994 ues of the scope are: acls, alerts, configuration, constraints,
995 crm_config, fencing-topology, nodes, op_defaults, resources,
996 rsc_defaults, tags. --config is the same as scope=configuration.
997 Use of --config is recommended. Do not specify a scope if you
998 need to push the whole CIB or be warned in the case of outdated
999 CIB.
1000 If --wait is specified wait up to 'n' seconds for changes to be
1001 applied.
1002 WARNING: the selected scope of the CIB will be overwritten by
1003 the current content of the specified file.
1004
1005 Example:
1006 pcs cluster cib > original.xml
1007 cp original.xml new.xml
1008 pcs -f new.xml constraint location apache prefers node2
1009 pcs cluster cib-push new.xml diff-against=original.xml
1010
1011 cib-upgrade
1012 Upgrade the CIB to conform to the latest version of the document
1013 schema.
1014
1015 edit [scope=<scope> | --config]
1016 Edit the cib in the editor specified by the $EDITOR environment
1017 variable and push out any changes upon saving. Specify scope to
1018 edit a specific section of the CIB. Valid values of the scope
1019 are: acls, alerts, configuration, constraints, crm_config, fenc‐
1020 ing-topology, nodes, op_defaults, resources, rsc_defaults, tags.
1021 --config is the same as scope=configuration. Use of --config is
1022 recommended. Do not specify a scope if you need to edit the
1023 whole CIB or be warned in the case of outdated CIB.
1024
1025 node add <node name> [addr=<node address>]... [watchdog=<watchdog
1026 path>] [device=<SBD device path>]... [--start [--wait[=<n>]]] [--en‐
1027 able] [--no-watchdog-validation]
1028 Add the node to the cluster and synchronize all relevant config‐
1029 uration files to the new node. This command can only be run on
1030 an existing cluster node.
1031
1032 The new node is specified by its name and optionally its ad‐
1033 dresses. If no addresses are specified for the node, pcs will
1034 configure corosync to communicate with the node using an address
1035 provided in 'pcs host auth' command. Otherwise, pcs will config‐
1036 ure corosync to communicate with the node using the specified
1037 addresses.
1038
1039 Use 'watchdog' to specify a path to a watchdog on the new node,
1040 when SBD is enabled in the cluster. If SBD is configured with
1041 shared storage, use 'device' to specify path to shared device(s)
1042 on the new node.
1043
1044 If --start is specified also start cluster on the new node, if
1045 --wait is specified wait up to 'n' seconds for the new node to
1046 start. If --enable is specified configure cluster to start on
1047 the new node on boot. If --no-watchdog-validation is specified,
1048 validation of watchdog will be skipped.
1049
1050 WARNING: By default, it is tested whether the specified watchdog
1051 is supported. This may cause a restart of the system when a
1052 watchdog with no-way-out-feature enabled is present. Use
1053 --no-watchdog-validation to skip watchdog validation.
1054
1055 node delete <node name> [<node name>]...
1056 Shutdown specified nodes and remove them from the cluster.
1057
1058 node remove <node name> [<node name>]...
1059 Shutdown specified nodes and remove them from the cluster.
1060
1061 node add-remote <node name> [<node address>] [options] [op <operation
1062 action> <operation options> [<operation action> <operation op‐
1063 tions>]...] [meta <meta options>...] [--wait[=<n>]]
1064 Add the node to the cluster as a remote node. Sync all relevant
1065 configuration files to the new node. Start the node and config‐
1066 ure it to start the cluster on boot. Options are port and recon‐
1067 nect_interval. Operations and meta belong to an underlying con‐
1068 nection resource (ocf:pacemaker:remote). If node address is not
1069 specified for the node, pcs will configure pacemaker to communi‐
1070 cate with the node using an address provided in 'pcs host auth'
1071 command. Otherwise, pcs will configure pacemaker to communicate
1072 with the node using the specified addresses. If --wait is speci‐
1073 fied, wait up to 'n' seconds for the node to start.
1074
1075 node delete-remote <node identifier>
1076 Shutdown specified remote node and remove it from the cluster.
1077 The node-identifier can be the name of the node or the address
1078 of the node.
1079
1080 node remove-remote <node identifier>
1081 Shutdown specified remote node and remove it from the cluster.
1082 The node-identifier can be the name of the node or the address
1083 of the node.
1084
1085 node add-guest <node name> <resource id> [options] [--wait[=<n>]]
1086 Make the specified resource a guest node resource. Sync all rel‐
1087 evant configuration files to the new node. Start the node and
1088 configure it to start the cluster on boot. Options are re‐
1089 mote-addr, remote-port and remote-connect-timeout. If re‐
1090 mote-addr is not specified for the node, pcs will configure
1091 pacemaker to communicate with the node using an address provided
1092 in 'pcs host auth' command. Otherwise, pcs will configure pace‐
1093 maker to communicate with the node using the specified ad‐
1094 dresses. If --wait is specified, wait up to 'n' seconds for the
1095 node to start.
1096
1097 node delete-guest <node identifier>
1098 Shutdown specified guest node and remove it from the cluster.
1099 The node-identifier can be the name of the node or the address
1100 of the node or id of the resource that is used as the guest
1101 node.
1102
1103 node remove-guest <node identifier>
1104 Shutdown specified guest node and remove it from the cluster.
1105 The node-identifier can be the name of the node or the address
1106 of the node or id of the resource that is used as the guest
1107 node.
1108
1109 node clear <node name>
1110 Remove specified node from various cluster caches. Use this if a
1111 removed node is still considered by the cluster to be a member
1112 of the cluster.
1113
1114 link add <node_name>=<node_address>... [options <link options>]
1115 Add a corosync link. One address must be specified for each
1116 cluster node. If no linknumber is specified, pcs will use the
1117 lowest available linknumber.
1118 Link options (documented in corosync.conf(5) man page) are:
1119 link_priority, linknumber, mcastport, ping_interval, ping_preci‐
1120 sion, ping_timeout, pong_count, transport (udp or sctp)
1121
1122 link delete <linknumber> [<linknumber>]...
1123 Remove specified corosync links.
1124
1125 link remove <linknumber> [<linknumber>]...
1126 Remove specified corosync links.
1127
1128 link update <linknumber> [<node_name>=<node_address>...] [options <link
1129 options>]
1130 Add, remove or change node addresses / link options of an exist‐
1131 ing corosync link. Use this if you cannot add / remove links
1132 which is the preferred way. Unspecified options will be kept un‐
1133 changed. If you wish to remove an option, set it to empty value,
1134 i.e. 'option_name='.
1135 Link options (documented in corosync.conf(5) man page) are:
1136 for knet transport: link_priority, mcastport, ping_interval,
1137 ping_precision, ping_timeout, pong_count, transport (udp or
1138 sctp)
1139 for udp and udpu transports: bindnetaddr, broadcast, mcastaddr,
1140 mcastport, ttl
1141
1142 uidgid List the current configured uids and gids of users allowed to
1143 connect to corosync.
1144
1145 uidgid add [uid=<uid>] [gid=<gid>]
1146 Add the specified uid and/or gid to the list of users/groups al‐
1147 lowed to connect to corosync.
1148
1149 uidgid delete [uid=<uid>] [gid=<gid>]
1150 Remove the specified uid and/or gid from the list of
1151 users/groups allowed to connect to corosync.
1152
1153 uidgid remove [uid=<uid>] [gid=<gid>]
1154 Remove the specified uid and/or gid from the list of
1155 users/groups allowed to connect to corosync.
1156
1157 corosync [node]
1158 Get the corosync.conf from the specified node or from the cur‐
1159 rent node if node not specified.
1160
1161 reload corosync
1162 Reload the corosync configuration on the current node.
1163
1164 destroy [--all] [--force]
1165 Permanently destroy the cluster on the current node, killing all
1166 cluster processes and removing all cluster configuration files.
1167 Using --all will attempt to destroy the cluster on all nodes in
1168 the local cluster.
1169
1170 WARNING: This command permanently removes any cluster configura‐
1171 tion that has been created. It is recommended to run 'pcs clus‐
1172 ter stop' before destroying the cluster. To prevent accidental
1173 running of this command, --force or interactive user response is
1174 required in order to proceed.
1175
1176 verify [--full] [-f <filename>]
1177 Checks the pacemaker configuration (CIB) for syntax and common
1178 conceptual errors. If no filename is specified the check is per‐
1179 formed on the currently running cluster. If --full is used more
1180 verbose output will be printed.
1181
1182 report [--from "YYYY-M-D H:M:S" [--to "YYYY-M-D H:M:S"]] <dest>
1183 Create a tarball containing everything needed when reporting
1184 cluster problems. If --from and --to are not used, the report
1185 will include the past 24 hours.
1186
1187 stonith
1188 [status [<resource id | tag id>] [node=<node>] [--hide-inactive]]
1189 Show status of all currently configured stonith devices. If
1190 --hide-inactive is specified, only show active stonith devices.
1191 If a resource or tag id is specified, only show status of the
1192 specified resource or resources in the specified tag. If node is
1193 specified, only show status of resources configured for the
1194 specified node.
1195
1196 config [--output-format text|cmd|json] [<stonith id>]...
1197 Show options of all currently configured stonith devices or if
1198 stonith device ids are specified show the options for the speci‐
1199 fied stonith device ids. There are 3 formats of output avail‐
1200 able: 'cmd', 'json' and 'text', default is 'text'. Format 'text'
1201 is a human friendly output. Format 'cmd' prints pcs commands
1202 which can be used to recreate the same configuration. Format
1203 'json' is a machine oriented output of the configuration.
1204
1205 list [filter] [--nodesc]
1206 Show list of all available stonith agents (if filter is provided
1207 then only stonith agents matching the filter will be shown). If
1208 --nodesc is used then descriptions of stonith agents are not
1209 printed.
1210
1211 describe <stonith agent> [--full]
1212 Show options for specified stonith agent. If --full is speci‐
1213 fied, all options including advanced and deprecated ones are
1214 shown.
1215
1216 create <stonith id> <stonith device type> [stonith device options] [op
1217 <operation action> <operation options> [<operation action> <operation
1218 options>]...] [meta <meta options>...] [--group <group id> [--before
1219 <stonith id> | --after <stonith id>]] [--disabled] [--wait[=n]]
1220 Create stonith device with specified type and options. If
1221 --group is specified the stonith device is added to the group
1222 named. You can use --before or --after to specify the position
1223 of the added stonith device relatively to some stonith device
1224 already existing in the group. If--disabled is specified the
1225 stonith device is not used. If --wait is specified, pcs will
1226 wait up to 'n' seconds for the stonith device to start and then
1227 return 0 if the stonith device is started, or 1 if the stonith
1228 device has not yet started. If 'n' is not specified it defaults
1229 to 60 minutes.
1230
1231 Example: Create a device for nodes node1 and node2
1232 pcs stonith create MyFence fence_virt pcmk_host_list=node1,node2
1233 Example: Use port p1 for node n1 and ports p2 and p3 for node n2
1234 pcs stonith create MyFence fence_virt
1235 'pcmk_host_map=n1:p1;n2:p2,p3'
1236
1237 update <stonith id> [stonith options] [op [<operation action> <opera‐
1238 tion options>]...] [meta <meta operations>...] [--wait[=n]]
1239 Add, remove or change options of specified stonith device. Un‐
1240 specified options will be kept unchanged. If you wish to remove
1241 an option, set it to empty value, i.e. 'option_name='.
1242
1243 If an operation (op) is specified it will update the first found
1244 operation with the same action on the specified stonith device.
1245 If no operation with that action exists then a new operation
1246 will be created. (WARNING: all existing options on the updated
1247 operation will be reset if not specified.) If you want to create
1248 multiple monitor operations you should use the 'op add' & 'op
1249 remove' commands.
1250
1251 If --wait is specified, pcs will wait up to 'n' seconds for the
1252 changes to take effect and then return 0 if the changes have
1253 been processed or 1 otherwise. If 'n' is not specified it de‐
1254 faults to 60 minutes.
1255
1256 update-scsi-devices <stonith id> (set <device-path> [<device-path>...])
1257 | (add <device-path> [<device-path>...] delete|remove <device-path>
1258 [<device-path>...] )
1259 Update scsi fencing devices without affecting other resources.
1260 You must specify either list of set devices or at least one de‐
1261 vice for add or delete/remove devices. Stonith resource must be
1262 running on one cluster node. Each device will be unfenced on
1263 each cluster node running cluster. Supported fence agents:
1264 fence_scsi, fence_mpath.
1265
1266 delete <stonith id>
1267 Remove stonith id from configuration.
1268
1269 remove <stonith id>
1270 Remove stonith id from configuration.
1271
1272 op add <stonith id> <operation action> [operation properties]
1273 Add operation for specified stonith device.
1274
1275 op delete <stonith id> <operation action> [<operation properties>...]
1276 Remove specified operation (note: you must specify the exact op‐
1277 eration properties to properly remove an existing operation).
1278
1279 op delete <operation id>
1280 Remove the specified operation id.
1281
1282 op remove <stonith id> <operation action> [<operation properties>...]
1283 Remove specified operation (note: you must specify the exact op‐
1284 eration properties to properly remove an existing operation).
1285
1286 op remove <operation id>
1287 Remove the specified operation id.
1288
1289 op defaults [config] [--all] [--full] [--no-check-expired]
1290 This command is an alias of 'resource op defaults [config]' com‐
1291 mand.
1292
1293 List currently configured default values for operations. If
1294 --all is specified, also list expired sets of values. If --full
1295 is specified, also list ids. If --no-expire-check is specified,
1296 do not evaluate whether sets of values are expired.
1297
1298 op defaults <name>=<value>...
1299 This command is an alias of 'resource op defaults' command.
1300
1301 Set default values for operations.
1302 NOTE: Defaults do not apply to resources / stonith devices which
1303 override them with their own defined values.
1304
1305 op defaults set create [<set options>] [meta [<name>=<value>]...] [rule
1306 [<expression>]]
1307 This command is an alias of 'resource op defaults set create'
1308 command.
1309
1310 Create a new set of default values for resource / stonith device
1311 operations. You may specify a rule describing resources /
1312 stonith devices and / or operations to which the set applies.
1313
1314 Set options are: id, score
1315
1316 Expression looks like one of the following:
1317 op <operation name> [interval=<interval>]
1318 resource [<standard>]:[<provider>]:[<type>]
1319 defined|not_defined <node attribute>
1320 <node attribute> lt|gt|lte|gte|eq|ne [string|integer|num‐
1321 ber|version] <value>
1322 date gt|lt <date>
1323 date in_range [<date>] to <date>
1324 date in_range <date> to duration <duration options>
1325 date-spec <date-spec options>
1326 <expression> and|or <expression>
1327 (<expression>)
1328
1329 You may specify all or any of 'standard', 'provider' and 'type'
1330 in a resource expression. For example: 'resource ocf::' matches
1331 all resources of 'ocf' standard, while 'resource ::Dummy'
1332 matches all resources of 'Dummy' type regardless of their stan‐
1333 dard and provider.
1334
1335 Dates are expected to conform to ISO 8601 format.
1336
1337 Duration options are: hours, monthdays, weekdays, yearsdays,
1338 months, weeks, years, weekyears, moon. Value for these options
1339 is an integer.
1340
1341 Date-spec options are: hours, monthdays, weekdays, yearsdays,
1342 months, weeks, years, weekyears, moon. Value for these options
1343 is an integer or a range written as integer-integer.
1344
1345 NOTE: Defaults do not apply to resources / stonith devices which
1346 override them with their own defined values.
1347
1348 op defaults set delete [<set id>]...
1349 This command is an alias of 'resource op defaults set delete'
1350 command.
1351
1352 Delete specified options sets.
1353
1354 op defaults set remove [<set id>]...
1355 This command is an alias of 'resource op defaults set delete'
1356 command.
1357
1358 Delete specified options sets.
1359
1360 op defaults set update <set id> [meta [<name>=<value>]...]
1361 This command is an alias of 'resource op defaults set update'
1362 command.
1363
1364 Add, remove or change values in specified set of default values
1365 for resource / stonith device operations. Unspecified options
1366 will be kept unchanged. If you wish to remove an option, set it
1367 to empty value, i.e. 'option_name='.
1368
1369 NOTE: Defaults do not apply to resources / stonith devices which
1370 override them with their own defined values.
1371
1372 op defaults update <name>=<value>...
1373 This command is an alias of 'resource op defaults update' com‐
1374 mand.
1375
1376 Add, remove or change default values for operations. This is a
1377 simplified command useful for cases when you only manage one set
1378 of default values. Unspecified options will be kept unchanged.
1379 If you wish to remove an option, set it to empty value, i.e.
1380 'option_name='.
1381
1382 NOTE: Defaults do not apply to resources / stonith devices which
1383 override them with their own defined values.
1384
1385 meta <stonith id> <meta options> [--wait[=n]]
1386 Add specified options to the specified stonith device. Meta op‐
1387 tions should be in the format of name=value, options may be re‐
1388 moved by setting an option without a value. If --wait is speci‐
1389 fied, pcs will wait up to 'n' seconds for the changes to take
1390 effect and then return 0 if the changes have been processed or 1
1391 otherwise. If 'n' is not specified it defaults to 60 minutes.
1392
1393 Example: pcs stonith meta test_stonith failure-timeout=50 re‐
1394 source-stickiness=
1395
1396 defaults [config] [--all] [--full] [--no-check-expired]
1397 This command is an alias of 'resource defaults [config]' com‐
1398 mand.
1399
1400 List currently configured default values for resources / stonith
1401 devices. If --all is specified, also list expired sets of val‐
1402 ues. If --full is specified, also list ids. If --no-expire-check
1403 is specified, do not evaluate whether sets of values are ex‐
1404 pired.
1405
1406 defaults <name>=<value>...
1407 This command is an alias of 'resource defaults' command.
1408
1409 Set default values for resources / stonith devices.
1410 NOTE: Defaults do not apply to resources / stonith devices which
1411 override them with their own defined values.
1412
1413 defaults set create [<set options>] [meta [<name>=<value>]...] [rule
1414 [<expression>]]
1415 This command is an alias of 'resource defaults set create' com‐
1416 mand.
1417
1418 Create a new set of default values for resources / stonith de‐
1419 vices. You may specify a rule describing resources / stonith de‐
1420 vices to which the set applies.
1421
1422 Set options are: id, score
1423
1424 Expression looks like one of the following:
1425 resource [<standard>]:[<provider>]:[<type>]
1426 date gt|lt <date>
1427 date in_range [<date>] to <date>
1428 date in_range <date> to duration <duration options>
1429 date-spec <date-spec options>
1430 <expression> and|or <expression>
1431 (<expression>)
1432
1433 You may specify all or any of 'standard', 'provider' and 'type'
1434 in a resource expression. For example: 'resource ocf::' matches
1435 all resources of 'ocf' standard, while 'resource ::Dummy'
1436 matches all resources of 'Dummy' type regardless of their stan‐
1437 dard and provider.
1438
1439 Dates are expected to conform to ISO 8601 format.
1440
1441 Duration options are: hours, monthdays, weekdays, yearsdays,
1442 months, weeks, years, weekyears, moon. Value for these options
1443 is an integer.
1444
1445 Date-spec options are: hours, monthdays, weekdays, yearsdays,
1446 months, weeks, years, weekyears, moon. Value for these options
1447 is an integer or a range written as integer-integer.
1448
1449 NOTE: Defaults do not apply to resources / stonith devices which
1450 override them with their own defined values.
1451
1452 defaults set delete [<set id>]...
1453 This command is an alias of 'resource defaults set delete' com‐
1454 mand.
1455
1456 Delete specified options sets.
1457
1458 defaults set remove [<set id>]...
1459 This command is an alias of 'resource defaults set delete' com‐
1460 mand.
1461
1462 Delete specified options sets.
1463
1464 defaults set update <set id> [meta [<name>=<value>]...]
1465 This command is an alias of 'resource defaults set update' com‐
1466 mand.
1467
1468 Add, remove or change values in specified set of default values
1469 for resources / stonith devices. Unspecified options will be
1470 kept unchanged. If you wish to remove an option, set it to empty
1471 value, i.e. 'option_name='.
1472
1473 NOTE: Defaults do not apply to resources / stonith devices which
1474 override them with their own defined values.
1475
1476 defaults update <name>=<value>...
1477 This command is an alias of 'resource defaults update' command.
1478
1479 Add, remove or change default values for resources / stonith de‐
1480 vices. This is a simplified command useful for cases when you
1481 only manage one set of default values. Unspecified options will
1482 be kept unchanged. If you wish to remove an option, set it to
1483 empty value, i.e. 'option_name='.
1484
1485 NOTE: Defaults do not apply to resources / stonith devices which
1486 override them with their own defined values.
1487
1488 cleanup [<resource id | stonith id>] [node=<node>] [operation=<opera‐
1489 tion> [interval=<interval>]] [--strict]
1490 This command is an alias of 'resource cleanup' command.
1491
1492 Make the cluster forget failed operations from history of the
1493 resource / stonith device and re-detect its current state. This
1494 can be useful to purge knowledge of past failures that have
1495 since been resolved.
1496
1497 If the named resource is part of a group, or one numbered in‐
1498 stance of a clone or bundled resource, the clean-up applies to
1499 the whole collective resource unless --strict is given.
1500
1501 If a resource id / stonith id is not specified then all re‐
1502 sources / stonith devices will be cleaned up.
1503
1504 If a node is not specified then resources / stonith devices on
1505 all nodes will be cleaned up.
1506
1507 refresh [<resource id | stonith id>] [node=<node>] [--strict]
1508 This command is an alias of 'resource refresh' command.
1509
1510 Make the cluster forget the complete operation history (includ‐
1511 ing failures) of the resource / stonith device and re-detect its
1512 current state. If you are interested in forgetting failed opera‐
1513 tions only, use the 'pcs resource cleanup' command.
1514
1515 If the named resource is part of a group, or one numbered in‐
1516 stance of a clone or bundled resource, the refresh applies to
1517 the whole collective resource unless --strict is given.
1518
1519 If a resource id / stonith id is not specified then all re‐
1520 sources / stonith devices will be refreshed.
1521
1522 If a node is not specified then resources / stonith devices on
1523 all nodes will be refreshed.
1524
1525 failcount [show [<resource id | stonith id>] [node=<node>] [opera‐
1526 tion=<operation> [interval=<interval>]]] [--full]
1527 This command is an alias of 'resource failcount show' command.
1528
1529 Show current failcount for resources and stonith devices, op‐
1530 tionally filtered by a resource / stonith device, node, opera‐
1531 tion and its interval. If --full is specified do not sum fail‐
1532 counts per resource / stonith device and node. Use 'pcs resource
1533 cleanup' or 'pcs resource refresh' to reset failcounts.
1534
1535 enable <stonith id>... [--wait[=n]]
1536 Allow the cluster to use the stonith devices. If --wait is spec‐
1537 ified, pcs will wait up to 'n' seconds for the stonith devices
1538 to start and then return 0 if the stonith devices are started,
1539 or 1 if the stonith devices have not yet started. If 'n' is not
1540 specified it defaults to 60 minutes.
1541
1542 disable <stonith id>... [--wait[=n]]
1543 Attempt to stop the stonith devices if they are running and dis‐
1544 allow the cluster to use them. If --wait is specified, pcs will
1545 wait up to 'n' seconds for the stonith devices to stop and then
1546 return 0 if the stonith devices are stopped or 1 if the stonith
1547 devices have not stopped. If 'n' is not specified it defaults to
1548 60 minutes.
1549
1550 level [config]
1551 Lists all of the fencing levels currently configured.
1552
1553 level add <level> <target> <stonith id> [stonith id]...
1554 Add the fencing level for the specified target with the list of
1555 stonith devices to attempt for that target at that level. Fence
1556 levels are attempted in numerical order (starting with 1). If a
1557 level succeeds (meaning all devices are successfully fenced in
1558 that level) then no other levels are tried, and the target is
1559 considered fenced. Target may be a node name <node_name> or
1560 %<node_name> or node%<node_name>, a node name regular expression
1561 regexp%<node_pattern> or a node attribute value at‐
1562 trib%<name>=<value>.
1563
1564 level delete <level> [target <target>] [stonith <stonith id>...]
1565 Removes the fence level for the level, target and/or devices
1566 specified. If no target or devices are specified then the fence
1567 level is removed. Target may be a node name <node_name> or
1568 %<node_name> or node%<node_name>, a node name regular expression
1569 regexp%<node_pattern> or a node attribute value at‐
1570 trib%<name>=<value>.
1571
1572 level remove <level> [target <target>] [stonith <stonith id>...]
1573 Removes the fence level for the level, target and/or devices
1574 specified. If no target or devices are specified then the fence
1575 level is removed. Target may be a node name <node_name> or
1576 %<node_name> or node%<node_name>, a node name regular expression
1577 regexp%<node_pattern> or a node attribute value at‐
1578 trib%<name>=<value>.
1579
1580 level clear [target <target> | stonith <stonith id>...]
1581 Clears the fence levels on the target (or stonith id) specified
1582 or clears all fence levels if a target/stonith id is not speci‐
1583 fied. Target may be a node name <node_name> or %<node_name> or
1584 node%<node_name>, a node name regular expression reg‐
1585 exp%<node_pattern> or a node attribute value at‐
1586 trib%<name>=<value>. Example: pcs stonith level clear stonith
1587 dev_a dev_b
1588
1589 level verify
1590 Verifies all fence devices and nodes specified in fence levels
1591 exist.
1592
1593 fence <node> [--off]
1594 Fence the node specified (if --off is specified, use the 'off'
1595 API call to stonith which will turn the node off instead of re‐
1596 booting it).
1597
1598 confirm <node> [--force]
1599 Confirm to the cluster that the specified node is powered off.
1600 This allows the cluster to recover from a situation where no
1601 stonith device is able to fence the node. This command should
1602 ONLY be used after manually ensuring that the node is powered
1603 off and has no access to shared resources.
1604
1605 WARNING: If this node is not actually powered off or it does
1606 have access to shared resources, data corruption/cluster failure
1607 can occur. To prevent accidental running of this command,
1608 --force or interactive user response is required in order to
1609 proceed.
1610
1611 NOTE: It is not checked if the specified node exists in the
1612 cluster in order to be able to work with nodes not visible from
1613 the local cluster partition.
1614
1615 history [show [<node>]]
1616 Show fencing history for the specified node or all nodes if no
1617 node specified.
1618
1619 history cleanup [<node>]
1620 Cleanup fence history of the specified node or all nodes if no
1621 node specified.
1622
1623 history update
1624 Update fence history from all nodes.
1625
1626 sbd enable [watchdog=<path>[@<node>]]... [device=<path>[@<node>]]...
1627 [<SBD_OPTION>=<value>]... [--no-watchdog-validation]
1628 Enable SBD in cluster. Default path for watchdog device is
1629 /dev/watchdog. Allowed SBD options: SBD_WATCHDOG_TIMEOUT (de‐
1630 fault: 5), SBD_DELAY_START (default: no), SBD_STARTMODE (de‐
1631 fault: always) and SBD_TIMEOUT_ACTION. SBD options are docu‐
1632 mented in sbd(8) man page. It is possible to specify up to 3 de‐
1633 vices per node. If --no-watchdog-validation is specified, vali‐
1634 dation of watchdogs will be skipped.
1635
1636 WARNING: Cluster has to be restarted in order to apply these
1637 changes.
1638
1639 WARNING: By default, it is tested whether the specified watchdog
1640 is supported. This may cause a restart of the system when a
1641 watchdog with no-way-out-feature enabled is present. Use
1642 --no-watchdog-validation to skip watchdog validation.
1643
1644 Example of enabling SBD in cluster with watchdogs on node1 will
1645 be /dev/watchdog2, on node2 /dev/watchdog1, /dev/watchdog0 on
1646 all other nodes, device /dev/sdb on node1, device /dev/sda on
1647 all other nodes and watchdog timeout will bet set to 10 seconds:
1648
1649 pcs stonith sbd enable watchdog=/dev/watchdog2@node1 watch‐
1650 dog=/dev/watchdog1@node2 watchdog=/dev/watchdog0 de‐
1651 vice=/dev/sdb@node1 device=/dev/sda SBD_WATCHDOG_TIMEOUT=10
1652
1653
1654 sbd disable
1655 Disable SBD in cluster.
1656
1657 WARNING: Cluster has to be restarted in order to apply these
1658 changes.
1659
1660 sbd device setup device=<path> [device=<path>]... [watchdog-time‐
1661 out=<integer>] [allocate-timeout=<integer>] [loop-timeout=<integer>]
1662 [msgwait-timeout=<integer>]
1663 Initialize SBD structures on device(s) with specified timeouts.
1664
1665 WARNING: All content on device(s) will be overwritten.
1666
1667 sbd device message <device-path> <node> <message-type>
1668 Manually set a message of the specified type on the device for
1669 the node. Possible message types (they are documented in sbd(8)
1670 man page): test, reset, off, crashdump, exit, clear
1671
1672 sbd status [--full]
1673 Show status of SBD services in cluster and local device(s) con‐
1674 figured. If --full is specified, also dump of SBD headers on de‐
1675 vice(s) will be shown.
1676
1677 sbd config
1678 Show SBD configuration in cluster.
1679
1680
1681 sbd watchdog list
1682 Show all available watchdog devices on the local node.
1683
1684 WARNING: Listing available watchdogs may cause a restart of the
1685 system when a watchdog with no-way-out-feature enabled is
1686 present.
1687
1688
1689 sbd watchdog test [<watchdog-path>]
1690 This operation is expected to force-reboot the local system
1691 without following any shutdown procedures using a watchdog. If
1692 no watchdog is specified, available watchdog will be used if
1693 only one watchdog device is available on the local system.
1694
1695
1696 acl
1697 [config]
1698 List all current access control lists.
1699
1700 enable Enable access control lists.
1701
1702 disable
1703 Disable access control lists.
1704
1705 role create <role id> [description=<description>] [((read | write |
1706 deny) (xpath <query> | id <id>))...]
1707 Create a role with the id and (optional) description specified.
1708 Each role can also have an unlimited number of permissions
1709 (read/write/deny) applied to either an xpath query or the id of
1710 a specific element in the cib.
1711 Permissions are applied to the selected XML element's entire XML
1712 subtree (all elements enclosed within it). Write permission
1713 grants the ability to create, modify, or remove the element and
1714 its subtree, and also the ability to create any "scaffolding"
1715 elements (enclosing elements that do not have attributes other
1716 than an ID). Permissions for more specific matches (more deeply
1717 nested elements) take precedence over more general ones. If mul‐
1718 tiple permissions are configured for the same match (for exam‐
1719 ple, in different roles applied to the same user), any deny per‐
1720 mission takes precedence, then write, then lastly read.
1721 An xpath may include an attribute expression to select only ele‐
1722 ments that match the expression, but the permission still ap‐
1723 plies to the entire element (and its subtree), not to the attri‐
1724 bute alone. For example, using the xpath "//*[@name]" to give
1725 write permission would allow changes to the entirety of all ele‐
1726 ments that have a "name" attribute and everything enclosed by
1727 those elements. There is no way currently to give permissions
1728 for just one attribute of an element. That is to say, you can
1729 not define an ACL that allows someone to read just the dc-uuid
1730 attribute of the cib tag - that would select the cib element and
1731 give read access to the entire CIB.
1732
1733 role delete <role id>
1734 Delete the role specified and remove it from any users/groups it
1735 was assigned to.
1736
1737 role remove <role id>
1738 Delete the role specified and remove it from any users/groups it
1739 was assigned to.
1740
1741 role assign <role id> [to] [user|group] <username/group>
1742 Assign a role to a user or group already created with 'pcs acl
1743 user/group create'. If there is user and group with the same id
1744 and it is not specified which should be used, user will be pri‐
1745 oritized. In cases like this specify whenever user or group
1746 should be used.
1747
1748 role unassign <role id> [from] [user|group] <username/group>
1749 Remove a role from the specified user. If there is user and
1750 group with the same id and it is not specified which should be
1751 used, user will be prioritized. In cases like this specify when‐
1752 ever user or group should be used.
1753
1754 user create <username> [<role id>]...
1755 Create an ACL for the user specified and assign roles to the
1756 user.
1757
1758 user delete <username>
1759 Remove the user specified (and roles assigned will be unassigned
1760 for the specified user).
1761
1762 user remove <username>
1763 Remove the user specified (and roles assigned will be unassigned
1764 for the specified user).
1765
1766 group create <group> [<role id>]...
1767 Create an ACL for the group specified and assign roles to the
1768 group.
1769
1770 group delete <group>
1771 Remove the group specified (and roles assigned will be unas‐
1772 signed for the specified group).
1773
1774 group remove <group>
1775 Remove the group specified (and roles assigned will be unas‐
1776 signed for the specified group).
1777
1778 permission add <role id> ((read | write | deny) (xpath <query> | id
1779 <id>))...
1780 Add the listed permissions to the role specified. Permissions
1781 are applied to either an xpath query or the id of a specific el‐
1782 ement in the CIB.
1783 Permissions are applied to the selected XML element's entire XML
1784 subtree (all elements enclosed within it). Write permission
1785 grants the ability to create, modify, or remove the element and
1786 its subtree, and also the ability to create any "scaffolding"
1787 elements (enclosing elements that do not have attributes other
1788 than an ID). Permissions for more specific matches (more deeply
1789 nested elements) take precedence over more general ones. If mul‐
1790 tiple permissions are configured for the same match (for exam‐
1791 ple, in different roles applied to the same user), any deny per‐
1792 mission takes precedence, then write, then lastly read.
1793 An xpath may include an attribute expression to select only ele‐
1794 ments that match the expression, but the permission still ap‐
1795 plies to the entire element (and its subtree), not to the attri‐
1796 bute alone. For example, using the xpath "//*[@name]" to give
1797 write permission would allow changes to the entirety of all ele‐
1798 ments that have a "name" attribute and everything enclosed by
1799 those elements. There is no way currently to give permissions
1800 for just one attribute of an element. That is to say, you can
1801 not define an ACL that allows someone to read just the dc-uuid
1802 attribute of the cib tag - that would select the cib element and
1803 give read access to the entire CIB.
1804
1805 permission delete <permission id>
1806 Remove the permission id specified (permission id's are listed
1807 in parenthesis after permissions in 'pcs acl' output).
1808
1809 permission remove <permission id>
1810 Remove the permission id specified (permission id's are listed
1811 in parenthesis after permissions in 'pcs acl' output).
1812
1813 property
1814 [config [<property> | --all | --defaults]] | [--all | --defaults]
1815 List property settings (default: lists configured properties).
1816 If --defaults is specified will show all property defaults, if
1817 --all is specified, current configured properties will be shown
1818 with unset properties and their defaults. See pacemaker-con‐
1819 trold(7) and pacemaker-schedulerd(7) man pages for a description
1820 of the properties.
1821
1822 set <property>=[<value>] ... [--force]
1823 Set specific pacemaker properties (if the value is blank then
1824 the property is removed from the configuration). If a property
1825 is not recognized by pcs the property will not be created unless
1826 the --force is used. See pacemaker-controld(7) and pacemaker-
1827 schedulerd(7) man pages for a description of the properties.
1828
1829 unset <property> ...
1830 Remove property from configuration. See pacemaker-controld(7)
1831 and pacemaker-schedulerd(7) man pages for a description of the
1832 properties.
1833
1834 constraint
1835 [config] [--full] [--all]
1836 List all current constraints that are not expired. If --all is
1837 specified also show expired constraints. If --full is specified
1838 also list the constraint ids.
1839
1840 location <resource> prefers <node>[=<score>] [<node>[=<score>]]...
1841 Create a location constraint on a resource to prefer the speci‐
1842 fied node with score (default score: INFINITY). Resource may be
1843 either a resource id <resource_id> or %<resource_id> or re‐
1844 source%<resource_id>, or a resource name regular expression reg‐
1845 exp%<resource_pattern>.
1846
1847 location <resource> avoids <node>[=<score>] [<node>[=<score>]]...
1848 Create a location constraint on a resource to avoid the speci‐
1849 fied node with score (default score: INFINITY). Resource may be
1850 either a resource id <resource_id> or %<resource_id> or re‐
1851 source%<resource_id>, or a resource name regular expression reg‐
1852 exp%<resource_pattern>.
1853
1854 location <resource> rule [id=<rule id>] [resource-discovery=<option>]
1855 [role=Promoted|Unpromoted] [constraint-id=<id>] [score=<score> |
1856 score-attribute=<attribute>] <expression>
1857 Creates a location constraint with a rule on the specified re‐
1858 source where expression looks like one of the following:
1859 defined|not_defined <node attribute>
1860 <node attribute> lt|gt|lte|gte|eq|ne [string|integer|num‐
1861 ber|version] <value>
1862 date gt|lt <date>
1863 date in_range <date> to <date>
1864 date in_range <date> to duration <duration options>...
1865 date-spec <date spec options>...
1866 <expression> and|or <expression>
1867 ( <expression> )
1868 where duration options and date spec options are: hours, month‐
1869 days, weekdays, yeardays, months, weeks, years, weekyears, moon.
1870 Resource may be either a resource id <resource_id> or %<re‐
1871 source_id> or resource%<resource_id>, or a resource name regular
1872 expression regexp%<resource_pattern>. If score is omitted it de‐
1873 faults to INFINITY. If id is omitted one is generated from the
1874 resource id. If resource-discovery is omitted it defaults to
1875 'always'.
1876
1877 location [config [resources [<resource>...]] | [nodes [<node>...]]]
1878 [--full] [--all]
1879 List all the current location constraints that are not expired.
1880 If 'resources' is specified, location constraints are displayed
1881 per resource (default). If 'nodes' is specified, location con‐
1882 straints are displayed per node. If specific nodes or resources
1883 are specified then we only show information about them. Resource
1884 may be either a resource id <resource_id> or %<resource_id> or
1885 resource%<resource_id>, or a resource name regular expression
1886 regexp%<resource_pattern>. If --full is specified show the in‐
1887 ternal constraint id's as well. If --all is specified show the
1888 expired constraints.
1889
1890 location add <id> <resource> <node> <score> [resource-discovery=<op‐
1891 tion>]
1892 Add a location constraint with the appropriate id for the speci‐
1893 fied resource, node name and score. Resource may be either a re‐
1894 source id <resource_id> or %<resource_id> or resource%<re‐
1895 source_id>, or a resource name regular expression regexp%<re‐
1896 source_pattern>.
1897
1898 location delete <id>
1899 Remove a location constraint with the appropriate id.
1900
1901 location remove <id>
1902 Remove a location constraint with the appropriate id.
1903
1904 order [config] [--full]
1905 List all current ordering constraints (if --full is specified
1906 show the internal constraint id's as well).
1907
1908 order [action] <resource id> then [action] <resource id> [options]
1909 Add an ordering constraint specifying actions (start, stop, pro‐
1910 mote, demote) and if no action is specified the default action
1911 will be start. Available options are kind=Optional/Manda‐
1912 tory/Serialize, symmetrical=true/false, require-all=true/false
1913 and id=<constraint-id>.
1914
1915 order set <resource1> [resourceN]... [options] [set <resourceX> ...
1916 [options]] [setoptions [constraint_options]]
1917 Create an ordered set of resources. Available options are se‐
1918 quential=true/false, require-all=true/false and ac‐
1919 tion=start/promote/demote/stop. Available constraint_options are
1920 id=<constraint-id>, kind=Optional/Mandatory/Serialize and sym‐
1921 metrical=true/false.
1922
1923 order delete <resource1> [resourceN]...
1924 Remove resource from any ordering constraint
1925
1926 order remove <resource1> [resourceN]...
1927 Remove resource from any ordering constraint
1928
1929 colocation [config] [--full]
1930 List all current colocation constraints (if --full is specified
1931 show the internal constraint id's as well).
1932
1933 colocation add [<role>] <source resource id> with [<role>] <target re‐
1934 source id> [score] [options] [id=constraint-id]
1935 Request <source resource> to run on the same node where pace‐
1936 maker has determined <target resource> should run. Positive
1937 values of score mean the resources should be run on the same
1938 node, negative values mean the resources should not be run on
1939 the same node. Specifying 'INFINITY' (or '-INFINITY') for the
1940 score forces <source resource> to run (or not run) with <target
1941 resource> (score defaults to "INFINITY"). A role can be: 'Pro‐
1942 moted', 'Unpromoted', 'Started', 'Stopped' (if no role is speci‐
1943 fied, it defaults to 'Started').
1944
1945 colocation set <resource1> [resourceN]... [options] [set <resourceX>
1946 ... [options]] [setoptions [constraint_options]]
1947 Create a colocation constraint with a resource set. Available
1948 options are sequential=true/false and role=Stopped/Started/Pro‐
1949 moted/Unpromoted. Available constraint_options are id and either
1950 of: score, score-attribute, score-attribute-mangle.
1951
1952 colocation delete <source resource id> <target resource id>
1953 Remove colocation constraints with specified resources.
1954
1955 colocation remove <source resource id> <target resource id>
1956 Remove colocation constraints with specified resources.
1957
1958 ticket [config] [--full]
1959 List all current ticket constraints (if --full is specified show
1960 the internal constraint id's as well).
1961
1962 ticket add <ticket> [<role>] <resource id> [<options>] [id=<con‐
1963 straint-id>]
1964 Create a ticket constraint for <resource id>. Available option
1965 is loss-policy=fence/stop/freeze/demote. A role can be Promoted,
1966 Unpromoted, Started or Stopped.
1967
1968 ticket set <resource1> [<resourceN>]... [<options>] [set <resourceX>
1969 ... [<options>]] setoptions <constraint_options>
1970 Create a ticket constraint with a resource set. Available op‐
1971 tions are role=Stopped/Started/Promoted/Unpromoted. Required
1972 constraint option is ticket=<ticket>. Optional constraint op‐
1973 tions are id=<constraint-id> and loss-pol‐
1974 icy=fence/stop/freeze/demote.
1975
1976 ticket delete <ticket> <resource id>
1977 Remove all ticket constraints with <ticket> from <resource id>.
1978
1979 ticket remove <ticket> <resource id>
1980 Remove all ticket constraints with <ticket> from <resource id>.
1981
1982 delete <constraint id>...
1983 Remove constraint(s) or constraint rules with the specified
1984 id(s).
1985
1986 remove <constraint id>...
1987 Remove constraint(s) or constraint rules with the specified
1988 id(s).
1989
1990 ref <resource>...
1991 List constraints referencing specified resource.
1992
1993 rule add <constraint id> [id=<rule id>] [role=Promoted|Unpromoted]
1994 [score=<score>|score-attribute=<attribute>] <expression>
1995 Add a rule to a location constraint specified by 'constraint id'
1996 where the expression looks like one of the following:
1997 defined|not_defined <node attribute>
1998 <node attribute> lt|gt|lte|gte|eq|ne [string|integer|num‐
1999 ber|version] <value>
2000 date gt|lt <date>
2001 date in_range <date> to <date>
2002 date in_range <date> to duration <duration options>...
2003 date-spec <date spec options>...
2004 <expression> and|or <expression>
2005 ( <expression> )
2006 where duration options and date spec options are: hours, month‐
2007 days, weekdays, yeardays, months, weeks, years, weekyears, moon.
2008 If score is omitted it defaults to INFINITY. If id is omitted
2009 one is generated from the constraint id.
2010
2011 rule delete <rule id>
2012 Remove a rule from its location constraint and if it's the last
2013 rule, the constraint will also be removed.
2014
2015 rule remove <rule id>
2016 Remove a rule from its location constraint and if it's the last
2017 rule, the constraint will also be removed.
2018
2019 qdevice
2020 status <device model> [--full] [<cluster name>]
2021 Show runtime status of specified model of quorum device
2022 provider. Using --full will give more detailed output. If
2023 <cluster name> is specified, only information about the speci‐
2024 fied cluster will be displayed.
2025
2026 setup model <device model> [--enable] [--start]
2027 Configure specified model of quorum device provider. Quorum de‐
2028 vice then can be added to clusters by running "pcs quorum device
2029 add" command in a cluster. --start will also start the
2030 provider. --enable will configure the provider to start on
2031 boot.
2032
2033 destroy <device model>
2034 Disable and stop specified model of quorum device provider and
2035 delete its configuration files.
2036
2037 start <device model>
2038 Start specified model of quorum device provider.
2039
2040 stop <device model>
2041 Stop specified model of quorum device provider.
2042
2043 kill <device model>
2044 Force specified model of quorum device provider to stop (per‐
2045 forms kill -9). Note that init system (e.g. systemd) can detect
2046 that the qdevice is not running and start it again. If you want
2047 to stop the qdevice, run "pcs qdevice stop" command.
2048
2049 enable <device model>
2050 Configure specified model of quorum device provider to start on
2051 boot.
2052
2053 disable <device model>
2054 Configure specified model of quorum device provider to not start
2055 on boot.
2056
2057 quorum
2058 [config]
2059 Show quorum configuration.
2060
2061 status Show quorum runtime status.
2062
2063 device add [<generic options>] model <device model> [<model options>]
2064 [heuristics <heuristics options>]
2065 Add a quorum device to the cluster. Quorum device should be con‐
2066 figured first with "pcs qdevice setup". It is not possible to
2067 use more than one quorum device in a cluster simultaneously.
2068 Currently the only supported model is 'net'. It requires model
2069 options 'algorithm' and 'host' to be specified. Options are doc‐
2070 umented in corosync-qdevice(8) man page; generic options are
2071 'sync_timeout' and 'timeout', for model net options check the
2072 quorum.device.net section, for heuristics options see the quo‐
2073 rum.device.heuristics section. Pcs automatically creates and
2074 distributes TLS certificates and sets the 'tls' model option to
2075 the default value 'on'.
2076 Example: pcs quorum device add model net algorithm=lms
2077 host=qnetd.internal.example.com
2078
2079 device heuristics delete
2080 Remove all heuristics settings of the configured quorum device.
2081
2082 device heuristics remove
2083 Remove all heuristics settings of the configured quorum device.
2084
2085 device delete
2086 Remove a quorum device from the cluster.
2087
2088 device remove
2089 Remove a quorum device from the cluster.
2090
2091 device status [--full]
2092 Show quorum device runtime status. Using --full will give more
2093 detailed output.
2094
2095 device update [<generic options>] [model <model options>] [heuristics
2096 <heuristics options>]
2097 Add, remove or change quorum device options. Unspecified options
2098 will be kept unchanged. If you wish to remove an option, set it
2099 to empty value, i.e. 'option_name='. Requires the cluster to be
2100 stopped. Model and options are all documented in corosync-qde‐
2101 vice(8) man page; for heuristics options check the quorum.de‐
2102 vice.heuristics subkey section, for model options check the quo‐
2103 rum.device.<device model> subkey sections.
2104
2105 WARNING: If you want to change "host" option of qdevice model
2106 net, use "pcs quorum device remove" and "pcs quorum device add"
2107 commands to set up configuration properly unless old and new
2108 host is the same machine.
2109
2110 expected-votes <votes>
2111 Set expected votes in the live cluster to specified value. This
2112 only affects the live cluster, not changes any configuration
2113 files.
2114
2115 unblock [--force]
2116 Cancel waiting for all nodes when establishing quorum. Useful
2117 in situations where you know the cluster is inquorate, but you
2118 are confident that the cluster should proceed with resource man‐
2119 agement regardless. This command should ONLY be used when nodes
2120 which the cluster is waiting for have been confirmed to be pow‐
2121 ered off and to have no access to shared resources.
2122
2123 WARNING: If the nodes are not actually powered off or they do
2124 have access to shared resources, data corruption/cluster failure
2125 can occur. To prevent accidental running of this command,
2126 --force or interactive user response is required in order to
2127 proceed.
2128
2129 update [auto_tie_breaker=[0|1]] [last_man_standing=[0|1]]
2130 [last_man_standing_window=[<time in ms>]] [wait_for_all=[0|1]]
2131 Add, remove or change quorum options. At least one option must
2132 be specified. Unspecified options will be kept unchanged. If you
2133 wish to remove an option, set it to empty value, i.e. 'op‐
2134 tion_name='. Options are documented in corosync's votequorum(5)
2135 man page. Requires the cluster to be stopped.
2136
2137 booth
2138 setup sites <address> <address> [<address>...] [arbitrators <address>
2139 ...] [--force]
2140 Write new booth configuration with specified sites and arbitra‐
2141 tors. Total number of peers (sites and arbitrators) must be
2142 odd. When the configuration file already exists, command fails
2143 unless --force is specified.
2144
2145 destroy
2146 Remove booth configuration files.
2147
2148 ticket add <ticket> [<name>=<value> ...]
2149 Add new ticket to the current configuration. Ticket options are
2150 specified in booth manpage.
2151
2152 ticket delete <ticket>
2153 Remove the specified ticket from the current configuration.
2154
2155 ticket remove <ticket>
2156 Remove the specified ticket from the current configuration.
2157
2158 config [<node>]
2159 Show booth configuration from the specified node or from the
2160 current node if node not specified.
2161
2162 create ip <address>
2163 Make the cluster run booth service on the specified ip address
2164 as a cluster resource. Typically this is used to run booth
2165 site.
2166
2167 delete Remove booth resources created by the "pcs booth create" com‐
2168 mand.
2169
2170 remove Remove booth resources created by the "pcs booth create" com‐
2171 mand.
2172
2173 restart
2174 Restart booth resources created by the "pcs booth create" com‐
2175 mand.
2176
2177 ticket grant <ticket> [<site address>]
2178 Grant the ticket to the site specified by the address, hence to
2179 the booth formation this site is a member of. When this specifi‐
2180 cation is omitted, site address that has been specified with
2181 'pcs booth create' command is used. Specifying site address is
2182 therefore mandatory when running this command at a host in an
2183 arbitrator role.
2184 Note that the ticket must not be already granted in given booth
2185 formation; for an ad-hoc (and, in the worst case, abrupt, for a
2186 lack of a direct atomicity) change of this preference baring di‐
2187 rect interventions at the sites, the ticket needs to be revoked
2188 first, only then it can be granted at another site again.
2189
2190 ticket revoke <ticket> [<site address>]
2191 Revoke the ticket in the booth formation as identified with one
2192 of its member sites specified by the address. When this specifi‐
2193 cation is omitted, site address that has been specified with a
2194 prior 'pcs booth create' command is used. Specifying site ad‐
2195 dress is therefore mandatory when running this command at a host
2196 in an arbitrator role.
2197
2198 status Print current status of booth on the local node.
2199
2200 pull <node>
2201 Pull booth configuration from the specified node.
2202
2203 sync [--skip-offline]
2204 Send booth configuration from the local node to all nodes in the
2205 cluster.
2206
2207 enable Enable booth arbitrator service.
2208
2209 disable
2210 Disable booth arbitrator service.
2211
2212 start Start booth arbitrator service.
2213
2214 stop Stop booth arbitrator service.
2215
2216 status
2217 [status] [--full] [--hide-inactive]
2218 View all information about the cluster and resources (--full
2219 provides more details, --hide-inactive hides inactive re‐
2220 sources).
2221
2222 resources [<resource id | tag id>] [node=<node>] [--hide-inactive]
2223 Show status of all currently configured resources. If --hide-in‐
2224 active is specified, only show active resources. If a resource
2225 or tag id is specified, only show status of the specified re‐
2226 source or resources in the specified tag. If node is specified,
2227 only show status of resources configured for the specified node.
2228
2229 cluster
2230 View current cluster status.
2231
2232 corosync
2233 View current membership information as seen by corosync.
2234
2235 quorum View current quorum status.
2236
2237 qdevice <device model> [--full] [<cluster name>]
2238 Show runtime status of specified model of quorum device
2239 provider. Using --full will give more detailed output. If
2240 <cluster name> is specified, only information about the speci‐
2241 fied cluster will be displayed.
2242
2243 booth Print current status of booth on the local node.
2244
2245 nodes [corosync | both | config]
2246 View current status of nodes from pacemaker. If 'corosync' is
2247 specified, view current status of nodes from corosync instead.
2248 If 'both' is specified, view current status of nodes from both
2249 corosync & pacemaker. If 'config' is specified, print nodes from
2250 corosync & pacemaker configuration.
2251
2252 pcsd [<node>]...
2253 Show current status of pcsd on nodes specified, or on all nodes
2254 configured in the local cluster if no nodes are specified.
2255
2256 xml View xml version of status (output from crm_mon -r -1 -X).
2257
2258 config
2259 [show] View full cluster configuration.
2260
2261 backup [filename]
2262 Creates the tarball containing the cluster configuration files.
2263 If filename is not specified the standard output will be used.
2264
2265 restore [--local] [filename]
2266 Restores the cluster configuration files on all nodes from the
2267 backup. If filename is not specified the standard input will be
2268 used. If --local is specified only the files on the current
2269 node will be restored.
2270
2271 checkpoint
2272 List all available configuration checkpoints.
2273
2274 checkpoint view <checkpoint_number>
2275 Show specified configuration checkpoint.
2276
2277 checkpoint diff <checkpoint_number> <checkpoint_number>
2278 Show differences between the two specified checkpoints. Use
2279 checkpoint number 'live' to compare a checkpoint to the current
2280 live configuration.
2281
2282 checkpoint restore <checkpoint_number>
2283 Restore cluster configuration to specified checkpoint.
2284
2285 pcsd
2286 certkey <certificate file> <key file>
2287 Load custom certificate and key files for use in pcsd.
2288
2289 status [<node>]...
2290 Show current status of pcsd on nodes specified, or on all nodes
2291 configured in the local cluster if no nodes are specified.
2292
2293 sync-certificates
2294 Sync pcsd certificates to all nodes in the local cluster.
2295
2296 deauth [<token>]...
2297 Delete locally stored authentication tokens used by remote sys‐
2298 tems to connect to the local pcsd instance. If no tokens are
2299 specified all tokens will be deleted. After this command is run
2300 other nodes will need to re-authenticate against this node to be
2301 able to connect to it.
2302
2303 host
2304 auth (<host name> [addr=<address>[:<port>]])... [-u <username>] [-p
2305 <password>]
2306 Authenticate local pcs/pcsd against pcsd on specified hosts. It
2307 is possible to specify an address and a port via which pcs/pcsd
2308 will communicate with each host. If an address is not specified
2309 a host name will be used. If a port is not specified 2224 will
2310 be used.
2311
2312 deauth [<host name>]...
2313 Delete authentication tokens which allow pcs/pcsd on the current
2314 system to connect to remote pcsd instances on specified host
2315 names. If the current system is a member of a cluster, the to‐
2316 kens will be deleted from all nodes in the cluster. If no host
2317 names are specified all tokens will be deleted. After this com‐
2318 mand is run this node will need to re-authenticate against other
2319 nodes to be able to connect to them.
2320
2321 node
2322 attribute [[<node>] [--name <name>] | <node> <name>=<value> ...]
2323 Manage node attributes. If no parameters are specified, show
2324 attributes of all nodes. If one parameter is specified, show
2325 attributes of specified node. If --name is specified, show
2326 specified attribute's value from all nodes. If more parameters
2327 are specified, set attributes of specified node. Attributes can
2328 be removed by setting an attribute without a value.
2329
2330 maintenance [--all | <node>...] [--wait[=n]]
2331 Put specified node(s) into maintenance mode, if no nodes or op‐
2332 tions are specified the current node will be put into mainte‐
2333 nance mode, if --all is specified all nodes will be put into
2334 maintenance mode. If --wait is specified, pcs will wait up to
2335 'n' seconds for the node(s) to be put into maintenance mode and
2336 then return 0 on success or 1 if the operation not succeeded
2337 yet. If 'n' is not specified it defaults to 60 minutes.
2338
2339 unmaintenance [--all | <node>...] [--wait[=n]]
2340 Remove node(s) from maintenance mode, if no nodes or options are
2341 specified the current node will be removed from maintenance
2342 mode, if --all is specified all nodes will be removed from main‐
2343 tenance mode. If --wait is specified, pcs will wait up to 'n'
2344 seconds for the node(s) to be removed from maintenance mode and
2345 then return 0 on success or 1 if the operation not succeeded
2346 yet. If 'n' is not specified it defaults to 60 minutes.
2347
2348 standby [--all | <node>...] [--wait[=n]]
2349 Put specified node(s) into standby mode (the node specified will
2350 no longer be able to host resources), if no nodes or options are
2351 specified the current node will be put into standby mode, if
2352 --all is specified all nodes will be put into standby mode. If
2353 --wait is specified, pcs will wait up to 'n' seconds for the
2354 node(s) to be put into standby mode and then return 0 on success
2355 or 1 if the operation not succeeded yet. If 'n' is not specified
2356 it defaults to 60 minutes.
2357
2358 unstandby [--all | <node>...] [--wait[=n]]
2359 Remove node(s) from standby mode (the node specified will now be
2360 able to host resources), if no nodes or options are specified
2361 the current node will be removed from standby mode, if --all is
2362 specified all nodes will be removed from standby mode. If --wait
2363 is specified, pcs will wait up to 'n' seconds for the node(s) to
2364 be removed from standby mode and then return 0 on success or 1
2365 if the operation not succeeded yet. If 'n' is not specified it
2366 defaults to 60 minutes.
2367
2368 utilization [[<node>] [--name <name>] | <node> <name>=<value> ...]
2369 Add specified utilization options to specified node. If node is
2370 not specified, shows utilization of all nodes. If --name is
2371 specified, shows specified utilization value from all nodes. If
2372 utilization options are not specified, shows utilization of
2373 specified node. Utilization option should be in format
2374 name=value, value has to be integer. Options may be removed by
2375 setting an option without a value. Example: pcs node utiliza‐
2376 tion node1 cpu=4 ram=
2377
2378 alert
2379 [config]
2380 Show all configured alerts.
2381
2382 create path=<path> [id=<alert-id>] [description=<description>] [options
2383 [<option>=<value>]...] [meta [<meta-option>=<value>]...]
2384 Define an alert handler with specified path. Id will be automat‐
2385 ically generated if it is not specified.
2386
2387 update <alert-id> [path=<path>] [description=<description>] [options
2388 [<option>=<value>]...] [meta [<meta-option>=<value>]...]
2389 Update an existing alert handler with specified id. Unspecified
2390 options will be kept unchanged. If you wish to remove an option,
2391 set it to empty value, i.e. 'option_name='.
2392
2393 delete <alert-id> ...
2394 Remove alert handlers with specified ids.
2395
2396 remove <alert-id> ...
2397 Remove alert handlers with specified ids.
2398
2399 recipient add <alert-id> value=<recipient-value> [id=<recipient-id>]
2400 [description=<description>] [options [<option>=<value>]...] [meta
2401 [<meta-option>=<value>]...]
2402 Add new recipient to specified alert handler.
2403
2404 recipient update <recipient-id> [value=<recipient-value>] [descrip‐
2405 tion=<description>] [options [<option>=<value>]...] [meta [<meta-op‐
2406 tion>=<value>]...]
2407 Update an existing recipient identified by its id. Unspecified
2408 options will be kept unchanged. If you wish to remove an option,
2409 set it to empty value, i.e. 'option_name='.
2410
2411 recipient delete <recipient-id> ...
2412 Remove specified recipients.
2413
2414 recipient remove <recipient-id> ...
2415 Remove specified recipients.
2416
2417 client
2418 local-auth [<pcsd-port>] [-u <username>] [-p <password>]
2419 Authenticate current user to local pcsd. This is required to run
2420 some pcs commands which may require permissions of root user
2421 such as 'pcs cluster start'.
2422
2423 dr
2424 config Display disaster-recovery configuration from the local node.
2425
2426 status [--full] [--hide-inactive]
2427 Display status of the local and the remote site cluster (--full
2428 provides more details, --hide-inactive hides inactive re‐
2429 sources).
2430
2431 set-recovery-site <recovery site node>
2432 Set up disaster-recovery with the local cluster being the pri‐
2433 mary site. The recovery site is defined by a name of one of its
2434 nodes.
2435
2436 destroy
2437 Permanently destroy disaster-recovery configuration on all
2438 sites.
2439
2440 tag
2441 [config|list [<tag id>...]]
2442 Display configured tags.
2443
2444 create <tag id> <id> [<id>]...
2445 Create a tag containing the specified ids.
2446
2447 delete <tag id>...
2448 Delete specified tags.
2449
2450 remove <tag id>...
2451 Delete specified tags.
2452
2453 update <tag id> [add <id> [<id>]... [--before <id> | --after <id>]]
2454 [remove <id> [<id>]...]
2455 Update a tag using the specified ids. Ids can be added, removed
2456 or moved in a tag. You can use --before or --after to specify
2457 the position of the added ids relatively to some id already ex‐
2458 isting in the tag. By adding ids to a tag they are already in
2459 and specifying --after or --before you can move the ids in the
2460 tag.
2461
2463 Show all resources
2464 # pcs resource config
2465
2466 Show options specific to the 'VirtualIP' resource
2467 # pcs resource config VirtualIP
2468
2469 Create a new resource called 'VirtualIP' with options
2470 # pcs resource create VirtualIP ocf:heartbeat:IPaddr2
2471 ip=192.168.0.99 cidr_netmask=32 nic=eth2 op monitor interval=30s
2472
2473 Create a new resource called 'VirtualIP' with options
2474 # pcs resource create VirtualIP IPaddr2 ip=192.168.0.99
2475 cidr_netmask=32 nic=eth2 op monitor interval=30s
2476
2477 Change the ip address of VirtualIP and remove the nic option
2478 # pcs resource update VirtualIP ip=192.168.0.98 nic=
2479
2480 Delete the VirtualIP resource
2481 # pcs resource delete VirtualIP
2482
2483 Create the MyStonith stonith fence_virt device which can fence host
2484 'f1'
2485 # pcs stonith create MyStonith fence_virt pcmk_host_list=f1
2486
2487 Set the stonith-enabled property to false on the cluster (which dis‐
2488 ables stonith)
2489 # pcs property set stonith-enabled=false
2490
2492 Various pcs commands accept the --force option. Its purpose is to over‐
2493 ride some of checks that pcs is doing or some of errors that may occur
2494 when a pcs command is run. When such error occurs, pcs will print the
2495 error with a note it may be overridden. The exact behavior of the op‐
2496 tion is different for each pcs command. Using the --force option can
2497 lead into situations that would normally be prevented by logic of pcs
2498 commands and therefore its use is strongly discouraged unless you know
2499 what you are doing.
2500
2502 EDITOR
2503 Path to a plain-text editor. This is used when pcs is requested
2504 to present a text for the user to edit.
2505
2506 no_proxy, https_proxy, all_proxy, NO_PROXY, HTTPS_PROXY, ALL_PROXY
2507 These environment variables (listed according to their priori‐
2508 ties) control how pcs handles proxy servers when connecting to
2509 cluster nodes. See curl(1) man page for details.
2510
2512 This section summarizes the most important changes in commands done in
2513 pcs-0.11.x compared to pcs-0.10.x. For detailed description of current
2514 commands see above.
2515
2516 Legacy role names
2517 Roles 'Master' and 'Slave' are deprecated and should not be used any‐
2518 more. Instead use 'Promoted' and 'Unpromoted' respectively. Similarly,
2519 --master has been deprecated and replaced with --promoted.
2520
2521 cluster
2522 uidgid rm
2523 This command has been replaced with 'pcs cluster uidgid delete'
2524 and 'pcs cluster uidgid remove'.
2525
2526 resource
2527 move The 'pcs resource move' now automatically removes location con‐
2528 straint used for moving a resource. It is equivalent of 'pcs re‐
2529 source move --autodelete' from pcs-0.10.9. Legacy functionality
2530 of the 'resource move' command is still available as 'resource
2531 move-with-constraint <resource id>'.
2532
2533 show --full
2534 This command has been replaced with 'pcs resource config'.
2535
2536 show --groups
2537 This command has been replaced with 'pcs resource group list'.
2538
2539 show This command has been replaced with 'pcs resource status'.
2540
2541 stonith
2542 show --full
2543 This command has been replaced with 'pcs stonith config'.
2544
2545 show This command has been replaced with 'pcs stonith status'.
2546
2548 This section summarizes the most important changes in commands done in
2549 pcs-0.10.x compared to pcs-0.9.x. For detailed description of current
2550 commands see above.
2551
2552 acl
2553 show The 'pcs acl show' command has been deprecated and will be re‐
2554 moved. Please use 'pcs acl config' instead. Applicable in
2555 pcs-0.10.9 and newer.
2556
2557 alert
2558 show The 'pcs alert show' command has been deprecated and will be re‐
2559 moved. Please use 'pcs alert config' instead. Applicable in
2560 pcs-0.10.9 and newer.
2561
2562 cluster
2563 auth The 'pcs cluster auth' command only authenticates nodes in a lo‐
2564 cal cluster and does not accept a node list. The new command for
2565 authentication is 'pcs host auth'. It allows one to specify host
2566 names, addresses and pcsd ports.
2567
2568 node add
2569 Custom node names and Corosync 3.x with knet are fully supported
2570 now, therefore the syntax has been completely changed.
2571 The --device and --watchdog options have been replaced with 'de‐
2572 vice' and 'watchdog' options, respectively.
2573
2574 pcsd-status
2575 The 'pcs cluster pcsd-status' command has been deprecated and
2576 will be removed. Please use 'pcs pcsd status' or 'pcs status
2577 pcsd' instead. Applicable in pcs-0.10.9 and newer.
2578
2579 quorum This command has been replaced with 'pcs quorum'.
2580
2581 remote-node add
2582 This command has been replaced with 'pcs cluster node
2583 add-guest'.
2584
2585 remote-node remove
2586 This command has been replaced with 'pcs cluster node
2587 delete-guest' and its alias 'pcs cluster node remove-guest'.
2588
2589 setup Custom node names and Corosync 3.x with knet are fully supported
2590 now, therefore the syntax has been completely changed.
2591 The --name option has been removed. The first parameter of the
2592 command is the cluster name now.
2593 The --local option has been replaced with --corosync_conf
2594 <path>.
2595
2596 standby
2597 This command has been replaced with 'pcs node standby'.
2598
2599 uidgid rm
2600 This command has been deprecated, use 'pcs cluster uidgid
2601 delete' or 'pcs cluster uidgid remove' instead.
2602
2603 unstandby
2604 This command has been replaced with 'pcs node unstandby'.
2605
2606 verify The -V option has been replaced with --full.
2607 To specify a filename, use the -f option.
2608
2609 constraint
2610 list The 'pcs constraint list' command, as well as its variants 'pcs
2611 constraint [location | colocation | order | ticket] list', has
2612 been deprecated and will be removed. Please use 'pcs constraint
2613 [location | colocation | order | ticket] config' instead. Appli‐
2614 cable in pcs-0.10.9 and newer.
2615
2616 show The 'pcs constraint show' command, as well as its variants 'pcs
2617 constraint [location | colocation | order | ticket] show', has
2618 been deprecated and will be removed. Please use 'pcs constraint
2619 [location | colocation | order | ticket] config' instead. Appli‐
2620 cable in pcs-0.10.9 and newer.
2621
2622 pcsd
2623 clear-auth
2624 This command has been replaced with 'pcs host deauth' and 'pcs
2625 pcsd deauth'.
2626
2627 property
2628 list The 'pcs property list' command has been deprecated and will be
2629 removed. Please use 'pcs property config' instead. Applicable in
2630 pcs-0.10.9 and newer.
2631
2632 set The --node option is no longer supported. Use the 'pcs node at‐
2633 tribute' command to set node attributes.
2634
2635 show The --node option is no longer supported. Use the 'pcs node at‐
2636 tribute' command to view node attributes.
2637 The 'pcs property show' command has been deprecated and will be
2638 removed. Please use 'pcs property config' instead. Applicable in
2639 pcs-0.10.9 and newer.
2640
2641 unset The --node option is no longer supported. Use the 'pcs node at‐
2642 tribute' command to unset node attributes.
2643
2644 resource
2645 create The 'master' keyword has been changed to 'promotable'.
2646
2647 failcount reset
2648 The command has been removed as 'pcs resource cleanup' is doing
2649 exactly the same job.
2650
2651 master This command has been replaced with 'pcs resource promotable'.
2652
2653 show Previously, this command displayed either status or configura‐
2654 tion of resources depending on the parameters specified. This
2655 was confusing, therefore the command was replaced by several new
2656 commands. To display resources status, run 'pcs resource' or
2657 'pcs resource status'. To display resources configuration, run
2658 'pcs resource config' or 'pcs resource config <resource name>'.
2659 To display configured resource groups, run 'pcs resource group
2660 list'.
2661
2662 status
2663 groups This command has been replaced with 'pcs resource group list'.
2664
2665 stonith
2666 level add | clear | delete | remove
2667 Delimiting stonith devices with a comma is deprecated, use a
2668 space instead. Applicable in pcs-0.10.9 and newer.
2669
2670 level clear
2671 Syntax of the command has been fixed so that it is not ambiguous
2672 any more. New syntax is 'pcs stonith level clear [target <tar‐
2673 get> | stonith <stonith id>...]'. Old syntax 'pcs stonith level
2674 clear [<target> | <stonith ids>]' is deprecated but still func‐
2675 tional in pcs-0.10.x. Applicable in pcs-0.10.9 and newer.
2676
2677 level delete | remove
2678 Syntax of the command has been fixed so that it is not ambiguous
2679 any more. New syntax is 'pcs stonith level delete | remove [tar‐
2680 get <target>] [stonith <stonith id>]...'. Old syntax 'pcs
2681 stonith level delete | remove [<target>] [<stonith id>]...' is
2682 deprecated but still functional in pcs-0.10.x. Applicable in
2683 pcs-0.10.9 and newer.
2684
2685 sbd device setup
2686 The --device option has been replaced with the 'device' option.
2687
2688 sbd enable
2689 The --device and --watchdog options have been replaced with 'de‐
2690 vice' and 'watchdog' options, respectively.
2691
2692 show Previously, this command displayed either status or configura‐
2693 tion of stonith resources depending on the parameters specified.
2694 This was confusing, therefore the command was replaced by sev‐
2695 eral new commands. To display stonith resources status, run 'pcs
2696 stonith' or 'pcs stonith status'. To display stonith resources
2697 configuration, run 'pcs stonith config' or 'pcs stonith config
2698 <stonith name>'.
2699
2700 tag
2701 list The 'pcs tag list' command has been deprecated and will be re‐
2702 moved. Please use 'pcs tag config' instead. Applicable in
2703 pcs-0.10.9 and newer.
2704
2706 http://clusterlabs.org/doc/
2707
2708 pcsd(8), pcs_snmp_agent(8)
2709
2710 corosync_overview(8), votequorum(5), corosync.conf(5), corosync-qde‐
2711 vice(8), corosync-qdevice-tool(8), corosync-qnetd(8),
2712 corosync-qnetd-tool(8)
2713
2714 pacemaker-controld(7), pacemaker-fenced(7), pacemaker-schedulerd(7),
2715 crm_mon(8), crm_report(8), crm_simulate(8)
2716
2717 boothd(8), sbd(8)
2718
2719
2720
2721pcs 0.11.3 2022-06-23 PCS(8)