1LVMLOCKD(8) LVMLOCKD(8)
2
3
4
6 lvmlockd — LVM locking daemon
7
8
10 LVM commands use lvmlockd to coordinate access to shared storage.
11 When LVM is used on devices shared by multiple hosts, locks will:
12
13 · coordinate reading and writing of LVM metadata
14 · validate caching of LVM metadata
15 · prevent conflicting activation of logical volumes
16
17 lvmlockd uses an external lock manager to perform basic locking.
18 Lock manager (lock type) options are:
19
20 · sanlock: places locks on disk within LVM storage.
21 · dlm: uses network communication and a cluster manager.
22
23
25 lvmlockd [options]
26
27 For default settings, see lvmlockd -h.
28
29 --help | -h
30 Show this help information.
31
32 --version | -V
33 Show version of lvmlockd.
34
35 --test | -T
36 Test mode, do not call lock manager.
37
38 --foreground | -f
39 Don't fork.
40
41 --daemon-debug | -D
42 Don't fork and print debugging to stdout.
43
44 --pid-file | -p path
45 Set path to the pid file.
46
47 --socket-path | -s path
48 Set path to the socket to listen on.
49
50 --syslog-priority | -S err|warning|debug
51 Write log messages from this level up to syslog.
52
53 --gl-type | -g sanlock|dlm
54 Set global lock type to be sanlock or dlm.
55
56 --host-id | -i num
57 Set the local sanlock host id.
58
59 --host-id-file | -F path
60 A file containing the local sanlock host_id.
61
62 --sanlock-timeout | -o seconds
63 Override the default sanlock I/O timeout.
64
65 --adopt | -A 0|1
66 Adopt locks from a previous instance of lvmlockd.
67
68
69
71 Initial set up
72 Setting up LVM to use lvmlockd and a shared VG for the first time
73 includes some one time set up steps:
74
75
76 1. choose a lock manager
77 dlm
78 If dlm (or corosync) are already being used by other cluster software,
79 then select dlm. dlm uses corosync which requires additional configu‐
80 ration beyond the scope of this document. See corosync and dlm docu‐
81 mentation for instructions on configuration, set up and usage.
82
83 sanlock
84 Choose sanlock if dlm/corosync are not otherwise required. sanlock
85 does not depend on any clustering software or configuration.
86
87
88 2. configure hosts to use lvmlockd
89 On all hosts running lvmlockd, configure lvm.conf:
90 locking_type = 1
91 use_lvmlockd = 1
92
93 sanlock
94 Assign each host a unique host_id in the range 1-2000 by setting
95 /etc/lvm/lvmlocal.conf local/host_id
96
97
98 3. start lvmlockd
99 Start the lvmlockd daemon.
100 Use systemctl, a cluster resource agent, or run directly, e.g.
101 systemctl start lvm2-lvmlockd
102
103
104 4. start lock manager
105 sanlock
106 Start the sanlock and wdmd daemons.
107 Use systemctl or run directly, e.g.
108 systemctl start wdmd sanlock
109
110 dlm
111 Start the dlm and corosync daemons.
112 Use systemctl, a cluster resource agent, or run directly, e.g.
113 systemctl start corosync dlm
114
115
116 5. create VG on shared devices
117 vgcreate --shared <vgname> <devices>
118
119 The shared option sets the VG lock type to sanlock or dlm depending on
120 which lock manager is running. LVM commands acquire locks from lvm‐
121 lockd, and lvmlockd uses the chosen lock manager.
122
123
124 6. start VG on all hosts
125 vgchange --lock-start
126
127 Shared VGs must be started before they are used. Starting the VG per‐
128 forms lock manager initialization that is necessary to begin using
129 locks (i.e. creating and joining a lockspace). Starting the VG may
130 take some time, and until the start completes the VG may not be modi‐
131 fied or activated.
132
133
134 7. create and activate LVs
135 Standard lvcreate and lvchange commands are used to create and activate
136 LVs in a shared VG.
137
138 An LV activated exclusively on one host cannot be activated on another.
139 When multiple hosts need to use the same LV concurrently, the LV can be
140 activated with a shared lock (see lvchange options -aey vs -asy.)
141 (Shared locks are disallowed for certain LV types that cannot be used
142 from multiple hosts.)
143
144
145
146 Normal start up and shut down
147 After initial set up, start up and shut down include the following
148 steps. They can be performed directly or may be automated using sys‐
149 temd or a cluster resource manager/agents.
150
151 · start lvmlockd
152 · start lock manager
153 · vgchange --lock-start
154 · activate LVs in shared VGs
155
156 The shut down sequence is the reverse:
157
158 · deactivate LVs in shared VGs
159 · vgchange --lock-stop
160 · stop lock manager
161 · stop lvmlockd
162
163
165 Protecting VGs on shared devices
166 The following terms are used to describe the different ways of access‐
167 ing VGs on shared devices.
168
169 shared VG
170
171 A shared VG exists on shared storage that is visible to multiple hosts.
172 LVM acquires locks through lvmlockd to coordinate access to shared VGs.
173 A shared VG has lock_type "dlm" or "sanlock", which specifies the lock
174 manager lvmlockd will use.
175
176 When the lock manager for the lock type is not available (e.g. not
177 started or failed), lvmlockd is unable to acquire locks for LVM com‐
178 mands. In this situation, LVM commands are only allowed to read and
179 display the VG; changes and activation will fail.
180
181 local VG
182
183 A local VG is meant to be used by a single host. It has no lock type
184 or lock type "none". A local VG typically exists on local (non-shared)
185 devices and cannot be used concurrently from different hosts.
186
187 If a local VG does exist on shared devices, it should be owned by a
188 single host by having the system ID set, see lvmsystemid(7). The host
189 with a matching system ID can use the local VG and other hosts will
190 ignore it. A VG with no lock type and no system ID should be excluded
191 from all but one host using lvm.conf filters. Without any of these
192 protections, a local VG on shared devices can be easily damaged or
193 destroyed.
194
195 clvm VG
196
197 A clvm VG (or clustered VG) is a VG on shared storage (like a shared
198 VG) that requires clvmd for clustering and locking. See below for con‐
199 verting a clvm/clustered VG to a shared VG.
200
201
202
203 shared VGs from hosts not using lvmlockd
204 Hosts that do not use shared VGs will not be running lvmlockd. In this
205 case, shared VGs that are still visible to the host will be ignored
206 (like foreign VGs, see lvmsystemid(7).)
207
208 The --shared option for reporting and display commands causes shared
209 VGs to be displayed on a host not using lvmlockd, like the --foreign
210 option does for foreign VGs.
211
212
213
214 creating the first sanlock VG
215 Creating the first sanlock VG is not protected by locking, so it
216 requires special attention. This is because sanlock locks exist on
217 storage within the VG, so they are not available until after the VG is
218 created. The first sanlock VG that is created will automatically con‐
219 tain the "global lock". Be aware of the following special considera‐
220 tions:
221
222
223 · The first vgcreate command needs to be given the path to a device
224 that has not yet been initialized with pvcreate. The pvcreate ini‐
225 tialization will be done by vgcreate. This is because the pvcreate
226 command requires the global lock, which will not be available until
227 after the first sanlock VG is created.
228
229
230 · Because the first sanlock VG will contain the global lock, this VG
231 needs to be accessible to all hosts that will use sanlock shared VGs.
232 All hosts will need to use the global lock from the first sanlock VG.
233
234
235 · The device and VG name used by the initial vgcreate will not be pro‐
236 tected from concurrent use by another vgcreate on another host.
237
238 See below for more information about managing the sanlock global
239 lock.
240
241
242
243 using shared VGs
244 There are some special considerations when using shared VGs.
245
246 When use_lvmlockd is first enabled in lvm.conf, and before the first
247 shared VG is created, no global lock will exist. In this initial
248 state, LVM commands try and fail to acquire the global lock, producing
249 a warning, and some commands are disallowed. Once the first shared VG
250 is created, the global lock will be available, and LVM will be fully
251 operational.
252
253 When a new shared VG is created, its lockspace is automatically started
254 on the host that creates it. Other hosts need to run 'vgchange
255 --lock-start' to start the new VG before they can use it.
256
257 From the 'vgs' command, shared VGs are indicated by "s" (for shared) in
258 the sixth attr field, and by "shared" in the "--options shared" report
259 field. The specific lock type and lock args for a shared VG can be
260 displayed with 'vgs -o+locktype,lockargs'.
261
262 Shared VGs need to be "started" and "stopped", unlike other types of
263 VGs. See the following section for a full description of starting and
264 stopping.
265
266 Removing a shared VG will fail if other hosts have the VG started. Run
267 vgchange --lock-stop <vgname> on all other hosts before vgremove. (It
268 may take several seconds before vgremove recognizes that all hosts have
269 stopped a sanlock VG.)
270
271
272 starting and stopping VGs
273 Starting a shared VG (vgchange --lock-start) causes the lock manager to
274 start (join) the lockspace for the VG on the host where it is run.
275 This makes locks for the VG available to LVM commands on the host.
276 Before a VG is started, only LVM commands that read/display the VG are
277 allowed to continue without locks (and with a warning).
278
279 Stopping a shared VG (vgchange --lock-stop) causes the lock manager to
280 stop (leave) the lockspace for the VG on the host where it is run.
281 This makes locks for the VG inaccessible to the host. A VG cannot be
282 stopped while it has active LVs.
283
284 When using the lock type sanlock, starting a VG can take a long time
285 (potentially minutes if the host was previously shut down without
286 cleanly stopping the VG.)
287
288 A shared VG can be started after all the following are true:
289 · lvmlockd is running
290 · the lock manager is running
291 · the VG's devices are visible on the system
292
293 A shared VG can be stopped if all LVs are deactivated.
294
295 All shared VGs can be started/stopped using:
296 vgchange --lock-start
297 vgchange --lock-stop
298
299
300 Individual VGs can be started/stopped using:
301 vgchange --lock-start <vgname> ...
302 vgchange --lock-stop <vgname> ...
303
304 To make vgchange not wait for start to complete:
305 vgchange --lock-start --lock-opt nowait ...
306
307 lvmlockd can be asked directly to stop all lockspaces:
308 lvmlockctl -S--stop-lockspaces
309
310 To start only selected shared VGs, use the lvm.conf activa‐
311 tion/lock_start_list. When defined, only VG names in this list are
312 started by vgchange. If the list is not defined (the default), all
313 visible shared VGs are started. To start only "vg1", use the following
314 lvm.conf configuration:
315
316 activation {
317 lock_start_list = [ "vg1" ]
318 ...
319 }
320
321
322
323 automatic starting and automatic activation
324 When system-level scripts/programs automatically start VGs, they should
325 use the "auto" option. This option indicates that the command is being
326 run automatically by the system:
327
328 vgchange --lock-start --lock-opt auto [<vgname> ...]
329
330 The "auto" option causes the command to follow the lvm.conf activa‐
331 tion/auto_lock_start_list. If auto_lock_start_list is undefined, all
332 VGs are started, just as if the auto option was not used.
333
334 When auto_lock_start_list is defined, it lists the shared VGs that
335 should be started by the auto command. VG names that do not match an
336 item in the list will be ignored by the auto start command.
337
338 (The lock_start_list is also still used to filter VG names from all
339 start commands, i.e. with or without the auto option. When the
340 lock_start_list is defined, only VGs matching a list item can be
341 started with vgchange.)
342
343 The auto_lock_start_list allows a user to select certain shared VGs
344 that should be automatically started by the system (or indirectly,
345 those that should not).
346
347
348
349 internal command locking
350 To optimize the use of LVM with lvmlockd, be aware of the three kinds
351 of locks and when they are used:
352
353 Global lock
354
355 The global lock s associated with global information, which is informa‐
356 tion not isolated to a single VG. This includes:
357
358 · The global VG namespace.
359 · The set of orphan PVs and unused devices.
360 · The properties of orphan PVs, e.g. PV size.
361
362 The global lock is acquired in shared mode by commands that read this
363 information, or in exclusive mode by commands that change it. For
364 example, the command 'vgs' acquires the global lock in shared mode
365 because it reports the list of all VG names, and the vgcreate command
366 acquires the global lock in exclusive mode because it creates a new VG
367 name, and it takes a PV from the list of unused PVs.
368
369 When an LVM command is given a tag argument, or uses select, it must
370 read all VGs to match the tag or selection, which causes the global
371 lock to be acquired.
372
373 VG lock
374
375 A VG lock is associated with each shared VG. The VG lock is acquired
376 in shared mode to read the VG and in exclusive mode to change the VG or
377 activate LVs. This lock serializes access to a VG with all other LVM
378 commands accessing the VG from all hosts.
379
380 The command 'vgs <vgname>' does not acquire the global lock (it does
381 not need the list of all VG names), but will acquire the VG lock on
382 each VG name argument.
383
384 LV lock
385
386 An LV lock is acquired before the LV is activated, and is released
387 after the LV is deactivated. If the LV lock cannot be acquired, the LV
388 is not activated. (LV locks are persistent and remain in place when
389 the activation command is done. Global and VG locks are transient, and
390 are held only while an LVM command is running.)
391
392 lock retries
393
394 If a request for a Global or VG lock fails due to a lock conflict with
395 another host, lvmlockd automatically retries for a short time before
396 returning a failure to the LVM command. If those retries are insuffi‐
397 cient, the LVM command will retry the entire lock request a number of
398 times specified by global/lvmlockd_lock_retries before failing. If a
399 request for an LV lock fails due to a lock conflict, the command fails
400 immediately.
401
402
403
404 managing the global lock in sanlock VGs
405 The global lock exists in one of the sanlock VGs. The first sanlock VG
406 created will contain the global lock. Subsequent sanlock VGs will each
407 contain a disabled global lock that can be enabled later if necessary.
408
409 The VG containing the global lock must be visible to all hosts using
410 sanlock VGs. For this reason, it can be useful to create a small san‐
411 lock VG, visible to all hosts, and dedicated to just holding the global
412 lock. While not required, this strategy can help to avoid difficulty
413 in the future if VGs are moved or removed.
414
415 The vgcreate command typically acquires the global lock, but in the
416 case of the first sanlock VG, there will be no global lock to acquire
417 until the first vgcreate is complete. So, creating the first sanlock
418 VG is a special case that skips the global lock.
419
420 vgcreate determines that it's creating the first sanlock VG when no
421 other sanlock VGs are visible on the system. It is possible that other
422 sanlock VGs do exist, but are not visible when vgcreate checks for
423 them. In this case, vgcreate will create a new sanlock VG with the
424 global lock enabled. When the another VG containing a global lock
425 appears, lvmlockd will then see more than one VG with a global lock
426 enabled. LVM commands will report that there are duplicate global
427 locks.
428
429 If the situation arises where more than one sanlock VG contains a
430 global lock, the global lock should be manually disabled in all but one
431 of them with the command:
432
433 lvmlockctl --gl-disable <vgname>
434
435 (The one VG with the global lock enabled must be visible to all hosts.)
436
437 An opposite problem can occur if the VG holding the global lock is
438 removed. In this case, no global lock will exist following the vgre‐
439 move, and subsequent LVM commands will fail to acquire it. In this
440 case, the global lock needs to be manually enabled in one of the
441 remaining sanlock VGs with the command:
442
443 lvmlockctl --gl-enable <vgname>
444
445 (Using a small sanlock VG dedicated to holding the global lock can
446 avoid the case where the global lock must be manually enabled after a
447 vgremove.)
448
449
450
451 internal lvmlock LV
452 A sanlock VG contains a hidden LV called "lvmlock" that holds the san‐
453 lock locks. vgreduce cannot yet remove the PV holding the lvmlock LV.
454 To remove this PV, change the VG lock type to "none", run vgreduce,
455 then change the VG lock type back to "sanlock". Similarly, pvmove can‐
456 not be used on a PV used by the lvmlock LV.
457
458 To place the lvmlock LV on a specific device, create the VG with only
459 that device, then use vgextend to add other devices.
460
461
462
463 LV activation
464 In a shared VG, LV activation involves locking through lvmlockd, and
465 the following values are possible with lvchange/vgchange -a:
466
467
468 y|ey The command activates the LV in exclusive mode, allowing a sin‐
469 gle host to activate the LV. Before activating the LV, the com‐
470 mand uses lvmlockd to acquire an exclusive lock on the LV. If
471 the lock cannot be acquired, the LV is not activated and an
472 error is reported. This would happen if the LV is active on
473 another host.
474
475
476 sy The command activates the LV in shared mode, allowing multiple
477 hosts to activate the LV concurrently. Before activating the
478 LV, the command uses lvmlockd to acquire a shared lock on the
479 LV. If the lock cannot be acquired, the LV is not activated and
480 an error is reported. This would happen if the LV is active
481 exclusively on another host. If the LV type prohibits shared
482 access, such as a snapshot, the command will report an error and
483 fail. The shared mode is intended for a multi-host/cluster
484 application or file system. LV types that cannot be used con‐
485 currently from multiple hosts include thin, cache, raid, and
486 snapshot.
487
488
489 n The command deactivates the LV. After deactivating the LV, the
490 command uses lvmlockd to release the current lock on the LV.
491
492
493
494 manually repairing a shared VG
495 Some failure conditions may not be repairable while the VG has a shared
496 lock type. In these cases, it may be possible to repair the VG by
497 forcibly changing the lock type to "none". This is done by adding
498 "--lock-opt force" to the normal command for changing the lock type:
499 vgchange --lock-type none VG. The VG lockspace should first be stopped
500 on all hosts, and be certain that no hosts are using the VG before this
501 is done.
502
503
504
505 recover from lost PV holding sanlock locks
506 In a sanlock VG, the sanlock locks are held on the hidden "lvmlock" LV.
507 If the PV holding this LV is lost, a new lvmlock LV needs to be cre‐
508 ated. To do this, ensure no hosts are using the VG, then forcibly
509 change the lock type to "none" (see above). Then change the lock type
510 back to "sanlock" with the normal command for changing the lock type:
511 vgchange --lock-type sanlock VG. This recreates the internal lvmlock
512 LV with the necessary locks.
513
514
515
516 locking system failures
517 lvmlockd failure
518
519 If lvmlockd fails or is killed while holding locks, the locks are
520 orphaned in the lock manager. lvmlockd can be restarted with an option
521 to adopt locks in the lock manager that had been held by the previous
522 instance.
523
524 dlm/corosync failure
525
526 If dlm or corosync fail, the clustering system will fence the host
527 using a method configured within the dlm/corosync clustering environ‐
528 ment.
529
530 LVM commands on other hosts will be blocked from acquiring any locks
531 until the dlm/corosync recovery process is complete.
532
533 sanlock lease storage failure
534
535 If the PV under a sanlock VG's lvmlock LV is disconnected, unresponsive
536 or too slow, sanlock cannot renew the lease for the VG's locks. After
537 some time, the lease will expire, and locks that the host owns in the
538 VG can be acquired by other hosts. The VG must be forcibly deactivated
539 on the host with the expiring lease before other hosts can acquire its
540 locks.
541
542 When the sanlock daemon detects that the lease storage is lost, it runs
543 the command lvmlockctl --kill <vgname>. This command emits a syslog
544 message stating that lease storage is lost for the VG, and LVs must be
545 immediately deactivated.
546
547 If no LVs are active in the VG, then the lockspace with an expiring
548 lease will be removed, and errors will be reported when trying to use
549 the VG. Use the lvmlockctl --drop command to clear the stale lockspace
550 from lvmlockd.
551
552 If the VG has active LVs when the lock storage is lost, the LVs must be
553 quickly deactivated before the lockspace lease expires. After all LVs
554 are deactivated, run lvmlockctl --drop <vgname> to clear the expiring
555 lockspace from lvmlockd. If all LVs in the VG are not deactivated
556 within about 40 seconds, sanlock uses wdmd and the local watchdog to
557 reset the host. The machine reset is effectively a severe form of
558 "deactivating" LVs before they can be activated on other hosts. The
559 reset is considered a better alternative than having LVs used by multi‐
560 ple hosts at once, which could easily damage or destroy their content.
561
562 In the future, the lvmlockctl kill command may automatically attempt to
563 forcibly deactivate LVs before the sanlock lease expires. Until then,
564 the user must notice the syslog message and manually deactivate the VG
565 before sanlock resets the machine.
566
567 sanlock daemon failure
568
569 If the sanlock daemon fails or exits while a lockspace is started, the
570 local watchdog will reset the host. This is necessary to protect any
571 application resources that depend on sanlock leases.
572
573
574
575 changing dlm cluster name
576 When a dlm VG is created, the cluster name is saved in the VG metadata.
577 To use the VG, a host must be in the named dlm cluster. If the dlm
578 cluster name changes, or the VG is moved to a new cluster, the dlm
579 cluster name saved in the VG must also be changed.
580
581 To see the dlm cluster name saved in the VG, use the command:
582 vgs -o+locktype,lockargs <vgname>
583
584 To change the dlm cluster name in the VG when the VG is still used by
585 the original cluster:
586
587
588 · Start the VG on the host changing the lock type
589 vgchange --lock-start <vgname>
590
591
592 · Stop the VG on all other hosts:
593 vgchange --lock-stop <vgname>
594
595
596 · Change the VG lock type to none on the host where the VG is started:
597 vgchange --lock-type none <vgname>
598
599
600 · Change the dlm cluster name on the hosts or move the VG to the new
601 cluster. The new dlm cluster must now be running on the host. Ver‐
602 ify the new name by:
603 cat /sys/kernel/config/dlm/cluster/cluster_name
604
605
606 · Change the VG lock type back to dlm which sets the new cluster name:
607 vgchange --lock-type dlm <vgname>
608
609
610 · Start the VG on hosts to use it:
611 vgchange --lock-start <vgname>
612
613
614 To change the dlm cluster name in the VG when the dlm cluster name has
615 already been changed on the hosts, or the VG has already moved to a
616 different cluster:
617
618
619 · Ensure the VG is not being used by any hosts.
620
621
622 · The new dlm cluster must be running on the host making the change.
623 The current dlm cluster name can be seen by:
624 cat /sys/kernel/config/dlm/cluster/cluster_name
625
626
627 · Change the VG lock type to none:
628 vgchange --lock-type none --lock-opt force <vgname>
629
630
631 · Change the VG lock type back to dlm which sets the new cluster name:
632 vgchange --lock-type dlm <vgname>
633
634
635 · Start the VG on hosts to use it:
636 vgchange --lock-start <vgname>
637
638
639
640 changing a local VG to a shared VG
641 All LVs must be inactive to change the lock type.
642
643 lvmlockd must be configured and running as described in USAGE.
644
645
646 · Change a local VG to a shared VG with the command:
647 vgchange --lock-type sanlock|dlm <vgname>
648
649
650 · Start the VG on hosts to use it:
651 vgchange --lock-start <vgname>
652
653
654 changing a shared VG to a local VG
655 All LVs must be inactive to change the lock type.
656
657
658 · Start the VG on the host making the change:
659 vgchange --lock-start <vgname>
660
661
662 · Stop the VG on all other hosts:
663 vgchange --lock-stop <vgname>
664
665
666 · Change the VG lock type to none on the host where the VG is started:
667 vgchange --lock-type none <vgname>
668
669
670 If the VG cannot be started with the previous lock type, then the lock
671 type can be forcibly changed to none with:
672
673 vgchange --lock-type none --lock-opt force <vgname>
674
675 To change a VG from one lock type to another (i.e. between sanlock and
676 dlm), first change it to a local VG, then to the new type.
677
678
679
680 changing a clvm/clustered VG to a shared VG
681 All LVs must be inactive to change the lock type.
682
683 First change the clvm/clustered VG to a local VG. Within a running
684 clvm cluster, change a clustered VG to a local VG with the command:
685
686 vgchange -cn <vgname>
687
688 If the clvm cluster is no longer running on any nodes, then extra
689 options can be used to forcibly make the VG local. Caution: this is
690 only safe if all nodes have stopped using the VG:
691
692 vgchange --lock-type none --lock-opt force <vgname>
693
694 After the VG is local, follow the steps described in "changing a local
695 VG to a shared VG".
696
697
698
699 limitations of shared VGs
700 Things that do not yet work in shared VGs:
701 · using external origins for thin LVs
702 · splitting snapshots from LVs
703 · splitting mirrors in sanlock VGs
704 · pvmove of entire PVs, or under LVs activated with shared locks
705 · vgsplit and vgmerge (convert to a local VG to do this)
706
707
708
709 lvmlockd changes from clvmd
710 (See above for converting an existing clvm VG to a shared VG.)
711
712 While lvmlockd and clvmd are entirely different systems, LVM command
713 usage remains similar. Differences are more notable when using lvm‐
714 lockd's sanlock option.
715
716 Visible usage differences between shared VGs (using lvmlockd) and
717 clvm/clustered VGs (using clvmd):
718
719
720 · lvm.conf must be configured to use either lvmlockd (use_lvmlockd=1)
721 or clvmd (locking_type=3), but not both.
722
723
724 · vgcreate --shared creates a shared VG, and vgcreate --clustered y
725 creates a clvm/clustered VG.
726
727
728 · lvmlockd adds the option of using sanlock for locking, avoiding the
729 need for network clustering.
730
731
732 · lvmlockd defaults to the exclusive activation mode whenever the acti‐
733 vation mode is unspecified, i.e. -ay means -aey, not -asy.
734
735
736 · lvmlockd commands always apply to the local host, and never have an
737 effect on a remote host. (The activation option 'l' is not used.)
738
739
740 · lvmlockd works with thin and cache pools and LVs.
741
742
743 · lvmlockd works with lvmetad.
744
745
746 · lvmlockd saves the cluster name for a shared VG using dlm. Only
747 hosts in the matching cluster can use the VG.
748
749
750 · lvmlockd requires starting/stopping shared VGs with vgchange
751 --lock-start and --lock-stop.
752
753
754 · vgremove of a sanlock VG may fail indicating that all hosts have not
755 stopped the VG lockspace. Stop the VG on all hosts using vgchange
756 --lock-stop.
757
758
759 · vgreduce or pvmove of a PV in a sanlock VG will fail if it holds the
760 internal "lvmlock" LV that holds the sanlock locks.
761
762
763 · lvmlockd uses lock retries instead of lock queueing, so high lock
764 contention may require increasing global/lvmlockd_lock_retries to
765 avoid transient lock failures.
766
767
768 · lvmlockd includes VG reporting options lock_type and lock_args, and
769 LV reporting option lock_args to view the corresponding metadata
770 fields.
771
772
773 · In the 'vgs' command's sixth VG attr field, "s" for "shared" is dis‐
774 played for shared VGs.
775
776
777 · If lvmlockd fails or is killed while in use, locks it held remain but
778 are orphaned in the lock manager. lvmlockd can be restarted with an
779 option to adopt the orphan locks from the previous instance of lvm‐
780 lockd.
781
782
783Red Hat, Inc LVM TOOLS 2.02.180(2)-RHEL7 (2018-07-20) LVMLOCKD(8)