[ClusterLabs] Antw: PSA: pvmove -i0 and LVM resource

Ulrich Windl Ulrich.Windl at rz.uni-regensburg.de
Wed Mar 28 06:03:49 UTC 2018


>>> "Hayden,Robert" <RHAYDEN at CERNER.COM> schrieb am 27.03.2018 um 20:12 in
Nachricht
<BL2PR01MB1924DAF2F8C81AD1C44276C0D5AC0 at BL2PR01MB1924.prod.exchangelabs.com>:

> Thought I would share an experience with the community.  We have RHEL 7.4 
> clusters that uses the heartbeat LVM resource (HA-LVM volume group).  The LVM 
> resource does a "vgscan --cache" command as part of its monitoring routine.
> 
> We have found that the pvmove command option "-i0" will block the vgscan 
> command (most likely any LVM command).  The pvmove command just needs to be 
> executed on any physical volume and not specifically on one being managed by 
> RHCS.   In our case, the node where the pvmove was being executed was evicted 
> from the cluster.

Hi!

Some random thoughts on this:

1) pvmove being I/O intensive is best done during some maintenace interval. Such operation is needed verey rarely.

2) AFAIK Linux LVM's pvmove does the whole move as one transaction (as opposed to former HP-UX LVM, where is was done block-wise. That means the relocated block was being used as soon as it was moved). So maybe that's the reason while LVM is seemingly unable to perform another command while pvmove is being active.

3) I think pvmove does not really care to which LVs the PEs (Physical Extents) belong while moving.

4) Monitoring the state of LVM via vgscan is probably a bad idea (in HP-UX you only needed vgscan if you wanted to "import" a VG or PVs already present on a disk, but unknown to the OS)

5) Maybe somthing like /proc/lvmstat (like /proc/mdstat) is needed.

Sorry if I couldn't provide a silver bullet...

Regards,
Ulrich

> 
> Blocking Command:   pvmove -v -i0 -n /dev/testvg/testlv00 /dev/mapper/mpathd1 
> /dev/mapper/mpaths1
> 
> When testing without the -i0 option or with the -iX where X is non-zero, the 
> pvmove did not block vgscan commands.
> 
> Associated errors in /var/log/messages:
> 
> Mar 26 14:03:27 nodeapp1 lvmpolld: W: LVMPOLLD: polling for output of the 
> lvm cmd (PID 74134) has timed out
> 
> <skipped lines>
> Mar 26 14:04:27 nodeapp1 lvmpolld: W: LVMPOLLD: polling for output of the 
> lvm cmd (PID 74134) has timed out
> Mar 26 14:04:32 nodeapp1 lrmd[81636]: warning: share1_vg_monitor_60000 
> process (PID 77254) timed out
> Mar 26 14:04:32 nodeapp1 lrmd[81636]: warning: share1_vg_monitor_60000:77254 
> - timed out after 90000ms
> Mar 26 14:04:32 nodeapp1 crmd[81641]:   error: Result of monitor operation 
> for share1_vg on nodeapp1: Timed Out
> Mar 26 14:04:32 nodeapp1 crmd[81641]:  notice: State transition S_IDLE -> 
> S_POLICY_ENGINE
> 
> <skipped lines>
> Mar 26 14:05:27 nodeapp1 LVM(share1_vg)[88723]: INFO: 0 logical volume(s) in 
> volume group "share1vg" now active
> Mar 26 14:05:27 nodeapp1 lvmpolld: W: LVMPOLLD: polling for output of the 
> lvm cmd (PID 74134) has timed out
> Mar 26 14:05:27 nodeapp1 lvmpolld[74130]: LVMPOLLD: LVM2 cmd is unresponsive 
> too long (PID 74134) (no output for 180 seconds)
> 
> <skipped lines>
> Mar 26 14:05:55 nodeapp1 lrmd[81636]: warning: share1_vg_stop_0 process (PID 
> 88723) timed out
> Mar 26 14:05:55 nodeapp1 lrmd[81636]: warning: share1_vg_stop_0:88723 - timed 
> out after 30000ms
> Mar 26 14:05:55 nodeapp1 crmd[81641]:   error: Result of stop operation for 
> share1_vg on nodeapp1: Timed Out
> Mar 26 14:05:55 nodeapp1 crmd[81641]: warning: Action 6 (share1_vg_stop_0) 
> on nodeapp1 failed (target: 0 vs. rc: 1): Error
> Mar 26 14:05:55 nodeapp1 crmd[81641]:  notice: Transition aborted by 
> operation share1_vg_stop_0 'modify' on nodeapp1: Event failed
> Mar 26 14:05:55 nodeapp1 crmd[81641]: warning: Action 6 (share1_vg_stop_0) 
> on nodeapp1 failed (target: 0 vs. rc: 1): Error
> 
> <skipped lines>
> Mar 26 14:05:55 nodeapp1 pengine[81639]: warning: Processing failed op stop 
> for share1_vg on nodeapp1: unknown error (1)
> Mar 26 14:05:55 nodeapp1 pengine[81639]: warning: Processing failed op stop 
> for share1_vg on nodeapp1: unknown error (1)
> Mar 26 14:05:55 nodeapp1 pengine[81639]: warning: Cluster node nodeapp1 will 
> be fenced: share1_vg failed there
> Mar 26 14:05:55 nodeapp1 pengine[81639]: warning: Forcing share1_vg away 
> from nodeapp1 after 1000000 failures (max=1000000)
> Mar 26 14:05:55 nodeapp1 pengine[81639]: warning: Scheduling Node nodeapp1 
> for STONITH
> 
> Hope this helps someone down the line.
> 
> 
> Robert
> 
> 
> 
> CONFIDENTIALITY NOTICE This message and any included attachments are from 
> Cerner Corporation and are intended only for the addressee. The information 
> contained in this message is confidential and may constitute inside or 
> non-public information under international, federal, or state securities 
> laws. Unauthorized forwarding, printing, copying, distribution, or use of 
> such information is strictly prohibited and may be unlawful. If you are not 
> the addressee, please promptly delete this message and notify the sender of 
> the delivery error by e-mail or you may call Cerner's corporate offices in 
> Kansas City, Missouri, U.S.A at (+1) (816)221-1024.





More information about the Users mailing list