[PVE-User] LVM autoactivation failed with multipath over iSCSI

nada nada at verdnatura.es
Wed Jan 15 10:55:06 CET 2020


On 2020-01-14 19:46, Stefan M. Radman via pve-user wrote:
> Hi Nada
> What's the output of "systemctl --failed" and "systemctl status 
> lvm2-activation-net.service".
> Stefan

Hi Stefan
thank you for your response !
the output of "systemctl --failed" was claiming devices from SAN during 
boot,
which were activated by rc-local.service after boot
i do NOT have "lvm2-activation-net.service"
and find masked status of multipath-tools-boot.service < is this ok ???
but i find some mistake in multipath.conf eventhough it is running
i am going to reconfigure it and reboot this afternoon
following are details
Nada

root at mox11:~# systemctl --failed --all
   UNIT                      LOAD   ACTIVE SUB    DESCRIPTION
   lvm2-pvscan at 253:7.service loaded failed failed LVM event activation on 
device 253:7
   lvm2-pvscan at 253:8.service loaded failed failed LVM event activation on 
device 253:8

LOAD   = Reflects whether the unit definition was properly loaded.
ACTIVE = The high-level unit activation state, i.e. generalization of 
SUB.
SUB    = The low-level unit activation state, values depend on unit 
type.

2 loaded units listed.
To show all installed unit files use 'systemctl list-unit-files'.

root at mox11:~# dmsetup ls
san2020jan-vm--903--disk--0	(253:18)
santest-santestpool	(253:12)
3600c0ff000195f8e7d0a855701000000	(253:7)
pve-data-tpool	(253:4)
pve-data_tdata	(253:3)
pve-zfs	(253:6)
santest-santestpool-tpool	(253:11)
santest-santestpool_tdata	(253:10)
pve-data_tmeta	(253:2)
san2020jan-san2020janpool	(253:17)
santest-santestpool_tmeta	(253:9)
pve-swap	(253:0)
pve-root	(253:1)
pve-data	(253:5)
3600c0ff000195f8ec3f01d5e01000000	(253:8)
san2020jan-san2020janpool-tpool	(253:16)
san2020jan-san2020janpool_tdata	(253:15)
san2020jan-san2020janpool_tmeta	(253:14)

root at mox11:~# pvs -a
   PV                                            VG         Fmt  Attr 
PSize   PFree
   /dev/mapper/3600c0ff000195f8e7d0a855701000000 santest    lvm2 a--   
<9.31g 292.00m
   /dev/mapper/3600c0ff000195f8ec3f01d5e01000000 san2020jan lvm2 a--  
<93.13g  <2.95g
   /dev/mapper/san2020jan-vm--903--disk--0                       ---      
  0       0
   /dev/sdb                                                      ---      
  0       0
   /dev/sdc2                                                     ---      
  0       0
   /dev/sdc3                                     pve        lvm2 a--   
67.73g   6.97g
root at mox11:~# vgs -a
   VG         #PV #LV #SN Attr   VSize   VFree
   pve          1   4   0 wz--n-  67.73g   6.97g
   san2020jan   1   2   0 wz--n- <93.13g  <2.95g
   santest      1   1   0 wz--n-  <9.31g 292.00m
root at mox11:~# lvs -a
   LV                     VG         Attr       LSize  Pool           
Origin Data%  Meta%  Move Log Cpy%Sync Convert
   data                   pve        twi-aotz--  9.98g                    
    0.00   10.61
   [data_tdata]           pve        Twi-ao----  9.98g
   [data_tmeta]           pve        ewi-ao---- 12.00m
   [lvol0_pmspare]        pve        ewi------- 12.00m
   root                   pve        -wi-ao---- 16.75g
   swap                   pve        -wi-ao----  4.00g
   zfs                    pve        -wi-ao---- 30.00g
   [lvol0_pmspare]        san2020jan ewi------- 92.00m
   san2020janpool         san2020jan twi-aotz-- 90.00g                    
    0.86   10.84
   [san2020janpool_tdata] san2020jan Twi-ao---- 90.00g
   [san2020janpool_tmeta] san2020jan ewi-ao---- 92.00m
   vm-903-disk-0          san2020jan Vwi-aotz--  2.50g san2020janpool     
    30.99
   [lvol0_pmspare]        santest    ewi------- 12.00m
   santestpool            santest    twi-aotz--  9.00g                    
    0.00   10.58
   [santestpool_tdata]    santest    Twi-ao----  9.00g
   [santestpool_tmeta]    santest    ewi-ao---- 12.00m

root at mox11:~# multipathd -k"show maps"
Jan 15 10:50:02 | /etc/multipath.conf line 24, duplicate keyword: wwid
name                              sysfs uuid
3600c0ff000195f8e7d0a855701000000 dm-7  
3600c0ff000195f8e7d0a855701000000
3600c0ff000195f8ec3f01d5e01000000 dm-8  
3600c0ff000195f8ec3f01d5e01000000

root at mox11:~# multipathd -k"show paths"
Jan 15 10:50:07 | /etc/multipath.conf line 24, duplicate keyword: wwid
hcil     dev dev_t pri dm_st  chk_st dev_st  next_check
6:0:0:3  sde 8:64  10  active ready  running XX........ 2/8
9:0:0:3  sdn 8:208 10  active ready  running XX........ 2/8
7:0:0:3  sdh 8:112 50  active ready  running XX........ 2/8
5:0:0:3  sdd 8:48  10  active ready  running XXXXXX.... 5/8
11:0:0:3 sdp 8:240 50  active ready  running X......... 1/8
10:0:0:3 sdl 8:176 50  active ready  running XXXXXXXXXX 8/8
8:0:0:6  sdk 8:160 50  active ready  running XXXXXX.... 5/8
8:0:0:3  sdj 8:144 50  active ready  running XXXXXXXXXX 8/8
9:0:0:6  sdo 8:224 10  active ready  running X......... 1/8
6:0:0:6  sdg 8:96  10  active ready  running XXXXXX.... 5/8
5:0:0:6  sdf 8:80  10  active ready  running XXXXXX.... 5/8
10:0:0:6 sdm 8:192 50  active ready  running XXXXXXX... 6/8
11:0:0:6 sdq 65:0  50  active ready  running XXXXXXXX.. 7/8
7:0:0:6  sdi 8:128 50  active ready  running XXXXXXX... 6/8

root at mox11:~# cat /etc/multipath.conf
defaults {
         polling_interval        2
         uid_attribute           ID_SERIAL
	no_path_retry		queue
	find_multipaths		yes
}
blacklist {
	wwid .*
	# BECAREFULL @mox11 blacklit sda,sdb,sdc
	devnode "^sd[a-c]"
}
blacklist_exceptions {
	# 25G v_multitest
	# wwid "3600c0ff000195f8e2172de5d01000000"
	# 10G prueba
	wwid "3600c0ff000195f8e7d0a855701000000"
	# 100G sanmox11
	wwid "3600c0ff000195f8ec3f01d5e01000000"
}
multipaths {
   multipath {
         # wwid "3600c0ff000195f8e2172de5d01000000"
	wwid "3600c0ff000195f8e7d0a855701000000"
	wwid "3600c0ff000195f8ec3f01d5e01000000"
   }
}
devices {
   device {
	#### the following 6 lines do NOT change
	vendor "HP"
	product "P2000 G3 FC|P2000G3 FC/iSCSI|P2000 G3 SAS|P2000 G3 iSCSI"
	# getuid_callout "/lib/udev/scsi_id -g -u -s /block/%n"
	path_grouping_policy "group_by_prio"
	prio "alua"
	failback "immediate"
	no_path_retry 18
	####
	hardware_handler "0"
	path_selector "round-robin 0"
	rr_weight uniform
	rr_min_io 100
	path_checker tur
   }
}


root at mox11:~# systemctl status multipath-tools-boot
   multipath-tools-boot.service
    Loaded: masked (Reason: Unit multipath-tools-boot.service is masked.)
    Active: inactive (dead)

root at mox11:~# pveversion -V
proxmox-ve: 6.1-2 (running kernel: 5.3.13-1-pve)
pve-manager: 6.1-5 (running version: 6.1-5/9bf06119)
pve-kernel-5.3: 6.1-1
pve-kernel-helper: 6.1-1
pve-kernel-4.15: 5.4-12
pve-kernel-5.3.13-1-pve: 5.3.13-1
pve-kernel-4.15.18-24-pve: 4.15.18-52
pve-kernel-4.15.18-21-pve: 4.15.18-48
pve-kernel-4.15.18-11-pve: 4.15.18-34
ceph-fuse: 12.2.11+dfsg1-2.1+b1
corosync: 3.0.2-pve4
criu: 3.11-3
glusterfs-client: 5.5-3
ifupdown: 0.8.35+pve1
ksm-control-daemon: 1.3-1
libjs-extjs: 6.0.1-10
libknet1: 1.13-pve1
libpve-access-control: 6.0-5
libpve-apiclient-perl: 3.0-2
libpve-common-perl: 6.0-9
libpve-guest-common-perl: 3.0-3
libpve-http-server-perl: 3.0-3
libpve-storage-perl: 6.1-3
libqb0: 1.0.5-1
libspice-server1: 0.14.2-4~pve6+1
lvm2: 2.03.02-pve3
lxc-pve: 3.2.1-1
lxcfs: 3.0.3-pve60
novnc-pve: 1.1.0-1
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.1-1
pve-cluster: 6.1-2
pve-container: 3.0-15
pve-docs: 6.1-3
pve-edk2-firmware: 2.20191127-1
pve-firewall: 4.0-9
pve-firmware: 3.0-4
pve-ha-manager: 3.0-8
pve-i18n: 2.0-3
pve-qemu-kvm: 4.1.1-2
pve-xtermjs: 3.13.2-1
qemu-server: 6.1-4
smartmontools: 7.0-pve2
spiceterm: 3.1-1
vncterm: 1.6-1
zfsutils-linux: 0.8.2-pve2


More information about the pve-user mailing list