Hell David, >>> On 2018/10/8 at 23:00, in message <20181008150016.GB21471@xxxxxxxxxx>, David Teigland <teigland@xxxxxxxxxx> wrote: > On Mon, Oct 08, 2018 at 04:23:27AM -0600, Gang He wrote: >> Hello List >> >> The system uses lvm based on raid1. >> It seems that the PV of the raid1 is found also on the single disks that > build the raid1 device: >> [ 147.121725] linux-472a dracut-initqueue[391]: WARNING: PV > qG1QRz-Ivm1-QVwq-uaHV-va9w-wwXh-lIIOhV on /dev/sda2 was already found on > /dev/md1. >> [ 147.123427] linux-472a dracut-initqueue[391]: WARNING: PV > qG1QRz-Ivm1-QVwq-uaHV-va9w-wwXh-lIIOhV on /dev/sdb2 was already found on > /dev/md1. >> [ 147.369863] linux-472a dracut-initqueue[391]: WARNING: PV > qG1QRz-Ivm1-QVwq-uaHV-va9w-wwXh-lIIOhV prefers device /dev/md1 because device > size is correct. >> [ 147.370597] linux-472a dracut-initqueue[391]: WARNING: PV > qG1QRz-Ivm1-QVwq-uaHV-va9w-wwXh-lIIOhV prefers device /dev/md1 because device > size is correct. >> [ 147.371698] linux-472a dracut-initqueue[391]: Cannot activate LVs in VG > vghome while PVs appear on duplicate devices. > > Do these warnings only appear from "dracut-initqueue"? Can you run and > send 'vgs -vvvv' from the command line? If they don't appear from the > command line, then is "dracut-initqueue" using a different lvm.conf? > lvm.conf settings can effect this (filter, md_component_detection, > external_device_info_source). mdadm --detail --scan -vvv /dev/md/linux:0: Version : 1.0 Creation Time : Sun Jul 22 22:49:21 2012 Raid Level : raid1 Array Size : 513012 (500.99 MiB 525.32 MB) Used Dev Size : 513012 (500.99 MiB 525.32 MB) Raid Devices : 2 Total Devices : 2 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Mon Jul 16 00:29:19 2018 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Consistency Policy : bitmap Name : linux:0 UUID : 160998c8:7e21bcff:9cea0bbc:46454716 Events : 469 Number Major Minor RaidDevice State 0 8 17 0 active sync /dev/sdb1 1 8 33 1 active sync /dev/sdc1 /dev/md/linux:1: Version : 1.0 Creation Time : Sun Jul 22 22:49:22 2012 Raid Level : raid1 Array Size : 1953000312 (1862.53 GiB 1999.87 GB) Used Dev Size : 1953000312 (1862.53 GiB 1999.87 GB) Raid Devices : 2 Total Devices : 2 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Fri Oct 12 20:16:25 2018 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Consistency Policy : bitmap Name : linux:1 UUID : 17426969:03d7bfa7:5be33b0b:8171417a Events : 326248 Number Major Minor RaidDevice State 0 8 18 0 active sync /dev/sdb2 1 8 34 1 active sync /dev/sdc2 Thanks Gang > >> This is a regression bug? since the user did not encounter this problem with > lvm2 v2.02.177. > > It could be, since the new scanning changed how md detection works. The > md superblock version effects how lvm detects this. md superblock 1.0 (at > the end of the device) is not detected as easily as newer md versions > (1.1, 1.2) where the superblock is at the beginning. Do you know which > this is? _______________________________________________ linux-lvm mailing list linux-lvm@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-lvm read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/