Brian Elliott Finley [finley@anl.gov] wrote: > Malahal, > > Ok, thanks -- that helps me to understand what the LVs are actually using. > > So, how/why is LVM choosing to use a /dev/sd* device, when I have > filtered them out? My filter is: > > filter = [ "a|^/dev/md.*|", > "a|^/dev/mapper/zimbra-mb.*-t.*-v.*_.*-lun.*|", "r|.*|" ] > > Or, is there a flaw in my filter? > The filter seems alright and in fact it looked like LVM picked multipath devices from the "lvs ..." command output. It didn't match your 'dmsetup table' output though. I wonder how that can happen. Try "de-actvating and reactivating" LVM group and see if that changes "dmsetup table" output. --Malahal. > > > malahal@us.ibm.com wrote: > > Brain, > > > > Your 'dmsetup table' output didn't seem to correspond to this 'lvs' > > command output. Your 'dmsetup table' command has this entry: > > > > mb1_t1-bin: 0 41943040 striped 4 128 8:0 384 8:16 384 8:32 384 8:48 384 > > > > That clearly says your 'mb1_t1/bin' volume is triped across devices with > > major:minor numbers as 8:0, 8:16, 8:32, 8:48. The major number 8 > > normally belongs to sd driver that controls single paths. Your 'lvs ..." > > output shows that it should be stripped across > > zimbra-mb1-t1-v1_fujitsu1-lun55 and others. > > > > Your "pvs" showed multipath PV names in your first email. I believe, > > something changed when you posted your "dmsetup ls" and "dmsetup table" > > commands. > > > > --Malahal. > > PS: Does your 'dmsetup table' now show the same output as before? > > > > > > Brian Elliott Finley [finley@anl.gov] wrote: > >> Malahal, > >> > >> Here is the output from the command you suggested (I'm sure I'll be > >> using that one again -- quite handy): > >> > >> root@zimbra-mb1:~# lvs -o lv_name,vg_name,devices > >> LV VG Devices > >> > >> > >> > >> bin mb1_t1 > >> /dev/mapper/zimbra-mb1-t1-v1_fujitsu1-lun55(0),/dev/mapper/zimbra-mb1-t1-v2_fujitsu1-lun56(0),/dev/mapper/zimbra-mb1-t1-v3_fujitsu1-lun57(0),/dev/mapper/zimbra-mb1-t1-v4_fujitsu1-lun58(0) > >> > >> db mb1_t1 > >> /dev/mapper/zimbra-mb1-t1-v1_fujitsu1-lun55(2560),/dev/mapper/zimbra-mb1-t1-v2_fujitsu1-lun56(2560),/dev/mapper/zimbra-mb1-t1-v3_fujitsu1-lun57(2560),/dev/mapper/zimbra-mb1-t1-v4_fujitsu1-lun58(2560) > >> > >> index mb1_t1 > >> /dev/mapper/zimbra-mb1-t1-v1_fujitsu1-lun55(5760),/dev/mapper/zimbra-mb1-t1-v2_fujitsu1-lun56(5760),/dev/mapper/zimbra-mb1-t1-v3_fujitsu1-lun57(5760),/dev/mapper/zimbra-mb1-t1-v4_fujitsu1-lun58(5760) > >> > >> log mb1_t1 > >> /dev/mapper/zimbra-mb1-t1-v1_fujitsu1-lun55(1280),/dev/mapper/zimbra-mb1-t1-v2_fujitsu1-lun56(1280),/dev/mapper/zimbra-mb1-t1-v3_fujitsu1-lun57(1280),/dev/mapper/zimbra-mb1-t1-v4_fujitsu1-lun58(1280) > >> > >> redo mb1_t1 > >> /dev/mapper/zimbra-mb1-t1-v1_fujitsu1-lun55(12160),/dev/mapper/zimbra-mb1-t1-v2_fujitsu1-lun56(12160),/dev/mapper/zimbra-mb1-t1-v3_fujitsu1-lun57(12160),/dev/mapper/zimbra-mb1-t1-v4_fujitsu1-lun58(12160) > >> store mb1_t1 > >> /dev/mapper/zimbra-mb1-t1-v1_fujitsu1-lun55(21760),/dev/mapper/zimbra-mb1-t1-v2_fujitsu1-lun56(21760),/dev/mapper/zimbra-mb1-t1-v3_fujitsu1-lun57(21760),/dev/mapper/zimbra-mb1-t1-v4_fujitsu1-lun58(21760) > >> hsm_store1 mb1_t2 /dev/mapper/zimbra-mb1-t2-v1_fujitsu1-lun63(0) > >> > >> > >> > >> hsm_store2 mb1_t2 /dev/mapper/zimbra-mb1-t2-v2_fujitsu1-lun64(0) > >> > >> > >> > >> hsm_store2 mb1_t2 /dev/mapper/zimbra-mb1-t2-v5_fujitsu1-lun67(0) > >> > >> > >> > >> hsm_store3 mb1_t2 > >> /dev/mapper/zimbra-mb1-t2-v5_fujitsu1-lun67(3001) > >> > >> > >> hsm_store3 mb1_t2 /dev/mapper/zimbra-mb1-t2-v6_fujitsu1-lun68(0) > >> > >> > >> > >> home zimbra-mb1 /dev/md1(832) > >> > >> > >> > >> root zimbra-mb1 /dev/md1(512) > >> > >> > >> > >> swap zimbra-mb1 /dev/md1(416) > >> > >> > >> > >> tmp zimbra-mb1 /dev/md1(1152) > >> > >> > >> > >> var zimbra-mb1 /dev/md1(1472) > >> > >> > >> -Brian > >> > >> > >> malahal@us.ibm.com wrote: > >>> Try running "lvs -o lv_name,vg_name,devices". It will list your logical > >>> volumes and their backing devices. I also found that your stripes are on > >>> single path devices as Chandra found. > >>> > >>> Maybe a problem with your filter??? > >>> > >>> --Malahal. > >>> > >>> Chandra Seetharaman [sekharan@us.ibm.com] wrote: > >>>> Hi Brian, > >>>> > >>>> which of these are your LVM volumes ? > >>>> > >>>> Perusal of "dmsetup table" shows you do not have your LVM volumes on top > >>>> of multipath devices. (I an elaborate on it once I get your answer for > >>>> the question above) > >>>> > >>>> chandra > >>>> On Wed, 2009-07-15 at 17:22 -0500, Brian E. Finley wrote: > >>>>> Chandra, > >>>>> > >>>>> root@zimbra-mb1:~# dmsetup ls > >>>>> mb1_t1-redo (254, 4) > >>>>> zimbra-mb1-t2-v6_fujitsu1-lun68 (254, 29) > >>>>> zimbra--mb1-home (254, 19) > >>>>> zimbra-mb1-t2-v5_fujitsu1-lun67 (254, 24) > >>>>> mb1_t1-bin (254, 0) > >>>>> zimbra-mb1-t2-v4_fujitsu1-lun66 (254, 6) > >>>>> zimbra--mb1-var (254, 21) > >>>>> mb1_t2-hsm_store3 (254, 12) > >>>>> zimbra-mb1-t2-v3_fujitsu1-lun65 (254, 28) > >>>>> zimbra-mb1-t2-v2_fujitsu1-lun64 (254, 26) > >>>>> mb1_t2-hsm_store2 (254, 10) > >>>>> zimbra--mb1-swap (254, 17) > >>>>> zimbra-mb1-t2-v1_fujitsu1-lun63 (254, 25) > >>>>> zimbra--mb1-root (254, 18) > >>>>> 35000c5000b36aa2b (254, 11) > >>>>> mb1_t2-hsm_store1 (254, 7) > >>>>> mb1_t1-store (254, 5) > >>>>> mb1_t1-db (254, 2) > >>>>> mb1_t1-log (254, 1) > >>>>> 35000c5000b15fe7b-part2 (254, 14) > >>>>> 35000c5000b15fe7b-part1 (254, 13) > >>>>> zimbra-mb1-t1-v4_fujitsu1-lun58 (254, 9) > >>>>> zimbra-mb1-t1-v3_fujitsu1-lun57 (254, 22) > >>>>> zimbra-mb1-t1-v2_fujitsu1-lun56 (254, 27) > >>>>> zimbra-mb1-t1-v1_fujitsu1-lun55 (254, 23) > >>>>> 35000c5000b36aa2b-part2 (254, 16) > >>>>> 35000c5000b15fe7b (254, 8) > >>>>> 35000c5000b36aa2b-part1 (254, 15) > >>>>> zimbra--mb1-tmp (254, 20) > >>>>> mb1_t1-index (254, 3) > >>>>> > >>>>> > >>>>> root@zimbra-mb1:~# dmsetup table > >>>>> mb1_t1-redo: 0 314572800 striped 4 128 8:0 99615104 8:16 99615104 8:32 99615104 8:48 99615104 > >>>>> zimbra-mb1-t2-v6_fujitsu1-lun68: 0 491911168 multipath 1 queue_if_no_path 0 1 1 round-robin 0 4 1 66:112 1000 8:144 1000 65:48 1000 65:208 1000 > >>>>> zimbra--mb1-home: 0 20971520 linear 9:1 54526336 > >>>>> zimbra-mb1-t2-v5_fujitsu1-lun67: 0 743178240 multipath 1 queue_if_no_path 0 1 1 round-robin 0 4 1 65:32 1000 8:128 1000 65:192 1000 66:96 1000 > >>>>> mb1_t1-bin: 0 41943040 striped 4 128 8:0 384 8:16 384 8:32 384 8:48 384 > >>>>> zimbra-mb1-t2-v4_fujitsu1-lun66: 0 409600000 multipath 1 queue_if_no_path 0 1 1 round-robin 0 4 1 8:112 1000 65:16 1000 65:176 1000 66:80 1000 > >>>>> zimbra--mb1-var: 0 41943040 linear 9:1 96469376 > >>>>> mb1_t2-hsm_store3: 0 718585856 linear 65:192 24584576 > >>>>> mb1_t2-hsm_store3: 718585856 329990144 linear 65:208 384 > >>>>> zimbra-mb1-t2-v3_fujitsu1-lun65: 0 409600000 multipath 1 queue_if_no_path 0 1 1 round-robin 0 4 1 8:96 1000 65:0 1000 65:160 1000 66:64 1000 > >>>>> zimbra-mb1-t2-v2_fujitsu1-lun64: 0 1024000000 multipath 1 queue_if_no_path 0 1 1 round-robin 0 4 1 66:48 1000 8:80 1000 8:240 1000 65:144 1000 > >>>>> mb1_t2-hsm_store2: 0 1023991808 linear 8:240 384 > >>>>> mb1_t2-hsm_store2: 1023991808 24584192 linear 65:192 384 > >>>>> zimbra--mb1-swap: 0 6291456 linear 9:1 27263360 > >>>>> zimbra-mb1-t2-v1_fujitsu1-lun63: 0 1056964608 multipath 1 queue_if_no_path 0 1 1 round-robin 0 4 1 8:64 1000 8:224 1000 65:128 1000 66:32 1000 > >>>>> zimbra--mb1-root: 0 20971520 linear 9:1 33554816 > >>>>> 35000c5000b36aa2b: 0 286739329 multipath 0 0 1 1 round-robin 0 1 1 66:144 1000 > >>>>> mb1_t2-hsm_store1: 0 1048576000 linear 8:224 384 > >>>>> mb1_t1-store: 0 419430400 striped 4 128 8:0 178258304 8:16 178258304 8:32 178258304 8:48 178258304 > >>>>> mb1_t1-db: 0 104857600 striped 4 128 8:0 20971904 8:16 20971904 8:32 20971904 8:48 20971904 > >>>>> mb1_t1-log: 0 41943040 striped 4 128 8:0 10486144 8:16 10486144 8:32 10486144 8:48 10486144 > >>>>> 35000c5000b15fe7b-part2: 0 286326495 linear 254:8 401625 > >>>>> 35000c5000b15fe7b-part1: 0 401562 linear 254:8 63 > >>>>> zimbra-mb1-t1-v4_fujitsu1-lun58: 0 409600000 multipath 1 queue_if_no_path 0 1 1 round-robin 0 4 1 8:48 1000 8:208 1000 65:112 1000 66:16 1000 > >>>>> zimbra-mb1-t1-v3_fujitsu1-lun57: 0 409600000 multipath 1 queue_if_no_path 0 1 1 round-robin 0 4 1 65:96 1000 8:32 1000 8:192 1000 66:0 1000 > >>>>> zimbra-mb1-t1-v2_fujitsu1-lun56: 0 409600000 multipath 1 queue_if_no_path 0 1 1 round-robin 0 4 1 8:176 1000 8:16 1000 65:80 1000 65:240 1000 > >>>>> zimbra-mb1-t1-v1_fujitsu1-lun55: 0 409600000 multipath 1 queue_if_no_path 0 1 1 round-robin 0 4 1 8:0 1000 8:160 1000 65:64 1000 65:224 1000 > >>>>> 35000c5000b36aa2b-part2: 0 286326495 linear 254:11 401625 > >>>>> 35000c5000b15fe7b: 0 286739329 multipath 0 0 1 1 round-robin 0 1 1 66:128 1000 > >>>>> 35000c5000b36aa2b-part1: 0 401562 linear 254:11 63 > >>>>> zimbra--mb1-tmp: 0 20971520 linear 9:1 75497856 > >>>>> mb1_t1-index: 0 209715200 striped 4 128 8:0 47186304 8:16 47186304 8:32 47186304 8:48 47186304 > >>>>> > >>>>> > >>>>> Thanks, -Brian > >>>>> > >>>>> > >>>>> ----- Original Message ----- > >>>>> From: "Chandra Seetharaman" <sekharan@us.ibm.com> > >>>>> To: "Brian Elliott Finley" <finley@anl.gov>, "LVM general discussion and development" <linux-lvm@redhat.com> > >>>>> Sent: Wednesday, July 15, 2009 5:11:47 PM GMT -06:00 US/Canada Central > >>>>> Subject: Re: multipath works; lvm on multipath does not > >>>>> > >>>>> Can you send the o/p of "dmsetup ls" and "dmsetup table" > >>>>> > >>>>> On Wed, 2009-07-15 at 16:53 -0500, finley@anl.gov wrote: > >>>>>> How can I tell what is causing LVM on multipath volumes to fail, while direct multipath volumes are working? > >>>>>> > >>>>>> > >>>>>> The details: > >>>>>> > >>>>>> I have several volumes that are multipathed over fibre channel (2 qlogic cards in the host) from a Fujitsu Eternus 4000 (4 controllers). > >>>>>> > >>>>>> When using a file system mounted on the dm-multipath device directly (Ie.: generating I/O via IOZone), it fails over gracefully when I pull out either of the fibre pairs, and the file system continues to operate. This device is called /dev/mapper/zimbra-mb1-t2-v4_fujitsu1-lun66. > >>>>>> > >>>>>> However, when I access an LVM volume created with sister devices as PVs, the LVM presented volume does not fail over. Rather the LV device goes inaccessible and the file system (ext3) re-mounts itself as read-only. If I try to access any of the underlying PVs, they still respond as available (Ie.: fdisk -l $PV). > >>>>>> > >>>>>> I am not seeing duplicate devices, and am using the following filter in /etc/lvm/lvm.conf: > >>>>>> > >>>>>> filter = [ "a|^/dev/md.*|", "a|^/dev/mapper/zimbra-mb.*-t.*-v.*_.*-lun.*|", "r|.*|" ] > >>>>>> > >>>>>> > >>>>>> I have also added the following, but dm-multipath devices seem usable by LVM with or without this setting: > >>>>>> > >>>>>> types = [ "device-mapper", 1 ] > >>>>>> > >>>>>> > >>>>>> I have tried removing the /etc/lvm/cache/.cache file, but that seems to have had no effect. I've also tried re-building the initrd after modifying the lvm.conf file, also with no effect. > >>>>>> > >>>>>> Additional info: > >>>>>> root@zimbra-mb1:~# pvs > >>>>>> PV VG Fmt Attr PSize PFree > >>>>>> /dev/mapper/zimbra-mb1-t1-v1_fujitsu1-lun55 mb1_t1 lvm2 a- 195.31G 60.31G > >>>>>> /dev/mapper/zimbra-mb1-t1-v2_fujitsu1-lun56 mb1_t1 lvm2 a- 195.31G 60.31G > >>>>>> /dev/mapper/zimbra-mb1-t1-v3_fujitsu1-lun57 mb1_t1 lvm2 a- 195.31G 60.31G > >>>>>> /dev/mapper/zimbra-mb1-t1-v4_fujitsu1-lun58 mb1_t1 lvm2 a- 195.31G 60.31G > >>>>>> /dev/mapper/zimbra-mb1-t2-v1_fujitsu1-lun63 mb1_t2 lvm2 a- 504.00G 4.00G > >>>>>> /dev/mapper/zimbra-mb1-t2-v2_fujitsu1-lun64 mb1_t2 lvm2 a- 488.28G 0 > >>>>>> /dev/mapper/zimbra-mb1-t2-v3_fujitsu1-lun65 mb1_t2 lvm2 a- 195.31G 195.31G > >>>>>> /dev/mapper/zimbra-mb1-t2-v5_fujitsu1-lun67 mb1_t2 lvm2 a- 354.37G 0 > >>>>>> /dev/mapper/zimbra-mb1-t2-v6_fujitsu1-lun68 mb1_t2 lvm2 a- 234.56G 77.21G > >>>>>> /dev/md1 zimbra-mb1 lvm2 a- 136.50G 83.50G > >>>>>> > >>>>>> > >>>>>> > >>>>>> kernel version: 2.6.24-24-server > >>>>>> distoro: Ubuntu Hardy > >>>>>> multipath-tools: 0.4.8-7ubuntu2 > >>>>>> lvm2: 2.02.26-1ubuntu9 > >>>>>> > >>>>>> Thanks! > >>>>>> > >>>>>> -Brian > >>>>>> > >>>>>> _______________________________________________ > >>>>>> linux-lvm mailing list > >>>>>> linux-lvm@redhat.com > >>>>>> https://www.redhat.com/mailman/listinfo/linux-lvm > >>>>>> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/ > >>>> _______________________________________________ > >>>> linux-lvm mailing list > >>>> linux-lvm@redhat.com > >>>> https://www.redhat.com/mailman/listinfo/linux-lvm > >>>> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/ > >>> _______________________________________________ > >>> linux-lvm mailing list > >>> linux-lvm@redhat.com > >>> https://www.redhat.com/mailman/listinfo/linux-lvm > >>> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/ > >> -- > >> Brian Elliott Finley > >> Deputy Manager, Unix, Storage, and Operations > >> Computing and Information Systems > >> Argonne National Laboratory > >> Office: 630.252.4742 > >> Mobile: 630.447.9108 > >> > >> _______________________________________________ > >> linux-lvm mailing list > >> linux-lvm@redhat.com > >> https://www.redhat.com/mailman/listinfo/linux-lvm > >> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/ > > > > _______________________________________________ > > linux-lvm mailing list > > linux-lvm@redhat.com > > https://www.redhat.com/mailman/listinfo/linux-lvm > > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/ > > -- > Brian Elliott Finley > Deputy Manager, Unix, Storage, and Operations > Computing and Information Systems > Argonne National Laboratory > Office: 630.252.4742 > Mobile: 630.447.9108 > > _______________________________________________ > linux-lvm mailing list > linux-lvm@redhat.com > https://www.redhat.com/mailman/listinfo/linux-lvm > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/ _______________________________________________ linux-lvm mailing list linux-lvm@redhat.com https://www.redhat.com/mailman/listinfo/linux-lvm read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/