I just tried this myself.... 1. First create a new PV on a whole disk, VG and LV # pvcreate /dev/sdc # vgcreate test /dev/sdc # lvcreate -L2G -n testlv test 2. Format the LV, mount it and copy some data to it (just a random tarball) # mke2fs -j /dev/test/testlv # mount /dev/test/testlv /mnt # tar -C /mnt -xvzf ~/iscsitarget-0.4.17.tar.gz # umount /mnt # e2fsck /dev/test/testlv e2fsck 1.39 (29-May-2006) /dev/test/testlv: clean, 87/262144 files, 25554/524288 blocks 3. So the LV is OK. Now I'll make sure there's a config backup, then wipe the PV label... # vgcfgbackup test Volume group "test" successfully backed up. # vgchange -an test 0 logical volume(s) in volume group "test" now active [root@europa ~]# pvremove -ff /dev/sdc Really WIPE LABELS from physical volume "/dev/sdc" of volume group "test" [y/n]? y WARNING: Wiping physical volume label from /dev/sdc of volume group "test" Labels on physical volume "/dev/sdc" successfully wiped 4. Now, I'll try to recreate the PV using the backup data, and see if the contents are intact. # pvcreate --uuid="A0LDgs-KMlm-QEBR-sGNW-7Rlf-j3aU-x2JUKY" --restorefile=/etc/lvm/backup/test /dev/sdc Couldn't find device with uuid 'A0LDgs-KMlm-QEBR-sGNW-7Rlf-j3aU-x2JUKY'. Physical volume "/dev/sdc" successfully created # vgcfgrestore -f /etc/lvm/backup/test test Restored volume group test 5. Check to see if we can see the previously created LV, and mount it # lvs LV VG Attr LSize Origin Snap% Move Log Copy% Convert testlv test -wi--- 2.00G # vgchange -ay test 1 logical volume(s) in volume group "test" now active # e2fsck /dev/test/testlv e2fsck 1.39 (29-May-2006) /dev/test/testlv: clean, 87/262144 files, 25554/524288 blocks # mount /dev/test/testlv /mnt # ls /mnt iscsitarget-0.4.17 lost+found So, YMMV but it appears from my experiments that this operation should be safe, and should recover your volumes. I am concerned though about the news that the RHEL 5.3->5.4 upgrade may have caused this, as we're looking at making the same upgrade before not too long. Do you have any suspicion as to why this may have happened ? Have you filed a bug with Red Hat ? Regards, -Mark -----Original Message----- From: linux-lvm-bounces@redhat.com [mailto:linux-lvm-bounces@redhat.com] On Behalf Of Julie Ashworth Sent: 24 September 2009 11:06 To: linux-lvm@redhat.com Subject: missing physical volumes after upgrade to rhel 5.4 I apologize for the cross-posting (to rhelv5-list). The lvm list is a more relevant list for my problem, and I'm sorry I didn't realize this sooner. After an upgrade from rhel5.3 -> rhel5.4 (and reboot) I can no longer see PVs for 3 fibre-channel storage devices. The operating system still see the disk: ---------------------- # multipath -l mpath2 (2001b4d28000064db) dm-1 JetStor,Volume Set # 00 [size=12T][features=0][hwhandler=0][rw] \_ round-robin 0 [prio=0][active] \_ 11:0:1:0 sdj 8:144 [active][undef] mpath16 (1ACNCorp_FF01000113200019) dm-2 ACNCorp,R_LogVol-despo [size=15T][features=0][hwhandler=0][rw] \_ round-robin 0 [prio=0][active] \_ 11:0:2:0 sdk 8:160 [active][undef] mpath7 (32800001b4d00cf5b) dm-0 JetStor,Volume Set 416F [size=12T][features=0][hwhandler=0][rw] \_ round-robin 0 [prio=0][active] \_ 11:0:0:0 sdi 8:128 [active][undef] ---------------------- There are files in /etc/lvm/backup/ that contain the original volume information, e.g. ---------------------- jetstor642 { id = "0e53Q3-evHX-I5f9-CWqf-NPcw-IqmC-0fVcTO" seqno = 2 status = ["RESIZEABLE", "READ", "WRITE"] flags = [] extent_size = 8192 # 4 Megabytes max_lv = 0 max_pv = 0 physical_volumes { pv0 { id = "5wJCEA-IDC1-5GhI-jnEs-EpYF-8Uf3-sqPL4O" device = "/dev/dm-7" # Hint only status = ["ALLOCATABLE"] flags = [] dev_size = 31214845952 # 14.5355 Terabytes pe_start = 384 pe_count = 3810405 # 14.5355 Terabytes } } ---------------------- The devices were formatted using parted on the entire disk, i.e. I didn't create a partition. The partition table is "gpt" (possible label types are "bsd", "dvh", "gpt", "loop", "mac", "msdos", "pc98" or "sun".) partition table information for one of the devices is below: -------------------------- # parted /dev/sdi GNU Parted 1.8.1 Using /dev/sdi Welcome to GNU Parted! Type 'help' to view a list of commands. (parted) print Model: JetStor Volume Set 416F (scsi) Disk /dev/sdi: 13.0TB Sector size (logical/physical): 512B/512B Partition Table: gpt Number Start End Size File system Name Flags -------------------------- output of some commands: $ pvdisplay returns nothing (no error) $ lvs -a -o +devices returns nothing (no error) $ pvck -vvvvv /dev/sdb #lvmcmdline.c:915 Processing: pvck -vvvvv /dev/sdb #lvmcmdline.c:918 O_DIRECT will be used #config/config.c:950 Setting global/locking_type to 3 #locking/locking.c:245 Cluster locking selected. #locking/cluster_locking.c:83 connect() failed on local socket: Connection +refused #config/config.c:955 locking/fallback_to_local_locking not found in +config: defaulting to 1 WARNING: Falling back to local file-based locking. Volume Groups with the clustered attribute will be inaccessible. #config/config.c:927 Setting global/locking_dir to /var/lock/lvm #pvck.c:32 Scanning /dev/sdb #device/dev-cache.c:260 /dev/sdb: Added to device cache #device/dev-io.c:439 Opened /dev/sdb RO #device/dev-io.c:260 /dev/sdb: size is 25395814912 sectors #device/dev-io.c:134 /dev/sdb: block size is 4096 bytes #filters/filter.c:124 /dev/sdb: Skipping: Partition table signature +found #device/dev-io.c:485 Closed /dev/sdb #metadata/metadata.c:2337 Device /dev/sdb not found (or ignored by filtering). ------------------------- from doing google searches, I found this gem to restore a PV: pvcreate --uuid "cqH4SD-VrCw-jMsN-GcwH-omCq-ThpE-dO9KmJ" --restorefile /etc/lvm/backup/vg_04 /dev/sdd1 however, the man page says to 'use with care'. I don't want to lose data. Can anybody comment on how safe it would be to run this? Thanks in advance, Julie Ashworth -- Julie Ashworth <julie.ashworth@berkeley.edu> Computational Infrastructure for Research Labs, UC Berkeley http://cirl.berkeley.edu/ PGP Key ID: 0x17F013D2 _______________________________________________ linux-lvm mailing list linux-lvm@redhat.com https://www.redhat.com/mailman/listinfo/linux-lvm read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/ _______________________________________________ linux-lvm mailing list linux-lvm@redhat.com https://www.redhat.com/mailman/listinfo/linux-lvm read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/