Rodrigo, which LVM version is this ? Because we fixed some snapshot bugs in 1.0.7 you should give that version a try. In order to get your "vg1" back, you should try to remove the snapshots with "vgscan -r". Regards, Heinz -- The LVM Guy -- On Thu, Mar 13, 2003 at 04:56:37PM +0100, Rodrigo de Salazar wrote: > Ok, I will paste the output of the relevant commands, if you want more info > just ask... and thanks for the quick answer, this machine is destined to be > an important server, and it should be ready ASAP, aka if we can't get it to > work with LVM soon, we'll have to use a traditional scheme, superior's orders > /: > So here it goes: > > volatil:~# vgck > vgck -- VGDA of "vg1" in lvmtab is consistent > vgck -- VGDA of "vg1" on physical volume is consistent > vgck -- VGDA of "vg2" in lvmtab is consistent > vgck -- VGDA of "vg2" on physical volume is consistent > > volatil:~# pvscan > pvscan -- reading all physical volumes (this may take a while...) > pvscan -- inactive PV "/dev/ide/host0/bus1/target0/lun0/disc" is in no VG > [8.03 GB] > pvscan -- ACTIVE PV "/dev/ide/host0/bus1/target0/lun0/part2" of VG "vg2" > [74.05 GB / 10.05 GB free] > pvscan -- inactive PV "/dev/ide/host0/bus0/target0/lun0/part4" of VG "vg1" > [35.74 GB / 2.79 GB free] > pvscan -- total: 3 [117.83 GB] / in use: 2 [109.80 GB] / in no VG: 1 [8.03 GB] > > The 8.03 GB PV makes no sense to me, it is supposedly /dev/hdc, but that disk > (80 GB) is partitioned, first partition 512 MB swap and the second is the LVM > partition (the rest of the disk). > > volatil:~# vgscan > vgscan -- reading all physical volumes (this may take a while...) > vgscan -- found active volume group "vg2" > vgscan -- found inactive volume group "vg1" > vgscan -- "/etc/lvmtab" and "/etc/lvmtab.d" successfully created > vgscan -- WARNING: This program does not do a VGDA backup of your volume > groups > > volatil:~# lvscan > lvscan -- volume group "vg1" is NOT active; try -D > lvscan -- ACTIVE Original "/dev/vg2/home" [45 GB] > lvscan -- ACTIVE "/dev/vg2/usr" [10 GB] > lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-16" [1008 MB] of /dev/vg2/home > lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-19" [1008 MB] of /dev/vg2/home > lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-10" [1008 MB] of /dev/vg2/home > lvscan -- ACTIVE Snapshot "/dev/vg2/home-hora-13" [1008 MB] of /dev/vg2/home > lvscan -- ACTIVE "/dev/vg2/local" [5 GB] > lvscan -- 7 logical volumes with 63.94 GB total in 1 volume group > lvscan -- 7 active logical volumes > > volatil:~# lvscan -D > lvscan -- reading all physical volumes (this may take a while...) > lvscan -- inactive "/dev/vg1/tmp" [1 GB] > lvscan -- inactive "/dev/vg1/var" [15 GB] > lvscan -- inactive Original "/dev/vg1/www" [15 GB] > lvscan -- inactive Snapshot "/dev/vg1/www-hora-16" [0] of /dev/vg1/www > lvscan -- inactive Snapshot "/dev/vg1/www-hora-19" [0] of /dev/vg1/www > lvscan -- inactive Snapshot "/dev/vg1/www-hora-10" [0] of /dev/vg1/www > lvscan -- inactive Snapshot "/dev/vg1/www-hora-13" [0] of /dev/vg1/www > lvscan -- inactive Original "/dev/vg2/home" [45 GB] > lvscan -- inactive "/dev/vg2/usr" [10 GB] > lvscan -- inactive Snapshot "/dev/vg2/home-hora-16" [0] of /dev/vg2/home > lvscan -- inactive Snapshot "/dev/vg2/home-hora-19" [0] of /dev/vg2/home > lvscan -- inactive Snapshot "/dev/vg2/home-hora-10" [0] of /dev/vg2/home > lvscan -- inactive Snapshot "/dev/vg2/home-hora-13" [0] of /dev/vg2/home > lvscan -- inactive "/dev/vg2/local" [5 GB] > lvscan -- 14 logical volumes with 91 GB total in 2 volume groups > lvscan -- 14 inactive logical volumes > > As you can see, there are 4 daily snapshots of /home and /var/www, maybe it > is too much, but those are the directories where people work and well, files > get accidentally deleted during the day and so on and so forth (: > All were available when this happened. > > volatil:~# vgchange -a y > vgchange -- ERROR "Bad address" activating volume group "vg1" > vgchange -- volume group "vg2" already active > > volatil:~# pvdisplay /dev/hda4 > --- Physical volume --- > PV Name /dev/ide/host0/bus0/target0/lun0/part4 > VG Name vg1 > PV Size 35.75 GB [74970000 secs] / NOT usable 4.19 MB [LVM: 163 > KB] > PV# 1 > PV Status available > Allocatable yes > Cur LV 7 > PE Size (KByte) 4096 > Total PE 9150 > Free PE 714 > Allocated PE 8436 > PV UUID kvIRnL-ZNQm-Ab0g-TSqe-rasN-Hg1C-pHZIkF > > volatil:~# vgdisplay -D > --- Volume group --- > VG Name vg2 > VG Access read/write > VG Status NOT available/resizable > VG # 1 > MAX LV 256 > Cur LV 7 > Open LV 0 > MAX LV Size 255.99 GB > Max PV 256 > Cur PV 1 > Act PV 1 > VG Size 74.05 GB > PE Size 4 MB > Total PE 18956 > Alloc PE / Size 16384 / 64 GB > Free PE / Size 2572 / 10.05 GB > VG UUID Nm7brW-Hk8I-hYkt-1LvU-6COW-G5ds-FduaFe > > --- Volume group --- > VG Name vg1 > VG Access read/write > VG Status NOT available/resizable > VG # 0 > MAX LV 256 > Cur LV 7 > Open LV 0 > MAX LV Size 255.99 GB > Max PV 256 > Cur PV 1 > Act PV 1 > VG Size 35.74 GB > PE Size 4 MB > Total PE 9150 > Alloc PE / Size 8436 / 32.95 GB > Free PE / Size 714 / 2.79 GB > VG UUID GNeKh6-W3TL-AMa6-3E7q-291d-pjm2-av59ou > > Another thing, originally /dev/vg1/var was 30 GB, it was resized with: > > e2fsadm -L-15G /dev/vg1/var > > (only 9 GB were occupied) and then the /dev/vg1/www was created... > Also, they have ext3 filesystems on them, but with the journal disabled > (working as ext2). It all worked fine even on reboot, a few days before the > "accident". > > Don't know what else to include... if I missed out something, just ask for > it, and thanks for the help (: > > _______________________________________________ > linux-lvm mailing list > linux-lvm@sistina.com > http://lists.sistina.com/mailman/listinfo/linux-lvm > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/ *** Software bugs are stupid. Nevertheless it needs not so stupid people to solve them *** =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=- Heinz Mauelshagen Sistina Software Inc. Senior Consultant/Developer Am Sonnenhang 11 56242 Marienrachdorf Germany Mauelshagen@Sistina.com +49 2626 141200 FAX 924446 =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=- _______________________________________________ linux-lvm mailing list linux-lvm@sistina.com http://lists.sistina.com/mailman/listinfo/linux-lvm read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/