inconsistend volume group after pvmove

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

This is our setup: We have 3 Linux servers (2.6.18 Centos 5), clustered, with a clvmd running one “big” volume group (15 SCSI disks a 69,9 GB).

After we got an hardware I/O error on one disk out gfs filesystem began to loop.

So we stopped all services and we determined the corrupted disk (/dev/sdh) and my intention was to do the following:

-          pvmove /dev/sdh

-          vgreduce my_volumegroup /dev/sdh

-          do an intensive hardware check on the volume

 

 

But: that’s what happened during pvmove –v /dev/sdh:

…….

/dev/sdh: Moved: 78,6%

  /dev/sdh: Moved: 79,1%

  /dev/sdh: Moved: 79,7%

  /dev/sdh: Moved: 80,0%

    Updating volume group metadata

    Creating volume group backup "/etc/lvm/backup/myvol_vg" (seqno 46).

  Error locking on node server1: device-mapper: reload ioctl failed: Das Argument ist ungültig

  Unable to reactivate logical volume "pvmove0"

  ABORTING: Segment progression failed.

    Removing temporary pvmove LV

    Writing out final volume group after pvmove

    Creating volume group backup "/etc/lvm/backup/myvol_vg" (seqno 48).

[root@hpserver1 ~]# pvscan

  PV /dev/cciss/c0d0p2   VG VolGroup00   lvm2 [33,81 GB / 0    free]

  PV /dev/sda            VG fit_vg       lvm2 [68,36 GB / 0    free]

  PV /dev/sdb            VG fit_vg       lvm2 [68,36 GB / 0    free]

  PV /dev/sdc            VG fit_vg       lvm2 [68,36 GB / 0    free]

  PV /dev/sdd            VG fit_vg       lvm2 [68,36 GB / 0    free]

  PV /dev/sde            VG fit_vg       lvm2 [66,75 GB / 46,75 GB free]

  PV /dev/sdf            VG fit_vg       lvm2 [68,36 GB / 0    free]

  PV /dev/sdg            VG fit_vg       lvm2 [68,36 GB / 0    free]

  PV /dev/sdh            VG fit_vg       lvm2 [68,36 GB / 58,36 GB free]

  PV /dev/sdj            VG fit_vg       lvm2 [68,36 GB / 54,99 GB free]

  PV /dev/sdi            VG fit_vg       lvm2 [68,36 GB / 15,09 GB free]

  PV /dev/sdk1           VG fit_vg       lvm2 [68,36 GB / 55,09 GB free]

  Total: 12 [784,20 GB] / in use: 12 [784,20 GB] / in no VG: 0 [0   ]

 

That sounded bad, and I didn’t have any idea what to do, but read, that pvmove can start at the point it was, so I started pvmove againg and now pvmove could move all data.

pvscan and vgscan -vvv showed me, that all data were moved from the /dev/sdh volume to the other volumes.

 

To be sure I restarted my cluster nodes, but they encountered problems mounting the gfs filesystems.

I got this error:

 

[root@server1 ~]# /etc/init.d/clvmd stop

Deactivating VG myvol_vg:   Volume group "myvol_vg" inconsistent

  WARNING: Inconsistent metadata found for VG myvol_vg - updating to use version 148

  0 logical volume(s) in volume group "myvol_vg" now active

                                                           [  OK  ]

Stopping clvm:                                             [  OK  ]

[root@server1 ~]# /etc/init.d/clvmd start

Starting clvmd:                                            [  OK  ]

Activating VGs:   2 logical volume(s) in volume group "VolGroup00" now active

  Volume group "myvol_vg" inconsistent

  WARNING: Inconsistent metadata found for VG myvol_vg - updating to use version 151

  Error locking on node server1: Volume group for uuid not found: tGRfaK5aW00pFRXcLtrdHAw5a4GNDVBtuFZZe8QKoX8sVA0XRTNoDQVWVftk8cSa

  Error locking on node server1: Volume group for uuid not found: tGRfaK5aW00pFRXcLtrdHAw5a4GNDVBtqDfFtrJTFTGuju8nNjwtCdPGnzP3hh8k

  Error locking on node server1: Volume group for uuid not found: tGRfaK5aW00pFRXcLtrdHAw5a4GNDVBtc22hBY40phdVvVdFBFX28PvfF7JrlIYz

  Error locking on node server1: Volume group for uuid not found: tGRfaK5aW00pFRXcLtrdHAw5a4GNDVBtWfJ1EqXJ309gO3Gx0ZvpNekrmHFo9u2V

  Error locking on node server1: Volume group for uuid not found: tGRfaK5aW00pFRXcLtrdHAw5a4GNDVBtCP6czghnQFEjNdv9DF6bsUmnK3eJ5vKp

  Error locking on node server1: Volume group for uuid not found: tGRfaK5aW00pFRXcLtrdHAw5a4GNDVBt0KNlnblpwOfcnqIjk4GJ662dxOsL70GF

  0 logical volume(s) in volume group "myvol_vg" now active

                                                           [  OK  ]

 

As I take a look at it, these 6 volumes are exactly the LVs which should be found and where all datas are stored.

 

The next step was in the beginning step by step and in the end stupid try and error.

This was one of the first actions:

 

[root@hpserver1 ~]# vgreduce --removemissing myvol_vg

    Logging initialised at Tue Jul  1 10:00:52 2008

    Set umask to 0077

    Finding volume group "myvol_vg"

    Wiping cache of LVM-capable devices

  WARNING: Inconsistent metadata found for VG myvol_vg - updating to use version 229

  Volume group "myvol_vg" is already consistent

 

We tried to deactivate the volume via vgchange –n y myvol_vg, we tried to “removemissing” and sadly after a few concurrent tries (dmsetup info –c, dmsetup mknodes and vgchange –n y myvol_vg) we can access our LVs, but we still get this message and we don’t know why:

 

  Volume group "myvol_vg" inconsistent

  WARNING: Inconsistent metadata found for VG myvol_vg - updating to use version 228

 

I’m a little bit worried about our data,

 

Regards

Andreas

 

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux