Re: lvm raid failure

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




------------------------------

Message: 6
Date: Mon, 10 Dec 2007 21:42:45 -0500
From: "Null EXE" <nullexe@gmail.com>
Subject: lvm raid failure
To: linux-lvm@redhat.com
Message-ID:
       <1f31f75c0712101842q4d71c03bu75926f1deb026a26@mail.gmail.com">1f31f75c0712101842q4d71c03bu75926f1deb026a26@mail.gmail.com>
Content-Type: text/plain; charset="iso-8859-1"

Hi everyone,


I have been running an Edgy fileserver for about 6 months now. When I came
home from work last week I found my files all inaccessible. After
investigation I found issues with my PV. I would like to recover my data and
replace the drives if possible any help is greatly appreciated!

I have one system drive and 6 storage drive, all raid1. The storage layout
is
2x160GB /dev/md0
2x250GB /dev/md1
2x250GB /dev/md2
After investigation I found problems with my lvm. pvdisplay showed:
Couldn't find device with uuid 'xjWU5M-G3WB-tZjB-Tyrk-Q7rN-Yide-1FEVh3'.

I ran the command:
sudo pvcreate --uuid "xjWU5M-G3WB-tZjB-Tyrk-Q7rN-Yide-1FEVh3" --restorefile
/etc/lvm/archive/fileserver_00006.vg /dev/md2

This seemed to recover /dev/md2. I re-ran pvdisplay and got
Couldn't find device with uuid '1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo'.

I figured I could run the same command on md1,
sudo pvcreate --uuid "1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo" --restorefile
/etc/lvm/archive/fileserver_00006.vg /dev/md1

and got the message:
Couldn't find device with uuid '1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo'.
Device /dev/md1 not found (or ignored by filtering).

After re-executing the above command with -vvv at the bottom of the output I
get the following message:
#device/dev-io.c:439         Opened /dev/md1 RO
#device/dev-io.c:264       /dev/md1: size is 0 sectors
#filters/filter.c:106         /dev/md1: Skipping: Too small to hold a PV
#device/dev-io.c:485         Closed /dev/md1
#pvcreate.c:81   Device /dev/md1 not found (or ignored by filtering).

Here is my pvdisplay. Again, any help is Greatly appreciated
***START***

Couldn't find device with uuid '1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo'.
Couldn't find device with uuid '1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo'.
--- Physical volume ---
PV Name /dev/md0
VG Name fileserver
PV Size 149.05 GB / not usable 0
Allocatable yes
PE Size (KByte) 4096
Total PE 38156
Free PE 9996
Allocated PE 28160
PV UUID pZV1Ff-Y7fu-S8m1-tVFn-fOMJ-VRls-fLFEov

--- Physical volume ---
PV Name unknown device
VG Name fileserver
PV Size 232.88 GB / not usable 0
Allocatable yes (but full)
PE Size (KByte) 4096
Total PE 59618
Free PE 0
Allocated PE 59618
PV UUID 1ATm8s-oxKG-nz0p-z1QA-a0s4-od9T-AMZRoo

--- Physical volume ---
PV Name /dev/md2
VG Name fileserver
PV Size 232.88 GB / not usable 0
Allocatable yes
PE Size (KByte) 4096
Total PE 59618
Free PE 9156
Allocated PE 50462
PV UUID xjWU5M-G3WB-tZjB-Tyrk-Q7rN-Yide-1FEVh3
-------------- next part --------------
An HTML attachment was scrubbed...
URL: https://www.redhat.com/archives/linux-lvm/attachments/20071210/1c987df9/attachment.html

------------------------------

Message: 7
Date: Tue, 11 Dec 2007 09:20:29 +0100
From: Luca Berra < bluca@comedia.it>
Subject: Re: lvm raid failure
To: linux-lvm@redhat.com
Message-ID: <20071211082029.GA1535@percy.comedia.it">20071211082029.GA1535@percy.comedia.it >
Content-Type: text/plain; charset=us-ascii; format=flowed

On Mon, Dec 10, 2007 at 09:42:45PM -0500, Null EXE wrote:
>get the following message:
>#device/dev-io.c:439         Opened /dev/md1 RO
>#device/dev-io.c:264       /dev/md1: size is 0 sectors
>#filters/filter.c:106         /dev/md1: Skipping: Too small to hold a PV
>#device/dev-io.c:485         Closed /dev/md1
>#pvcreate.c:81   Device /dev/md1 not found (or ignored by filtering).
>
>Here is my pvdisplay. Again, any help is Greatly appreciated
you should investigate what's wrong at the md layer, lvm seems to be
just a victim.

check
/proc/mdstat
kernel-logs
mdadm.conf
mdadm -Es

L.

--
Luca Berra -- bluca@comedia.it
       Communication Media & Services S.r.l.
 /"\
 \ /     ASCII RIBBON CAMPAIGN
 X        AGAINST HTML MAIL
 / \


  ***
/proc/mdstat
Personalities : [raid1]
md2 : active raid1 dm-5[1]
      244195904 blocks [2/1] [_U]
     
md1 : inactive hdj1[0]
      244195904 blocks super non-persistent
      
md0 : active raid1 hdc1[0]
      156288256 blocks [2/1] [U_]
     
unused devices: <none>

/should my dm-5 be displayed or should it be a /dev/hd[a-z] device

***
mdadm -Es
ARRAY /dev/md0 level=raid1 num-devices=2 UUID=eeb0befe:9e50d574:974f5eae:ccb8e527
ARRAY /dev/md2 level=raid1 num-devices=2 UUID=1e5c1ab5:a5a06f84:d722583d:ca0f8cad
ARRAY /dev/md1 level=raid1 num-devices=2 UUID=41fc507e:7293a47c:7f18baeb:89dd5958

***
mdadm.conf
DEVICE partitions

/proc/partitions
   3     0   10022040 hda
   3     1    9574708 hda1
   3     2          1 hda2
   3     5     441756 hda5
  22     0  156290904 hdc
  22     1  156288321 hdc1
  22    64  156290904 hdd
  22    65  156288321 hdd1
  56     0  244198584 hdi
  56     1  244196001 hdi1
  56    64  244198584 hdj
  56    65  244196001 hdj1
  57     0  244198584 hdk
  57     1  244196001 hdk1
  57    64  244198584 hdl
  57    65  244196001 hdl1
   9     0  156288256 md0
   9     2  244195904 md2
 253     0    9574708 dm-0
 253     1     441756 dm-1
 253     2  156288321 dm-2
 253     3  156288321 dm-3
 253     4  244196001 dm-4
 253     5  244196001 dm-5
 253     6  244196001 dm-6
 253     7  244196001 dm-7


***
kernel logs
Nov 28 22:58:46 ark-server kernel: [42949384.540000 ] raid1: raid set md1 active with 1 out of 2 mirrors
Nov 28 22:58:46 ark-server kernel: [42949384.560000] md: md2 stopped.
Nov 28 22:58:46 ark-server kernel: [42949384.560000] md: bind<hdj1>
Nov 28 22:58:46 ark-server kernel: [ 42949384.560000] md: hdl1 has same UUID but different superblock to hdj1
Nov 28 22:58:46 ark-server kernel: [42949384.560000] md: hdl1 has different UUID to hdj1
Nov 28 22:58:46 ark-server kernel: [42949384.560000] md: export_rdev(hdl1)

***
Looking at all of this.  When I set up the array I remember my devices ordered with hd[cdefgh] now I'm seeing md1 trying to use hd[jl].  Is this a problem that it put assigned the drives new letter automatically?

_______________________________________________
linux-lvm mailing list
linux-lvm@redhat.com
https://www.redhat.com/mailman/listinfo/linux-lvm
read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/

[Index of Archives]     [Gluster Users]     [Kernel Development]     [Linux Clusters]     [Device Mapper]     [Security]     [Bugtraq]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]

  Powered by Linux