RE: mounting of unknown filesystem type on RAID5 array

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Neither of those tools will work as a RAID array doesn't have a
partition table, just a filesystem.

I've discovered that the filesystem is an LVM filesystem and I've done a
'vgscan' which found two volume groups from 2 of the RAID arrays I'm
trying to recover:

[root@masterbackend root]# vgdisplay -D
--- Volume group ---
VG Name               vgroup00
VG Access             read/write
VG Status             NOT available/resizable
VG #                  0
MAX LV                256
Cur LV                1
Open LV               0
MAX LV Size           1023.97 GB
Max PV                256
Cur PV                1
Act PV                1
VG Size               832.28 GB
PE Size               16 MB
Total PE              53266
Alloc PE / Size       53266 / 832.28 GB
Free  PE / Size       0 / 0
VG UUID               oizRKm-JFUq-hMiZ-rN6F-1M7u-mRDc-vqqy1p

--- Volume group ---
VG Name               logdev
VG Access             read/write
VG Status             NOT available/resizable
VG #                  1
MAX LV                256
Cur LV                2
Open LV               0
MAX LV Size           255.99 GB
Max PV                256
Cur PV                1
Act PV                1
VG Size               1.46 GB
PE Size               4 MB
Total PE              375
Alloc PE / Size       138 / 552 MB
Free  PE / Size       237 / 948 MB
VG UUID               nCpyXh-5bn4-Qh2W-UlAc-3dyh-zQOT-i33ow8

So far I'm not understanding how to make the VG Status 'available' and
how to mount them.  I now have the following devices:

/dev/vgroup00/storage1 block special (58/2)
/dev/vgroup00/group character specail (109/0)
/dev/logdev/storage1 block special (58/1)
/dev/logdev/syslog block special (58/0)
/dev/logdev/group character special (109/1)

I believe these are XFS filesystems based on examination of some of the
raw data in the /dev/md's.  But I still can't mount them via:

[root@masterbackend root]# mount /dev/vgroup00/storage1 /mnt/array/ -t
xfs
mount: wrong fs type, bad option, bad superblock on
/dev/vgroup00/storage1,
       or too many mounted file systems

Any ideas?  I'm not familiar with LVM, but have been googling it.  I
think its time to post this over on the linux-lvm maillist.

Tim

> -----Original Message-----
> From: linux-raid-owner@xxxxxxxxxxxxxxx [mailto:linux-raid-
> owner@xxxxxxxxxxxxxxx] On Behalf Of M K
> Sent: Wednesday, June 02, 2004 2:58 PM
> To: Tim Harvey; linux-raid@xxxxxxxxxxxxxxx
> Subject: Re: mounting of unknown filesystem type on RAID5 array
> 
> would it be possible to examine the fs using
> parted or fdisk ?
> --- Tim Harvey <tharvey@xxxxxxxxxxxxxxxxxx> wrote:
> > Greetings,
> >
> > I'm trying to mount a RAID5 array created on a
> > NASAS-2040 which uses an
> > embedded Linux OS which has failed.  I'm having
> > trouble determining the
> > filesystem type used.  I've physically connected 3
> > out of the 4 drives,
> > so as to start it in degraded mode and pull the data
> > off.  Here are
> > sections of my logs:
> >
> > Jun  2 13:58:16 localhost kernel: hda:
> > IBM-DJNA-351520, ATA DISK drive
> > Jun  2 13:58:16 localhost kernel: hdb: Maxtor
> > 5A300J0, ATA DISK drive
> > Jun  2 13:58:16 localhost kernel: blk: queue
> > c03fcf00, I/O limit 4095Mb
> > (mask 0xffffffff)
> > Jun  2 13:58:16 localhost kernel: blk: queue
> > c03fd040, I/O limit 4095Mb
> > (mask 0xffffffff)
> > Jun  2 13:58:16 localhost kernel: hdc: Maxtor
> > 5A300J0, ATA DISK drive
> > Jun  2 13:58:16 localhost kernel: hdd: Maxtor
> > 5A300J0, ATA DISK drive
> > Jun  2 13:58:16 localhost kernel: blk: queue
> > c03fd35c, I/O limit 4095Mb
> > (mask 0xffffffff)
> > Jun  2 13:58:16 localhost kernel: blk: queue
> > c03fd49c, I/O limit 4095Mb
> > (mask 0xffffffff)
> > Jun  2 13:58:16 localhost kernel: ide0 at
> > 0x1f0-0x1f7,0x3f6 on irq 14
> > Jun  2 13:58:16 localhost kernel: ide1 at
> > 0x170-0x177,0x376 on irq 15
> > Jun  2 13:58:16 localhost kernel: hda: attached
> > ide-disk driver.
> > Jun  2 13:58:16 localhost kernel: hda: host
> > protected area => 1
> > Jun  2 13:58:16 localhost kernel: hda: 30033360
> > sectors (15377 MB)
> > w/430KiB Cache, CHS=1869/255/63
> > Jun  2 13:58:16 localhost kernel: hdb: attached
> > ide-disk driver.
> > Jun  2 13:58:16 localhost kernel: hdb: host
> > protected area => 1
> > Jun  2 13:58:16 localhost kernel: hdb: 585940320
> > sectors (300001 MB)
> > w/2048KiB Cache, CHS=36473/255/63, UDMA(133)
> > Jun  2 13:58:16 localhost kernel: hdc: attached
> > ide-disk driver.
> > Jun  2 13:58:16 localhost kernel: hdc: host
> > protected area => 1
> > Jun  2 13:58:16 localhost kernel: hdc: 585940320
> > sectors (300001 MB)
> > w/2048KiB Cache, CHS=36473/255/63, UDMA(133)
> > Jun  2 13:58:16 localhost kernel: hdd: attached
> > ide-disk driver.
> > Jun  2 13:58:16 localhost kernel: hdd: host
> > protected area => 1
> > Jun  2 13:58:16 localhost kernel: hdd: 585940320
> > sectors (300001 MB)
> > w/2048KiB Cache, CHS=36473/255/63, UDMA(133)
> > Jun  2 13:58:16 localhost kernel: Partition check:
> > Jun  2 13:58:16 localhost kernel:  hda: hda1 hda2
> > hda3
> > Jun  2 13:58:16 localhost kernel:  hdb: hdb1 hdb2
> > hdb3
> > Jun  2 13:58:16 localhost kernel:  hdc: hdc1 hdc2
> > hdc3
> > Jun  2 13:58:16 localhost kernel:  hdd: hdd1 hdd2
> > hdd3
> > Jun  2 13:58:16 localhost kernel: ide: late
> > registration of driver.
> > Jun  2 13:58:16 localhost kernel: md: md driver
> > 0.90.0 MAX_MD_DEVS=256,
> > MD_SB_DISKS=27
> > Jun  2 13:58:16 localhost kernel: md: Autodetecting
> > RAID arrays.
> > Jun  2 13:58:16 localhost kernel:  [events:
> > 00000008]
> > Jun  2 13:58:17 localhost last message repeated 2
> > times
> > Jun  2 13:58:17 localhost kernel: md: autorun ...
> > Jun  2 13:58:17 localhost kernel: md: considering
> > hdd1 ...
> > Jun  2 13:58:17 localhost kernel: md:  adding hdd1
> > ...
> > Jun  2 13:58:17 localhost kernel: md:  adding hdc1
> > ...
> > Jun  2 13:58:17 localhost kernel: md:  adding hdb1
> > ...
> > Jun  2 13:58:17 localhost kernel: md: created md0
> > Jun  2 13:58:17 localhost kernel: md: bind<hdb1,1>
> > Jun  2 13:58:17 localhost kernel: md: bind<hdc1,2>
> > Jun  2 13:58:17 localhost kernel: md: bind<hdd1,3>
> > Jun  2 13:58:17 localhost kernel: md: running:
> > <hdd1><hdc1><hdb1>
> > Jun  2 13:58:17 localhost kernel: md: hdd1's event
> > counter: 00000008
> > Jun  2 13:58:17 localhost kernel: md: hdc1's event
> > counter: 00000008
> > Jun  2 13:58:17 localhost kernel: md: hdb1's event
> > counter: 00000008
> > Jun  2 13:58:17 localhost kernel: kmod: failed to
> > exec /sbin/modprobe -s
> > -k md-personality-4, errno = 2
> > Jun  2 13:58:17 localhost kernel: md: personality 4
> > is not loaded!
> > Jun  2 13:58:17 localhost kernel: md :do_md_run()
> > returned -22
> > Jun  2 13:58:17 localhost kernel: md: md0 stopped.
> > Jun  2 13:58:17 localhost kernel: md: unbind<hdd1,2>
> > Jun  2 13:58:17 localhost kernel: md:
> > export_rdev(hdd1)
> > Jun  2 13:58:17 localhost kernel: md: unbind<hdc1,1>
> > Jun  2 13:58:17 localhost kernel: md:
> > export_rdev(hdc1)
> > Jun  2 13:58:17 localhost kernel: md: unbind<hdb1,0>
> > Jun  2 13:58:17 localhost kernel: md:
> > export_rdev(hdb1)
> > Jun  2 13:58:17 localhost kernel: md: ... autorun
> > DONE.
> > ...
> > Jun  2 14:01:59 localhost kernel:  [events:
> > 00000008]
> > Jun  2 14:01:59 localhost kernel: md: bind<hdc1,1>
> > Jun  2 14:01:59 localhost kernel:  [events:
> > 00000008]
> > Jun  2 14:01:59 localhost kernel: md: bind<hdd1,2>
> > Jun  2 14:01:59 localhost kernel:  [events:
> > 00000008]
> > Jun  2 14:01:59 localhost kernel: md: bind<hdb1,3>
> > Jun  2 14:01:59 localhost kernel: md: hdb1's event
> > counter: 00000008
> > Jun  2 14:01:59 localhost kernel: md: hdd1's event
> > counter: 00000008
> > Jun  2 14:01:59 localhost kernel: md: hdc1's event
> > counter: 00000008
> > Jun  2 14:01:59 localhost kernel: raid5: measuring
> > checksumming speed
> > Jun  2 14:01:59 localhost kernel:    8regs     :
> > 2060.800 MB/sec
> > Jun  2 14:01:59 localhost kernel:    32regs    :
> > 1369.200 MB/sec
> > Jun  2 14:01:59 localhost kernel:    pIII_sse  :
> > 3178.800 MB/sec
> > Jun  2 14:01:59 localhost kernel:    pII_mmx   :
> > 3168.800 MB/sec
> > Jun  2 14:01:59 localhost kernel:    p5_mmx    :
> > 4057.600 MB/sec
> > Jun  2 14:01:59 localhost kernel: raid5: using
> > function: pIII_sse
> > (3178.800 MB/sec)
> > Jun  2 14:01:59 localhost kernel: md: raid5
> > personality registered as nr
> > 4
> > Jun  2 14:01:59 localhost kernel: md0: max total
> > readahead window set to
> > 744k
> > Jun  2 14:01:59 localhost kernel: md0: 3 data-disks,
> > max readahead per
> > data-disk: 248k
> > Jun  2 14:01:59 localhost kernel: raid5: device hdb1
> > operational as raid
> > disk 1
> > Jun  2 14:01:59 localhost kernel: raid5: device hdd1
> > operational as raid
> > disk 3
> > Jun  2 14:01:59 localhost kernel: raid5: device hdc1
> > operational as raid
> > disk 2
> > Jun  2 14:01:59 localhost kernel: raid5: md0, not
> > all disks are
> > operational -- trying to recover array
> > Jun  2 14:01:59 localhost kernel: raid5: allocated
> > 4334kB for md0
> > Jun  2 14:01:59 localhost kernel: raid5: raid level
> > 5 set md0 active
> > with 3 out of 4 devices, algorithm 2
> > Jun  2 14:01:59 localhost kernel: RAID5 conf
> > printout:
> > Jun  2 14:01:59 localhost kernel:  --- rd:4 wd:3
> > fd:1
> > Jun  2 14:01:59 localhost kernel:  disk 0, s:0, o:0,
> > n:0 rd:0 us:1
> > dev:[dev 00:00]
> > Jun  2 14:01:59 localhost kernel:  disk 1, s:0, o:1,
> > n:1 rd:1 us:1
> > dev:hdb1
> > Jun  2 14:01:59 localhost kernel:  disk 2, s:0, o:1,
> > n:2 rd:2 us:1
> > dev:hdc1
> > Jun  2 14:01:59 localhost kernel:  disk 3, s:0, o:1,
> > n:3 rd:3 us:1
> > dev:hdd1
> > Jun  2 14:01:59 localhost kernel: RAID5 conf
> > printout:
> > Jun  2 14:01:59 localhost kernel:  --- rd:4 wd:3
> > fd:1
> >
> === message truncated ===
> 
> 
> 
> 
> 
> __________________________________
> Do you Yahoo!?
> Friends.  Fun.  Try the all-new Yahoo! Messenger.
> http://messenger.yahoo.com/
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid"
in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux