Re: large filesystem corruptions

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 03/12/2010 03:01 PM, Kapetanakis Giannis wrote:
Hi,

The last few days I'm trying to setup a large filesystem
14TB and is always getting corrupted...

This is my latest non-working setup. Sorry for long message
but I want to make clear my  actual setup

2.6.18-164.11.1.el5PAE (x86)
4GB RAM

CONFIG_EFI_VARS=y
CONFIG_EFI=y

/dev/sdb - 6 x 1.5TB SATA drives in hardware RAID 5 (256 chunk size)
/dev/sdc - 6 x 1.5TB SATA drives in hardware RAID 5 (256 chunk size)

-- Both hardware raids are GPT labeled

Model: Adaptec ARRAY01 (scsi)
Disk /dev/sdb: 7489GB
Sector size (logical/physical): 512B/512B
Partition Table: gpt

Number  Start   End     Size    File system  Name     Flags
 1      17.4kB  7489GB  7489GB               ARRAY01

Model: Adaptec ARRAY02 (scsi)
Disk /dev/sdc: 7489GB
Sector size (logical/physical): 512B/512B
Partition Table: gpt

Number  Start   End     Size    File system  Name     Flags
 1      17.4kB  7489GB  7489GB               ARRAY02

-- /dev/md_d0 software raid0 on /dev/sdb1 and /dev/sdc1
-- (doing on /dev/sdb and /dev/sdc also corrupted)
-- this raid0 is also GPT labeled (256 chunk size)

mdadm --create /dev/md_d0 -a p1 -c 256 -l 0 -n 2 /dev/sdb1 /dev/sdc1

  md_d0 : active raid0 sdb1[0] sdc1[1]
      14627614208 blocks 256k chunks

Model: Unknown (unknown)
Disk /dev/md_d0: 15.0TB
Sector size (logical/physical): 512B/512B
Partition Table: gpt

Number  Start   End     Size    File system  Name     Flags
 1      17.4kB  15.0TB  15.0TB               primary


-- LVM2 on top of /dev/md_d0p1

 --- Physical volume ---
  PV Name               /dev/md_d0p1
  VG Name               vgshare
  PV Size               13.62 TB / not usable 3.47 MB
  PE Size (KByte)       4096

  --- Volume group ---
  VG Name               vgshare
  System ID
  Format                lvm2
  VG Size               13.62 TB
  PE Size               4.00 MB

  --- Logical volume ---
  LV Name                /dev/vgshare/share
  VG Name                vgshare
  LV UUID                Aoj27F-kf4U-i6XE-eNWg-hMLX-MS1h-s3oArp
  LV Write Access        read/write
  LV Status              available
  # open                 0
  LV Size                7.00 TB
  Current LE             1835008
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     2048
  Block device           253:4

-- /dev/vgshare/share is ext4 formated with
mkfs.ext4 -b 4096 -E stride=64,stripe-width=128 /dev/md_d0p1

All well so far. And I start write data...

Then I create a new 2TB test lv
and format it.

/share_7TB is lost, /test_2TB is there....

Mar 12 21:13:28 server kernel: EXT4-fs error (device dm-4): ext4_mb_generate_buddy: EXT4-fs: group 0: 32768 blocks in bitmap, 3248 in gd Mar 12 21:13:30 server kernel: EXT4-fs error (device dm-4): ext4_mb_generate_buddy: EXT4-fs: group 1648: 24544 blocks in bitmap, 153 in gd
Mar 12 21:13:31 server kernel: attempt to access beyond end of device
Mar 12 21:13:31 server kernel: dm-4: rw=2, want=15493450520, limit=15032385536
Mar 12 21:13:31 server kernel: attempt to access beyond end of device
--snip
Mar 12 21:17:49 server kernel: EXT4-fs error (device dm-4): ext4_mb_release_inode_pa: free 1802, pa_free 1458 Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 93430033 blocks 705745 reqs (482734 success) Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 327298 extents scanned, 241152 goal hits, 219206 2^N hits, 0 breaks, 0 lost Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 9561 generated and it took 2012010656 Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 85047925 preallocated, 30759591 discarded Mar 12 21:18:09 server kernel: EXT4-fs: ext4_check_descriptors: Inode table for group 0 not in group (block 1936681314)!
Mar 12 21:18:09 server kernel: EXT4-fs: group descriptors corrupted!

GFS instead of ext4 also corrupted.

ext4 on top of /dev/md0 (non partitioned) on top of /dev/sdb /dev/sdc (without GPT) also corrupted.

I want to use software raid0 on top of the two hardware radi5
for better performance.

I understood that labeling GPT would solve this problem.
Is it x86 problem? Something fishy with my setup probably
but can't figure it out.

thanks and sorry for long message,
but I can't find a way to get this mirror server up
on it's feet again after this upgrade....

regards,

Giannis


This is probably an issue with the early version of ext4 you are using - note that the support for ext4 > 16TB is still gated by some work done up in the tools chain.

Have you tried xfs?

regards,

Ric

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux