RE: Problem in creating RAID5 MD array with kernel 2.6.15

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Neil,

When I try the following steps on kernel 2.6.16 I am getting the same errors
as on kernel 2.6.15.

# mdadm -Cv /dev/md0 --assume-clean --force --bitmap=/tmp/bitmap.txt
--bitmap-chunk=1024 -l5 -n4 /dev/sd{a,b,c,d}

Array gets created successfully.

# mdadm /dev/md0 -f /dev/sda
mdadm: set /dev/sda faulty in /dev/md0

# mdadm /dev/md0 -r /dev/sda
mdadm: hot remove failed for /dev/sda: Device or resource busy

# cat /proc/mdstat shows
Personalities : [raid5] [raid4]
md0 : active raid5 sdd[3] sdc[2] sdb[1] sda[0](F)
      1465159488 blocks level 5, 64k chunk, algorithm 2 [4/3] [_UUU]
      bitmap: 0/233 pages [0KB], 1024KB chunk, file: /tmp/bitmap.txt

unused devices: <none>

# mdadm -D /dev/md0 shows

/dev/md0:
        Version : 00.90.03
  Creation Time : Tue Apr 11 03:39:28 2006
     Raid Level : raid5
     Array Size : 1465159488 (1397.29 GiB 1500.32 GB)
    Device Size : 488386496 (465.76 GiB 500.11 GB)
   Raid Devices : 4
  Total Devices : 4
Preferred Minor : 0
    Persistence : Superblock is persistent

  Intent Bitmap : /tmp/bitmap.txt

    Update Time : Tue Apr 11 03:39:28 2006
          State : clean, degraded
 Active Devices : 3
Working Devices : 3
 Failed Devices : 1
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 64K

           UUID : ee25e1fa:d1ee78d5:d63869c6:dd7cff82
         Events : 0.1

    Number   Major   Minor   RaidDevice State
       0       8        0        0      faulty spare rebuilding   /dev/sda
       1       8       16        1      active sync   /dev/sdb
       2       8       32        2      active sync   /dev/sdc
       3       8       48        3      active sync   /dev/sdd


# tail -f /var/log/messages shows

Apr 11 03:40:47 localhost kernel:  <4>md: cannot remove active disk sda from
md0 ...
Apr 11 03:41:29 localhost kernel: raid5: Disk failure on sda, disabling
device. Operation continuing on 3 devices
Apr 11 03:41:35 localhost kernel: md: cannot remove active disk sda from md0
...


-----Original Message-----
From: Neil Brown [mailto:neilb@xxxxxxx] 
Sent: Tuesday, April 11, 2006 1:52 PM
To: Yogesh Pahilwan
Cc: linux-raid@xxxxxxxxxxxxxxx
Subject: RE: Problem in creating RAID5 MD array with kernel 2.6.15

On Tuesday April 11, pahilwan.yogesh@xxxxxxxxxxxxxxx wrote:
> Hi Neil,
> 
> Actually I want to calculate the performance of a RAID5 MD array in
rebuild
> state.
> 
> For doing this I do the following steps:
> 
> # mdadm -f /dev/md0 /dev/sda
> mdadm: set /dev/sda faulty in /dev/md0
> 
> # mdadm -r /dev/md0 /dev/sda
> mdadm: hot remove failed for /dev/sda: Device or resource busy

Hmmm.... That shouldn't happen.  I think you have found a bug :-(

However I cannot trivially reproduce it.

 - Can you reproduce this behaviour (mdadm -r failing) ?
    -  If so, can you list the steps?
    -  Can you reproduce on 2.6.16?
    -  Can you reproduce it without use a bitmap?

If you cannot reproduce it, please tell me as much as possible about
what led up to this situation.  Do you add/fail other drives? Did you
create or mount a filesystem, etc.

Thanks,
NeilBrown


> 
> # tail -f /var/log/messages shows 
> 
> Apr 11 01:48:11 localhost kernel:  <1>raid5: Disk failure on sda,
disabling
> device. Operation continuing on 3 devices
> Apr 11 01:48:24 localhost ntpd[3540]: synchronized to LOCAL(0), stratum 10
> Apr 11 01:48:24 localhost ntpd[3540]: kernel time sync disabled 0041
> Apr 11 01:48:26 localhost kernel: md: cannot remove active disk sda from
md0
> ...
> Apr 11 01:49:26 localhost ntpd[3540]: synchronized to 10.8.0.8, stratum 3
> Apr 11 01:50:51 localhost kernel: md: cannot remove active disk sda from
md0
> ...
> Apr 11 01:51:58 localhost kernel: md: cannot remove active disk sda from
md0
> ...
> Apr 11 01:54:16 localhost kernel: md: cannot remove active disk sda from
md0
> ...
> Apr 11 01:57:11 localhost kernel: md: cannot remove active disk sda from
md0
> .
> 
> I am not getting why I am not able to hot remove /dev/sda from /dev/md0?
> 
> 
> # mdadm -D /dev/md0
> /dev/md0:
>         Version : 00.90.03
>   Creation Time : Tue Apr 11 01:47:20 2006
>      Raid Level : raid5
>      Array Size : 1465159488 (1397.29 GiB 1500.32 GB)
>     Device Size : 488386496 (465.76 GiB 500.11 GB)
>    Raid Devices : 4
>   Total Devices : 4
> Preferred Minor : 0
>     Persistence : Superblock is persistent
> 
>   Intent Bitmap : /tmp/bitmap.txt
> 
>     Update Time : Tue Apr 11 01:47:20 2006
>           State : clean, degraded
>  Active Devices : 3
> Working Devices : 3
>  Failed Devices : 1
>   Spare Devices : 0
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>            UUID : 5ce49b71:e6083c2a:121b9ac2:cb675771
>          Events : 0.1
> 
>     Number   Major   Minor   RaidDevice State
>        0       8        0        0      faulty spare rebuilding   /dev/sda
>        1       8       16        1      active sync   /dev/sdb
>        2       8       32        2      active sync   /dev/sdc
>        3       8       48        3      active sync   /dev/sdd
> 
> This output shows that RAID5 /dev/md0 is in the degraded mode?
> 
> How should I rebuild this RAID5 so that I can calculate I/O performance
> while rebuilding RAID5 MD Array?
> 
> Thanks,
> Yogesh
> 
> 
> 
> 
> 
> 
> -----Original Message-----
> From: Neil Brown [mailto:neilb@xxxxxxx] 
> Sent: Tuesday, April 11, 2006 1:01 PM
> To: Yogesh Pahilwan
> Cc: linux-raid@xxxxxxxxxxxxxxx
> Subject: RE: Problem in creating RAID5 MD array with kernel 2.6.15
> 
> On Tuesday April 11, pahilwan.yogesh@xxxxxxxxxxxxxxx wrote:
> > Hi Neil,
> > 
> > I have set --bitmap-chunk=1024 and RAID5 gets created successfully.
> 
> Good.
> 
> > 
> > But why I will have to set --bitmap-chunk for big size devices such as
> 500GB
> > each in my case?
> > 
> > What is the default value of --bitmap-chunk?
> 
> 4, which is probably too low.
> 
> For every 2048 chunks, md potentially needs to allocate one page.
> md also needs to allocate a table to hold all these pages.
> 
> At a chunk size of 4K, your 500GB would use 125million chunks.
> That's 64000 pages - but these are only allocated on demand, and we can
> survive failure.
> However the table would need 4 bytes per page, or 250K
> Allocating a 250K stable is unlikely to succeed due to memory
> fragmentation.
> 
> With 1024K chunks you only need 1K, which is easy.
> 
> You could safely go down to 256K chunks but I'm not sure it would gain
> much.
> 
> I have put a note on my mdadm todo list to choose a more sensible
> default chunk size which limits the number of chunks to 2million.
> 
> NeilBrown

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux