Re: RAID5 / 6 Growth

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



"Leslie Rhorer" <lrhorer@xxxxxxxxxxx> writes:

>> > High enough?  Wouldn't a higher speed limit mean more stress on the
>> systems?
>> > Its value is 1000.
>> 
>> A higher min value will block more normal IO (if there is
>> any).
>
> 	No, not really.  All the system that is currently being grown does
> is incrementally rsync the data from the main server every morning.  The
> main server, which will be grown after this server is done, is another
> matter.
>
>> Raising min is usefull to ensure the job gets done in a certain
>> time, to not let normal IO slow down a rebuild too much.
>
> 	OK, but like I said, for the most part there isn't any other I/O.
>  
>> >> MB/s) along with /proc/sys/dev/raid/speed_limit_max?
>> >
>> > It's 200,000
>> 
>> I only ever had to tune this once when too much IO would deadlock an
>> external enclosure. Otherwise keep this really high so it uses all
>> idle IO there is.
>
> 	200MBps is far more than the system can handle.  These are consumer
> class drives on a relatively inexpensive 4 port controller feeding a Port
> Multiplier chassis.
>
>> As to your initial question: Being able to keep the filesystem mounted
>> and used is the whole point of having online growing of the raid
>> system. If that weren't save then there would be no point to it as you
>> could just as well stop the raid if you already umounted it and grow
>> it offline.
>
> 	I know that is the point of the utility.  My question boils down to,
> "How safe is it to avail one's self of the capability if it is not essential
> to have the array mounted for the duration?"  I don't particularly like
> having the array unavailable (especially not for nearly 5 days), but I
> prefer that to risking data loss, or especially risking irretrievably losing
> the entire array.  The question is particularly pertinent given the fact the
> growth is going to take nearly 5 days (a lot can happen in 5 days), and the
> fact the system was having the rather squirrelly issue a few days back which
> seems - emphasis on SEEMS - to have been resolved by disabling NCQ.  What
> happens if the system kicks a couple of drives, especially if one drive gets
> kicked, a bunch of data gets written and then a few minutes later another
> drive gets kicked?  In particular, what if neither of the two drives that
> get kicked are the new drive?

It should just keep going. I don't see how having the filesystem in
use will have any influence there other than adding a little more load
on the disks. Reshaping is heavy on the drive. Lots of reads, writes
and seeks. A daily rsync probably doesn't even register as extra load.

MfG
        Goswin
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux