Re: AFR comments. Maximizing free space use when using mirroring.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 7/25/07, DeeDee Park <deedee6905@xxxxxxxxxxx> wrote:
> Here is my 2c on AFR.
>
> When I setup file servers, the first priority is always to get it up and
> running, and then
> later the next priority is to add mirrors/high availability. Somtimes due to
> business concerns
> the second priority sometimes does not happen until something drastic
> happens. By
> the time that there is budget to get additional drives, typically the drive
> sizes that are
> available are also much bigger (remember, they double every 2 years or so).
> So when
> I've bought drives, I've gotten 40GB, 80GB, 120Gb, 160GB, 200GB, 250GB,
> 300GB,
> 500GB, 750GB... So what I'm saying is that when new drives are bought to
> either expand
> the total file server size, or to add additional replicas, the new drives
> are most of the time bigger than the original drives purchased.
>
> The In the current implementation of AFR, the second brick (in a non well
> manged environment)
> will most likely be bigger than the first brick, thus underutilizing
> additional storage space due to mismatch in disk sizes.
>
> The idea I have is that I want to use as many available commodity parts that
> I can find and
> build a largest file server for my customer's needs and reallocating the
> remaining space for
> replicas. I still have a lot of these 120GB drives sitting around from a few
> years ago, and I've
> got 500/750GB drives. It seems to be a difficult task to match each 120GB
> drive with another
> 120GB drive to optimize disk usage for AFR purposes. I could have 2 500GB
> drives for
> replication *:2, but if I want to move to *:3 in the future, most likely
> I'll have some 750GB
> drives laying around. Using a 750GB as my third brick would most likely
> waste the remaining 250GB.

You can just put the bigger drive brick as the first subvolume in the subvolume
list. This should fix the problem right?

>
> Just as RR or ALU puts files anywhere. I envisioned originally that AFR also
> did the same. If my dataset is larger than the largest possible RAID I can
> afford, then one brick will never carry all the
> files.

No, That would complicate the functionality of AFR and its self-heal feature.

>
> What I think would be cool would be to have the AFR on top of the unify so
> that if the dataset is spread across X drives, that is fine, the remote
> mirrors would not require the same hardware, and I would just need to
> purchase the approximate 2X hard drive space at the new co-lo. I can just
> ask a client "How much disk space are you currently using?". If they say
> 20TB all using 200GB drives (=100 drives), then I can setup the additional
> glusterfs replica to utilize 20TB using 750GB drives. I would like to have
> to buy 27 750GB drives to make up my 20TB, instead of having to buy 100
> 750GB drives to replicate the existing 100GB drives. (It doesn't make sense
> to buy 200GB drives when larger drives are available for purchase).

I just tried afr over unify, there is some problem, I shall look into it.
However we need to see if it is advisable to use this kind of setup.

>
> Also, I have the premise that 100% of the dataset is critical (It is all
> user data), and I cannot say which file extensions should be replicated or
> should not be replicated. The example that *.c is more critical than *.o
> probably true, but I know users have told me that they have .o files from
> systems that are no longer available, so those .o files for that user are
> critical. Since I cannot specify *.c:2,*.o:1 for some users and *.c:2,*.o:2
> for others (nor would I really want to get that involved in the user data
> details or think I'll have that much free time to investigate that level of
> detail), it only makes sense to replicate everything eg: *:2 or *:3. It is a
> cool feature to have. But also if a user specifies *.c:2,*.o:1, then that
> assumes (with the current implementation of AFR), that the 2nd brick should
> be smaller than the first brick (Then I have questions as to what happens
> when there isn't enough space etc).

That is right, second brick should be smaller or equal to fist disk.

Regards,
Krishna




[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux