problem with missing bricks

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Can you post the output of df -h? It looks like you have mounted the
drives as /gluster/brick[0-9]? If that is correct then you can
mitigate that by create a sub directory on that mount points accross
all the bricks, and do writes only inside that sub-dir.

On Sat, Dec 31, 2011 at 8:50 AM, Todd Pfaff <pfaff at rhpcs.mcmaster.ca> wrote:
> Gluster-user folks,
>
> I'm trying to use gluster in a way that may be a considered an unusual use
> case for gluster. ?Feel free to let me know if you think what I'm doing
> is dumb. ?It just feels very comfortable doing this with gluster.
> I have been using gluster in other, more orthodox configurations, for
> several years.
>
> I have a single system with 45 inexpensive sata drives - it's a self-built
> backblaze similar to that documented at this url but with some upgrades
> and substitutions:
>
> ?http://blog.backblaze.com/2009/09/01/petabytes-on-a-budget-how-to-build-cheap-cloud-storage/
>
> We use this system for disk-to-disk backups only, no primary storage,
> nothing mission critical.
>
> For the past two years I have been using this system with linux software
> raid, with the drives organized as multiple raid 5/6/10 sets of 5 drives
> per set. ?This has worked ok but I have suffered enough multiple
> simultaneous drive failures to prompt me to explore alternatives to raid.
> Yes, I know, that's what I get for using cheap sata drives.
>
> What I'm experimenting with now is creating gluster distributed-replicated
> volumes on some of these drives, and maybe all in the future if this works
> reasonably well.
>
> At this point I am using 10 of the drives configured as shown here:
>
> ?Volume Name: volume1
> ?Type: Distributed-Replicate
> ?Status: Started
> ?Number of Bricks: 5 x 2 = 10
> ?Transport-type: tcp
> ?Bricks:
> ?Brick1: host:/gluster/brick01
> ?Brick2: host:/gluster/brick06
> ?Brick3: host:/gluster/brick02
> ?Brick4: host:/gluster/brick07
> ?Brick5: host:/gluster/brick03
> ?Brick6: host:/gluster/brick08
> ?Brick7: host:/gluster/brick04
> ?Brick8: host:/gluster/brick09
> ?Brick9: host:/gluster/brick05
> ?Brick10: host:/gluster/brick10
> ?Options Reconfigured:
> ?auth.allow: 127.0.0.1,10.10.10.10
>
>
> For the most part this is working fine so far. ?The problem I have run
> into several times now is that when a drive fails and the system is
> rebooted, the volume comes up without that brick. ?Gluster happily writes
> to the missing brick's mount point, thereby eventually filling up the root
> filesystem. ?Once the root filesystem is full and processes writing to
> gluster space are hung, I can never recover from this state without
> rebooting.
>
> Is there any way to avoid this problem of gluster writing to a brick
> path that isn't really populated by the intended brick filesystem?
> Does gluster not create any sort of signature or meta-data that
> indicates whether or not a path is really a gluster brick?
> How do others deal with missing bricks?
>
> I realize that ultimately I should get the bricks replaced as soon as
> possible but there may be times when I want to continue running for some
> time with a "degraded" volume if you will.
>
> Any and all ideas, suggestions, comments, criticisms are welcome.
>
> Cheers and Happy New Year,
> Todd
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux