On Wed, 2014-10-08 at 07:59 +1000, Dan Mons wrote: > I've not used ZFS in production. Outside of GlusterFS, how does it > handle running at 90%+ allocated? A number of our bricks are at or close to 95% and we've not seem any massive slow downs, we have gluster set to leave 5% free. I'm not saying it isn't slower than when the disks were empty but production doesn't appear to be suffering with disk levels above 90%. Latency for lookups is sometimes a problem, files are not always where they are supposed to be. ZOL isn't the fastest filesystem out there, so maybe coming off a lower base level hides the impact of fragmentation somewhat. > > Again, speaking outside of GlusterFS, ext3, ext4 and XFS all > classically slow down when they reach high usage rates. Not > noticeable at all on my home media server in a family of 5, but hugely > noticeable at work when hundreds of artists and render nodes on a mix > of 1GbE and 10GbE are smashing a NAS. We used ext3/ext4 with hardware RAID and NFS before moving to ZOL + Gluster, so far (touching wood) it's an improvement on all counts. What are you using for RAID? Judging from the video on your website, we are using similar hardware. > > GlusterFS alleviates the problem somewhat (in combination due to it's > distribution, and the fact that not all bricks hit ~90% at the same > time). But it's still felt. > Not sure where the 90% figure comes from, surely it depends on the disk capacity??? > -Dan > > ---------------- > Dan Mons > Unbreaker of broken things > Cutting Edge > http://cuttingedge.com.au > > > On 7 October 2014 14:56, Franco Broi <franco.broi@xxxxxxxxxx> wrote: > > Our bricks are 50TB, running ZOL, 16 disks raidz2. Works OK with Gluster > > now that they fixed xattrs. > > > > 8k writes with fsync 170MB/Sec, reads 335MB/Sec. > > > > On Tue, 2014-10-07 at 14:24 +1000, Dan Mons wrote: > >> We have 6 nodes with one brick per node (2x3 replicate-distribute). > >> 35TB per brick, for 107TB total usable. > >> > >> Not sure if our low brick count (or maybe large brick per node?) > >> contributes to the slowdown when full. > >> > >> We're looking to add more nodes by the end of the year. After that, > >> I'll look this thread up and comment on what that's changed, > >> performance wise. > >> > >> -Dan > >> > >> ---------------- > >> Dan Mons > >> Unbreaker of broken things > >> Cutting Edge > >> http://cuttingedge.com.au > >> > >> > >> On 7 October 2014 14:16, Franco Broi <franco.broi@xxxxxxxxxx> wrote: > >> > > >> > Not an issue for us, were at 92% on an 800TB distributed volume, 16 > >> > bricks spread across 4 servers. Lookups can be a bit slow but raw IO > >> > hasn't changed. > >> > > >> > On Tue, 2014-10-07 at 09:16 +1000, Dan Mons wrote: > >> >> On 7 October 2014 08:56, Jeff Darcy <jdarcy@xxxxxxxxxx> wrote: > >> >> > I can't think of a good reason for such a steep drop-off in GlusterFS. > >> >> > Sure, performance should degrade somewhat due to fragmenting, but not > >> >> > suddenly. It's not like Lustre, which would do massive preallocation > >> >> > and fall apart when there was no longer enough space to do that. It > >> >> > might be worth measuring average latency at the local-FS level, to see > >> >> > if the problem is above or below that line. > >> >> > >> >> Happens like clockwork for us. The moment we get alerts saying the > >> >> file system has hit 90%, we get a flood of support tickets about > >> >> performance. > >> >> > >> >> It happens to a lesser degree on standard CentOS NAS units running XFS > >> >> we have around the place. But again, I see the same sort of thing on > >> >> any file system (vendor supplied, self-built, OS and FS agnostic). > >> >> And yes, it's measurable (Munin graphs show it off nicely). > >> >> > >> >> -Dan > >> >> _______________________________________________ > >> >> Gluster-users mailing list > >> >> Gluster-users@xxxxxxxxxxx > >> >> http://supercolony.gluster.org/mailman/listinfo/gluster-users > >> > > >> > > > > > _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://supercolony.gluster.org/mailman/listinfo/gluster-users