[Gluster-users] Perfomance issue on a 90+% full file system
Franco Broi
franco.broi at iongeo.com
Tue Oct 7 23:40:12 UTC 2014
On Wed, 2014-10-08 at 07:59 +1000, Dan Mons wrote:
> I've not used ZFS in production. Outside of GlusterFS, how does it
> handle running at 90%+ allocated?
A number of our bricks are at or close to 95% and we've not seem any
massive slow downs, we have gluster set to leave 5% free. I'm not saying
it isn't slower than when the disks were empty but production doesn't
appear to be suffering with disk levels above 90%. Latency for lookups
is sometimes a problem, files are not always where they are supposed to
be.
ZOL isn't the fastest filesystem out there, so maybe coming off a lower
base level hides the impact of fragmentation somewhat.
>
> Again, speaking outside of GlusterFS, ext3, ext4 and XFS all
> classically slow down when they reach high usage rates. Not
> noticeable at all on my home media server in a family of 5, but hugely
> noticeable at work when hundreds of artists and render nodes on a mix
> of 1GbE and 10GbE are smashing a NAS.
We used ext3/ext4 with hardware RAID and NFS before moving to ZOL +
Gluster, so far (touching wood) it's an improvement on all counts.
What are you using for RAID? Judging from the video on your website, we
are using similar hardware.
>
> GlusterFS alleviates the problem somewhat (in combination due to it's
> distribution, and the fact that not all bricks hit ~90% at the same
> time). But it's still felt.
>
Not sure where the 90% figure comes from, surely it depends on the disk
capacity???
> -Dan
>
> ----------------
> Dan Mons
> Unbreaker of broken things
> Cutting Edge
> http://cuttingedge.com.au
>
>
> On 7 October 2014 14:56, Franco Broi <franco.broi at iongeo.com> wrote:
> > Our bricks are 50TB, running ZOL, 16 disks raidz2. Works OK with Gluster
> > now that they fixed xattrs.
> >
> > 8k writes with fsync 170MB/Sec, reads 335MB/Sec.
> >
> > On Tue, 2014-10-07 at 14:24 +1000, Dan Mons wrote:
> >> We have 6 nodes with one brick per node (2x3 replicate-distribute).
> >> 35TB per brick, for 107TB total usable.
> >>
> >> Not sure if our low brick count (or maybe large brick per node?)
> >> contributes to the slowdown when full.
> >>
> >> We're looking to add more nodes by the end of the year. After that,
> >> I'll look this thread up and comment on what that's changed,
> >> performance wise.
> >>
> >> -Dan
> >>
> >> ----------------
> >> Dan Mons
> >> Unbreaker of broken things
> >> Cutting Edge
> >> http://cuttingedge.com.au
> >>
> >>
> >> On 7 October 2014 14:16, Franco Broi <franco.broi at iongeo.com> wrote:
> >> >
> >> > Not an issue for us, were at 92% on an 800TB distributed volume, 16
> >> > bricks spread across 4 servers. Lookups can be a bit slow but raw IO
> >> > hasn't changed.
> >> >
> >> > On Tue, 2014-10-07 at 09:16 +1000, Dan Mons wrote:
> >> >> On 7 October 2014 08:56, Jeff Darcy <jdarcy at redhat.com> wrote:
> >> >> > I can't think of a good reason for such a steep drop-off in GlusterFS.
> >> >> > Sure, performance should degrade somewhat due to fragmenting, but not
> >> >> > suddenly. It's not like Lustre, which would do massive preallocation
> >> >> > and fall apart when there was no longer enough space to do that. It
> >> >> > might be worth measuring average latency at the local-FS level, to see
> >> >> > if the problem is above or below that line.
> >> >>
> >> >> Happens like clockwork for us. The moment we get alerts saying the
> >> >> file system has hit 90%, we get a flood of support tickets about
> >> >> performance.
> >> >>
> >> >> It happens to a lesser degree on standard CentOS NAS units running XFS
> >> >> we have around the place. But again, I see the same sort of thing on
> >> >> any file system (vendor supplied, self-built, OS and FS agnostic).
> >> >> And yes, it's measurable (Munin graphs show it off nicely).
> >> >>
> >> >> -Dan
> >> >> _______________________________________________
> >> >> Gluster-users mailing list
> >> >> Gluster-users at gluster.org
> >> >> http://supercolony.gluster.org/mailman/listinfo/gluster-users
> >> >
> >> >
> >
> >
More information about the Gluster-users
mailing list