[Gluster-users] Extremely slow file listing in folders with many files
Felix Kölzow
felix.koelzow at gmx.de
Thu Apr 30 16:01:44 UTC 2020
Dear Artem,
can you also provide some information w.r.t your xfs filesystem, i.e.
xfs_info of your block device?
Regards,
Felix
On 30/04/2020 17:27, Artem Russakovskii wrote:
> Hi Strahil, in the original email I included both the times for the
> first and subsequent reads on the fuse mounted gluster volume as well
> as the xfs filesystem the gluster data resides on (this is the brick,
> right?).
>
> On Thu, Apr 30, 2020, 7:44 AM Strahil Nikolov <hunter86_bg at yahoo.com
> <mailto:hunter86_bg at yahoo.com>> wrote:
>
> On April 30, 2020 4:24:23 AM GMT+03:00, Artem Russakovskii
> <archon810 at gmail.com <mailto:archon810 at gmail.com>> wrote:
> >Hi all,
> >
> >We have 500GB and 10TB 4x1 replicate xfs-based gluster volumes,
> and the
> >10TB one especially is extremely slow to do certain things with (and
> >has
> >been since gluster 3.x when we started). We're currently on 5.13.
> >
> >The number of files isn't even what I'd consider that great - under
> >100k
> >per dir.
> >
> >Here are some numbers to look at:
> >
> >On gluster volume in a dir of 45k files:
> >The first time
> >
> >time find | wc -l
> >45423
> >real 8m44.819s
> >user 0m0.459s
> >sys 0m0.998s
> >
> >And again
> >
> >time find | wc -l
> >45423
> >real 0m34.677s
> >user 0m0.291s
> >sys 0m0.754s
> >
> >
> >If I run the same operation on the xfs block device itself:
> >The first time
> >
> >time find | wc -l
> >45423
> >real 0m13.514s
> >user 0m0.144s
> >sys 0m0.501s
> >
> >And again
> >
> >time find | wc -l
> >45423
> >real 0m0.197s
> >user 0m0.088s
> >sys 0m0.106s
> >
> >
> >I'd expect a performance difference here but just as it was several
> >years
> >ago when we started with gluster, it's still huge, and simple file
> >listings
> >are incredibly slow.
> >
> >At the time, the team was looking to do some optimizations, but
> I'm not
> >sure this has happened.
> >
> >What can we do to try to improve performance?
> >
> >Thank you.
> >
> >
> >
> >Some setup values follow.
> >
> >xfs_info /mnt/SNIP_block1
> >meta-data=/dev/sdc isize=512 agcount=103,
> >agsize=26214400
> >blks
> > = sectsz=512 attr=2, projid32bit=1
> > = crc=1 finobt=1, sparse=0,
> rmapbt=0
> > = reflink=0
> >data = bsize=4096 blocks=2684354560,
> >imaxpct=25
> > = sunit=0 swidth=0 blks
> >naming =version 2 bsize=4096 ascii-ci=0, ftype=1
> >log =internal log bsize=4096 blocks=51200, version=2
> > = sectsz=512 sunit=0 blks,
> lazy-count=1
> >realtime =none extsz=4096 blocks=0, rtextents=0
> >
> >Volume Name: SNIP_data1
> >Type: Replicate
> >Volume ID: SNIP
> >Status: Started
> >Snapshot Count: 0
> >Number of Bricks: 1 x 4 = 4
> >Transport-type: tcp
> >Bricks:
> >Brick1: nexus2:/mnt/SNIP_block1/SNIP_data1
> >Brick2: forge:/mnt/SNIP_block1/SNIP_data1
> >Brick3: hive:/mnt/SNIP_block1/SNIP_data1
> >Brick4: citadel:/mnt/SNIP_block1/SNIP_data1
> >Options Reconfigured:
> >cluster.quorum-count: 1
> >cluster.quorum-type: fixed
> >network.ping-timeout: 5
> >network.remote-dio: enable
> >performance.rda-cache-limit: 256MB
> >performance.readdir-ahead: on
> >performance.parallel-readdir: on
> >network.inode-lru-limit: 500000
> >performance.md-cache-timeout: 600
> >performance.cache-invalidation: on
> >performance.stat-prefetch: on
> >features.cache-invalidation-timeout: 600
> >features.cache-invalidation: on
> >cluster.readdir-optimize: on
> >performance.io-thread-count: 32
> >server.event-threads: 4
> >client.event-threads: 4
> >performance.read-ahead: off
> >cluster.lookup-optimize: on
> >performance.cache-size: 1GB
> >cluster.self-heal-daemon: enable
> >transport.address-family: inet
> >nfs.disable: on
> >performance.client-io-threads: on
> >cluster.granular-entry-heal: enable
> >cluster.data-self-heal-algorithm: full
> >
> >Sincerely,
> >Artem
> >
> >--
> >Founder, Android Police <http://www.androidpolice.com>, APK Mirror
> ><http://www.apkmirror.com/>, Illogical Robot LLC
> >beerpla.net <http://beerpla.net> | @ArtemR
> <http://twitter.com/ArtemR>
>
> Hi Artem,
>
> Have you checked the same on brick level ? How big is the difference ?
>
> Best Regards,
> Strahil Nikolov
>
>
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://bluejeans.com/441850968
>
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200430/ee340afd/attachment.html>
More information about the Gluster-users
mailing list