[Gluster-users] State of Gluster project
Gionatan Danti
g.danti at assyoma.it
Sun Jun 21 17:43:24 UTC 2020
Il 2020-06-21 14:20 Strahil Nikolov ha scritto:
> With every community project , you are in the position of a Betta
> Tester - no matter Fedora, Gluster or CEPH. So far , I had
> issues with upstream projects only diring and immediately after
> patching - but this is properly mitigated with a reasonable
> patching strategy (patch test environment and several months later
> patch prod with the same repos).
> Enterprise Linux breaks (and alot) having 10-times more users and
> use cases, so you cannot expect to start to use Gluster and assume
> that a free peoject won't break at all.
> Our part in this project is to help the devs to create a test case for
> our workload , so regressions will be reduced to minimum.
Well, this is true, and both devs & community deserve a big thanks for
all the work done.
> In the past 2 years, we got 2 major issues with VMware VSAN and 1
> major issue with a Enterprise Storage cluster (both solutions are
> quite expensive) - so I always recommend proper testing of your
> software .
Interesting, I am almost tempted to ask you what issue you had with
vSAN, but this is not the right mailing list ;)
> From my observations, almost nobody is complaining about Ganesha in
> the mailing list -> 50% are having issues with geo replication,20%
> are having issues with small file performance and the rest have
> issues with very old version of gluster -> v5 or older.
Mmm, I would swear to have read quite a few posts where the problem was
solved by migrating away from NFS Ganesha. Still, for hyperconverged
setup a problem remains: NFS on loopback/localhost is not 100% supported
(or, at least, RH is not willing to declare it supportable/production
ready [1]). A fuse mount would be the more natural way to access the
underlying data.
> I can't say that a replace-brick on a 'replica 3' volume is more
> riskier than a rebuild of a raid, but I have noticed that nobody is
> following Red Hat's guide to use either:
> - a Raid6 of 12 Disks (2-3 TB big)
> - a Raid10 of 12 Disks (2-3 TB big)
> - JBOD disks in 'replica 3' mode (i'm not sure about the size RH
> recommends, most probably 2-3 TB)
> So far, I didn' have the opportunity to run on JBODs.
For the RAID6/10 setup, I found no issues: simply replace the broken
disk without involing Gluster at all. However, this also means facing
the "iops wall" I described earlier for single-brick node. Going
full-Guster with JBODs would be interesting from a performance
standpoint, but this complicate eventual recovery from bad disks.
Does someone use Gluster in JBOD mode? If so, can you share your
experience?
Thanks.
[1] https://access.redhat.com/solutions/22231 (accound required)
[2] https://bugzilla.redhat.com/show_bug.cgi?id=489889 (old, but I can
not find anything newer)
--
Danti Gionatan
Supporto Tecnico
Assyoma S.r.l. - www.assyoma.it [1]
email: g.danti at assyoma.it - info at assyoma.it
GPG public key ID: FF5F32A8
More information about the Gluster-users
mailing list