[Gluster-Maintainers] [Gluster-users] Proposal to mark few features as Deprecated / SunSet from Version 5.0
Jim Kinney
jim.kinney at gmail.com
Tue Mar 19 19:52:39 UTC 2019
This python will fail when writing to a file in a glusterfs fuse
mounted directory.
import mmap # write a simple example filewith open("hello.txt", "wb")
as f: f.write("Hello Python!\n")
with open("hello.txt", "r+b") as f: # memory-map the file, size 0
means whole file mm = mmap.mmap(f.fileno(), 0) # read content via
standard file methods print mm.readline() # prints "Hello
Python!" # read content via slice notation print mm[:5] # prints
"Hello" # update content using slice notation; # note that new
content must have same size mm[6:] = " world!\n" # ... and read
again using standard file methods mm.seek(0) print
mm.readline() # prints "Hello world!" # close the
map mm.close()
On Tue, 2019-03-19 at 12:06 -0400, Jim Kinney wrote:
> Native mount issue with multiple clients (centos7 glusterfs 3.12).
> Seems to hit python 2.7 and 3+. User tries to open file(s) for write
> on long process and system eventually times out.
> Switching to NFS stops the error.
> No bug notice yet. Too many pans on the fire :-(
> On Tue, 2019-03-19 at 18:42 +0530, Amar Tumballi Suryanarayan wrote:
> > Hi Jim,
> >
> > On Tue, Mar 19, 2019 at 6:21 PM Jim Kinney <jim.kinney at gmail.com>
> > wrote:
> > >
> > >
> > >
> > > Issues with glusterfs fuse mounts cause issues with python file
> > > open for write. We have to use nfs to avoid this.
> > >
> > > Really want to see better back-end tools to facilitate cleaning
> > > up of glusterfs failures. If system is going to use hard linked
> > > ID, need a mapping of id to file to fix things. That option is
> > > now on for all exports. It should be the default If a host is
> > > down and users delete files by the thousands, gluster _never_
> > > catches up. Finding path names for ids across even a 40TB mount,
> > > much less the 200+TB one, is a slow process. A network outage of
> > > 2 minutes and one system didn't get the call to recursively
> > > delete several dozen directories each with several thousand
> > > files.
> > >
> > >
> >
> > Are you talking about some issues in geo-replication module or some
> > other application using native mount? Happy to take the discussion
> > forward about these issues.
> > Are there any bugs open on this?
> > Thanks,Amar
> > > nfsOn March 19, 2019 8:09:01 AM EDT, Hans Henrik Happe <
> > > happe at nbi.dk> wrote:
> > > > Hi,
> > > > Looking into something else I fell over this proposal.
> > > > Being a
> > > > shop that are going into "Leaving GlusterFS" mode, I
> > > > thought I
> > > > would give my two cents.
> > > >
> > > >
> > > > While being partially an HPC shop with a few Lustre
> > > > filesystems,
> > > > we chose GlusterFS for an archiving solution (2-3 PB),
> > > > because we
> > > > could find files in the underlying ZFS filesystems if
> > > > GlusterFS
> > > > went sour.
> > > > We have used the access to the underlying files plenty,
> > > > because
> > > > of the continuous instability of GlusterFS'. Meanwhile,
> > > > Lustre
> > > > have been almost effortless to run and mainly for that
> > > > reason we
> > > > are planning to move away from GlusterFS.
> > > > Reading this proposal kind of underlined that "Leaving
> > > > GluserFS"
> > > > is the right thing to do. While I never understood why
> > > > GlusterFS
> > > > has been in feature crazy mode instead of stabilizing
> > > > mode, taking
> > > > away crucial features I don't get. With RoCE, RDMA is
> > > > getting
> > > > mainstream. Quotas are very useful, even though the
> > > > current
> > > > implementation are not perfect. Tiering also makes so
> > > > much sense,
> > > > but, for large files, not on a per-file level.
> > > > To be honest we only use quotas. We got scared of trying
> > > > out new
> > > > performance features that potentially would open up a new
> > > > back of
> > > > issues.
> > > > Sorry for being such a buzzkill. I really wanted it to be
> > > > different.
> > > >
> > > >
> > > > Cheers,
> > > >
> > > > Hans Henrik
> > > >
> > > >
> > > > On 19/07/2018 08.56, Amar Tumballi
> > > > wrote:
> > > >
> > > >
> > > >
> > > > >
> > > > >
> > > > > Hi all,
> > > > > Over last 12 years of Gluster, we have developed
> > > > > many features, and continue to support most of it till now.
> > > > > But along the way, we have figured out better methods of
> > > > > doing things. Also we are not actively maintaining some of
> > > > > these features.
> > > > > We are now thinking of cleaning up some of these
> > > > > ‘unsupported’ features, and mark them as ‘SunSet’ (i.e.,
> > > > > would be totally taken out of codebase in following releases)
> > > > > in next upcoming release, v5.0. The release notes will
> > > > > provide options for smoothly migrating to the supported
> > > > > configurations.
> > > > > If you are using any of these features, do let us
> > > > > know, so that we can help you with ‘migration’.. Also, we are
> > > > > happy to guide new developers to work on those components
> > > > > which are not actively being maintained by current set of
> > > > > developers.
> > > > > List of features hitting sunset:
> > > > > ‘cluster/stripe’ translator:
> > > > > This translator was developed very early in the
> > > > > evolution of GlusterFS, and addressed one of the very common
> > > > > question of Distributed FS, which is “What happens if one of
> > > > > my file is bigger than the available brick. Say, I have 2 TB
> > > > > hard drive, exported in glusterfs, my file is 3 TB”. While it
> > > > > solved the purpose, it was very hard to handle failure
> > > > > scenarios, and give a real good experience to our users with
> > > > > this feature. Over the time, Gluster solved the problem with
> > > > > it’s ‘Shard’ feature, which solves the problem in much better
> > > > > way, and provides much better solution with existing well
> > > > > supported stack. Hence the proposal for Deprecation.
> > > > > If you are using this feature, then do write to us,
> > > > > as it needs a proper migration from existing volume to a new
> > > > > full supported volume type before you upgrade.
> > > > > ‘storage/bd’ translator:
> > > > > This feature got into the code base 5 years back
> > > > > with this patch[1]. Plan was to use a block device directly
> > > > > as a brick, which would help to handle disk-image storage
> > > > > much easily in glusterfs.
> > > > > As the feature is not getting more contribution,
> > > > > and we are not seeing any user traction on this, would like
> > > > > to propose for Deprecation.
> > > > > If you are using the feature, plan to move to a
> > > > > supported gluster volume configuration, and have your setup
> > > > > ‘supported’ before upgrading to your new gluster version.
> > > > > ‘RDMA’ transport support:
> > > > > Gluster started supporting RDMA while ib-verbs was
> > > > > still new, and very high-end infra around that time were
> > > > > using Infiniband. Engineers did work with Mellanox, and got
> > > > > the technology into GlusterFS for better data migration, data
> > > > > copy. While current day kernels support very good speed with
> > > > > IPoIB module itself, and there are no more bandwidth for
> > > > > experts in these area to maintain the feature, we recommend
> > > > > migrating over to TCP (IP based) network for your volume.
> > > > > If you are successfully using RDMA transport, do
> > > > > get in touch with us to prioritize the migration plan for
> > > > > your volume. Plan is to work on this after the release, so by
> > > > > version 6.0, we will have a cleaner transport code, which
> > > > > just needs to support one type.
> > > > > ‘Tiering’ feature
> > > > > Gluster’s tiering feature which was planned to be
> > > > > providing an option to keep your ‘hot’ data in different
> > > > > location than your cold data, so one can get better
> > > > > performance. While we saw some users for the feature, it
> > > > > needs much more attention to be completely bug free. At the
> > > > > time, we are not having any active maintainers for the
> > > > > feature, and hence suggesting to take it out of the
> > > > > ‘supported’ tag.
> > > > > If you are willing to take it up, and maintain it,
> > > > > do let us know, and we are happy to assist you.
> > > > > If you are already using tiering feature, before
> > > > > upgrading, make sure to do gluster volume tier detach all the
> > > > > bricks before upgrading to next release. Also, we recommend
> > > > > you to use features like dmcache on your LVM setup to get
> > > > > best performance from bricks.
> > > > > ‘Quota’
> > > > > This is a call out for ‘Quota’ feature, to let you
> > > > > all know that it will be ‘no new development’ state. While
> > > > > this feature is ‘actively’ in use by many people, the
> > > > > challenges we have in accounting mechanisms involved, has
> > > > > made it hard to achieve good performance with the feature.
> > > > > Also, the amount of extended attribute get/set operations
> > > > > while using the feature is not very ideal. Hence we recommend
> > > > > our users to move towards setting quota on backend bricks
> > > > > directly (ie, XFS project quota), or to use different volumes
> > > > > for different directories etc.
> > > > > As the feature wouldn’t be deprecated immediately,
> > > > > the feature doesn’t need a migration plan when you upgrade to
> > > > > newer version, but if you are a new user, we wouldn’t
> > > > > recommend setting quota feature. By the release dates, we
> > > > > will be publishing our best alternatives guide for gluster’s
> > > > > current quota feature.
> > > > > Note that if you want to contribute to the feature,
> > > > > we have project quota based issue open[2] Happy to get
> > > > > contributions, and help in getting a newer approach to Quota.
> > > > >
> > > > >
> > > > >
> > > > > These are our set of initial features which we
> > > > > propose to take out of ‘fully’ supported features. While we
> > > > > are in the process of making the user/developer experience of
> > > > > the project much better with providing well maintained
> > > > > codebase, we may come up with few more set of features which
> > > > > we may possibly consider to move out of support, and hence
> > > > > keep watching this space.
> > > > > [1] - http://review.gluster.org/4809
> > > > > [2] -
> > > > > https://github.com/gluster/glusterfs/issues/184
> > > > >
> > > > > Regards,
> > > > >
> > > > > Vijay, Shyam, Amar
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > _______________________________________________Gluster-
> > > > > users mailing listGluster-users at gluster.org
> > > > > https://lists.glus
> > > > > --
> > > > > Sent from my Android device with K-9 Mail. All tyopes are
> > > > > thumb related and reflect
> > > > > authenticity.ter.org/mailman/listinfo/gluster-users
> > > > >
> > > >
> > > >
> >
> >
> --
> James P. Kinney III
> Every time you stop a school, you will have to build a jail. What
> yougain at one end you lose at the other. It's like feeding a dog on
> hisown tail. It won't fatten the dog.- Speech 11/23/1900 Mark Twain
> http://heretothereideas.blogspot.com/
--
James P. Kinney III
Every time you stop a school, you will have to build a jail. What you
gain at one end you lose at the other. It's like feeding a dog on his
own tail. It won't fatten the dog.
- Speech 11/23/1900 Mark Twain
http://heretothereideas.blogspot.com/
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/maintainers/attachments/20190319/44831235/attachment-0001.html>
More information about the maintainers
mailing list