[Gluster-users] df does not show full volume capacity after update to 3.12.4

Nithya Balachandran nbalacha at redhat.com
Wed Jan 31 17:09:19 UTC 2018


Hi Eva,

I'm sorry but I need to get in touch with another developer to check about
the changes here and he will be available only tomorrow. Is there someone
else I could work with while you are away?

Regards,
Nithya

On 31 January 2018 at 22:00, Freer, Eva B. <freereb at ornl.gov> wrote:

> Nithya,
>
>
>
> I will be out of the office for ~10 days starting tomorrow. Is there any
> way we could possibly resolve it today?
>
>
>
> Thanks,
>
> Eva     (865) 574-6894
>
>
>
> *From: *Nithya Balachandran <nbalacha at redhat.com>
> *Date: *Wednesday, January 31, 2018 at 11:26 AM
>
> *To: *Eva Freer <freereb at ornl.gov>
> *Cc: *"Greene, Tami McFarlin" <greenet at ornl.gov>, "
> gluster-users at gluster.org" <gluster-users at gluster.org>, Amar Tumballi <
> atumball at redhat.com>
> *Subject: *Re: [Gluster-users] df does not show full volume capacity
> after update to 3.12.4
>
>
>
>
>
> On 31 January 2018 at 21:50, Freer, Eva B. <freereb at ornl.gov> wrote:
>
> The values for shared-brick-count are still the same. I did not re-start
> the volume after setting the cluster.min-free-inodes to 6%. Do I need to
> restart it?
>
>
>
> That is not necessary. Let me get back to you on this tomorrow.
>
>
>
> Regards,
>
> Nithya
>
>
>
>
>
> Thanks,
>
> Eva     (865) 574-6894
>
>
>
> *From: *Nithya Balachandran <nbalacha at redhat.com>
> *Date: *Wednesday, January 31, 2018 at 11:14 AM
> *To: *Eva Freer <freereb at ornl.gov>
> *Cc: *"Greene, Tami McFarlin" <greenet at ornl.gov>, "
> gluster-users at gluster.org" <gluster-users at gluster.org>, Amar Tumballi <
> atumball at redhat.com>
>
>
> *Subject: *Re: [Gluster-users] df does not show full volume capacity
> after update to 3.12.4
>
>
>
>
>
>
>
> On 31 January 2018 at 21:34, Freer, Eva B. <freereb at ornl.gov> wrote:
>
> Nithya,
>
>
>
> Responding to an earlier question: Before the upgrade, we were at 3.103 on
> these servers, but some of the clients were 3.7.6. From below, does this
> mean that “shared-brick-count” needs to be set to 1 for all bricks.
>
>
>
> All of the bricks are on separate xfs partitions composed hardware of RAID
> 6 volumes. LVM is not used. The current setting for cluster.min-free-inodes
> was 5%. I changed it to 6% per your instructions below. The df output is
> still the same, but I haven’t done the
>
> *find /var/lib/glusterd/vols -type f|xargs sed -i -e 's/option
> shared-brick-count [0-9]*/option shared-brick-count 1/g'*
>
> Should I go ahead and do this?
>
>
>
> Can you check if the values have been changed in the .vol files before you
> try this?
>
>
>
> These files will be regenerated every time the volume is changed so
> changing them directly may not be permanent. I was hoping setting the
> cluster.min-free-inodes would have corrected this automatically and helped
> us figure out what was happening as we have not managed to reproduce this
> issue yet.
>
>
>
>
>
>
>
>
>
> Output of stat –f for all the bricks:
>
>
>
> [root at jacen ~]# stat -f /bricks/data_A*
>
>   File: "/bricks/data_A1"
>
>     ID: 80100000000 Namelen: 255     Type: xfs
>
> Block size: 4096       Fundamental block size: 4096
>
> Blocks: Total: 15626471424 Free: 4530515093 Available: 4530515093
>
> Inodes: Total: 1250159424 Free: 1250028064
>
>   File: "/bricks/data_A2"
>
>     ID: 81100000000 Namelen: 255     Type: xfs
>
> Block size: 4096       Fundamental block size: 4096
>
> Blocks: Total: 15626471424 Free: 3653183901 Available: 3653183901
>
> Inodes: Total: 1250159424 Free: 1250029262
>
>   File: "/bricks/data_A3"
>
>     ID: 82100000000 Namelen: 255     Type: xfs
>
> Block size: 4096       Fundamental block size: 4096
>
> Blocks: Total: 15626471424 Free: 15134840607 Available: 15134840607
>
> Inodes: Total: 1250159424 Free: 1250128031
>
>   File: "/bricks/data_A4"
>
>     ID: 83100000000 Namelen: 255     Type: xfs
>
> Block size: 4096       Fundamental block size: 4096
>
> Blocks: Total: 15626471424 Free: 15626461604 Available: 15626461604
>
> Inodes: Total: 1250159424 Free: 1250153857
>
>
>
> [root at jaina dataeng]# stat -f /bricks/data_B*
>
>   File: "/bricks/data_B1"
>
>     ID: 80100000000 Namelen: 255     Type: xfs
>
> Block size: 4096       Fundamental block size: 4096
>
> Blocks: Total: 15626471424 Free: 5689640723 Available: 5689640723
>
> Inodes: Total: 1250159424 Free: 1250047934
>
>   File: "/bricks/data_B2"
>
>     ID: 81100000000 Namelen: 255     Type: xfs
>
> Block size: 4096       Fundamental block size: 4096
>
> Blocks: Total: 15626471424 Free: 6623312785 Available: 6623312785
>
> Inodes: Total: 1250159424 Free: 1250048131
>
>   File: "/bricks/data_B3"
>
>     ID: 82100000000 Namelen: 255     Type: xfs
>
> Block size: 4096       Fundamental block size: 4096
>
> Blocks: Total: 15626471424 Free: 15106888485 Available: 15106888485
>
> Inodes: Total: 1250159424 Free: 1250122139
>
>   File: "/bricks/data_B4"
>
>     ID: 83100000000 Namelen: 255     Type: xfs
>
> Block size: 4096       Fundamental block size: 4096
>
> Blocks: Total: 15626471424 Free: 15626461604 Available: 15626461604
>
> Inodes: Total: 1250159424 Free: 1250153857
>
>
>
>
>
> Thanks,
>
> Eva     (865) 574-6894
>
>
>
> *From: *Nithya Balachandran <nbalacha at redhat.com>
> *Date: *Wednesday, January 31, 2018 at 10:46 AM
> *To: *Eva Freer <freereb at ornl.gov>, "Greene, Tami McFarlin" <
> greenet at ornl.gov>
> *Cc: *Amar Tumballi <atumball at redhat.com>
>
>
> *Subject: *Re: [Gluster-users] df does not show full volume capacity
> after update to 3.12.4
>
>
>
> Thank you Eva.
>
>
>
> From the files you sent:
>
> dataeng.jacen.bricks-data_A1-dataeng.vol:    option shared-brick-count 2
>
> dataeng.jacen.bricks-data_A2-dataeng.vol:    option shared-brick-count 2
>
> dataeng.jacen.bricks-data_A3-dataeng.vol:    option shared-brick-count 1
>
> dataeng.jacen.bricks-data_A4-dataeng.vol:    option shared-brick-count 1
>
> dataeng.jaina.bricks-data_B1-dataeng.vol:    option shared-brick-count 0
>
> dataeng.jaina.bricks-data_B2-dataeng.vol:    option shared-brick-count 0
>
> dataeng.jaina.bricks-data_B3-dataeng.vol:    option shared-brick-count 0
>
> dataeng.jaina.bricks-data_B4-dataeng.vol:    option shared-brick-count 0
>
>
>
>
>
> Are all of these bricks on separate Filesystem partitions? If yes, can you
> please try running the following on one of the gluster nodes and see if the
> df output works post that?
>
>
>
> gluster v set dataeng cluster.min-free-inodes 6%
>
>
>
>
>
> If it doesn;t work, please send us the stat -f output for each brick.
>
>
>
> Regards,
>
> Nithya
>
>
>
> On 31 January 2018 at 20:41, Freer, Eva B. <freereb at ornl.gov> wrote:
>
> Nithya,
>
>
>
> The file for one of the servers is attached.
>
>
>
> Thanks,
>
> Eva     (865) 574-6894
>
>
>
> *From: *Nithya Balachandran <nbalacha at redhat.com>
> *Date: *Wednesday, January 31, 2018 at 1:17 AM
> *To: *Eva Freer <freereb at ornl.gov>
> *Cc: *"gluster-users at gluster.org" <gluster-users at gluster.org>, "Greene,
> Tami McFarlin" <greenet at ornl.gov>
> *Subject: *Re: [Gluster-users] df does not show full volume capacity
> after update to 3.12.4
>
>
>
> I found this on the mailing list:
>
>
>
>
>
>
>
>
>
> *I found the issue. The CentOS 7 RPMs, upon upgrade, modifies the .vol
> files. Among other things, it adds "option shared-brick-count \d", using
> the number of bricks in the volume. This gives you an average free space
> per brick, instead of total free space in the volume. When I create a new
> volume, the value of "shared-brick-count" is "1". find
> /var/lib/glusterd/vols -type f|xargs sed -i -e 's/option shared-brick-count
> [0-9]*/option shared-brick-count 1/g'*
>
>
>
>
>
>
>
> Eva, can you send me the contents of the /var/lib/glusterd/<volname>
> folder from any one node so I can confirm if this is the problem?
>
>
>
> Regards,
>
> Nithya
>
>
>
>
>
> On 31 January 2018 at 10:47, Nithya Balachandran <nbalacha at redhat.com>
> wrote:
>
> Hi Eva,
>
>
>
> One more question. What version of gluster were you running before the
> upgrade?
>
>
>
> Thanks,
>
> Nithya
>
>
>
> On 31 January 2018 at 09:52, Nithya Balachandran <nbalacha at redhat.com>
> wrote:
>
> Hi Eva,
>
>
>
> Can you send us the following:
>
>
>
> gluster volume info
>
> gluster volume status
>
>
>
> The log files and tcpdump for df on a fresh mount point for that volume.
>
>
>
> Thanks,
>
> Nithya
>
>
>
>
>
> On 31 January 2018 at 07:17, Freer, Eva B. <freereb at ornl.gov> wrote:
>
> After OS update to CentOS 7.4 or RedHat 6.9 and update to Gluster 3.12.4,
> the ‘df’ command shows only part of the available space on the mount point
> for multi-brick volumes. All nodes are at 3.12.4. This occurs on both
> servers and clients.
>
>
>
> We have 2 different server configurations.
>
>
>
> Configuration 1: A distributed volume of 8 bricks with 4 on each server.
> The initial configuration had 4 bricks of 59TB each with 2 on each server.
> Prior to the update to CentOS 7.4 and gluster 3.12.4, ‘df’ correctly showed
> the size for the volume as 233TB. After the update, we added 2 bricks with
> 1 on each server, but the output of ‘df’ still only listed 233TB for the
> volume. We added 2 more bricks, again with 1 on each server. The output of
> ‘df’ now shows 350TB, but the aggregate of 8 – 59TB bricks should be ~466TB.
>
>
>
> Configuration 2: A distributed, replicated volume with 9 bricks on each
> server for a total of ~350TB of storage. After the server update to RHEL
> 6.9 and gluster 3.12.4, the volume now shows as having 50TB with ‘df’. No
> changes were made to this volume after the update.
>
>
>
> In both cases, examining the bricks shows that the space and files are
> still there, just not reported correctly with ‘df’. All machines have been
> rebooted and the problem persists.
>
>
>
> Any help/advice you can give on this would be greatly appreciated.
>
>
>
> Thanks in advance.
>
> Eva Freer
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
>
>
>
>
>
>
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20180131/0d17aa22/attachment.html>


More information about the Gluster-users mailing list