[Gluster-users] different free disk space size on distributed replicated
Pranith Kumar Karampuri
pkarampu at redhat.com
Tue Jan 12 08:45:24 UTC 2016
On 01/12/2016 01:48 PM, Patrick Kaiser wrote:
> Hi,
>
> thanks for your feedback. I've figured out, that a brick was not
> working anymore. Only after restarting the whole server with the
> failed brick, the
> volume now has identical sizes.
Thanks for the confirmation. Thought there could be some bug in gluster :-)
Pranith
>
> Thanks
>
>
> Mit freundlichen Grüßen
>
> Patrick Kaiser
> VNC - Virtual Network Consult GmbH
>
> ------------------------------------------------------------------------
> *From: *"Pranith Kumar Karampuri" <pkarampu at redhat.com>
> *To: *"Patrick Kaiser" <patrick.kaiser at vnc.biz>, gluster-users at gluster.org
> *Sent: *Tuesday, January 12, 2016 4:42:22 AM
> *Subject: *Re: [Gluster-users] different free disk space size on
> distributed replicated
>
>
>
> On 01/08/2016 06:30 PM, Patrick Kaiser wrote:
>
> hi,
>
> I am running a distributed replicated gluster fs setup with 4 nodes.
> currently i have no problems but i was wondering when i am running
> gluster volume status
> and seeing different free disk space on every node.
> I am wondering if I should not have the same free and used size on
> gluster00 and gluster01
> and also on gluster02 and gluster03 (as they are the replicated ones)
>
> It doesn't look right to me either. Do you have any self-heals that
> need to happen on the first replica subvolume? "gluster volume heal
> <volname> info"
>
> Pranith
>
>
>
>
> 1.
> root at gluster0:~# gluster volume status GV01 detail
> 2.
> Status of volume: GV01
> 3.
> ------------------------------------------------------------------------------
> 4.
> Brick : Brick gluster00.storage.domain:/brick/gv01
> 5.
> Port : 49163
> 6.
> Online : Y
> 7.
> Pid : 3631
> 8.
> File System : xfs
> 9.
> Device : /dev/mapper/vg--gluster0-DATA
> 10.
> Mount Options : rw,relatime,attr2,delaylog,noquota
> 11.
> Inode Size : 256
> 12.
> Disk Space Free : 5.7TB
> 13.
> Total Disk Space : 13.6TB
> 14.
> Inode Count : 2923388928
> 15.
> Free Inodes : 2922850330
> 16.
> ------------------------------------------------------------------------------
> 17.
> Brick : Brick gluster01.storage.domain:/brick/gv01
> 18.
> Port : 49163
> 19.
> Online : Y
> 20.
> Pid : 2976
> 21.
> File System : xfs
> 22.
> Device : /dev/mapper/vg--gluster1-DATA
> 23.
> Mount Options : rw,relatime,attr2,delaylog,noquota
> 24.
> Inode Size : 256
> 25.
> Disk Space Free : 4.4TB
> 26.
> Total Disk Space : 13.6TB
> 27.
> Inode Count : 2923388928
> 28.
> Free Inodes : 2922826116
> 29.
> ------------------------------------------------------------------------------
> 30.
> Brick : Brick gluster02.storage.domain:/brick/gv01
> 31.
> Port : 49163
> 32.
> Online : Y
> 33.
> Pid : 3051
> 34.
> File System : xfs
> 35.
> Device : /dev/mapper/vg--gluster2-DATA
> 36.
> Mount Options : rw,relatime,attr2,delaylog,noquota
> 37.
> Inode Size : 256
> 38.
> Disk Space Free : 6.4TB
> 39.
> Total Disk Space : 13.6TB
> 40.
> Inode Count : 2923388928
> 41.
> Free Inodes : 2922851020
> 42.
> ------------------------------------------------------------------------------
> 43.
> Brick : Brick gluster03.storage.domain:/brick/gv01
> 44.
> Port : N/A
> 45.
> Online : N
> 46.
> Pid : 29822
> 47.
> File System : xfs
> 48.
> Device : /dev/mapper/vg--gluster3-DATA
> 49.
> Mount Options : rw,relatime,attr2,delaylog,noquota
> 50.
> Inode Size : 256
> 51.
> Disk Space Free : 6.2TB
> 52.
> Total Disk Space : 13.6TB
> 53.
> Inode Count : 2923388928
> 54.
> Free Inodes : 2922847631
>
>
> friendly regards,
> Patrick
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160112/13dd5ee2/attachment.html>
More information about the Gluster-users
mailing list