[Gluster-users] different free disk space size on distributed replicated
    Patrick Kaiser 
    patrick.kaiser at vnc.biz
       
    Tue Jan 12 08:18:46 UTC 2016
    
    
  
Hi, 
thanks for your feedback. I've figured out, that a brick was not working anymore. Only after restarting the whole server with the failed brick, the 
volume now has identical sizes. 
Thanks 
Mit freundlichen Grüßen 
Patrick Kaiser 
VNC - Virtual Network Consult GmbH 
From: "Pranith Kumar Karampuri" <pkarampu at redhat.com> 
To: "Patrick Kaiser" <patrick.kaiser at vnc.biz>, gluster-users at gluster.org 
Sent: Tuesday, January 12, 2016 4:42:22 AM 
Subject: Re: [Gluster-users] different free disk space size on distributed replicated 
On 01/08/2016 06:30 PM, Patrick Kaiser wrote: 
hi, 
I am running a distributed replicated gluster fs setup with 4 nodes. 
currently i have no problems but i was wondering when i am running gluster volume status 
and seeing different free disk space on every node. 
I am wondering if I should not have the same free and used size on gluster00 and gluster01 
and also on gluster02 and gluster03 (as they are the replicated ones) 
It doesn't look right to me either. Do you have any self-heals that need to happen on the first replica subvolume? "gluster volume heal <volname> info" 
Pranith 
BQ_BEGIN
    1. 
root at gluster0:~# gluster volume status GV01 detail 
    2. 
Status of volume: GV01 
    3. 
------------------------------------------------------------------------------ 
    4. 
Brick : Brick gluster00.storage.domain:/brick/gv01 
    5. 
Port : 49163 
    6. 
Online : Y 
    7. 
Pid : 3631 
    8. 
File System : xfs 
    9. 
Device : /dev/mapper/vg--gluster0-DATA 
    10. 
Mount Options : rw,relatime,attr2,delaylog,noquota 
    11. 
Inode Size : 256 
    12. 
Disk Space Free : 5.7TB 
    13. 
Total Disk Space : 13.6TB 
    14. 
Inode Count : 2923388928 
    15. 
Free Inodes : 2922850330 
    16. 
------------------------------------------------------------------------------ 
    17. 
Brick : Brick gluster01.storage.domain:/brick/gv01 
    18. 
Port : 49163 
    19. 
Online : Y 
    20. 
Pid : 2976 
    21. 
File System : xfs 
    22. 
Device : /dev/mapper/vg--gluster1-DATA 
    23. 
Mount Options : rw,relatime,attr2,delaylog,noquota 
    24. 
Inode Size : 256 
    25. 
Disk Space Free : 4.4TB 
    26. 
Total Disk Space : 13.6TB 
    27. 
Inode Count : 2923388928 
    28. 
Free Inodes : 2922826116 
    29. 
------------------------------------------------------------------------------ 
    30. 
Brick : Brick gluster02.storage.domain:/brick/gv01 
    31. 
Port : 49163 
    32. 
Online : Y 
    33. 
Pid : 3051 
    34. 
File System : xfs 
    35. 
Device : /dev/mapper/vg--gluster2-DATA 
    36. 
Mount Options : rw,relatime,attr2,delaylog,noquota 
    37. 
Inode Size : 256 
    38. 
Disk Space Free : 6.4TB 
    39. 
Total Disk Space : 13.6TB 
    40. 
Inode Count : 2923388928 
    41. 
Free Inodes : 2922851020 
    42. 
------------------------------------------------------------------------------ 
    43. 
Brick : Brick gluster03.storage.domain:/brick/gv01 
    44. 
Port : N/A 
    45. 
Online : N 
    46. 
Pid : 29822 
    47. 
File System : xfs 
    48. 
Device : /dev/mapper/vg--gluster3-DATA 
    49. 
Mount Options : rw,relatime,attr2,delaylog,noquota 
    50. 
Inode Size : 256 
    51. 
Disk Space Free : 6.2TB 
    52. 
Total Disk Space : 13.6TB 
    53. 
Inode Count : 2923388928 
    54. 
Free Inodes : 2922847631 
friendly regards, 
Patrick 
_______________________________________________
Gluster-users mailing list Gluster-users at gluster.org http://www.gluster.org/mailman/listinfo/gluster-users 
BQ_END
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160112/a214651f/attachment.html>
    
    
More information about the Gluster-users
mailing list