[Gluster-users] Erroneous "No space left on device." messages

Pat Haley phaley at mit.edu
Tue Mar 10 18:45:29 UTC 2020


Hi,

I get the following

[root at mseas-data2 bricks]# gluster  volume get data-volume all | grep 
cluster.min-free
cluster.min-free-disk 10%
cluster.min-free-inodes 5%


On 3/10/20 2:34 PM, Strahil Nikolov wrote:
> On March 10, 2020 8:14:41 PM GMT+02:00, Pat Haley <phaley at mit.edu> wrote:
>> HI,
>>
>> After some more poking around in the logs (specifically the brick logs)
>>
>>   * brick1 & brick2 have both been recording "No space left on device"
>>     messages today (as recently at 15 minutes ago)
>>   * brick3 last recorded a "No space left on device" message last night
>>     around 10:30pm
>>   * brick4 has no such messages in its log file
>>
>> Note brick1 & brick2 are on one server, brick3 and brick4 are on the
>> second server.
>>
>> Pat
>>
>>
>> On 3/10/20 11:51 AM, Pat Haley wrote:
>>> Hi,
>>>
>>> We have developed a problem with Gluster reporting "No space left on
>>> device." even though "df" of both the gluster filesystem and the
>>> underlying bricks show space available (details below).  Our inode
>>> usage is between 1-3%.  We are running gluster 3.7.11 in a
>> distributed
>>> volume across 2 servers (2 bricks each). We have followed the thread
>>>
>> https://lists.gluster.org/pipermail/gluster-users/2020-March/037821.html
>>
>>> but haven't found a solution yet.
>>>
>>> Last night we ran a rebalance which appeared successful (and have
>>> since cleared up some more space which seems to have mainly been on
>>> one brick).  There were intermittent erroneous "No space..." messages
>>> last night, but they have become much more frequent today.
>>>
>>> Any help would be greatly appreciated.
>>>
>>> Thanks
>>>
>>> ---------------------------
>>> [root at mseas-data2 ~]# df -h
>>> ---------------------------
>>> Filesystem      Size  Used Avail Use% Mounted on
>>> /dev/sdb        164T  164T  324G 100% /mnt/brick2
>>> /dev/sda        164T  164T  323G 100% /mnt/brick1
>>> ---------------------------
>>> [root at mseas-data2 ~]# df -i
>>> ---------------------------
>>> Filesystem         Inodes    IUsed      IFree IUse% Mounted on
>>> /dev/sdb       1375470800 31207165 1344263635    3% /mnt/brick2
>>> /dev/sda       1384781520 28706614 1356074906    3% /mnt/brick1
>>>
>>> ---------------------------
>>> [root at mseas-data3 ~]# df -h
>>> ---------------------------
>>> /dev/sda               91T   91T  323G 100% /export/sda/brick3
>>> /dev/mapper/vg_Data4-lv_Data4
>>>                         91T   88T  3.4T  97% /export/sdc/brick4
>>> ---------------------------
>>> [root at mseas-data3 ~]# df -i
>>> ---------------------------
>>> /dev/sda              679323496  9822199  669501297    2%
>>> /export/sda/brick3
>>> /dev/mapper/vg_Data4-lv_Data4
>>>                       3906272768 11467484 3894805284    1%
>>> /export/sdc/brick4
>>>
>>>
>>>
>>> ---------------------------------------
>>> [root at mseas-data2 ~]# gluster --version
>>> ---------------------------------------
>>> glusterfs 3.7.11 built on Apr 27 2016 14:09:22
>>> Repository revision: git://git.gluster.com/glusterfs.git
>>> Copyright (c) 2006-2011 Gluster Inc. <http://www.gluster.com>
>>> GlusterFS comes with ABSOLUTELY NO WARRANTY.
>>> You may redistribute copies of GlusterFS under the terms of the GNU
>>> General Public License.
>>>
>>>
>>>
>>> -----------------------------------------
>>> [root at mseas-data2 ~]# gluster volume info
>>> -----------------------------------------
>>> Volume Name: data-volume
>>> Type: Distribute
>>> Volume ID: c162161e-2a2d-4dac-b015-f31fd89ceb18
>>> Status: Started
>>> Number of Bricks: 4
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: mseas-data2:/mnt/brick1
>>> Brick2: mseas-data2:/mnt/brick2
>>> Brick3: mseas-data3:/export/sda/brick3
>>> Brick4: mseas-data3:/export/sdc/brick4
>>> Options Reconfigured:
>>> nfs.export-volumes: off
>>> nfs.disable: on
>>> performance.readdir-ahead: on
>>> diagnostics.brick-sys-log-level: WARNING
>>> nfs.exports-auth-enable: on
>>> server.allow-insecure: on
>>> auth.allow: *
>>> disperse.eager-lock: off
>>> performance.open-behind: off
>>> performance.md-cache-timeout: 60
>>> network.inode-lru-limit: 50000
>>> diagnostics.client-log-level: ERROR
>>>
>>>
>>>
>>> --------------------------------------------------------------
>>> [root at mseas-data2 ~]# gluster volume status data-volume detail
>>> --------------------------------------------------------------
>>> Status of volume: data-volume
>>>
>> ------------------------------------------------------------------------------
>>
>>> Brick                : Brick mseas-data2:/mnt/brick1
>>> TCP Port             : 49154
>>> RDMA Port            : 0
>>> Online               : Y
>>> Pid                  : 4601
>>> File System          : xfs
>>> Device               : /dev/sda
>>> Mount Options        : rw
>>> Inode Size           : 256
>>> Disk Space Free      : 318.8GB
>>> Total Disk Space     : 163.7TB
>>> Inode Count          : 1365878288
>>> Free Inodes          : 1337173596
>>>
>> ------------------------------------------------------------------------------
>>
>>> Brick                : Brick mseas-data2:/mnt/brick2
>>> TCP Port             : 49155
>>> RDMA Port            : 0
>>> Online               : Y
>>> Pid                  : 7949
>>> File System          : xfs
>>> Device               : /dev/sdb
>>> Mount Options        : rw
>>> Inode Size           : 256
>>> Disk Space Free      : 319.8GB
>>> Total Disk Space     : 163.7TB
>>> Inode Count          : 1372421408
>>> Free Inodes          : 1341219039
>>>
>> ------------------------------------------------------------------------------
>>
>>> Brick                : Brick mseas-data3:/export/sda/brick3
>>> TCP Port             : 49153
>>> RDMA Port            : 0
>>> Online               : Y
>>> Pid                  : 4650
>>> File System          : xfs
>>> Device               : /dev/sda
>>> Mount Options        : rw
>>> Inode Size           : 512
>>> Disk Space Free      : 325.3GB
>>> Total Disk Space     : 91.0TB
>>> Inode Count          : 692001992
>>> Free Inodes          : 682188893
>>>
>> ------------------------------------------------------------------------------
>>
>>> Brick                : Brick mseas-data3:/export/sdc/brick4
>>> TCP Port             : 49154
>>> RDMA Port            : 0
>>> Online               : Y
>>> Pid                  : 23772
>>> File System          : xfs
>>> Device               : /dev/mapper/vg_Data4-lv_Data4
>>> Mount Options        : rw
>>> Inode Size           : 256
>>> Disk Space Free      : 3.4TB
>>> Total Disk Space     : 90.9TB
>>> Inode Count          : 3906272768
>>> Free Inodes          : 3894809903
>>>
> Hi Pat,
>
> What is the output of:
> gluster  volume get data-volume all | grep cluster.min-free
>
> 1% of 164 T is  1640G , but in your case you have only 324G  which is way lower.
>
> Best Regards,
> Strahil Nikolov

-- 

-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
Pat Haley                          Email:  phaley at mit.edu
Center for Ocean Engineering       Phone:  (617) 253-6824
Dept. of Mechanical Engineering    Fax:    (617) 253-8125
MIT, Room 5-213                    http://web.mit.edu/phaley/www/
77 Massachusetts Avenue
Cambridge, MA  02139-4301



More information about the Gluster-users mailing list