[Gluster-users] Metadata filesystem XFS gluster 3.6
Jorick Astrego
j.astrego at netbulae.eu
Tue Mar 31 11:58:43 UTC 2015
On 03/31/2015 09:05 AM, Félix de Lelelis wrote:
> Hi,
>
> I had a problem with a filesystem xfs on gluster. The filesystem
> metadata was filled:
>
> Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: space map
> metadata: unable to allocate new metadata block
> Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin: 252:2:
> metadata operation 'dm_thin_insert_block' failed: error = -28
> Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin: 252:2:
> aborting current metadata transaction
> Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin: 252:2:
> switching pool to read-only mode
> Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): metadata I/O error:
> block 0x701830 ("xfs_buf_iodone_callbacks") error 5 numblks 8
> Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of
> device
> Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
> Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
> process_bio_read_only: dm_thin_find_block() failed: error = -5
> Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of
> device
> Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
> Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
> process_bio_read_only: dm_thin_find_block() failed: error = -5
> Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): metadata I/O error:
> block 0x68047c ("xlog_iodone") error 5 numblks 64
> Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4):
> xfs_do_force_shutdown(0x2) called from line 1170 of file
> fs/xfs/xfs_log.c. Return address = 0xffffffffa012a4c1
> Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): Log I/O Error
> Detected. Shutting down filesystem
> Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): Please umount the
> filesystem and rectify the problem(s)
> Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of
> device
> Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
> Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
> process_bio_read_only: dm_thin_find_block() failed: error = -5
> Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): metadata I/O error:
> block 0x6804bc ("xlog_iodone") error 5 numblks 64
> Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4):
> xfs_do_force_shutdown(0x2) called from line 1170 of file
> fs/xfs/xfs_log.c. Return address = 0xffffffffa012a4c1
> Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of
> device
> Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): xfs_log_force: error
> 5 returned.
> Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
> Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
> process_bio_read_only: dm_thin_find_block() failed: error = -5
> Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of
> device
> Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
> Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
> process_bio_read_only: dm_thin_find_block() failed: error = -5
> Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of
> device
> Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
> Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
> process_bio_read_only: dm_thin_find_block() failed: error = -5
> Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): metadata I/O error:
> block 0x6804fc ("xlog_iodone") error 5 numblks 64
> Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4):
> xfs_do_force_shutdown(0x2) called from line 1170 of file
> fs/xfs/xfs_log.c. Return address = 0xffffffffa012a4c
>
>
>
> After that, gluster was shutdown and with it the 2 server are shtudown
> too. The lvm partition was missing and so far I haven't been able
> restore the file system. All data is missing??
> I don't understand the situation and I don't know if it's due a xfs
> filesystem or glusterfs fail. Someone it has been this situation?
>
> Thanks.
>
>
Hi,
Do you run thin provisioned on LVM?
There are some discussions about this:
> https://www.redhat.com/archives/linux-lvm/2014-December/msg00015.html
> You definitely ran out of metadata space. Which version of the kernel
> and lvm2 userspace are you using?
>
> See the "Metadata space exhaustion" section of the lvmthin manpage in a
> recent lvm2 release to guide you on how to recover.
>
> Also, once you've gotten past ths you really should configure lvm2 to
> autoextend the thin-pool (both data and metadata) as needed in response
> to low watermark, etc. See "Automatically extend thin pool LV" in
> lvmthin manpage.
Maybe related to:
> https://bugzilla.redhat.com/show_bug.cgi?id=1097948
Hope this helps.
Met vriendelijke groet, With kind regards,
Jorick Astrego
Netbulae Virtualization Experts
----------------
Tel: 053 20 30 270 info at netbulae.eu Staalsteden 4-3A KvK 08198180
Fax: 053 20 30 271 www.netbulae.eu 7547 TA Enschede BTW NL821234584B01
----------------
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150331/e5db366f/attachment.html>
More information about the Gluster-users
mailing list