[Gluster-users] gluster 3.7.8 page allocation failure

David Robinson david.robinson at corvidtec.com
Thu Feb 11 23:58:03 UTC 2016


I am sorting a fairly large file (27-million lines) and the output is 
being written to my gluster storage.  This seems to crash glusterfsd for 
3.7.8 as noted below.
Can anyone help?

David


[Thu Feb 11 18:25:24 2016] glusterfsd: page allocation failure. order:5, 
mode:0x20
[Thu Feb 11 18:25:24 2016] Pid: 17868, comm: glusterfsd Not tainted 
2.6.32-573.12.1.el6.x86_64 #1
[Thu Feb 11 18:25:24 2016] Call Trace:
[Thu Feb 11 18:25:24 2016]  [<ffffffff811376ac>] ? 
__alloc_pages_nodemask+0x7dc/0x950
[Thu Feb 11 18:25:24 2016]  [<ffffffffa02cba00>] ? 
mlx4_ib_post_send+0x6c0/0x1f90 [mlx4_ib]
[Thu Feb 11 18:25:24 2016]  [<ffffffffa037076c>] ? 
xfs_iext_bno_to_ext+0x8c/0x170 [xfs]
[Thu Feb 11 18:25:24 2016]  [<ffffffff81176f92>] ? 
kmem_getpages+0x62/0x170
[Thu Feb 11 18:25:24 2016]  [<ffffffff81177baa>] ? 
fallback_alloc+0x1ba/0x270
[Thu Feb 11 18:25:24 2016]  [<ffffffff811775ff>] ? 
cache_grow+0x2cf/0x320
[Thu Feb 11 18:25:24 2016]  [<ffffffff81177929>] ? 
____cache_alloc_node+0x99/0x160
[Thu Feb 11 18:25:24 2016]  [<ffffffff8145fdb2>] ? 
pskb_expand_head+0x62/0x280
[Thu Feb 11 18:25:24 2016]  [<ffffffff81178579>] ? __kmalloc+0x199/0x230
[Thu Feb 11 18:25:24 2016]  [<ffffffff8145fdb2>] ? 
pskb_expand_head+0x62/0x280
[Thu Feb 11 18:25:24 2016]  [<ffffffff812761c2>] ? 
get_request+0x302/0x3c0
[Thu Feb 11 18:25:24 2016]  [<ffffffff8146069a>] ? 
__pskb_pull_tail+0x2aa/0x360
[Thu Feb 11 18:25:24 2016]  [<ffffffff8146f9e9>] ? 
harmonize_features+0x29/0x70
[Thu Feb 11 18:25:24 2016]  [<ffffffff81470054>] ? 
dev_hard_start_xmit+0x1c4/0x490
[Thu Feb 11 18:25:24 2016]  [<ffffffff8148d53a>] ? 
sch_direct_xmit+0x15a/0x1c0
[Thu Feb 11 18:25:24 2016]  [<ffffffff814705c8>] ? 
dev_queue_xmit+0x228/0x320
[Thu Feb 11 18:25:24 2016]  [<ffffffff81476cbd>] ? 
neigh_connected_output+0xbd/0x100
[Thu Feb 11 18:25:24 2016]  [<ffffffff814ac217>] ? 
ip_finish_output+0x287/0x360
[Thu Feb 11 18:25:24 2016]  [<ffffffff814ac3a8>] ? ip_output+0xb8/0xc0
[Thu Feb 11 18:25:24 2016]  [<ffffffff814ab635>] ? 
ip_local_out+0x25/0x30
[Thu Feb 11 18:25:24 2016]  [<ffffffff814abb30>] ? 
ip_queue_xmit+0x190/0x420
[Thu Feb 11 18:25:24 2016]  [<ffffffff81136ff9>] ? 
__alloc_pages_nodemask+0x129/0x950
[Thu Feb 11 18:25:24 2016]  [<ffffffff814c1204>] ? 
tcp_transmit_skb+0x4b4/0x8b0
[Thu Feb 11 18:25:24 2016]  [<ffffffff814c374a>] ? 
tcp_write_xmit+0x1da/0xa90
[Thu Feb 11 18:25:24 2016]  [<ffffffff81178dbd>] ? 
__kmalloc_node+0x4d/0x60
[Thu Feb 11 18:25:24 2016]  [<ffffffff814c4030>] ? 
tcp_push_one+0x30/0x40
[Thu Feb 11 18:25:24 2016]  [<ffffffff814b46bc>] ? 
tcp_sendmsg+0x9cc/0xa20
[Thu Feb 11 18:25:24 2016]  [<ffffffff814589eb>] ? 
sock_aio_write+0x19b/0x1c0
[Thu Feb 11 18:25:24 2016]  [<ffffffff81458850>] ? 
sock_aio_write+0x0/0x1c0
[Thu Feb 11 18:25:24 2016]  [<ffffffff8119179b>] ? 
do_sync_readv_writev+0xfb/0x140
[Thu Feb 11 18:25:24 2016]  [<ffffffffa0345a66>] ? 
xfs_attr_get+0xb6/0xc0 [xfs]
[Thu Feb 11 18:25:24 2016]  [<ffffffffa039f7ef>] ? 
__xfs_xattr_get+0x2f/0x50 [xfs]
[Thu Feb 11 18:25:24 2016]  [<ffffffff810a1460>] ? 
autoremove_wake_function+0x0/0x40
[Thu Feb 11 18:25:24 2016]  [<ffffffff811ba34c>] ? getxattr+0x9c/0x170
[Thu Feb 11 18:25:24 2016]  [<ffffffff81231a16>] ? 
security_file_permission+0x16/0x20
[Thu Feb 11 18:25:24 2016]  [<ffffffff81192846>] ? 
do_readv_writev+0xd6/0x1f0
[Thu Feb 11 18:25:24 2016]  [<ffffffff811929a6>] ? vfs_writev+0x46/0x60
[Thu Feb 11 18:25:24 2016]  [<ffffffff81192ad1>] ? sys_writev+0x51/0xd0
[Thu Feb 11 18:25:24 2016]  [<ffffffff810e884e>] ? 
__audit_syscall_exit+0x25e/0x290
[Thu Feb 11 18:25:24 2016]  [<ffffffff8100b0d2>] ? 
system_call_fastpath+0x16/0x1b
[root at gfs02bkp ~]# gluster volume info
Volume Name: gfsbackup
Type: Distribute
Volume ID: e78d5123-d9bc-4d88-9c73-61d28abf0b41
Status: Started
Number of Bricks: 7
Transport-type: tcp
Bricks:
Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/gfsbackup
Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/gfsbackup
Brick3: gfsib02bkp.corvidtec.com:/data/brick01bkp/gfsbackup
Brick4: gfsib02bkp.corvidtec.com:/data/brick02bkp/gfsbackup
Brick5: gfsib02bkp.corvidtec.com:/data/brick03bkp/gfsbackup
Brick6: gfsib02bkp.corvidtec.com:/data/brick04bkp/gfsbackup
Brick7: gfsib02bkp.corvidtec.com:/data/brick05bkp/gfsbackup
Options Reconfigured:
nfs.disable: off
server.allow-insecure: on
storage.owner-gid: 100
server.manage-gids: on
cluster.lookup-optimize: on
server.event-threads: 8
client.event-threads: 8
changelog.changelog: off
storage.build-pgfid: on
performance.readdir-ahead: on
diagnostics.brick-log-level: WARNING
diagnostics.client-log-level: WARNING
[root at gfs02bkp ~]# rpm -qa | grep gluster
glusterfs-fuse-3.7.8-1.el6.x86_64
glusterfs-geo-replication-3.7.8-1.el6.x86_64
python-gluster-3.7.8-1.el6.noarch
glusterfs-client-xlators-3.7.8-1.el6.x86_64
glusterfs-server-3.7.8-1.el6.x86_64
glusterfs-api-devel-3.7.8-1.el6.x86_64
glusterfs-debuginfo-3.7.8-1.el6.x86_64
glusterfs-3.7.8-1.el6.x86_64
glusterfs-cli-3.7.8-1.el6.x86_64
glusterfs-devel-3.7.8-1.el6.x86_64
glusterfs-rdma-3.7.8-1.el6.x86_64
glusterfs-libs-3.7.8-1.el6.x86_64
glusterfs-extra-xlators-3.7.8-1.el6.x86_64
glusterfs-api-3.7.8-1.el6.x86_64
glusterfs-resource-agents-3.7.8-1.el6.noarch


========================



David F. Robinson, Ph.D.

President - Corvid Technologies

145 Overhill Drive

Mooresville, NC 28117

704.799.6944 x101   [Office]

704.252.1310           [Cell]

704.799.7974           [Fax]

david.robinson at corvidtec.com

http://www.corvidtec.com

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160211/92df504b/attachment.html>


More information about the Gluster-users mailing list