[Gluster-users] glusterfsd Call Trace Messages

Raghavendra Bhat rabhat at redhat.com
Wed Feb 3 19:09:07 UTC 2016


Hi,

Is your nfs client mounted on one of the gluster serves?


Regards,
Raghavendra

On Wed, Feb 3, 2016 at 10:08 AM, Taste-Of-IT <kontakt at taste-of-it.de> wrote:

> Hello,
>
> hope some expert can help. I have a 2 Brick 1 Volume Distributed GlusterFS
> in Version 3.7.6 on Debian. The volume is shared via nfs. If i copy via
> midnight commander large files (>30GB), i got following messages. I replace
> sata cable, checked memory but i didnt find an error. SMART Values on all
> disks seems ok. After 30-40 minutes i can copy again. Any Idea?
>
> Feb  3 12:46:31 gluster01 kernel: [11186.588367] [sched_delayed] sched: RT
> throttling activated
> Feb  3 12:56:09 gluster01 kernel: [11764.932749] glusterfsd      D
> ffff88040ca6d788     0  1150      1 0x00000000
> Feb  3 12:56:09 gluster01 kernel: [11764.932759]  ffff88040ca6d330
> 0000000000000082 0000000000012f00 ffff88040ad1bfd8
> Feb  3 12:56:09 gluster01 kernel: [11764.932767]  0000000000012f00
> ffff88040ca6d330 ffff88040ca6d330 ffff88040ad1be88
> Feb  3 12:56:09 gluster01 kernel: [11764.932773]  ffff88040e18d4b8
> ffff88040e18d4a0 ffffffff00000000 ffff88040e18d4a8
> Feb  3 12:56:09 gluster01 kernel: [11764.932780] Call Trace:
> Feb  3 12:56:09 gluster01 kernel: [11764.932796]  [<ffffffff81512cd5>] ?
> rwsem_down_write_failed+0x1d5/0x320
> Feb  3 12:56:09 gluster01 kernel: [11764.932807]  [<ffffffff812b7d13>] ?
> call_rwsem_down_write_failed+0x13/0x20
> Feb  3 12:56:09 gluster01 kernel: [11764.932816]  [<ffffffff812325b0>] ?
> proc_keys_show+0x3f0/0x3f0
> Feb  3 12:56:09 gluster01 kernel: [11764.932823]  [<ffffffff81512649>] ?
> down_write+0x29/0x40
> Feb  3 12:56:09 gluster01 kernel: [11764.932830]  [<ffffffff811592bc>] ?
> vm_mmap_pgoff+0x6c/0xc0
> Feb  3 12:56:09 gluster01 kernel: [11764.932838]  [<ffffffff8116ea4e>] ?
> SyS_mmap_pgoff+0x10e/0x250
> Feb  3 12:56:09 gluster01 kernel: [11764.932844]  [<ffffffff811a969a>] ?
> SyS_readv+0x6a/0xd0
> Feb  3 12:56:09 gluster01 kernel: [11764.932853]  [<ffffffff81513ccd>] ?
> system_call_fast_compare_end+0x10/0x15
> Feb  3 12:58:09 gluster01 kernel: [11884.979935] glusterfsd      D
> ffff88040ca6d788     0  1150      1 0x00000000
> Feb  3 12:58:09 gluster01 kernel: [11884.979945]  ffff88040ca6d330
> 0000000000000082 0000000000012f00 ffff88040ad1bfd8
> Feb  3 12:58:09 gluster01 kernel: [11884.979952]  0000000000012f00
> ffff88040ca6d330 ffff88040ca6d330 ffff88040ad1be88
> Feb  3 12:58:09 gluster01 kernel: [11884.979959]  ffff88040e18d4b8
> ffff88040e18d4a0 ffffffff00000000 ffff88040e18d4a8
> Feb  3 12:58:09 gluster01 kernel: [11884.979966] Call Trace:
> Feb  3 12:58:09 gluster01 kernel: [11884.979982]  [<ffffffff81512cd5>] ?
> rwsem_down_write_failed+0x1d5/0x320
> Feb  3 12:58:09 gluster01 kernel: [11884.979993]  [<ffffffff812b7d13>] ?
> call_rwsem_down_write_failed+0x13/0x20
> Feb  3 12:58:09 gluster01 kernel: [11884.980001]  [<ffffffff812325b0>] ?
> proc_keys_show+0x3f0/0x3f0
> Feb  3 12:58:09 gluster01 kernel: [11884.980008]  [<ffffffff81512649>] ?
> down_write+0x29/0x40
> Feb  3 12:58:09 gluster01 kernel: [11884.980015]  [<ffffffff811592bc>] ?
> vm_mmap_pgoff+0x6c/0xc0
> Feb  3 12:58:09 gluster01 kernel: [11884.980023]  [<ffffffff8116ea4e>] ?
> SyS_mmap_pgoff+0x10e/0x250
> Feb  3 12:58:09 gluster01 kernel: [11884.980030]  [<ffffffff811a969a>] ?
> SyS_readv+0x6a/0xd0
> Feb  3 12:58:09 gluster01 kernel: [11884.980038]  [<ffffffff81513ccd>] ?
> system_call_fast_compare_end+0x10/0x15
> Feb  3 12:58:09 gluster01 kernel: [11884.980351] mc              D
> ffff88040e6d8fb8     0  5119   1447 0x00000000
> Feb  3 12:58:09 gluster01 kernel: [11884.980358]  ffff88040e6d8b60
> 0000000000000082 0000000000012f00 ffff88040d5dbfd8
> Feb  3 12:58:09 gluster01 kernel: [11884.980365]  0000000000012f00
> ffff88040e6d8b60 ffff88041ec937b0 ffff88041efcc9e8
> Feb  3 12:58:09 gluster01 kernel: [11884.980371]  0000000000000002
> ffffffff8113ce00 ffff88040d5dbcb0 ffff88040d5dbd98
> Feb  3 12:58:09 gluster01 kernel: [11884.980377] Call Trace:
> Feb  3 12:58:09 gluster01 kernel: [11884.980385]  [<ffffffff8113ce00>] ?
> wait_on_page_read+0x60/0x60
> Feb  3 12:58:09 gluster01 kernel: [11884.980392]  [<ffffffff81510759>] ?
> io_schedule+0x99/0x120
> Feb  3 12:58:09 gluster01 kernel: [11884.980399]  [<ffffffff8113ce0a>] ?
> sleep_on_page+0xa/0x10
> Feb  3 12:58:09 gluster01 kernel: [11884.980405]  [<ffffffff81510adc>] ?
> __wait_on_bit+0x5c/0x90
> Feb  3 12:58:09 gluster01 kernel: [11884.980412]  [<ffffffff8113cbff>] ?
> wait_on_page_bit+0x7f/0x90
> Feb  3 12:58:09 gluster01 kernel: [11884.980420]  [<ffffffff810a7bd0>] ?
> autoremove_wake_function+0x30/0x30
> Feb  3 12:58:09 gluster01 kernel: [11884.980426]  [<ffffffff8114a17d>] ?
> pagevec_lookup_tag+0x1d/0x30
> Feb  3 12:58:09 gluster01 kernel: [11884.980433]  [<ffffffff8113cce0>] ?
> filemap_fdatawait_range+0xd0/0x160
> Feb  3 12:58:09 gluster01 kernel: [11884.980442]  [<ffffffff8113e7ca>] ?
> filemap_write_and_wait_range+0x3a/0x60
> Feb  3 12:58:09 gluster01 kernel: [11884.980461]  [<ffffffffa072363f>] ?
> nfs_file_fsync+0x7f/0x100 [nfs]
> Feb  3 12:58:09 gluster01 kernel: [11884.980476]  [<ffffffffa0723a2a>] ?
> nfs_file_write+0xda/0x1a0 [nfs]
> Feb  3 12:58:09 gluster01 kernel: [11884.980484]  [<ffffffff811a7e24>] ?
> new_sync_write+0x74/0xa0
> Feb  3 12:58:09 gluster01 kernel: [11884.980492]  [<ffffffff811a8562>] ?
> vfs_write+0xb2/0x1f0
> Feb  3 12:58:09 gluster01 kernel: [11884.980500]  [<ffffffff811a842d>] ?
> vfs_read+0xed/0x170
> Feb  3 12:58:09 gluster01 kernel: [11884.980505]  [<ffffffff811a90a2>] ?
> SyS_write+0x42/0xa0
> Feb  3 12:58:09 gluster01 kernel: [11884.980513]  [<ffffffff81513ccd>] ?
> system_call_fast_compare_end+0x10/0x15
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160203/04a2c2f8/attachment.html>


More information about the Gluster-users mailing list