[Bugs] [Bug 1422769] New: brick process crashes when glusterd is restarted

bugzilla at redhat.com bugzilla at redhat.com
Thu Feb 16 07:24:51 UTC 2017


https://bugzilla.redhat.com/show_bug.cgi?id=1422769

            Bug ID: 1422769
           Summary: brick process crashes when glusterd is restarted
           Product: GlusterFS
           Version: 3.10
         Component: glusterd
          Severity: high
          Assignee: bugs at gluster.org
          Reporter: kramdoss at redhat.com
                CC: bugs at gluster.org



Description of problem:
When glusterd was restarted with around 50 volumes in the cluster, glusterfsd
process crashed. 

Brick multiplexing was enabled on the cluster.

(gdb) bt
#0  0x00007fc1d59c97b0 in glusterfs_graph_attach (orig_graph=0x0,
path=<optimized out>) at graph.c:1085
#1  0x00007fc1d5e905da in glusterfs_handle_attach (req=0x7fc1c80034a0) at
glusterfsd-mgmt.c:842
#2  0x00007fc1d59ca6d0 in synctask_wrap (old_task=<optimized out>) at
syncop.c:375
#3  0x00007fc1d4088cf0 in ?? () from /lib64/libc.so.6
#4  0x0000000000000000 in ?? ()

gluster v status vol1
Status of volume: vol1
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick 10.70.47.29:/mnt/container_brick1/v1-
b1                                          49153     0          Y       15405
Brick 10.70.47.31:/mnt/container_brick1/v1-
b1                                          N/A       N/A        N       N/A  
Brick 10.70.46.128:/mnt/container_brick1/v1
-b1                                         49152     0          Y       336  
Brick 10.70.47.29:/mnt/container_brick1/v1-
b2                                          49153     0          Y       15405
Brick 10.70.47.31:/mnt/container_brick1/v1-
b2                                          N/A       N/A        N       11695
Brick 10.70.46.128:/mnt/container_brick1/v1
-b2                                         49152     0          Y       336  
Self-heal Daemon on localhost               N/A       N/A        Y       11704
Self-heal Daemon on 10.70.47.29             N/A       N/A        Y       18586
Self-heal Daemon on 10.70.46.128            N/A       N/A        Y       11382

Task Status of Volume vol1
------------------------------------------------------------------------------
There are no active volume tasks

 - glusterd was restarted on 10.70.47.31

Version-Release number of selected component (if applicable):
rpm -qa | grep 'gluster'
glusterfs-resource-agents-3.10.0rc-0.0.el7.centos.noarch
glusterfs-events-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-debuginfo-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-api-3.10.0rc-0.0.el7.centos.x86_64
python2-gluster-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-fuse-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-server-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-devel-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-api-devel-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-geo-replication-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-libs-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-client-xlators-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-extra-xlators-3.10.0rc-0.0.el7.centos.x86_64
glusterfs-cli-3.10.0rc-0.0.el7.centos.x86_64

How reproducible:
1/1

Steps to Reproduce:
1. set cluster.brick-multiplex  on
2. create 50 volumes (2x3 type volumes were created)
3. Mount all the volumes and run IOs
4. restart glusterd on one of the node

Actual results:
glusterfsd process crashed

Expected results:
No crashes and no other failures should be seen

Additional info:
logs shall be attached shortly

-- 
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.


More information about the Bugs mailing list