[Gluster-users] Upgrade 3.7.6 -> 3.7.13 one gluster server disconnected 1 of 3 volumes

Atin Mukherjee amukherj at redhat.com
Tue Aug 23 04:15:09 UTC 2016


On Tue, Aug 23, 2016 at 4:17 AM, Steve Dainard <sdainard at spd1.com> wrote:

> About 5 hours after upgrading gluster 3.7.6 -> 3.7.13 on Centos 7, one of
> my gluster servers disconnected its volume. The other two volumes this host
> serves were not affected.
>
> # gluster volume status storage
> Status of volume: storage
> Gluster process                             TCP Port  RDMA Port  Online
>  Pid
> ------------------------------------------------------------
> ------------------
> Brick 10.0.231.50:/mnt/raid6-storage/storage
>               49159     0          Y       30743
> Brick 10.0.231.51:/mnt/raid6-storage/storage
>               49159     0          Y       676
> *Brick 10.0.231.52:/mnt/raid6-storage/storag**e
>                 N/A       N/A        N       N/A  *
> Brick 10.0.231.53:/mnt/raid6-storage/storage
>               49154     0          Y       10253
> Brick 10.0.231.54:/mnt/raid6-storage/storage
>               49153     0          Y       2792
> Brick 10.0.231.55:/mnt/raid6-storage/storage
>               49153     0          Y       13590
> Brick 10.0.231.56:/mnt/raid6-storage/storage
>               49152     0          Y       9281
> NFS Server on localhost                     2049      0          Y
> 30775
> Quota Daemon on localhost                   N/A       N/A        Y
> 30781
> NFS Server on 10.0.231.54                   2049      0          Y
> 2817
> Quota Daemon on 10.0.231.54                 N/A       N/A        Y
> 2824
> NFS Server on 10.0.231.51                   2049      0          Y
> 710
> Quota Daemon on 10.0.231.51                 N/A       N/A        Y
> 719
> NFS Server on 10.0.231.52                   2049      0          Y
> 9090
> Quota Daemon on 10.0.231.52                 N/A       N/A        Y
> 9098
> NFS Server on 10.0.231.55                   2049      0          Y
> 13611
> Quota Daemon on 10.0.231.55                 N/A       N/A        Y
> 13619
> NFS Server on 10.0.231.56                   2049      0          Y
> 9303
> Quota Daemon on 10.0.231.56                 N/A       N/A        Y
> 9310
> NFS Server on 10.0.231.53                   2049      0          Y
> 26304
> Quota Daemon on 10.0.231.53                 N/A       N/A        Y
> 26320
>
> Task Status of Volume storage
> ------------------------------------------------------------
> ------------------
> There are no active volume tasks
>
> I see lots of logs related to trashcan (failed [file exists]), set xattrs
> (failed [no such file or directory]), quota (invalid arguments) in the
> brick logs, which I enabled as a feature after the upgrade this morning.
>

Could you let us know the time (in UTC) around which this issue was seen
such that we can look at the logs around that time and see if something
went wrong.


> After restarting glusterd on that host, the volume came back online.
>
> I've attached logs from that host if someone can take a look.
>
> # gluster volume info storage
>
> Volume Name: storage
> Type: Distribute
> Volume ID: 6f95525a-94d7-4174-bac4-e1a18fe010a2
> Status: Started
> Number of Bricks: 7
> Transport-type: tcp
> Bricks:
> Brick1: 10.0.231.50:/mnt/raid6-storage/storage
> Brick2: 10.0.231.51:/mnt/raid6-storage/storage
> Brick3: 10.0.231.52:/mnt/raid6-storage/storage
> Brick4: 10.0.231.53:/mnt/raid6-storage/storage
> Brick5: 10.0.231.54:/mnt/raid6-storage/storage
> Brick6: 10.0.231.55:/mnt/raid6-storage/storage
> Brick7: 10.0.231.56:/mnt/raid6-storage/storage
> Options Reconfigured:
> nfs.disable: no
> features.trash-max-filesize: 1GB
> features.trash: on
> features.quota-deem-statfs: on
> features.inode-quota: on
> features.quota: on
> performance.readdir-ahead: on
>
> # rpm -qa  |grep glusterfs
> glusterfs-fuse-3.7.13-1.el7.x86_64
> glusterfs-cli-3.7.13-1.el7.x86_64
> glusterfs-3.7.13-1.el7.x86_64
> glusterfs-server-3.7.13-1.el7.x86_64
> glusterfs-api-3.7.13-1.el7.x86_64
> glusterfs-libs-3.7.13-1.el7.x86_64
> glusterfs-client-xlators-3.7.13-1.el7.x86_64
>
>
> Thanks,
> Steve
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>



-- 

--Atin
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160823/d3d2eb63/attachment.html>


More information about the Gluster-users mailing list