[Gluster-users] 2 issues after upgrade 9.4 -> 10.1

Nikhil Ladha nladha at redhat.com
Tue Mar 1 10:40:35 UTC 2022


Hi Hu Bert,

Do you have a distributed-replicate volume and you followed an online
upgrade procedure? If so, then that is the reason that the ports are
different on only 1 server as you mentioned you did a reboot on it.
Secondly, the glusterfs mount issue is already fixed (check patch #3211
<https://github.com/gluster/glusterfs/pull/3211>) and will be available in
the next release.

--
Thanks and Regards,
*NiKHIL LADHA*


On Tue, Mar 1, 2022 at 3:10 PM Hu Bert <revirii at googlemail.com> wrote:

> Hey,
>
> ok, i think i found the reason for the port issue:
>
> https://docs.gluster.org/en/latest/release-notes/10.0/
> https://github.com/gluster/glusterfs/issues/786
>
> Should've look closer... mea culpa. But, quite interesting, that
> happened only on one server, i upgraded 6 servers in total.
>
> So only the issue with the glusterfs mount and backup-volfile-servers
> remains.
>
>
> Thx,
> Hubert
>
> Am Di., 1. März 2022 um 06:19 Uhr schrieb Hu Bert <revirii at googlemail.com
> >:
> >
> > Good morning,
> >
> > just did an upgrade of 3 gluster volumes and x clients from 9.4 to
> > 10.1. In principle the upgrade went fine, just 2 smaller issues
> > appeared.
> >
> > 1) on one of the servers the ports are screwed up.
> >
> > gluster volume status
> > Status of volume: workdata
> > Gluster process                             TCP Port  RDMA Port  Online
> Pid
> >
> ------------------------------------------------------------------------------
> > Brick glusterpub1:/gluster/md3/workdata     49152     0          Y
>  1452
> > Brick glusterpub2:/gluster/md3/workdata     49152     0          Y
>  1839
> > Brick glusterpub3:/gluster/md3/workdata     54105     0          Y
>  1974
> > Brick glusterpub1:/gluster/md4/workdata     49153     0          Y
>  1459
> > Brick glusterpub2:/gluster/md4/workdata     49153     0          Y
>  1849
> > Brick glusterpub3:/gluster/md4/workdata     58177     0          Y
>  1997
> > Brick glusterpub1:/gluster/md5/workdata     49154     0          Y
>  1468
> > Brick glusterpub2:/gluster/md5/workdata     49154     0          Y
>  1857
> > Brick glusterpub3:/gluster/md5/workdata     59071     0          Y
>  2003
> > Brick glusterpub1:/gluster/md6/workdata     49155     0          Y
>  1481
> > Brick glusterpub2:/gluster/md6/workdata     49155     0          Y
>  1868
> > Brick glusterpub3:/gluster/md6/workdata     53309     0          Y
>  2008
> > Brick glusterpub1:/gluster/md7/workdata     49156     0          Y
>  1490
> > Brick glusterpub2:/gluster/md7/workdata     49156     0          Y
>  1878
> > Brick glusterpub3:/gluster/md7/workdata     54310     0          Y
>  2027
> > Self-heal Daemon on localhost               N/A       N/A        Y
>  2108
> > Self-heal Daemon on glusterpub1             N/A       N/A        Y
>  1210749
> > Self-heal Daemon on glusterpub2             N/A       N/A        Y
>  950871
> >
> > Task Status of Volume workdata
> >
> ------------------------------------------------------------------------------
> > There are no active volume tasks
> >
> > glusterpub3 has different ports. I know, this is no problem, the
> > volume is good, but even after a reboot the ports stay like this.
> >
> > glustershd.log:
> > [2022-03-01 04:58:13.993349 +0000] I
> > [rpc-clnt.c:1969:rpc_clnt_reconfig] 0-workdata-client-0: changing port
> > to 49152 (from 0)
> > [2022-03-01 04:58:13.993410 +0000] I [socket.c:834:__socket_shutdown]
> > 0-workdata-client-0: intentional socket shutdown(13)
> > [............]
> > [2022-03-01 04:58:14.008111 +0000] I
> > [rpc-clnt.c:1969:rpc_clnt_reconfig] 0-workdata-client-1: changing port
> > to 49152 (from 0)
> > [2022-03-01 04:58:14.008148 +0000] I [socket.c:834:__socket_shutdown]
> > 0-workdata-client-1: intentional socket shutdown(14)
> > [............]
> > [2022-03-01 04:58:14.011416 +0000] I
> > [rpc-clnt.c:1969:rpc_clnt_reconfig] 0-workdata-client-2: changing port
> > to 54105 (from 0)
> > [2022-03-01 04:58:14.011469 +0000] I [socket.c:834:__socket_shutdown]
> > 0-workdata-client-2: intentional socket shutdown(13)
> >
> > same for the other 4 bricks. Probably some more related message,
> > unsure which ones to c+p. And some error messages like these (appear
> > on all servers):
> >
> > [2022-03-01 04:58:14.012523 +0000] E
> > [rpc-clnt.c:331:saved_frames_unwind] (-->
> >
> /lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x195)[0x7f4cec48c2a5]
> > (--> /lib/x86_64-linux-gnu/libgfrpc.so.0(+0x729c)[0x7f4cec42529c] (-->
> > /lib/
> >
> x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x10f)[0x7f4cec42d20f]
> > (--> /lib/x86_64-linux-gnu/libgfrpc.so.0(+0x10118)[0x7f4cec42e118]
> > (-->
> /lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_transport_notify+0x26)[0x7f4cec429646]
> > )))
> > )) 0-workdata-client-5: forced unwinding frame type(GF-DUMP)
> > op(DUMP(1)) called at 2022-03-01 04:58:14.011943 +0000 (xid=0x5)
> >
> > very strange.
> >
> > 2) when mounting on the clients (after upgrade):
> >
> > /sbin/mount.glusterfs: 90: [: glusterpub2 glusterpub3 SyntaxOK:
> > unexpected operator
> > /sbin/mount.glusterfs: 366: [: SyntaxOK: unexpected operator
> >
> > Syntax ok, but unexpected operator? Has the mount syntax changed?
> >
> > glusterpub1:/workdata /data/repository/shared/public glusterfs
> >
> defaults,_netdev,attribute-timeout=0,entry-timeout=0,backup-volfile-servers=glusterpub2:glusterpub3
> > 0 0
> >
> >
> > thx,
> > Hubert
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://meet.google.com/cpu-eiue-hvk
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20220301/4e326f6b/attachment.html>


More information about the Gluster-users mailing list