[Gluster-users] 2 issues after upgrade 9.4 -> 10.1
Hu Bert
revirii at googlemail.com
Tue Mar 1 05:19:44 UTC 2022
Good morning,
just did an upgrade of 3 gluster volumes and x clients from 9.4 to
10.1. In principle the upgrade went fine, just 2 smaller issues
appeared.
1) on one of the servers the ports are screwed up.
gluster volume status
Status of volume: workdata
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick glusterpub1:/gluster/md3/workdata 49152 0 Y 1452
Brick glusterpub2:/gluster/md3/workdata 49152 0 Y 1839
Brick glusterpub3:/gluster/md3/workdata 54105 0 Y 1974
Brick glusterpub1:/gluster/md4/workdata 49153 0 Y 1459
Brick glusterpub2:/gluster/md4/workdata 49153 0 Y 1849
Brick glusterpub3:/gluster/md4/workdata 58177 0 Y 1997
Brick glusterpub1:/gluster/md5/workdata 49154 0 Y 1468
Brick glusterpub2:/gluster/md5/workdata 49154 0 Y 1857
Brick glusterpub3:/gluster/md5/workdata 59071 0 Y 2003
Brick glusterpub1:/gluster/md6/workdata 49155 0 Y 1481
Brick glusterpub2:/gluster/md6/workdata 49155 0 Y 1868
Brick glusterpub3:/gluster/md6/workdata 53309 0 Y 2008
Brick glusterpub1:/gluster/md7/workdata 49156 0 Y 1490
Brick glusterpub2:/gluster/md7/workdata 49156 0 Y 1878
Brick glusterpub3:/gluster/md7/workdata 54310 0 Y 2027
Self-heal Daemon on localhost N/A N/A Y 2108
Self-heal Daemon on glusterpub1 N/A N/A Y 1210749
Self-heal Daemon on glusterpub2 N/A N/A Y 950871
Task Status of Volume workdata
------------------------------------------------------------------------------
There are no active volume tasks
glusterpub3 has different ports. I know, this is no problem, the
volume is good, but even after a reboot the ports stay like this.
glustershd.log:
[2022-03-01 04:58:13.993349 +0000] I
[rpc-clnt.c:1969:rpc_clnt_reconfig] 0-workdata-client-0: changing port
to 49152 (from 0)
[2022-03-01 04:58:13.993410 +0000] I [socket.c:834:__socket_shutdown]
0-workdata-client-0: intentional socket shutdown(13)
[............]
[2022-03-01 04:58:14.008111 +0000] I
[rpc-clnt.c:1969:rpc_clnt_reconfig] 0-workdata-client-1: changing port
to 49152 (from 0)
[2022-03-01 04:58:14.008148 +0000] I [socket.c:834:__socket_shutdown]
0-workdata-client-1: intentional socket shutdown(14)
[............]
[2022-03-01 04:58:14.011416 +0000] I
[rpc-clnt.c:1969:rpc_clnt_reconfig] 0-workdata-client-2: changing port
to 54105 (from 0)
[2022-03-01 04:58:14.011469 +0000] I [socket.c:834:__socket_shutdown]
0-workdata-client-2: intentional socket shutdown(13)
same for the other 4 bricks. Probably some more related message,
unsure which ones to c+p. And some error messages like these (appear
on all servers):
[2022-03-01 04:58:14.012523 +0000] E
[rpc-clnt.c:331:saved_frames_unwind] (-->
/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x195)[0x7f4cec48c2a5]
(--> /lib/x86_64-linux-gnu/libgfrpc.so.0(+0x729c)[0x7f4cec42529c] (-->
/lib/
x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x10f)[0x7f4cec42d20f]
(--> /lib/x86_64-linux-gnu/libgfrpc.so.0(+0x10118)[0x7f4cec42e118]
(--> /lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_transport_notify+0x26)[0x7f4cec429646]
)))
)) 0-workdata-client-5: forced unwinding frame type(GF-DUMP)
op(DUMP(1)) called at 2022-03-01 04:58:14.011943 +0000 (xid=0x5)
very strange.
2) when mounting on the clients (after upgrade):
/sbin/mount.glusterfs: 90: [: glusterpub2 glusterpub3 SyntaxOK:
unexpected operator
/sbin/mount.glusterfs: 366: [: SyntaxOK: unexpected operator
Syntax ok, but unexpected operator? Has the mount syntax changed?
glusterpub1:/workdata /data/repository/shared/public glusterfs
defaults,_netdev,attribute-timeout=0,entry-timeout=0,backup-volfile-servers=glusterpub2:glusterpub3
0 0
thx,
Hubert
More information about the Gluster-users
mailing list