[Gluster-users] Failure while upgrading gluster to 3.10.1

Atin Mukherjee amukherj at redhat.com
Thu May 25 16:24:17 UTC 2017


On Thu, 25 May 2017 at 19:11, Pawan Alwandi <pawan at platform.sh> wrote:

> Hello Atin,
>
> Yes, glusterd on other instances are up and running.  Below is the
> requested output on all the three hosts.
>
> Host 1
>
> # gluster peer status
> Number of Peers: 2
>
> Hostname: 192.168.0.7
> Uuid: 5ec54b4f-f60c-48c6-9e55-95f2bb58f633
> State: Peer in Cluster (Disconnected)
>

Glusterd is disconnected here.

>
>
> Hostname: 192.168.0.6
> Uuid: 83e9a0b9-6bd5-483b-8516-d8928805ed95
> State: Peer in Cluster (Disconnected)
>

Same as above

Can you please check what does glusterd log have to say here about these
disconnects?


>
> # gluster volume status
> Status of volume: shared
> Gluster process                             TCP Port  RDMA Port  Online
> Pid
>
> ------------------------------------------------------------------------------
> Brick 192.168.0.5:/data/exports/shared      49152     0          Y
> 2105
> NFS Server on localhost                     2049      0          Y
> 2089
> Self-heal Daemon on localhost               N/A       N/A        Y
> 2097
>

Volume status output does show all the bricks are up. So I'm not sure why
are you seeing the volume as read only. Can you please provide the mount
log?


>
> Task Status of Volume shared
>
> ------------------------------------------------------------------------------
> There are no active volume tasks
>
> Host 2
>
> # gluster peer status
> Number of Peers: 2
>
> Hostname: 192.168.0.7
> Uuid: 5ec54b4f-f60c-48c6-9e55-95f2bb58f633
> State: Peer in Cluster (Connected)
>
> Hostname: 192.168.0.5
> Uuid: 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073
> State: Peer in Cluster (Connected)
>
>
> # gluster volume status
> Status of volume: shared
> Gluster process                        Port    Online    Pid
>
> ------------------------------------------------------------------------------
> Brick 192.168.0.5:/data/exports/shared            49152    Y    2105
> Brick 192.168.0.6:/data/exports/shared            49152    Y    2188
> Brick 192.168.0.7:/data/exports/shared            49152    Y    2453
> NFS Server on localhost                    2049    Y    2194
> Self-heal Daemon on localhost                N/A    Y    2199
> NFS Server on 192.168.0.5                2049    Y    2089
> Self-heal Daemon on 192.168.0.5                N/A    Y    2097
> NFS Server on 192.168.0.7                2049    Y    2458
> Self-heal Daemon on 192.168.0.7                N/A    Y    2463
>
> Task Status of Volume shared
>
> ------------------------------------------------------------------------------
> There are no active volume tasks
>
> Host 3
>
> # gluster peer status
> Number of Peers: 2
>
> Hostname: 192.168.0.5
> Uuid: 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073
> State: Peer in Cluster (Connected)
>
> Hostname: 192.168.0.6
> Uuid: 83e9a0b9-6bd5-483b-8516-d8928805ed95
> State: Peer in Cluster (Connected)
>
> # gluster volume status
> Status of volume: shared
> Gluster process                        Port    Online    Pid
>
> ------------------------------------------------------------------------------
> Brick 192.168.0.5:/data/exports/shared            49152    Y    2105
> Brick 192.168.0.6:/data/exports/shared            49152    Y    2188
> Brick 192.168.0.7:/data/exports/shared            49152    Y    2453
> NFS Server on localhost                    2049    Y    2458
> Self-heal Daemon on localhost                N/A    Y    2463
> NFS Server on 192.168.0.6                2049    Y    2194
> Self-heal Daemon on 192.168.0.6                N/A    Y    2199
> NFS Server on 192.168.0.5                2049    Y    2089
> Self-heal Daemon on 192.168.0.5                N/A    Y    2097
>
> Task Status of Volume shared
>
> ------------------------------------------------------------------------------
> There are no active volume tasks
>
>
>
>
>
>
> On Wed, May 24, 2017 at 8:32 PM, Atin Mukherjee <amukherj at redhat.com>
> wrote:
>
>> Are the other glusterd instances are up? output of gluster peer status &
>> gluster volume status please?
>>
>> On Wed, May 24, 2017 at 4:20 PM, Pawan Alwandi <pawan at platform.sh> wrote:
>>
>>> Thanks Atin,
>>>
>>> So I got gluster downgraded to 3.7.9 on host 1 and now have the
>>> glusterfs and glusterfsd processes come up.  But I see the volume is
>>> mounted read only.
>>>
>>> I see these being logged every 3s:
>>>
>>> [2017-05-24 10:45:44.440435] W [socket.c:852:__socket_keepalive]
>>> 0-socket: failed to set keep idle -1 on socket 17, Invalid argument
>>> [2017-05-24 10:45:44.440475] E [socket.c:2966:socket_connect]
>>> 0-management: Failed to set keep-alive: Invalid argument
>>> [2017-05-24 10:45:44.440734] W [socket.c:852:__socket_keepalive]
>>> 0-socket: failed to set keep idle -1 on socket 20, Invalid argument
>>> [2017-05-24 10:45:44.440754] E [socket.c:2966:socket_connect]
>>> 0-management: Failed to set keep-alive: Invalid argument
>>> [2017-05-24 10:45:44.441354] E [rpc-clnt.c:362:saved_frames_unwind] (-->
>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x1a3)[0x7f767c46d483]
>>> (-->
>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_unwind+0x1cf)[0x7f767c2383af]
>>> (-->
>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f767c2384ce]
>>> (-->
>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e)[0x7f767c239c8e]
>>> (-->
>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x88)[0x7f767c23a4a8]
>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1))
>>> called at 2017-05-24 10:45:44.440945 (xid=0xbf)
>>> [2017-05-24 10:45:44.441505] W
>>> [glusterd-locks.c:681:glusterd_mgmt_v3_unlock]
>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.9/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4b)
>>> [0x7f767734dffb]
>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.9/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x14a)
>>> [0x7f7677357c6a]
>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.9/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x4c3)
>>> [0x7f76773f0ef3] ) 0-management: Lock for vol shared not held
>>> [2017-05-24 10:45:44.441660] E [rpc-clnt.c:362:saved_frames_unwind] (-->
>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x1a3)[0x7f767c46d483]
>>> (-->
>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_unwind+0x1cf)[0x7f767c2383af]
>>> (-->
>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f767c2384ce]
>>> (-->
>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e)[0x7f767c239c8e]
>>> (-->
>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x88)[0x7f767c23a4a8]
>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1))
>>> called at 2017-05-24 10:45:44.441086 (xid=0xbf)
>>> [2017-05-24 10:45:44.441790] W
>>> [glusterd-locks.c:681:glusterd_mgmt_v3_unlock]
>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.9/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4b)
>>> [0x7f767734dffb]
>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.9/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x14a)
>>> [0x7f7677357c6a]
>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.9/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x4c3)
>>> [0x7f76773f0ef3] ) 0-management: Lock for vol shared not held
>>>
>>> The heal info says this:
>>>
>>> # gluster volume heal shared info
>>> Brick 192.168.0.5:/data/exports/shared
>>> Number of entries: 0
>>>
>>> Brick 192.168.0.6:/data/exports/shared
>>> Status: Transport endpoint is not connected
>>>
>>> Brick 192.168.0.7:/data/exports/shared
>>> Status: Transport endpoint is not connected
>>>
>>> Any idea whats up here?
>>>
>>> Pawan
>>>
>>> On Mon, May 22, 2017 at 9:42 PM, Atin Mukherjee <amukherj at redhat.com>
>>> wrote:
>>>
>>>>
>>>>
>>>> On Mon, May 22, 2017 at 9:05 PM, Pawan Alwandi <pawan at platform.sh>
>>>> wrote:
>>>>
>>>>>
>>>>> On Mon, May 22, 2017 at 8:36 PM, Atin Mukherjee <amukherj at redhat.com>
>>>>> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Mon, May 22, 2017 at 7:51 PM, Atin Mukherjee <amukherj at redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Sorry Pawan, I did miss the other part of the attachments. So
>>>>>>> looking from the glusterd.info file from all the hosts, it looks
>>>>>>> like host2 and host3 do not have the correct op-version. Can you please set
>>>>>>> the op-version as "operating-version=30702" in host2 and host3 and restart
>>>>>>> glusterd instance one by one on all the nodes?
>>>>>>>
>>>>>>
>>>>>> Please ensure that all the hosts are upgraded to the same bits before
>>>>>> doing this change.
>>>>>>
>>>>>
>>>>> Having to upgrade all 3 hosts to newer version before gluster could
>>>>> work successfully on any of them means application downtime.  The
>>>>> applications running on these hosts are expected to be highly available.
>>>>> So with the way the things are right now, is an online upgrade possible?
>>>>> My upgrade steps are: (1) stop the applications (2) umount the gluster
>>>>> volume, and then (3) upgrade gluster one host at a time.
>>>>>
>>>>
>>>> One of the way to mitigate this is to first do an online upgrade to
>>>> glusterfs-3.7.9 (op-version:30707) given this bug was introduced in 3.7.10
>>>> and then come to 3.11.
>>>>
>>>>
>>>>> Our goal is to get gluster upgraded to 3.11 from 3.6.9, and to make
>>>>> this an online upgrade we are okay to take two steps 3.6.9 -> 3.7 and then
>>>>> 3.7 to 3.11.
>>>>>
>>>>>
>>>>>>
>>>>>>
>>>>>>>
>>>>>>> Apparently it looks like there is a bug which you have uncovered,
>>>>>>> during peer handshaking if one of the glusterd instance is running with old
>>>>>>> bits then during validating the handshake request there is a possibility
>>>>>>> that uuid received will be blank and the same was ignored however there was
>>>>>>> a patch http://review.gluster.org/13519 which had some additional
>>>>>>> changes which was always looking at this field and doing some extra checks
>>>>>>> which was causing the handshake to fail. For now, the above workaround
>>>>>>> should suffice. I'll be sending a patch pretty soon.
>>>>>>>
>>>>>>
>>>>>> Posted a patch https://review.gluster.org/#/c/17358 .
>>>>>>
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Mon, May 22, 2017 at 11:35 AM, Pawan Alwandi <pawan at platform.sh>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hello Atin,
>>>>>>>>
>>>>>>>> The tar's have the content of `/var/lib/glusterd` too for all 3
>>>>>>>> nodes, please check again.
>>>>>>>>
>>>>>>>> Thanks
>>>>>>>>
>>>>>>>> On Mon, May 22, 2017 at 11:32 AM, Atin Mukherjee <
>>>>>>>> amukherj at redhat.com> wrote:
>>>>>>>>
>>>>>>>>> Pawan,
>>>>>>>>>
>>>>>>>>> I see you have provided the log files from the nodes, however it'd
>>>>>>>>> be really helpful if you can provide me the content of /var/lib/glusterd
>>>>>>>>> from all the nodes to get to the root cause of this issue.
>>>>>>>>>
>>>>>>>>> On Fri, May 19, 2017 at 12:09 PM, Pawan Alwandi <pawan at platform.sh
>>>>>>>>> > wrote:
>>>>>>>>>
>>>>>>>>>> Hello Atin,
>>>>>>>>>>
>>>>>>>>>> Thanks for continued support.  I've attached requested files from
>>>>>>>>>> all 3 nodes.
>>>>>>>>>>
>>>>>>>>>> (I think we already verified the UUIDs to be correct, anyway let
>>>>>>>>>> us know if you find any more info in the logs)
>>>>>>>>>>
>>>>>>>>>> Pawan
>>>>>>>>>>
>>>>>>>>>> On Thu, May 18, 2017 at 11:45 PM, Atin Mukherjee <
>>>>>>>>>> amukherj at redhat.com> wrote:
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Thu, 18 May 2017 at 23:40, Atin Mukherjee <
>>>>>>>>>>> amukherj at redhat.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> On Wed, 17 May 2017 at 12:47, Pawan Alwandi <pawan at platform.sh>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hello Atin,
>>>>>>>>>>>>>
>>>>>>>>>>>>> I realized that these
>>>>>>>>>>>>> http://gluster.readthedocs.io/en/latest/Upgrade-Guide/upgrade_to_3.10/
>>>>>>>>>>>>> instructions only work for upgrades from 3.7, while we are running 3.6.2.
>>>>>>>>>>>>> Are there instructions/suggestion you have for us to upgrade from 3.6
>>>>>>>>>>>>> version?
>>>>>>>>>>>>>
>>>>>>>>>>>>> I believe upgrade from 3.6 to 3.7 and then to 3.10 would work,
>>>>>>>>>>>>> but I see similar errors reported when I upgraded to 3.7 too.
>>>>>>>>>>>>>
>>>>>>>>>>>>> For what its worth, I was able to set the op-version (gluster
>>>>>>>>>>>>> v set all cluster.op-version 30702) but that doesn't seem to help.
>>>>>>>>>>>>>
>>>>>>>>>>>>> [2017-05-17 06:48:33.700014] I [MSGID: 100030]
>>>>>>>>>>>>> [glusterfsd.c:2338:main] 0-/usr/sbin/glusterd: Started running
>>>>>>>>>>>>> /usr/sbin/glusterd version 3.7.20 (args: /usr/sbin/glusterd -p
>>>>>>>>>>>>> /var/run/glusterd.pid)
>>>>>>>>>>>>> [2017-05-17 06:48:33.703808] I [MSGID: 106478]
>>>>>>>>>>>>> [glusterd.c:1383:init] 0-management: Maximum allowed open file descriptors
>>>>>>>>>>>>> set to 65536
>>>>>>>>>>>>> [2017-05-17 06:48:33.703836] I [MSGID: 106479]
>>>>>>>>>>>>> [glusterd.c:1432:init] 0-management: Using /var/lib/glusterd as working
>>>>>>>>>>>>> directory
>>>>>>>>>>>>> [2017-05-17 06:48:33.708866] W [MSGID: 103071]
>>>>>>>>>>>>> [rdma.c:4594:__gf_rdma_ctx_create] 0-rpc-transport/rdma: rdma_cm event
>>>>>>>>>>>>> channel creation failed [No such device]
>>>>>>>>>>>>> [2017-05-17 06:48:33.709011] W [MSGID: 103055]
>>>>>>>>>>>>> [rdma.c:4901:init] 0-rdma.management: Failed to initialize IB Device
>>>>>>>>>>>>> [2017-05-17 06:48:33.709033] W
>>>>>>>>>>>>> [rpc-transport.c:359:rpc_transport_load] 0-rpc-transport: 'rdma'
>>>>>>>>>>>>> initialization failed
>>>>>>>>>>>>> [2017-05-17 06:48:33.709088] W
>>>>>>>>>>>>> [rpcsvc.c:1642:rpcsvc_create_listener] 0-rpc-service: cannot create
>>>>>>>>>>>>> listener, initing the transport failed
>>>>>>>>>>>>> [2017-05-17 06:48:33.709105] E [MSGID: 106243]
>>>>>>>>>>>>> [glusterd.c:1656:init] 0-management: creation of 1 listeners failed,
>>>>>>>>>>>>> continuing with succeeded transport
>>>>>>>>>>>>> [2017-05-17 06:48:35.480043] I [MSGID: 106513]
>>>>>>>>>>>>> [glusterd-store.c:2068:glusterd_restore_op_version] 0-glusterd: retrieved
>>>>>>>>>>>>> op-version: 30600
>>>>>>>>>>>>> [2017-05-17 06:48:35.605779] I [MSGID: 106498]
>>>>>>>>>>>>> [glusterd-handler.c:3640:glusterd_friend_add_from_peerinfo] 0-management:
>>>>>>>>>>>>> connect returned 0
>>>>>>>>>>>>> [2017-05-17 06:48:35.607059] I
>>>>>>>>>>>>> [rpc-clnt.c:1046:rpc_clnt_connection_init] 0-management: setting
>>>>>>>>>>>>> frame-timeout to 600
>>>>>>>>>>>>> [2017-05-17 06:48:35.607670] I
>>>>>>>>>>>>> [rpc-clnt.c:1046:rpc_clnt_connection_init] 0-management: setting
>>>>>>>>>>>>> frame-timeout to 600
>>>>>>>>>>>>> [2017-05-17 06:48:35.607025] I [MSGID: 106498]
>>>>>>>>>>>>> [glusterd-handler.c:3640:glusterd_friend_add_from_peerinfo] 0-management:
>>>>>>>>>>>>> connect returned 0
>>>>>>>>>>>>> [2017-05-17 06:48:35.608125] I [MSGID: 106544]
>>>>>>>>>>>>> [glusterd.c:159:glusterd_uuid_init] 0-management: retrieved UUID:
>>>>>>>>>>>>> 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>> Final graph:
>>>>>>>>>>>>>
>>>>>>>>>>>>> +------------------------------------------------------------------------------+
>>>>>>>>>>>>>   1: volume management
>>>>>>>>>>>>>   2:     type mgmt/glusterd
>>>>>>>>>>>>>   3:     option rpc-auth.auth-glusterfs on
>>>>>>>>>>>>>   4:     option rpc-auth.auth-unix on
>>>>>>>>>>>>>   5:     option rpc-auth.auth-null on
>>>>>>>>>>>>>   6:     option rpc-auth-allow-insecure on
>>>>>>>>>>>>>   7:     option transport.socket.listen-backlog 128
>>>>>>>>>>>>>   8:     option event-threads 1
>>>>>>>>>>>>>   9:     option ping-timeout 0
>>>>>>>>>>>>>  10:     option transport.socket.read-fail-log off
>>>>>>>>>>>>>  11:     option transport.socket.keepalive-interval 2
>>>>>>>>>>>>>  12:     option transport.socket.keepalive-time 10
>>>>>>>>>>>>>  13:     option transport-type rdma
>>>>>>>>>>>>>  14:     option working-directory /var/lib/glusterd
>>>>>>>>>>>>>  15: end-volume
>>>>>>>>>>>>>  16:
>>>>>>>>>>>>>
>>>>>>>>>>>>> +------------------------------------------------------------------------------+
>>>>>>>>>>>>> [2017-05-17 06:48:35.609868] I [MSGID: 101190]
>>>>>>>>>>>>> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread
>>>>>>>>>>>>> with index 1
>>>>>>>>>>>>> [2017-05-17 06:48:35.610839] W [socket.c:596:__socket_rwv]
>>>>>>>>>>>>> 0-management: readv on 192.168.0.7:24007 failed (No data
>>>>>>>>>>>>> available)
>>>>>>>>>>>>> [2017-05-17 06:48:35.611907] E
>>>>>>>>>>>>> [rpc-clnt.c:370:saved_frames_unwind] (-->
>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x1a3)[0x7fd6c2d70bb3]
>>>>>>>>>>>>> (-->
>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_unwind+0x1cf)[0x7fd6c2b3a2df]
>>>>>>>>>>>>> (-->
>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7fd6c2b3a3fe]
>>>>>>>>>>>>> (-->
>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x89)[0x7fd6c2b3ba39]
>>>>>>>>>>>>> (-->
>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x160)[0x7fd6c2b3c380]
>>>>>>>>>>>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1))
>>>>>>>>>>>>> called at 2017-05-17 06:48:35.609965 (xid=0x1)
>>>>>>>>>>>>> [2017-05-17 06:48:35.611928] E [MSGID: 106167]
>>>>>>>>>>>>> [glusterd-handshake.c:2091:__glusterd_peer_dump_version_cbk] 0-management:
>>>>>>>>>>>>> Error through RPC layer, retry again later
>>>>>>>>>>>>> [2017-05-17 06:48:35.611944] I [MSGID: 106004]
>>>>>>>>>>>>> [glusterd-handler.c:5201:__glusterd_peer_rpc_notify] 0-management: Peer
>>>>>>>>>>>>> <192.168.0.7> (<5ec54b4f-f60c-48c6-9e55-95f2bb58f633>), in state <Peer in
>>>>>>>>>>>>> Cluster>, has disconnected from glusterd.
>>>>>>>>>>>>> [2017-05-17 06:48:35.612024] W
>>>>>>>>>>>>> [glusterd-locks.c:681:glusterd_mgmt_v3_unlock]
>>>>>>>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4b)
>>>>>>>>>>>>> [0x7fd6bdc4912b]
>>>>>>>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x160)
>>>>>>>>>>>>> [0x7fd6bdc52dd0]
>>>>>>>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x4c3)
>>>>>>>>>>>>> [0x7fd6bdcef1b3] ) 0-management: Lock for vol shared not held
>>>>>>>>>>>>> [2017-05-17 06:48:35.612039] W [MSGID: 106118]
>>>>>>>>>>>>> [glusterd-handler.c:5223:__glusterd_peer_rpc_notify] 0-management: Lock not
>>>>>>>>>>>>> released for shared
>>>>>>>>>>>>> [2017-05-17 06:48:35.612079] W [socket.c:596:__socket_rwv]
>>>>>>>>>>>>> 0-management: readv on 192.168.0.6:24007 failed (No data
>>>>>>>>>>>>> available)
>>>>>>>>>>>>> [2017-05-17 06:48:35.612179] E
>>>>>>>>>>>>> [rpc-clnt.c:370:saved_frames_unwind] (-->
>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x1a3)[0x7fd6c2d70bb3]
>>>>>>>>>>>>> (-->
>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_unwind+0x1cf)[0x7fd6c2b3a2df]
>>>>>>>>>>>>> (-->
>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7fd6c2b3a3fe]
>>>>>>>>>>>>> (-->
>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x89)[0x7fd6c2b3ba39]
>>>>>>>>>>>>> (-->
>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x160)[0x7fd6c2b3c380]
>>>>>>>>>>>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1))
>>>>>>>>>>>>> called at 2017-05-17 06:48:35.610007 (xid=0x1)
>>>>>>>>>>>>> [2017-05-17 06:48:35.612197] E [MSGID: 106167]
>>>>>>>>>>>>> [glusterd-handshake.c:2091:__glusterd_peer_dump_version_cbk] 0-management:
>>>>>>>>>>>>> Error through RPC layer, retry again later
>>>>>>>>>>>>> [2017-05-17 06:48:35.612211] I [MSGID: 106004]
>>>>>>>>>>>>> [glusterd-handler.c:5201:__glusterd_peer_rpc_notify] 0-management: Peer
>>>>>>>>>>>>> <192.168.0.6> (<83e9a0b9-6bd5-483b-8516-d8928805ed95>), in state <Peer in
>>>>>>>>>>>>> Cluster>, has disconnected from glusterd.
>>>>>>>>>>>>> [2017-05-17 06:48:35.612292] W
>>>>>>>>>>>>> [glusterd-locks.c:681:glusterd_mgmt_v3_unlock]
>>>>>>>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4b)
>>>>>>>>>>>>> [0x7fd6bdc4912b]
>>>>>>>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x160)
>>>>>>>>>>>>> [0x7fd6bdc52dd0]
>>>>>>>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x4c3)
>>>>>>>>>>>>> [0x7fd6bdcef1b3] ) 0-management: Lock for vol shared not held
>>>>>>>>>>>>> [2017-05-17 06:48:35.613432] W [MSGID: 106118]
>>>>>>>>>>>>> [glusterd-handler.c:5223:__glusterd_peer_rpc_notify] 0-management: Lock not
>>>>>>>>>>>>> released for shared
>>>>>>>>>>>>> [2017-05-17 06:48:35.614317] E [MSGID: 106170]
>>>>>>>>>>>>> [glusterd-handshake.c:1051:gd_validate_mgmt_hndsk_req] 0-management:
>>>>>>>>>>>>> Request from peer 192.168.0.6:991 has an entry in peerinfo,
>>>>>>>>>>>>> but uuid does not match
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Apologies for delay. My initial suspect was correct. You have
>>>>>>>>>>>> an incorrect UUID in the peer file which is causing this. Can you please
>>>>>>>>>>>> provide me the
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Clicked the send button accidentally!
>>>>>>>>>>>
>>>>>>>>>>> Can you please send me the content of /var/lib/glusterd &
>>>>>>>>>>> glusterd log from all the nodes?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Mon, May 15, 2017 at 10:31 PM, Atin Mukherjee <
>>>>>>>>>>>>> amukherj at redhat.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Mon, 15 May 2017 at 11:58, Pawan Alwandi <pawan at platform.sh>
>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Hi Atin,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I see below error.  Do I require gluster to be upgraded on
>>>>>>>>>>>>>>> all 3 hosts for this to work?  Right now I have host 1 running 3.10.1 and
>>>>>>>>>>>>>>> host 2 & 3 running 3.6.2
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> # gluster v set all cluster.op-version 31001
>>>>>>>>>>>>>>> volume set: failed: Required op_version (31001) is not
>>>>>>>>>>>>>>> supported
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Yes you should given 3.6 version is EOLed.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Mon, May 15, 2017 at 3:32 AM, Atin Mukherjee <
>>>>>>>>>>>>>>> amukherj at redhat.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Sun, 14 May 2017 at 21:43, Atin Mukherjee <
>>>>>>>>>>>>>>>> amukherj at redhat.com> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Allright, I see that you haven't bumped up the op-version.
>>>>>>>>>>>>>>>>> Can you please execute:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> gluster v set all cluster.op-version 30101  and then
>>>>>>>>>>>>>>>>> restart glusterd on all the nodes and check the brick status?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> s/30101/31001
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Sun, May 14, 2017 at 8:55 PM, Pawan Alwandi <
>>>>>>>>>>>>>>>>> pawan at platform.sh> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hello Atin,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks for looking at this.  Below is the output you
>>>>>>>>>>>>>>>>>> requested for.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Again, I'm seeing those errors after upgrading gluster on
>>>>>>>>>>>>>>>>>> host 1.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Host 1
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> # cat /var/lib/glusterd/glusterd.info
>>>>>>>>>>>>>>>>>> UUID=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073
>>>>>>>>>>>>>>>>>> operating-version=30600
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> # cat /var/lib/glusterd/peers/*
>>>>>>>>>>>>>>>>>> uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633
>>>>>>>>>>>>>>>>>> state=3
>>>>>>>>>>>>>>>>>> hostname1=192.168.0.7
>>>>>>>>>>>>>>>>>> uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95
>>>>>>>>>>>>>>>>>> state=3
>>>>>>>>>>>>>>>>>> hostname1=192.168.0.6
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> # gluster --version
>>>>>>>>>>>>>>>>>> glusterfs 3.10.1
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Host 2
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> # cat /var/lib/glusterd/glusterd.info
>>>>>>>>>>>>>>>>>> UUID=83e9a0b9-6bd5-483b-8516-d8928805ed95
>>>>>>>>>>>>>>>>>> operating-version=30600
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> # cat /var/lib/glusterd/peers/*
>>>>>>>>>>>>>>>>>> uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633
>>>>>>>>>>>>>>>>>> state=3
>>>>>>>>>>>>>>>>>> hostname1=192.168.0.7
>>>>>>>>>>>>>>>>>> uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073
>>>>>>>>>>>>>>>>>> state=3
>>>>>>>>>>>>>>>>>> hostname1=192.168.0.5
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> # gluster --version
>>>>>>>>>>>>>>>>>> glusterfs 3.6.2 built on Jan 21 2015 14:23:44
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Host 3
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> # cat /var/lib/glusterd/glusterd.info
>>>>>>>>>>>>>>>>>> UUID=5ec54b4f-f60c-48c6-9e55-95f2bb58f633
>>>>>>>>>>>>>>>>>> operating-version=30600
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> # cat /var/lib/glusterd/peers/*
>>>>>>>>>>>>>>>>>> uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073
>>>>>>>>>>>>>>>>>> state=3
>>>>>>>>>>>>>>>>>> hostname1=192.168.0.5
>>>>>>>>>>>>>>>>>> uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95
>>>>>>>>>>>>>>>>>> state=3
>>>>>>>>>>>>>>>>>> hostname1=192.168.0.6
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> # gluster --version
>>>>>>>>>>>>>>>>>> glusterfs 3.6.2 built on Jan 21 2015 14:23:44
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Sat, May 13, 2017 at 6:28 PM, Atin Mukherjee <
>>>>>>>>>>>>>>>>>> amukherj at redhat.com> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> I have already asked for the following earlier:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Can you please provide output of following from all the
>>>>>>>>>>>>>>>>>>> nodes:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> cat /var/lib/glusterd/glusterd.info
>>>>>>>>>>>>>>>>>>> cat /var/lib/glusterd/peers/*
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On Sat, 13 May 2017 at 12:22, Pawan Alwandi
>>>>>>>>>>>>>>>>>>> <pawan at platform.sh> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Hello folks,
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Does anyone have any idea whats going on here?
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>>> Pawan
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On Wed, May 10, 2017 at 5:02 PM, Pawan Alwandi <
>>>>>>>>>>>>>>>>>>>> pawan at platform.sh> wrote:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Hello,
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> I'm trying to upgrade gluster from 3.6.2 to 3.10.1 but
>>>>>>>>>>>>>>>>>>>>> don't see the glusterfsd and glusterfs processes coming up.
>>>>>>>>>>>>>>>>>>>>> http://gluster.readthedocs.io/en/latest/Upgrade-Guide/upgrade_to_3.10/
>>>>>>>>>>>>>>>>>>>>> is the process that I'm trying to follow.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> This is a 3 node server setup with a replicated volume
>>>>>>>>>>>>>>>>>>>>> having replica count of 3.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Logs below:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:03.507959] I [MSGID: 100030]
>>>>>>>>>>>>>>>>>>>>> [glusterfsd.c:2460:main] 0-/usr/sbin/glusterd: Started running
>>>>>>>>>>>>>>>>>>>>> /usr/sbin/glusterd version 3.10.1 (args: /usr/sbin/glusterd -p
>>>>>>>>>>>>>>>>>>>>> /var/run/glusterd.pid)
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:03.512827] I [MSGID: 106478]
>>>>>>>>>>>>>>>>>>>>> [glusterd.c:1449:init] 0-management: Maximum allowed open file descriptors
>>>>>>>>>>>>>>>>>>>>> set to 65536
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:03.512855] I [MSGID: 106479]
>>>>>>>>>>>>>>>>>>>>> [glusterd.c:1496:init] 0-management: Using /var/lib/glusterd as working
>>>>>>>>>>>>>>>>>>>>> directory
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:03.520426] W [MSGID: 103071]
>>>>>>>>>>>>>>>>>>>>> [rdma.c:4590:__gf_rdma_ctx_create] 0-rpc-transport/rdma: rdma_cm event
>>>>>>>>>>>>>>>>>>>>> channel creation failed [No such device]
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:03.520452] W [MSGID: 103055]
>>>>>>>>>>>>>>>>>>>>> [rdma.c:4897:init] 0-rdma.management: Failed to initialize IB Device
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:03.520465] W
>>>>>>>>>>>>>>>>>>>>> [rpc-transport.c:350:rpc_transport_load] 0-rpc-transport: 'rdma'
>>>>>>>>>>>>>>>>>>>>> initialization failed
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:03.520518] W
>>>>>>>>>>>>>>>>>>>>> [rpcsvc.c:1661:rpcsvc_create_listener] 0-rpc-service: cannot create
>>>>>>>>>>>>>>>>>>>>> listener, initing the transport failed
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:03.520534] E [MSGID: 106243]
>>>>>>>>>>>>>>>>>>>>> [glusterd.c:1720:init] 0-management: creation of 1 listeners failed,
>>>>>>>>>>>>>>>>>>>>> continuing with succeeded transport
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:04.931764] I [MSGID: 106513]
>>>>>>>>>>>>>>>>>>>>> [glusterd-store.c:2197:glusterd_restore_op_version] 0-glusterd: retrieved
>>>>>>>>>>>>>>>>>>>>> op-version: 30600
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:04.964354] I [MSGID: 106544]
>>>>>>>>>>>>>>>>>>>>> [glusterd.c:158:glusterd_uuid_init] 0-management: retrieved UUID:
>>>>>>>>>>>>>>>>>>>>> 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:04.993944] I [MSGID: 106498]
>>>>>>>>>>>>>>>>>>>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] 0-management:
>>>>>>>>>>>>>>>>>>>>> connect returned 0
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:04.995864] I [MSGID: 106498]
>>>>>>>>>>>>>>>>>>>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] 0-management:
>>>>>>>>>>>>>>>>>>>>> connect returned 0
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:04.995879] W [MSGID: 106062]
>>>>>>>>>>>>>>>>>>>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] 0-glusterd:
>>>>>>>>>>>>>>>>>>>>> Failed to get tcp-user-timeout
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:04.995903] I
>>>>>>>>>>>>>>>>>>>>> [rpc-clnt.c:1059:rpc_clnt_connection_init] 0-management: setting
>>>>>>>>>>>>>>>>>>>>> frame-timeout to 600
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:04.996325] I
>>>>>>>>>>>>>>>>>>>>> [rpc-clnt.c:1059:rpc_clnt_connection_init] 0-management: setting
>>>>>>>>>>>>>>>>>>>>> frame-timeout to 600
>>>>>>>>>>>>>>>>>>>>> Final graph:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> +------------------------------------------------------------------------------+
>>>>>>>>>>>>>>>>>>>>>   1: volume management
>>>>>>>>>>>>>>>>>>>>>   2:     type mgmt/glusterd
>>>>>>>>>>>>>>>>>>>>>   3:     option rpc-auth.auth-glusterfs on
>>>>>>>>>>>>>>>>>>>>>   4:     option rpc-auth.auth-unix on
>>>>>>>>>>>>>>>>>>>>>   5:     option rpc-auth.auth-null on
>>>>>>>>>>>>>>>>>>>>>   6:     option rpc-auth-allow-insecure on
>>>>>>>>>>>>>>>>>>>>>   7:     option transport.socket.listen-backlog 128
>>>>>>>>>>>>>>>>>>>>>   8:     option event-threads 1
>>>>>>>>>>>>>>>>>>>>>   9:     option ping-timeout 0
>>>>>>>>>>>>>>>>>>>>>  10:     option transport.socket.read-fail-log off
>>>>>>>>>>>>>>>>>>>>>  11:     option transport.socket.keepalive-interval 2
>>>>>>>>>>>>>>>>>>>>>  12:     option transport.socket.keepalive-time 10
>>>>>>>>>>>>>>>>>>>>>  13:     option transport-type rdma
>>>>>>>>>>>>>>>>>>>>>  14:     option working-directory /var/lib/glusterd
>>>>>>>>>>>>>>>>>>>>>  15: end-volume
>>>>>>>>>>>>>>>>>>>>>  16:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> +------------------------------------------------------------------------------+
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:04.996310] W [MSGID: 106062]
>>>>>>>>>>>>>>>>>>>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] 0-glusterd:
>>>>>>>>>>>>>>>>>>>>> Failed to get tcp-user-timeout
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.000461] I [MSGID: 101190]
>>>>>>>>>>>>>>>>>>>>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread
>>>>>>>>>>>>>>>>>>>>> with index 1
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.001493] W
>>>>>>>>>>>>>>>>>>>>> [socket.c:593:__socket_rwv] 0-management: readv on
>>>>>>>>>>>>>>>>>>>>> 192.168.0.7:24007 failed (No data available)
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.001513] I [MSGID: 106004]
>>>>>>>>>>>>>>>>>>>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] 0-management: Peer
>>>>>>>>>>>>>>>>>>>>> <192.168.0.7> (<5ec54b4f-f60c-48c6-9e55-95f2bb58f633>), in state <Peer in
>>>>>>>>>>>>>>>>>>>>> Cluster>, h
>>>>>>>>>>>>>>>>>>>>> as disconnected from glusterd.
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.001677] W
>>>>>>>>>>>>>>>>>>>>> [glusterd-locks.c:675:glusterd_mgmt_v3_unlock]
>>>>>>>>>>>>>>>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559)
>>>>>>>>>>>>>>>>>>>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu
>>>>>>>>>>>>>>>>>>>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0)
>>>>>>>>>>>>>>>>>>>>> [0x7f0bf9d7dcf0]
>>>>>>>>>>>>>>>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3)
>>>>>>>>>>>>>>>>>>>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no
>>>>>>>>>>>>>>>>>>>>> t held
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.001696] W [MSGID: 106118]
>>>>>>>>>>>>>>>>>>>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] 0-management: Lock not
>>>>>>>>>>>>>>>>>>>>> released for shared
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.003099] E
>>>>>>>>>>>>>>>>>>>>> [rpc-clnt.c:365:saved_frames_unwind] (-->
>>>>>>>>>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x13c)[0x7f0bfeeca73c]
>>>>>>>>>>>>>>>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(s
>>>>>>>>>>>>>>>>>>>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (-->
>>>>>>>>>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de]
>>>>>>>>>>>>>>>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x
>>>>>>>>>>>>>>>>>>>>> 91)[0x7f0bfec91c21] (-->
>>>>>>>>>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710]
>>>>>>>>>>>>>>>>>>>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1))
>>>>>>>>>>>>>>>>>>>>> called at 2017-05-10 09:0
>>>>>>>>>>>>>>>>>>>>> 7:05.000627 (xid=0x1)
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.003129] E [MSGID: 106167]
>>>>>>>>>>>>>>>>>>>>> [glusterd-handshake.c:2181:__glusterd_peer_dump_version_cbk] 0-management:
>>>>>>>>>>>>>>>>>>>>> Error through RPC layer, retry again later
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.003251] W
>>>>>>>>>>>>>>>>>>>>> [socket.c:593:__socket_rwv] 0-management: readv on
>>>>>>>>>>>>>>>>>>>>> 192.168.0.6:24007 failed (No data available)
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.003267] I [MSGID: 106004]
>>>>>>>>>>>>>>>>>>>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] 0-management: Peer
>>>>>>>>>>>>>>>>>>>>> <192.168.0.6> (<83e9a0b9-6bd5-483b-8516-d8928805ed95>), in state <Peer in
>>>>>>>>>>>>>>>>>>>>> Cluster>, h
>>>>>>>>>>>>>>>>>>>>> as disconnected from glusterd.
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.003318] W
>>>>>>>>>>>>>>>>>>>>> [glusterd-locks.c:675:glusterd_mgmt_v3_unlock]
>>>>>>>>>>>>>>>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559)
>>>>>>>>>>>>>>>>>>>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu
>>>>>>>>>>>>>>>>>>>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0)
>>>>>>>>>>>>>>>>>>>>> [0x7f0bf9d7dcf0]
>>>>>>>>>>>>>>>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3)
>>>>>>>>>>>>>>>>>>>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no
>>>>>>>>>>>>>>>>>>>>> t held
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.003329] W [MSGID: 106118]
>>>>>>>>>>>>>>>>>>>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] 0-management: Lock not
>>>>>>>>>>>>>>>>>>>>> released for shared
>>>>>>>>>>>>>>>>>>>>> [2017-05-10 09:07:05.003457] E
>>>>>>>>>>>>>>>>>>>>> [rpc-clnt.c:365:saved_frames_unwind] (-->
>>>>>>>>>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x13c)[0x7f0bfeeca73c]
>>>>>>>>>>>>>>>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(s
>>>>>>>>>>>>>>>>>>>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (-->
>>>>>>>>>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de]
>>>>>>>>>>>>>>>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x
>>>>>>>>>>>>>>>>>>>>> 91)[0x7f0bfec91c21] (-->
>>>>>>>>>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710]
>>>>>>>>>>>>>>>>>>>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) op(DUMP(1))
>>>>>>>>>>>>>>>>>>>>> called at 2017-05-10 09:0
>>>>>>>>>>>>>>>>>>>>> 7:05.001407 (xid=0x1)
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> There are a bunch of errors reported but I'm not sure
>>>>>>>>>>>>>>>>>>>>> which is signal and which ones are noise.  Does anyone have any idea whats
>>>>>>>>>>>>>>>>>>>>> going on here?
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>>>> Pawan
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>>>>> Gluster-users mailing list
>>>>>>>>>>>>>>>>>>>> Gluster-users at gluster.org
>>>>>>>>>>>>>>>>>>>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>> - Atin (atinm)
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>> - Atin (atinm)
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> - Atin (atinm)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> --
>>>>>>>>>>>> - Atin (atinm)
>>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> - Atin (atinm)
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
> --
- Atin (atinm)
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170525/b1aa3e04/attachment.html>


More information about the Gluster-users mailing list