[Bugs] [Bug 1631248] If a node disconnects during volume delete, it assumes deleted volume as a freshly created volume when it is back online

bugzilla at redhat.com bugzilla at redhat.com
Mon Oct 22 10:45:41 UTC 2018


https://bugzilla.redhat.com/show_bug.cgi?id=1631248

hari gowtham <hgowtham at redhat.com> changed:

           What    |Removed                     |Added
----------------------------------------------------------------------------
                 CC|                            |hgowtham at redhat.com



--- Comment #26 from hari gowtham <hgowtham at redhat.com> ---
Hi,

>From the logs the following observations were made:

at 13:27:38 the glusterd restart has happened on node 10.70.35.38.

[2018-10-12 13:27:38.467525] I [MSGID: 100030] [glusterfsd.c:2504:main]
0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 3.12.2 (args:
/usr/sbin/glusterd -p /var/run/glusterd.pid --log-level INFO)

At this time the volumes have to be synced in the handshake.
The handshake started around 13:27:45

[2018-10-12 13:27:45.983464] I [MSGID: 106163]
[glusterd-handshake.c:1319:__glusterd_mgmt_hndsk_versions_ack] 0-management:
using the op-version 31302

and went till 13:27:59.876472 on 10.70.35.38

[2018-10-12 13:27:59.876472] I [MSGID: 106493]
[glusterd-rpc-ops.c:701:__glusterd_friend_update_cbk] 0-management: Received
ACC from uuid: 4382f67b-e759-409f-a579-f8c98ededd83


And the time the gluster volume delete for the volume iceberg-2-1-9-5 has been
issued:  
[2018-10-12 13:27:59.875956]  : v delete iceberg-2-1-9-5 : SUCCESS

>From the above observation the delete was issued while the handshake was
happening.

As the handshake started before delete and the iceberg-2-1-9-5 was supposed to
be available in the cross checked file for handshake. And this was being synced
in the restarted node (10.70.35.38).
Now delete was issued and it deleted the entry as the node has come up. After
the delete completed successfully and removed the entry, the handshake has
created the entry for the iceberg-2-1-9-5. This is why the entry has
stage_deleted as false (before delete was issued it is false).

And to confirm further the entry is available only on the node that went
through a reboot (10.70.35.38) and not in the other nodes.

-- 
You are receiving this mail because:
You are on the CC list for the bug.
Unsubscribe from this bug https://bugzilla.redhat.com/token.cgi?t=lXefACuTQB&a=cc_unsubscribe


More information about the Bugs mailing list