[Gluster-users] Unable to get lock for uuid peers
amukherj at redhat.com
Wed Jun 10 11:45:44 UTC 2015
On 06/10/2015 02:58 PM, Sergio Traldi wrote:
> On 06/10/2015 10:27 AM, Krishnan Parthasarathi wrote:
>>> Hi all,
>>> I two servers with 3.7.1 and have the same problem of this issue:
>>> My servers packages:
>>> # rpm -qa | grep gluster | sort
>>> # gluster volume status
>>> Another transaction is in progress. Please try again after sometime.
The problem is although you are running 3.7.1 binaries the cluster
op-version is set to 30501, because of glusterd still goes for acquiring
cluster lock instead of volume wise lock for every request. Command log
history indicates glusterD is getting multiple volume's status requests
and because of it fails to acquire cluster lock. Could you bump up your
cluster's op-version by the following command and recheck?
gluster volume set all cluster.op-version 30701
>>> In /var/log/gluster/etc-glusterfs-glusterd.vol.log I found:
>>> [2015-06-09 16:12:38.949842] E [glusterd-utils.c:164:glusterd_lock]
>>> 0-management: Unable to get lock for uuid:
>>> 99a41a2a-2ce5-461c-aec0-510bd5b37bf2, lock held by:
>>> [2015-06-09 16:12:38.949864] E
>>> 0-management: Unable to acquire lock
>>> I check the files:
>>> From server 1:
>>> # cat /var/lib/glusterd/peers/04a7d2bb-bdd9-4e0d-b460-87ad4adbe12c
>>> From server 2:
>>> # cat /var/lib/glusterd/peers/99a41a2a-2ce5-461c-aec0-510bd5b37bf2
>> Could you attach the complete glusterd log file and cmd-history.log
>> file under /var/log/glusterfs directory? Could you provide a more
>> detailed listing of things you did before hitting this issue?
> Hi Krishnan,
> thanks to a quick answer.
> In attach you can found the two log you request:
> We use the gluster volume as openstack nova, glance, cinder backend.
> The volume is configured using 2 bricks mounted by an iscsi device:
> [root at cld-stg-01 glusterfs]# gluster volume info volume-nova-prod
> Volume Name: volume-nova-prod
> Type: Distribute
> Volume ID: 4bbef4c8-0441-4e81-a2c5-559401adadc0
> Status: Started
> Number of Bricks: 2
> Transport-type: tcp
> Brick1: 192.168.61.100:/brickOpenstack/nova-prod/mpathb
> Brick2: 192.168.61.101:/brickOpenstack/nova-prod/mpathb
> Options Reconfigured:
> storage.owner-gid: 162
> storage.owner-uid: 162
> Last week we update openstack from havana to icehouse and we rename the
> storage hosts but we didn't change the IP.
> All volume have been created using ip addresses.
> So last week we stop all services (openstack gluster and also iscsi).
> We change the name in DNS of private ip of the 2 nics.
> We reboot the storage servers
> We start agian iscsi, multipath, glusterd process.
> We have to stop and start the volumes, but after that everything works
> Now we don't observe any other problems except this.
> We have a nagios probe which check the volume status each 5 minutes to
> ensure all gluster process is working fine and so we find this problem I
> Gluster-users mailing list
> Gluster-users at gluster.org
More information about the Gluster-users