[Gluster-users] Another transaction could be in progress
Franco Broi
franco.broi at iongeo.com
Tue Mar 18 08:21:07 UTC 2014
Sorry, didn't think to look in the log file, I can see I have bigger
problems. Last time I saw this was because I had changed an IP address
but this time all I did was reboot the server. I've checked all the
files in vols and everything looks good.
[2014-03-18 08:09:18.117040] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-0
[2014-03-18 08:09:18.117074] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-1
[2014-03-18 08:09:18.117087] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-2
[2014-03-18 08:09:18.117097] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-3
[2014-03-18 08:09:18.117107] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-4
[2014-03-18 08:09:18.117117] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-5
[2014-03-18 08:09:18.117128] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-6
[2014-03-18 08:09:18.117138] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-7
[2014-03-18 08:09:18.117148] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-8
[2014-03-18 08:09:18.117158] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-9
[2014-03-18 08:09:18.117168] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-10
[2014-03-18 08:09:18.117178] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-11
[2014-03-18 08:09:18.117196] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-12
[2014-03-18 08:09:18.117209] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-13
[2014-03-18 08:09:18.117219] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-14
[2014-03-18 08:09:18.117229] E [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key: brick-15
This is from another server
[root at nas1 bricks]# gluster vol status
Status of volume: data
Gluster process Port Online Pid
------------------------------------------------------------------------------
Brick nas1-10g:/data1/gvol 49152 Y 17331
Brick nas2-10g:/data5/gvol 49160 Y 3933
Brick nas1-10g:/data2/gvol 49153 Y 17340
Brick nas2-10g:/data6/gvol 49161 Y 3942
Brick nas1-10g:/data3/gvol 49154 Y 17350
Brick nas2-10g:/data7/gvol 49162 Y 3951
Brick nas1-10g:/data4/gvol 49155 Y 17360
Brick nas2-10g:/data8/gvol 49163 Y 3960
Brick nas3-10g:/data9/gvol 49156 Y 10076
Brick nas3-10g:/data10/gvol 49157 Y 10085
Brick nas3-10g:/data11/gvol 49158 Y 10094
Brick nas3-10g:/data12/gvol 49159 Y 10108
Brick nas4-10g:/data13/gvol N/A N 8879
Brick nas4-10g:/data14/gvol N/A N 8884
Brick nas4-10g:/data15/gvol N/A N 8888
Brick nas4-10g:/data16/gvol N/A N 8892
NFS Server on localhost 2049 Y 18725
NFS Server on nas3-10g 2049 Y 11667
NFS Server on nas2-10g 2049 Y 4980
NFS Server on nas4-10g N/A N N/A
There are no active volume tasks
Any ideas?
On Tue, 2014-03-18 at 12:39 +0530, Kaushal M wrote:
> The lock is an in-memory structure which isn't persisted. Restarting
> should reset the lock. You could possibly reset the lock by gdbing
> into the glusterd process.
>
> Since this is happening to you consistently, there is something else
> that is wrong. Could you please give more details on your cluster? And
> the glusterd logs of the misbehaving peer (if possible for all the
> peers). It would help in tracking it down.
>
>
>
> On Tue, Mar 18, 2014 at 12:24 PM, Franco Broi <franco.broi at iongeo.com> wrote:
> >
> > Restarted the glusterd daemons on all 4 servers, still the same.
> >
> > It only and always fails on the same server and it always works on the
> > other servers.
> >
> > I had to reboot the server in question this morning, perhaps it's got
> > itself in a funny state.
> >
> > Is the lock something that can be examined? And removed?
> >
> > On Tue, 2014-03-18 at 12:08 +0530, Kaushal M wrote:
> >> This mostly occurs when you run two gluster commands simultaneously.
> >> Gluster uses a lock on each peer to synchronize commands. Any command
> >> which would need to do operations on multiple peers, would first
> >> acquire this lock, and release it after doing the operation. If a
> >> command cannot acquire a lock because another command had the lock, it
> >> will fail with the above error message.
> >>
> >> It sometimes happens that a command could fail to release the lock on
> >> some peers. When this happens all further commands which need the lock
> >> will fail with the same error. In this case your only option is to
> >> restart glusterd on the peers which have the stale lock held. This
> >> will not cause any downtime as the brick processes are not affected by
> >> restarting glusterd.
> >>
> >> In your case, since you can run commands on other nodes, most likely
> >> you are running commands simultaneously or at least running a command
> >> before an old one finishes.
> >>
> >> ~kaushal
> >>
> >> On Tue, Mar 18, 2014 at 11:24 AM, Franco Broi <franco.broi at iongeo.com> wrote:
> >> >
> >> > What causes this error? And how do I get rid of it?
> >> >
> >> > [root at nas4 ~]# gluster vol status
> >> > Another transaction could be in progress. Please try again after sometime.
> >> >
> >> >
> >> > Looks normal on any other server.
> >> >
> >> > _______________________________________________
> >> > Gluster-users mailing list
> >> > Gluster-users at gluster.org
> >> > http://supercolony.gluster.org/mailman/listinfo/gluster-users
> >
> >
More information about the Gluster-users
mailing list