[Gluster-users] failing commits

lejeczek peljasz at yahoo.co.uk
Wed Feb 1 19:34:44 UTC 2017



On 01/02/17 19:30, lejeczek wrote:
>
>
> On 01/02/17 14:44, Atin Mukherjee wrote:
>> I think you have hit 
>> https://bugzilla.redhat.com/show_bug.cgi?id=1406411 which 
>> has been fixed in mainline and will be available in 
>> release-3.10 which is slated for next month.
>>
>> To prove you have hit the same problem can you please 
>> confirm the following:
>>
>> 1. Which Gluster version are you running?
>> 2. Was any of the existing brick down?
>> 2. Did you mounted the volume? If not you have two ways 
>> (1) bring up the brick and restart glusterd followed by 
>> add-brick or (2) if the existing brick(s) is bad for some 
>> reason, restarting glusterd and mounting the volume 
>> followed by a look up and then attempting add-brick 
>> should succeed.
>>
>>
> a chance to properly investigate it has been lost I think.
> I all started with one peer I missed was not migrated from 
> 3.7 to 3.8 and unfortunately it was a system I could not 
> tamper with until late evening, which is now.
> This problem though occurred after I already upgraded that 
> gluster to 3.8. I even removed that failing node's bricks 
> and detached it, re-attached it and still, those errors I 
> described earlier... until now when I restarted that one 
> last one peer... now all seems ok, well, at least I don't 
> see those errors any more.
>
> Should I now be looking at something particular more closely?
> b.w.
> L.
>
I realized  I might have not been clear there, that box 
where I missed that upgrade 3.7->3.8 was not the one where 
problem existed, it was a different box, if that 
changes/helps anything.


>
>> On Wed, Feb 1, 2017 at 7:49 PM, lejeczek 
>> <peljasz at yahoo.co.uk <mailto:peljasz at yahoo.co.uk>> wrote:
>>
>>     hi,
>>
>>     I have a four peers gluster and one is failing, well,
>>     kind of..
>>     If on a working peer I do:
>>
>>     $ gluster volume add-brick QEMU-VMs replica 3
>>     10.5.6.49:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-QEMU-VMs
>>     force
>>     volume add-brick: failed: Commit failed on whale.priv
>>     Please check log file for details.
>>
>>     but:
>>
>>     $ gluster vol info QEMU-VMs
>>     Volume Name: QEMU-VMs
>>     Type: Replicate
>>     Volume ID: 8709782a-daa5-4434-a816-c4e0aef8fef2
>>     Status: Started
>>     Snapshot Count: 0
>>     Number of Bricks: 1 x 3 = 3
>>     Transport-type: tcp
>>     Bricks:
>>     Brick1:
>>     10.5.6.100:/__.aLocalStorages/1/0-GLUSTERs/1GLUSTER-QEMU-VMs
>>     Brick2:
>>     10.5.6.17:/__.aLocalStorages/1/0-GLUSTERs/QEMU-VMs
>>     Brick3:
>>     10.5.6.49:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-QEMU-VMs
>>     # <= so it is here, also this command on that failing
>>     peers reports correctly.
>>
>>     Interestingly,
>>
>>     $ gluster volume remove-brick
>>
>>     removes no errors, but this change is not propagated
>>     to the failing peer. Vol info still reports its brick
>>     is part of the volume.
>>
>>     And the failing completely part: every command on
>>     failing peer reports:
>>
>>     $ gluster volume remove-brick QEMU-VMs replica 2
>>     10.5.6.49:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-QEMU-VMs
>>     force
>>     Removing brick(s) can result in data loss. Do you
>>     want to Continue? (y/n) y
>>     volume remove-brick commit force: failed: Commit
>>     failed on 10.5.6.32. Please check log file for details.
>>     Commit failed on rider.priv Please check log file for
>>     details.
>>     Commit failed on 10.5.6.17. Please check log file for
>>     details.
>>
>>     I've been watching logs but honestly, don't know
>>     which one(s) I should paste in here.
>>     b.w.
>>     L.
>>
>>
>>     _______________________________________________
>>     Gluster-users mailing list
>>     Gluster-users at gluster.org
>>     <mailto:Gluster-users at gluster.org>
>>     http://lists.gluster.org/mailman/listinfo/gluster-users
>>     <http://lists.gluster.org/mailman/listinfo/gluster-users>
>>
>>
>>
>>
>> -- 
>>
>> ~ Atin (atinm)
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170201/1c5bf399/attachment.html>


More information about the Gluster-users mailing list