[Gluster-users] add-brick: failed: Commit failed
David Cunningham
dcunningham at voisonics.com
Wed May 22 05:53:11 UTC 2019
Hi Ravi,
I'd already done exactly that before, where step 3 was a simple 'rm -rf
/nodirectwritedata/gluster/gvol0'. Have you another suggestion on what the
cleanup or reformat should be?
Thank you.
On Wed, 22 May 2019 at 13:56, Ravishankar N <ravishankar at redhat.com> wrote:
> Hmm, so the volume info seems to indicate that the add-brick was
> successful but the gfid xattr is missing on the new brick (as are the
> actual files, barring the .glusterfs folder, according to your previous
> mail).
>
> Do you want to try removing and adding it again?
>
> 1. `gluster volume remove-brick gvol0 replica 2
> gfs3:/nodirectwritedata/gluster/gvol0 force` from gfs1
>
> 2. Check that gluster volume info is now back to a 1x2 volume on all nodes
> and `gluster peer status` is connected on all nodes.
>
> 3. Cleanup or reformat '/nodirectwritedata/gluster/gvol0' on gfs3.
>
> 4. `gluster volume add-brick gvol0 replica 3 arbiter 1
> gfs3:/nodirectwritedata/gluster/gvol0` from gfs1.
>
> 5. Check that the files are getting healed on to the new brick.
> Thanks,
> Ravi
> On 22/05/19 6:50 AM, David Cunningham wrote:
>
> Hi Ravi,
>
> Certainly. On the existing two nodes:
>
> gfs1 # getfattr -d -m. -e hex /nodirectwritedata/gluster/gvol0
> getfattr: Removing leading '/' from absolute path names
> # file: nodirectwritedata/gluster/gvol0
> trusted.afr.dirty=0x000000000000000000000000
> trusted.afr.gvol0-client-2=0x000000000000000000000000
> trusted.gfid=0x00000000000000000000000000000001
> trusted.glusterfs.dht=0x000000010000000000000000ffffffff
> trusted.glusterfs.volume-id=0xfb5af69e1c3e41648b23c1d7bec9b1b6
>
> gfs2 # getfattr -d -m. -e hex /nodirectwritedata/gluster/gvol0
> getfattr: Removing leading '/' from absolute path names
> # file: nodirectwritedata/gluster/gvol0
> trusted.afr.dirty=0x000000000000000000000000
> trusted.afr.gvol0-client-0=0x000000000000000000000000
> trusted.afr.gvol0-client-2=0x000000000000000000000000
> trusted.gfid=0x00000000000000000000000000000001
> trusted.glusterfs.dht=0x000000010000000000000000ffffffff
> trusted.glusterfs.volume-id=0xfb5af69e1c3e41648b23c1d7bec9b1b6
>
> On the new node:
>
> gfs3 # getfattr -d -m. -e hex /nodirectwritedata/gluster/gvol0
> getfattr: Removing leading '/' from absolute path names
> # file: nodirectwritedata/gluster/gvol0
> trusted.afr.dirty=0x000000000000000000000001
> trusted.glusterfs.volume-id=0xfb5af69e1c3e41648b23c1d7bec9b1b6
>
> Output of "gluster volume info" is the same on all 3 nodes and is:
>
> # gluster volume info
>
> Volume Name: gvol0
> Type: Replicate
> Volume ID: fb5af69e-1c3e-4164-8b23-c1d7bec9b1b6
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1 x (2 + 1) = 3
> Transport-type: tcp
> Bricks:
> Brick1: gfs1:/nodirectwritedata/gluster/gvol0
> Brick2: gfs2:/nodirectwritedata/gluster/gvol0
> Brick3: gfs3:/nodirectwritedata/gluster/gvol0 (arbiter)
> Options Reconfigured:
> performance.client-io-threads: off
> nfs.disable: on
> transport.address-family: inet
>
>
> On Wed, 22 May 2019 at 12:43, Ravishankar N <ravishankar at redhat.com>
> wrote:
>
>> Hi David,
>> Could you provide the `getfattr -d -m. -e hex
>> /nodirectwritedata/gluster/gvol0` output of all bricks and the output of
>> `gluster volume info`?
>>
>> Thanks,
>> Ravi
>> On 22/05/19 4:57 AM, David Cunningham wrote:
>>
>> Hi Sanju,
>>
>> Here's what glusterd.log says on the new arbiter server when trying to
>> add the node:
>>
>> [2019-05-22 00:15:05.963059] I [run.c:242:runner_log]
>> (-->/usr/lib64/glusterfs/5.6/xlator/mgmt/glusterd.so(+0x3b2cd)
>> [0x7fe4ca9102cd]
>> -->/usr/lib64/glusterfs/5.6/xlator/mgmt/glusterd.so(+0xe6b85)
>> [0x7fe4ca9bbb85] -->/lib64/libglusterfs.so.0(runner_log+0x115)
>> [0x7fe4d5ecc955] ) 0-management: Ran script:
>> /var/lib/glusterd/hooks/1/add-brick/pre/S28Quota-enable-root-xattr-heal.sh
>> --volname=gvol0 --version=1 --volume-op=add-brick
>> --gd-workdir=/var/lib/glusterd
>> [2019-05-22 00:15:05.963177] I [MSGID: 106578]
>> [glusterd-brick-ops.c:1355:glusterd_op_perform_add_bricks] 0-management:
>> replica-count is set 3
>> [2019-05-22 00:15:05.963228] I [MSGID: 106578]
>> [glusterd-brick-ops.c:1360:glusterd_op_perform_add_bricks] 0-management:
>> arbiter-count is set 1
>> [2019-05-22 00:15:05.963257] I [MSGID: 106578]
>> [glusterd-brick-ops.c:1364:glusterd_op_perform_add_bricks] 0-management:
>> type is set 0, need to change it
>> [2019-05-22 00:15:17.015268] E [MSGID: 106053]
>> [glusterd-utils.c:13942:glusterd_handle_replicate_brick_ops] 0-management:
>> Failed to set extended attribute trusted.add-brick : Transport endpoint is
>> not connected [Transport endpoint is not connected]
>> [2019-05-22 00:15:17.036479] E [MSGID: 106073]
>> [glusterd-brick-ops.c:2595:glusterd_op_add_brick] 0-glusterd: Unable to add
>> bricks
>> [2019-05-22 00:15:17.036595] E [MSGID: 106122]
>> [glusterd-mgmt.c:299:gd_mgmt_v3_commit_fn] 0-management: Add-brick commit
>> failed.
>> [2019-05-22 00:15:17.036710] E [MSGID: 106122]
>> [glusterd-mgmt-handler.c:594:glusterd_handle_commit_fn] 0-management:
>> commit failed on operation Add brick
>>
>> As before gvol0-add-brick-mount.log said:
>>
>> [2019-05-22 00:15:17.005695] I [fuse-bridge.c:4267:fuse_init]
>> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.24 kernel
>> 7.22
>> [2019-05-22 00:15:17.005749] I [fuse-bridge.c:4878:fuse_graph_sync]
>> 0-fuse: switched to graph 0
>> [2019-05-22 00:15:17.010101] E [fuse-bridge.c:4336:fuse_first_lookup]
>> 0-fuse: first lookup on root failed (Transport endpoint is not connected)
>> [2019-05-22 00:15:17.014217] W [fuse-bridge.c:897:fuse_attr_cbk]
>> 0-glusterfs-fuse: 2: LOOKUP() / => -1 (Transport endpoint is not connected)
>> [2019-05-22 00:15:17.015097] W [fuse-resolve.c:127:fuse_resolve_gfid_cbk]
>> 0-fuse: 00000000-0000-0000-0000-000000000001: failed to resolve (Transport
>> endpoint is not connected)
>> [2019-05-22 00:15:17.015158] W [fuse-bridge.c:3294:fuse_setxattr_resume]
>> 0-glusterfs-fuse: 3: SETXATTR 00000000-0000-0000-0000-000000000001/1
>> (trusted.add-brick) resolution failed
>> [2019-05-22 00:15:17.035636] I [fuse-bridge.c:5144:fuse_thread_proc]
>> 0-fuse: initating unmount of /tmp/mntYGNbj9
>> [2019-05-22 00:15:17.035854] W [glusterfsd.c:1500:cleanup_and_exit]
>> (-->/lib64/libpthread.so.0(+0x7dd5) [0x7f7745ccedd5]
>> -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x55c81b63de75]
>> -->/usr/sbin/glusterfs(cleanup_and_exit+0x6b) [0x55c81b63dceb] ) 0-:
>> received signum (15), shutting down
>> [2019-05-22 00:15:17.035942] I [fuse-bridge.c:5914:fini] 0-fuse:
>> Unmounting '/tmp/mntYGNbj9'.
>> [2019-05-22 00:15:17.035966] I [fuse-bridge.c:5919:fini] 0-fuse: Closing
>> fuse connection to '/tmp/mntYGNbj9'.
>>
>> Here are the processes running on the new arbiter server:
>> # ps -ef | grep gluster
>> root 3466 1 0 20:13 ? 00:00:00 /usr/sbin/glusterfs -s
>> localhost --volfile-id gluster/glustershd -p
>> /var/run/gluster/glustershd/glustershd.pid -l
>> /var/log/glusterfs/glustershd.log -S
>> /var/run/gluster/24c12b09f93eec8e.socket --xlator-option
>> *replicate*.node-uuid=2069cfb3-c798-47e3-8cf8-3c584cf7c412 --process-name
>> glustershd
>> root 6832 1 0 May16 ? 00:02:10 /usr/sbin/glusterd -p
>> /var/run/glusterd.pid --log-level INFO
>> root 17841 1 0 May16 ? 00:00:58 /usr/sbin/glusterfs
>> --process-name fuse --volfile-server=gfs1 --volfile-id=/gvol0 /mnt/glusterfs
>>
>> Here are the files created on the new arbiter server:
>> # find /nodirectwritedata/gluster/gvol0 | xargs ls -ald
>> drwxr-xr-x 3 root root 4096 May 21 20:15 /nodirectwritedata/gluster/gvol0
>> drw------- 2 root root 4096 May 21 20:15
>> /nodirectwritedata/gluster/gvol0/.glusterfs
>>
>> Thank you for your help!
>>
>>
>> On Tue, 21 May 2019 at 00:10, Sanju Rakonde <srakonde at redhat.com> wrote:
>>
>>> David,
>>>
>>> can you please attach glusterd.logs? As the error message says, Commit
>>> failed on the arbitar node, we might be able to find some issue on that
>>> node.
>>>
>>> On Mon, May 20, 2019 at 10:10 AM Nithya Balachandran <
>>> nbalacha at redhat.com> wrote:
>>>
>>>>
>>>>
>>>> On Fri, 17 May 2019 at 06:01, David Cunningham <
>>>> dcunningham at voisonics.com> wrote:
>>>>
>>>>> Hello,
>>>>>
>>>>> We're adding an arbiter node to an existing volume and having an
>>>>> issue. Can anyone help? The root cause error appears to be
>>>>> "00000000-0000-0000-0000-000000000001: failed to resolve (Transport
>>>>> endpoint is not connected)", as below.
>>>>>
>>>>> We are running glusterfs 5.6.1. Thanks in advance for any assistance!
>>>>>
>>>>> On existing node gfs1, trying to add new arbiter node gfs3:
>>>>>
>>>>> # gluster volume add-brick gvol0 replica 3 arbiter 1
>>>>> gfs3:/nodirectwritedata/gluster/gvol0
>>>>> volume add-brick: failed: Commit failed on gfs3. Please check log file
>>>>> for details.
>>>>>
>>>>
>>>> This looks like a glusterd issue. Please check the glusterd logs for
>>>> more info.
>>>> Adding the glusterd dev to this thread. Sanju, can you take a look?
>>>>
>>>> Regards,
>>>> Nithya
>>>>
>>>>>
>>>>> On new node gfs3 in gvol0-add-brick-mount.log:
>>>>>
>>>>> [2019-05-17 01:20:22.689721] I [fuse-bridge.c:4267:fuse_init]
>>>>> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.24 kernel
>>>>> 7.22
>>>>> [2019-05-17 01:20:22.689778] I [fuse-bridge.c:4878:fuse_graph_sync]
>>>>> 0-fuse: switched to graph 0
>>>>> [2019-05-17 01:20:22.694897] E [fuse-bridge.c:4336:fuse_first_lookup]
>>>>> 0-fuse: first lookup on root failed (Transport endpoint is not connected)
>>>>> [2019-05-17 01:20:22.699770] W
>>>>> [fuse-resolve.c:127:fuse_resolve_gfid_cbk] 0-fuse:
>>>>> 00000000-0000-0000-0000-000000000001: failed to resolve (Transport endpoint
>>>>> is not connected)
>>>>> [2019-05-17 01:20:22.699834] W
>>>>> [fuse-bridge.c:3294:fuse_setxattr_resume] 0-glusterfs-fuse: 2: SETXATTR
>>>>> 00000000-0000-0000-0000-000000000001/1 (trusted.add-brick) resolution failed
>>>>> [2019-05-17 01:20:22.715656] I [fuse-bridge.c:5144:fuse_thread_proc]
>>>>> 0-fuse: initating unmount of /tmp/mntQAtu3f
>>>>> [2019-05-17 01:20:22.715865] W [glusterfsd.c:1500:cleanup_and_exit]
>>>>> (-->/lib64/libpthread.so.0(+0x7dd5) [0x7fb223bf6dd5]
>>>>> -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x560886581e75]
>>>>> -->/usr/sbin/glusterfs(cleanup_and_exit+0x6b) [0x560886581ceb] ) 0-:
>>>>> received signum (15), shutting down
>>>>> [2019-05-17 01:20:22.715926] I [fuse-bridge.c:5914:fini] 0-fuse:
>>>>> Unmounting '/tmp/mntQAtu3f'.
>>>>> [2019-05-17 01:20:22.715953] I [fuse-bridge.c:5919:fini] 0-fuse:
>>>>> Closing fuse connection to '/tmp/mntQAtu3f'.
>>>>>
>>>>> Processes running on new node gfs3:
>>>>>
>>>>> # ps -ef | grep gluster
>>>>> root 6832 1 0 20:17 ? 00:00:00 /usr/sbin/glusterd -p
>>>>> /var/run/glusterd.pid --log-level INFO
>>>>> root 15799 1 0 20:17 ? 00:00:00 /usr/sbin/glusterfs -s
>>>>> localhost --volfile-id gluster/glustershd -p
>>>>> /var/run/gluster/glustershd/glustershd.pid -l
>>>>> /var/log/glusterfs/glustershd.log -S
>>>>> /var/run/gluster/24c12b09f93eec8e.socket --xlator-option
>>>>> *replicate*.node-uuid=2069cfb3-c798-47e3-8cf8-3c584cf7c412 --process-name
>>>>> glustershd
>>>>> root 16856 16735 0 21:21 pts/0 00:00:00 grep --color=auto
>>>>> gluster
>>>>>
>>>>> --
>>>>> David Cunningham, Voisonics Limited
>>>>> http://voisonics.com/
>>>>> USA: +1 213 221 1092
>>>>> New Zealand: +64 (0)28 2558 3782
>>>>> _______________________________________________
>>>>> Gluster-users mailing list
>>>>> Gluster-users at gluster.org
>>>>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>>>
>>>>
>>>
>>> --
>>> Thanks,
>>> Sanju
>>>
>>
>>
>> --
>> David Cunningham, Voisonics Limited
>> http://voisonics.com/
>> USA: +1 213 221 1092
>> New Zealand: +64 (0)28 2558 3782
>>
>> _______________________________________________
>> Gluster-users mailing listGluster-users at gluster.orghttps://lists.gluster.org/mailman/listinfo/gluster-users
>>
>>
>
> --
> David Cunningham, Voisonics Limited
> http://voisonics.com/
> USA: +1 213 221 1092
> New Zealand: +64 (0)28 2558 3782
>
>
--
David Cunningham, Voisonics Limited
http://voisonics.com/
USA: +1 213 221 1092
New Zealand: +64 (0)28 2558 3782
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190522/366ca7f7/attachment-0001.html>
More information about the Gluster-users
mailing list