[Gluster-users] geo-replication fails on CentOS 6.5, gluster v 3.5.2

Kingsley gluster at gluster.dogwind.com
Fri Sep 26 23:15:59 UTC 2014


Hi,

I'm new to gluster so forgive me if I'm being an idiot. I've searched
the list archives back to May but haven't found the exact issue I've
come across, so I thought I'd ask on here.

Firstly, I'd like to thank the people working on this project. I've
found gluster to be pretty simple to get going and it seems to work
pretty well so far. It looks like it will be a good fit for the
application I have in mind, if we can get geo-replication to work
reliably.

Now on to my problem ...

I've set up an additional gluster volume and configured geo-replication
to replicate the master volume to it using the instructions here:

https://github.com/gluster/glusterfs/blob/master/doc/admin-guide/en-US/markdown/admin_distributed_geo_rep.md

To keep things simple while it was all new to me and I was just testing,
I didn't want to add confusion by thinking about using non-privileged
accounts and mountbroker and stuff so I just set everything up to use
root.

Anyway, I mounted the master volume and slave on a client machine (I
didn't modify the content of the slave volume, I just mounted it so that
I could check things were working).

When I manually create or delete a few files and wait 60 seconds for
replication to do its thing, it seems to work fine.

However, when I hit it with a script to simulate intense user activity,
geo-replication breaks. I deleted the geo-replication and removed the
slave volume, then re-created and re-enabled geo-replication several
times so that I could start again from scratch. Each time, my script
(which just creates, renames and deletes files in the master volume via
a glusterfs mount) runs for barely a minute before geo-replication
breaks.

I tried this with the slave volume containing just one brick, and also
with it containing 2 bricks replicating each other. Each time, it broke.

On the slave, I noticed that the geo-replication logs contained entries
like these:

[2014-09-26 16:32:23.995539] W [fuse-bridge.c:1214:fuse_err_cbk] 0-glusterfs-fuse: 6384: SETXATTR() /.gfid/5f9b6d20-a062-4168-9333-8d28f2ba2d57 => -1 (File exists)
[2014-09-26 16:32:23.995798] W [client-rpc-fops.c:256:client3_3_mknod_cbk] 0-gv2-slave-client-0: remote operation failed: File exists. Path: <gfid:855b5eda-f694-487c-adae-a4723fe6c316>/msg000002
[2014-09-26 16:32:23.996042] W [fuse-bridge.c:1214:fuse_err_cbk] 0-glusterfs-fuse: 6385: SETXATTR() /.gfid/855b5eda-f694-487c-adae-a4723fe6c316 => -1 (File exists)
[2014-09-26 16:32:24.550009] W [fuse-bridge.c:1911:fuse_create_cbk] 0-glusterfs-fuse: 6469: /.gfid/05a27020-5931-4890-9b74-a77cb1aca918 => -1 (Operation not permitted)
[2014-09-26 16:32:24.550533] W [defaults.c:1381:default_release] (-->/usr/lib64/glusterfs/3.5.2/xlator/mount/fuse.so(+0x1e7d0) [0x7fb2ebd1e7d0] (-->/usr/lib64/glusterfs/3.5.2/xlator/mount/fuse.so(free_fuse_state+0x93) [0x7fb2ebd07063] (-->/usr/lib64/libglusterfs.so.0(fd_unref+0x10e) [0x7fb2eef36fbe]))) 0-fuse: xlator does not implement release_cbk

I also noticed that at some point, rsync was returning error code 23.

Now ... I noted from the page I linked above that it requires rsync
version 3.0.7 and the version that ships with CentOS 6.5 is, wait for
it ... 3.0.6. Is this going to be the issue, or is the problem something
else?

If you need more logs, let me know. If you need a copy of my client
script that breaks it, let me know that and I'll send it along.

-- 
Cheers,
Kingsley.



More information about the Gluster-users mailing list