[Gluster-users] Replication delay
Fabio Rosati
fabio.rosati at geminformatica.it
Fri Jan 24 09:47:27 UTC 2014
Hi All,
in a distributed-replicated volume hosting some VMs disk images (GlusterFS 3.4.2 on CentOS 6.5, qemu-kvm with glusterfs native support, no fuse mount), I always get the same two files that need healing:
[root at networker ~]# gluster volume heal gv_pri info
Gathering Heal info on volume gv_pri has been successful
Brick nw1glus.gem.local:/glustexp/pri1/brick
Number of entries: 2
/alfresco.qc2
/remlog.qc2
Brick nw2glus.gem.local:/glustexp/pri1/brick
Number of entries: 2
/alfresco.qc2
/remlog.qc2
Brick nw3glus.gem.local:/glustexp/pri2/brick
Number of entries: 0
Brick nw4glus.gem.local:/glustexp/pri2/brick
Number of entries: 0
This is not a split-brain situation (I checked) and If I stop the two VMs that use these images, I get the two files healed/synced in about 15min. This is too much time, IMHO.
In this volume there are other VM images with (smaller) disk images replicated on the same bricks and they get synced "in real-time".
These are the volume's details, the host "networker" is nw1glus.gem.local :
[root at networker ~]# gluster volume info gv_pri
Volume Name: gv_pri
Type: Distributed-Replicate
Volume ID: 3d91b91e-4d72-484f-8655-e5ed8d38bb28
Status: Started
Number of Bricks: 2 x 2 = 4
Transport-type: tcp
Bricks:
Brick1: nw1glus.gem.local:/glustexp/pri1/brick
Brick2: nw2glus.gem.local:/glustexp/pri1/brick
Brick3: nw3glus.gem.local:/glustexp/pri2/brick
Brick4: nw4glus.gem.local:/glustexp/pri2/brick
Options Reconfigured:
server.allow-insecure: on
storage.owner-uid: 107
storage.owner-gid: 107
[root at networker ~]# gluster volume status gv_pri detail
Status of volume: gv_pri
------------------------------------------------------------------------------
Brick : Brick nw1glus.gem.local:/glustexp/pri1/brick
Port : 50178
Online : Y
Pid : 25721
File System : xfs
Device : /dev/mapper/vg_guests-lv_brick1
Mount Options : rw,noatime
Inode Size : 512
Disk Space Free : 168.4GB
Total Disk Space : 194.9GB
Inode Count : 102236160
Free Inodes : 102236130
------------------------------------------------------------------------------
Brick : Brick nw2glus.gem.local:/glustexp/pri1/brick
Port : 50178
Online : Y
Pid : 27832
File System : xfs
Device : /dev/mapper/vg_guests-lv_brick1
Mount Options : rw,noatime
Inode Size : 512
Disk Space Free : 168.4GB
Total Disk Space : 194.9GB
Inode Count : 102236160
Free Inodes : 102236130
------------------------------------------------------------------------------
Brick : Brick nw3glus.gem.local:/glustexp/pri2/brick
Port : 50182
Online : Y
Pid : 14571
File System : xfs
Device : /dev/mapper/vg_guests-lv_brick2
Mount Options : rw,noatime
Inode Size : 512
Disk Space Free : 418.3GB
Total Disk Space : 433.8GB
Inode Count : 227540992
Free Inodes : 227540973
------------------------------------------------------------------------------
Brick : Brick nw4glus.gem.local:/glustexp/pri2/brick
Port : 50181
Online : Y
Pid : 21942
File System : xfs
Device : /dev/mapper/vg_guests-lv_brick2
Mount Options : rw,noatime
Inode Size : 512
Disk Space Free : 418.3GB
Total Disk Space : 433.8GB
Inode Count : 227540992
Free Inodes : 227540973
fuse-mount of the gv_pri volume:
[root at networker ~]# ll -h /mnt/gluspri/
totale 37G
-rw-------. 1 qemu qemu 7,7G 24 gen 10:21 alfresco.qc2
-rw-------. 1 qemu qemu 4,2G 24 gen 10:22 check_mk-salmo.qc2
-rw-------. 1 qemu qemu 27M 23 gen 16:42 newnxserver.qc2
-rw-------. 1 qemu qemu 1,1G 23 gen 13:38 newubutest1.qc2
-rw-------. 1 qemu qemu 11G 24 gen 10:17 nxserver.qc2
-rw-------. 1 qemu qemu 8,1G 24 gen 10:17 remlog.qc2
-rw-------. 1 qemu qemu 5,6G 24 gen 10:19 ubutest1.qc2
Do you think this is the expected behaviour, maybe due to caching? What if the most updated node goes down while the VMs are running?
Thanks a lot,
Fabio Rosati
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140124/66d73551/attachment.html>
More information about the Gluster-users
mailing list