[Gluster-users] Healing issue

Miloš Kozák milos.kozak at lejmr.com
Mon Aug 17 06:28:44 UTC 2015


Hi,


Dne 17.8.2015 v 06:34 Ravishankar N napsal(a):
>
>
> On 08/16/2015 04:22 PM, Miloš Kozák wrote:
>> Hi, I have been running an glusterfs for a while, and everything 
>> works just fine even after one node failure.. However, I went for 
>> brick replacement due to my bricks were not thin-provisioned and I 
>> wanted to use snapshots. In short, whole volume went down due to heal 
>> daemon which tool all IO and all VMs running on top of that volume 
>> started to be unresponsive.
>>
>> In short, I am rebuilding the volume from scratch. I created new 
>> thinly provisioned bricks:
>> lvs:
>>   brick_s3-sata-10k                  vg_s3-sata-10k Vwi-aotz 931,25g 
>> s3-sata-10k_pool                     2,95
>>   s3-sata-10k_pool                   vg_s3-sata-10k twi-a-tz 931,25g
>>
>> vgs:
>>   vg_s3-sata-10k   1   3   0 wz--n- 931,51g 148,00m
>>
>> df:
>>   /dev/mapper/vg_s3--sata--10k-brick_s3--sata--10k 976009600 28383480 
>> 947626120   3% /gfs/s3-sata-10k
>>
>> and mounted. When I uploaded two images onto it I found there might 
>> be a problem. For the time being I run the volume in replica 2 mode 
>> on top of two servers. The files were copied from node1, and I think 
>> the files are OK on node1 only. However, the volume heal indicates 
>> everything is OK. 
> Hi,
> If I understand you correctly, are you saying that when you created a 
> new replica 2 volume with thin LVs for the bricks , mounted the volume 
> (fuse?) and copied 2 files (VM images?) via this mount, the files were 
> being written to only node1?

Mounted by fuse.
I would not say written to only one node because both nodes contain 
something, but I am quite suspicious about files on the node2.

>
> -Ravi
>> My symptoms are as follows:
>>
>> df information from both servers:
>> /dev/mapper/vg_s3--sata--10k-brick_s3--sata--10k 976009600 30754296 
>> 945255304   4% /gfs/s3-sata-10k
>> /dev/mapper/vg_s3--sata--10k-brick_s3--sata--10k 976009600 28383480 
>> 947626120   3% /gfs/s3-sata-10k
>>
>>
>>
>> [root at nodef01i ~]#  du /gfs/s3-sata-10k/
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/indices/xattrop
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/indices
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/changelogs/htime
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/changelogs/csnap
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/changelogs
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/00/00
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/00
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/landfill
>> 20480004    /gfs/s3-sata-10k/fs/.glusterfs/84/26
>> 20480004    /gfs/s3-sata-10k/fs/.glusterfs/84
>> 10240000    /gfs/s3-sata-10k/fs/.glusterfs/d0/ff
>> 10240000    /gfs/s3-sata-10k/fs/.glusterfs/d0
>> 30720008    /gfs/s3-sata-10k/fs/.glusterfs
>> 30720008    /gfs/s3-sata-10k/fs
>> 30720008    /gfs/s3-sata-10k/
>>
>> [root at nodef02i ~]# du /gfs/s3-sata-10k/
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/indices/xattrop
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/indices
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/changelogs/htime
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/changelogs/csnap
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/changelogs
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/00/00
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/00
>> 0    /gfs/s3-sata-10k/fs/.glusterfs/landfill
>> 18727172    /gfs/s3-sata-10k/fs/.glusterfs/84/26
>> 18727172    /gfs/s3-sata-10k/fs/.glusterfs/84
>> 9622016    /gfs/s3-sata-10k/fs/.glusterfs/d0/ff
>> 9622016    /gfs/s3-sata-10k/fs/.glusterfs/d0
>> 28349192    /gfs/s3-sata-10k/fs/.glusterfs
>> 28349192    /gfs/s3-sata-10k/fs
>> 28349192    /gfs/s3-sata-10k/
>>
>>
>> [root at nodef01i ~]# du /gfs/s3-sata-10k/fs/*
>> 20480004    /gfs/s3-sata-10k/fs/f1607f25aa52f4fb6f98f20ef0f3f9d7
>> 10240000    /gfs/s3-sata-10k/fs/3706a2cb0bb27ba5787b3c12388f4ebb
>>
>> [root at nodef02i ~]# du /gfs/s3-sata-10k/fs/*
>> 18727172    /gfs/s3-sata-10k/fs/f1607f25aa52f4fb6f98f20ef0f3f9d7
>> 9622016 /gfs/s3-sata-10k/fs/3706a2cb0bb27ba5787b3c12388f4ebb
>>
>>
>> [root at nodef01i ~]# ll /gfs/s3-sata-10k/fs/
>> celkem 30720004
>> -rw-r----- 2 oneadmin oneadmin 20971520512  3. srp 23.53 
>> f1607f25aa52f4fb6f98f20ef0f3f9d7
>> -rw-r----- 2 oneadmin oneadmin 10485760000 16. srp 11.23 
>> 3706a2cb0bb27ba5787b3c12388f4ebb
>>
>> [root at nodef02i ~]# ll /gfs/s3-sata-10k/fs/
>> celkem 28349188
>> -rw-r----- 2 oneadmin oneadmin 20971520512  3. srp 23.53 
>> f1607f25aa52f4fb6f98f20ef0f3f9d7
>> -rw-r----- 2 oneadmin oneadmin 10485760000 16. srp 11.22 
>> 3706a2cb0bb27ba5787b3c12388f4ebb
>>
>>
>> [root at nodef01i ~]# gluster volume heal ph-fs-0 info split-brain
>> Gathering list of split brain entries on volume ph-fs-0 has been 
>> successful
>>
>> Brick 10.11.100.1:/gfs/s3-sata-10k/fs
>> Number of entries: 0
>>
>> Brick 10.11.100.2:/gfs/s3-sata-10k/fs
>> Number of entries: 0
>>
>>
>> [root at nodef01i ~]# gluster volume heal ph-fs-0 info
>> Brick nodef01i.czprg:/gfs/s3-sata-10k/fs/
>> Number of entries: 0
>>
>> Brick nodef02i.czprg:/gfs/s3-sata-10k/fs/
>> Number of entries: 0
>>
>>
>> [root at nodef01i ~]# gluster volume status
>> Status of volume: ph-fs-0
>> Gluster process                        Port    Online    Pid
>> ------------------------------------------------------------------------------ 
>>
>> Brick 10.11.100.1:/gfs/s3-sata-10k/fs            49152    Y 3733
>> Brick 10.11.100.2:/gfs/s3-sata-10k/fs            49152    Y 64711
>> NFS Server on localhost                    2049    Y    3747
>> Self-heal Daemon on localhost                N/A    Y    3752
>> NFS Server on 10.11.100.2                2049    Y    64725
>> Self-heal Daemon on 10.11.100.2                N/A    Y    64730
>>
>> Task Status of Volume ph-fs-0
>> ------------------------------------------------------------------------------ 
>>
>> There are no active volume tasks
>>
>> [root at nodef02i ~]# gluster volume status
>> Status of volume: ph-fs-0
>> Gluster process                        Port    Online    Pid
>> ------------------------------------------------------------------------------ 
>>
>> Brick 10.11.100.1:/gfs/s3-sata-10k/fs            49152    Y 3733
>> Brick 10.11.100.2:/gfs/s3-sata-10k/fs            49152    Y 64711
>> NFS Server on localhost                    2049    Y    64725
>> Self-heal Daemon on localhost                N/A    Y    64730
>> NFS Server on 10.11.100.1                2049    Y    3747
>> Self-heal Daemon on 10.11.100.1                N/A    Y    3752
>>
>> Task Status of Volume ph-fs-0
>> ------------------------------------------------------------------------------ 
>>
>> There are no active volume tasks
>>
>>
>> [root at nodef02i ~]# rpm -qa | grep gluster
>> glusterfs-server-3.6.2-1.el6.x86_64
>> glusterfs-3.6.2-1.el6.x86_64
>> glusterfs-api-3.6.2-1.el6.x86_64
>> glusterfs-libs-3.6.2-1.el6.x86_64
>> glusterfs-cli-3.6.2-1.el6.x86_64
>> glusterfs-fuse-3.6.2-1.el6.x86_64
>>
>> What other information should I provide?
>>
>> Thanks Milos
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-users
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150817/27be29c2/attachment.html>


More information about the Gluster-users mailing list