[Bugs] [Bug 1222409] nfs-ganesha: HA failover happens but I/O does not move ahead when volume has two mounts and I/O going on both mounts

bugzilla at redhat.com bugzilla at redhat.com
Tue May 19 06:45:23 UTC 2015


https://bugzilla.redhat.com/show_bug.cgi?id=1222409



--- Comment #6 from Meghana <mmadhusu at redhat.com> ---
I saw cache_invalidation related errors in gfapi.log. This is what I did on my 
4 node set up.

1. Cloned gluster master ( no fix has gone except for Disable_ACL) and
installed it on all the machines.
2. Cloned NFS-Ganesha-2.3-dev-3 and applied Soumya's patch that is waiting to
merged on NFS-Ganesha upstream,
https://review.gerrithub.io/#/c/228614/
and installed NFS-Ganesha on all the nodes.

My four node set up has a volume testvol, that looks like this,
Volume Name: testvol
Type: Distribute
Volume ID: b021c59b-16ad-4eaf-9fc3-d0eb563a9ab0
Status: Started
Number of Bricks: 4
Transport-type: tcp
Bricks:
Brick1: rhs1:/brick/brick1
Brick2: rhs2:/brick/brick1
Brick3: rhs3:/brick/brick1
Brick4: rhs4:/brick/brick1
Options Reconfigured:
ganesha.enable: on
features.cache-invalidation: on
nfs.disable: on
performance.readdir-ahead: on
nfs-ganesha: enable


VIP of rhs1 : 10.70.40.173
VIP of rhs2 : 10.70.40. 174

I mounted the volume "testvol" on two different clients.

Client 1 :

10.70.40.173:/testvol on /mnt type nfs
(rw,vers=4,addr=10.70.40.173,clientaddr=10.70.43.78)

Client 2:
10.70.40.174:/testvol on /mnt type nfs
(rw,vers=4,addr=10.70.40.174,clientaddr=10.70.42.117)


And I created two directories dir and dir1 on the mount point.

On client1, I ran iozone by executing the following command,
cd /mnt/dir 
time iozone -a -f rhs1.ioz

On client2, I ran iozone by executing the following command,
time iozone -a -f rhs3.ioz

After a few seconds, I killed ganesha on rhs1 ( VIP 10.70.40.173 ),
I saw a delay for a minute and I/Os resumed after that.

pcs status looked like this,

Online: [ rhs1 rhs2 rhs3 rhs4 ]

Full list of resources:

 Clone Set: nfs-mon-clone [nfs-mon]
     Started: [ rhs1 rhs2 rhs3 rhs4 ]
 Clone Set: nfs-grace-clone [nfs-grace]
     Started: [ rhs1 rhs2 rhs3 rhs4 ]
 rhs1-cluster_ip-1    (ocf::heartbeat:IPaddr):    Started rhs4 
 rhs1-trigger_ip-1    (ocf::heartbeat:Dummy):    Started rhs4 
 rhs2-cluster_ip-1    (ocf::heartbeat:IPaddr):    Started rhs2 
 rhs2-trigger_ip-1    (ocf::heartbeat:Dummy):    Started rhs2 
 rhs3-cluster_ip-1    (ocf::heartbeat:IPaddr):    Started rhs3 
 rhs3-trigger_ip-1    (ocf::heartbeat:Dummy):    Started rhs3 
 rhs4-cluster_ip-1    (ocf::heartbeat:IPaddr):    Started rhs4 
 rhs4-trigger_ip-1    (ocf::heartbeat:Dummy):    Started rhs4 
 rhs1-dead_ip-1    (ocf::heartbeat:Dummy):    Started rhs1 


rhs1 had failed over to rhs4.

After around 9 minutes, iozone was successfully completed on both the clients.

Client 1 :
iozone test complete.

real    9m46.894s
user    0m3.679s
sys    2m3.828s


Client 2 :
iozone test complete.

real    7m9.126s
user    0m2.482s
sys    0m56.361s


After the tests finished, I checked the pcs status again,
Online: [ rhs1 rhs2 rhs3 rhs4 ]

Full list of resources:

 Clone Set: nfs-mon-clone [nfs-mon]
     Started: [ rhs1 rhs2 rhs3 rhs4 ]
 Clone Set: nfs-grace-clone [nfs-grace]
     Started: [ rhs1 rhs2 rhs3 rhs4 ]
 rhs1-cluster_ip-1    (ocf::heartbeat:IPaddr):    Started rhs4 
 rhs1-trigger_ip-1    (ocf::heartbeat:Dummy):    Started rhs4 
 rhs2-cluster_ip-1    (ocf::heartbeat:IPaddr):    Started rhs2 
 rhs2-trigger_ip-1    (ocf::heartbeat:Dummy):    Started rhs2 
 rhs3-cluster_ip-1    (ocf::heartbeat:IPaddr):    Started rhs3 
 rhs3-trigger_ip-1    (ocf::heartbeat:Dummy):    Started rhs3 
 rhs4-cluster_ip-1    (ocf::heartbeat:IPaddr):    Started rhs4 
 rhs4-trigger_ip-1    (ocf::heartbeat:Dummy):    Started rhs4 
 rhs1-dead_ip-1    (ocf::heartbeat:Dummy):    Started rhs1 

NFS-Ganesha was running on all the other three nodes.

-- 
You are receiving this mail because:
You are the assignee for the bug.


More information about the Bugs mailing list