[Gluster-users] [Nfs-ganesha-support] [SOLVED] volume start: gv01: failed: Quorum not met. Volume operation not allowed.
Daniel Gryniewicz
dang at redhat.com
Tue May 22 12:58:46 UTC 2018
Thanks, Tom. Good to know.
Daniel
On 05/22/2018 01:43 AM, TomK wrote:
> This list has been deprecated. Please subscribe to the new support list
> at lists.nfs-ganesha.org.
> Hey All,
>
> Appears I solved this one and NFS mounts now work on all my clients. No
> issues since fixing it a few hours back.
>
> RESOLUTION
>
> Auditd is to blame for the trouble. Noticed this in the logs on 2 of
> the 3 NFS servers (nfs01, nfs02, nfs03):
>
> type=AVC msg=audit(1526965320.850:4094): avc: denied { write } for
> pid=8714 comm="ganesha.nfsd" name="nfs_0" dev="dm-0" ino=201547689
> scontext=system_u:system_r:ganesha_t:s0
> tcontext=system_u:object_r:krb5_host_rcache_t:s0 tclass=file
> type=SYSCALL msg=audit(1526965320.850:4094): arch=c000003e syscall=2
> success=no exit=-13 a0=7f23b0003150 a1=2 a2=180 a3=2 items=0 ppid=1
> pid=8714 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0
> fsgid=0 tty=(none) ses=4294967295 comm="ganesha.nfsd"
> exe="/usr/bin/ganesha.nfsd" subj=system_u:system_r:ganesha_t:s0 key=(null)
> type=PROCTITLE msg=audit(1526965320.850:4094):
> proctitle=2F7573722F62696E2F67616E657368612E6E667364002D4C002F7661722F6C6F672F67616E657368612F67616E657368612E6C6F67002D66002F6574632F67616E657368612F67616E657368612E636F6E66002D4E004E49565F4556454E54
>
> type=AVC msg=audit(1526965320.850:4095): avc: denied { unlink } for
> pid=8714 comm="ganesha.nfsd" name="nfs_0" dev="dm-0" ino=201547689
> scontext=system_u:system_r:ganesha_t:s0
> tcontext=system_u:object_r:krb5_host_rcache_t:s0 tclass=file
> type=SYSCALL msg=audit(1526965320.850:4095): arch=c000003e syscall=87
> success=no exit=-13 a0=7f23b0004100 a1=7f23b0000050 a2=7f23b0004100 a3=5
> items=0 ppid=1 pid=8714 auid=4294967295 uid=0 gid=0 euid=0 suid=0
> fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295
> comm="ganesha.nfsd" exe="/usr/bin/ganesha.nfsd"
> subj=system_u:system_r:ganesha_t:s0 key=(null)
> type=PROCTITLE msg=audit(1526965320.850:4095):
> proctitle=2F7573722F62696E2F67616E657368612E6E667364002D4C002F7661722F6C6F672F67616E657368612F67616E657368612E6C6F67002D66002F6574632F67616E657368612F67616E657368612E636F6E66002D4E004E49565F4556454E54
>
>
> Fix was to adjust the SELinux rules using audit2allow.
>
> All the errors below including the one in the link below, were due to that.
>
> Turns out that when ever it worked, it hit the only working server in
> the system, nfs03. Whenever it didn't work, it was hitting the non
> working servers. So sometimes it worked, and other times it didn't. It
> looked like it was to do with Haproxy / Keepalived as well since I
> couldn't mount using the VIP but could using the host. But that wasn't
> the case either.
>
> I've also added the third brick to the Gluster FS, nfs03, trying to see
> if the backend FS was to blame since Gluster FS recommends 3 bricks
> minimum for replication, but that had no effect.
>
> In case anyone runs into this, I've added notes here as well:
>
> http://microdevsys.com/wp/kernel-nfs-nfs4_discover_server_trunking-unhandled-error-512-exiting-with-error-eio-and-mount-hangs/
>
>
> http://microdevsys.com/wp/nfs-reply-xid-3844308326-reply-err-20-auth-rejected-credentials-client-should-begin-new-session/
>
>
> The errors thrown included:
>
> NFS reply xid 3844308326 reply ERR 20: Auth Rejected Credentials (client
> should begin new session)
>
> kernel: NFS: nfs4_discover_server_trunking unhandled error -512. Exiting
> with error EIO and mount hangs
>
> + the kernel exception below.
>
More information about the Gluster-users
mailing list