[Gluster-users] gdeploy not starting all the daemons for NFS-ganesha :(

Manisha Saini msaini at redhat.com
Wed May 10 07:42:27 UTC 2017


Hi hvjunk

>From the logs it seems like the ganesha cluster is not yet created/running and performing refresh config is failing.As the cluster is not running and volume "ganesha" is not yet exported due to this performing refresh config on volume will fail.
For creating ganesha cluster,there are few pre-requisite of adding services like  nlm,nfs,rpc-bind,high-availability,mountd,rquota to firewalld.
After executing create-cluster block ,It will validate the HA Status of ganesha via gdeploy.If the ganesha cluster is up and running it will reflect HA status as "HEALTHY".
After then performing refresh-config on nfs exported volume will succeed.


Attaching the sample config file to create ganesha cluster and exporting ganesha volume to client -

[hosts]
dhcp37-102.lab.eng.blr.redhat.com
dhcp37-92.lab.eng.blr.redhat.com
dhcp37-119.lab.eng.blr.redhat.com
dhcp37-122.lab.eng.blr.redhat.com


[backend-setup]
devices=/dev/sdb,/dev/sdc,/dev/sdd
vgs=vg1,vg2,vg3
pools=pool1,pool2,pool3
lvs=lv1,lv2,lv3
mountpoints=/gluster/brick1,/gluster/brick2,/gluster/brick3
brick_dirs=/gluster/brick1/1,/gluster/brick2/1,/gluster/brick3/1

[yum]
action=install
repolist=
gpgcheck=no
update=no
packages=glusterfs-ganesha

[firewalld]
action=add
ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp
services=glusterfs,nlm,nfs,rpc-bind,high-availability,mountd,rquota

[volume]
action=create
volname=ganesha
transport=tcp
replica_count=2
force=yes

[nfs-ganesha]
action=create-cluster
ha-name=ganesha-ha-360
cluster-nodes=dhcp37-102.lab.eng.blr.redhat.com,dhcp37-92.lab.eng.blr.redhat.com,dhcp37-119.lab.eng.blr.redhat.com,dhcp37-122.lab.eng.blr.redhat.com
vip=10.70.36.217,10.70.36.218,10.70.36.219,10.70.36.220
volname=ganesha  #This will export volume via nfs-ganesha

[clients]
action=mount
volname=10.70.37.153:/ganesha  #VIP from which the volume needs to be mounted to client
hosts=10.70.46.30    #Client IP
client_mount_points=/mnt/ganesha
fstype=nfs
options=rw,relatime,vers=4.0,rsize=1048576,wsize=1048576,namlen=255,hard,proto=tcp,port=0,timeo=600,retrans=2,sec=sys,clientaddr=10.70.44.153,local_lock=none,addr=10.70.46.30

Regards,
Manisha Saini

----- Original Message -----
From: "Soumya Koduri" <skoduri at redhat.com>
To: "hvjunk" <hvjunk at gmail.com>, gluster-users at gluster.org, "Sachidananda URS" <surs at redhat.com>, "Manisha Saini" <msaini at redhat.com>, "Arthy Loganathan" <aloganat at redhat.com>
Sent: Wednesday, May 10, 2017 10:45:57 AM
Subject: Re: [Gluster-users] gdeploy not starting all the daemons for NFS-ganesha :(

CCin Sac, Manisha, Arthy who could help with troubleshooting.

Thanks,
Soumya


On 05/09/2017 08:31 PM, hvjunk wrote:
> Hi there,
>
> Given the following config file, what am I doing wrong causing the error at the bottom  & no mounted /gluster_shared_storage?
>
> Hendrik
>
> [root at linked-clone-of-centos-linux ~]# cat t.conf
> [hosts]
> 10.10.10.11
> 10.10.10.12
> 10.10.10.13
>
> [backend-setup]
> devices=/dev/sdb
> mountpoints=/gluster/brick1
> brick_dirs=/gluster/brick1/one
> pools=pool1
>
> #Installing nfs-ganesha
> [yum]
> action=install
> repolist=
> gpgcheck=no
> update=no
> packages=glusterfs-ganesha
>
> [service]
> action=start
> service=glusterd
>
> [service]
> action=enable
> service=glusterd
>
> #This will create a volume. Skip this section if your volume already exists
> [volume]
> action=create
> volname=ganesha
> transport=tcp
> replica_count=3
> arbiter_count=1
> force=yes
>
> [clients]
> action=mount
> volname=ganesha
> hosts=10.10.10.11,10.10.10.12,10.10.10.13
> fstype=glusterfs
> client_mount_points=/mnt/ganesha_mnt
>
>
> #Creating a high availability cluster and exporting the volume
> [nfs-ganesha]
> action=create-cluster
> ha-name=ganesha-ha-360
> cluster-nodes=10.10.10.11,10.10.10.12
> vip=10.10.10.31,10.10.10.41
> volname=ganesha
>
>
> [nfs-ganesha]
> action=export-volume
> volname=ganesha
>
> [nfs-ganesha]
> action=refresh-config
> volname=ganesha
> [root at linked-clone-of-centos-linux ~]#
>
>
> [root at linked-clone-of-centos-linux ~]# gdeploy -c t.conf -k
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Clean up filesystem signature] ***********************************************************************
> skipping: [10.10.10.11] => (item=/dev/sdb)
> skipping: [10.10.10.12] => (item=/dev/sdb)
> skipping: [10.10.10.13] => (item=/dev/sdb)
>
> TASK [Create Physical Volume] ******************************************************************************
> changed: [10.10.10.13] => (item=/dev/sdb)
> changed: [10.10.10.11] => (item=/dev/sdb)
> changed: [10.10.10.12] => (item=/dev/sdb)
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create volume group on the disks] ********************************************************************
> changed: [10.10.10.12] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.13] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.11] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create logical volume named metadata] ****************************************************************
> changed: [10.10.10.12] => (item=GLUSTER_vg1)
> changed: [10.10.10.13] => (item=GLUSTER_vg1)
> changed: [10.10.10.11] => (item=GLUSTER_vg1)
>
> TASK [create data LV that has a size which is a multiple of stripe width] **********************************
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
>
> TASK [Convert the logical volume] **************************************************************************
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
>
> TASK [create stripe-aligned thin volume] *******************************************************************
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
>
> TASK [Change the attributes of the logical volume] *********************************************************
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=5    changed=5    unreachable=0    failed=0
> 10.10.10.12                : ok=5    changed=5    unreachable=0    failed=0
> 10.10.10.13                : ok=5    changed=5    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create an xfs filesystem] ****************************************************************************
> changed: [10.10.10.13] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
> changed: [10.10.10.12] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
> changed: [10.10.10.11] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create the backend disks, skips if present] **********************************************************
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
>
> TASK [Mount the volumes] ***********************************************************************************
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.12                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.13                : ok=2    changed=2    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Set SELinux labels on the bricks] ********************************************************************
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
>
> TASK [Restore the SELinux context] *************************************************************************
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.12                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.13                : ok=2    changed=2    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create/Enabling yum repos] ***************************************************************************
> skipping: [10.10.10.12] => (item=)
> skipping: [10.10.10.11] => (item=)
> skipping: [10.10.10.13] => (item=)
>
> TASK [Clean up the metadata] *******************************************************************************
> skipping: [10.10.10.12]
> skipping: [10.10.10.11]
> skipping: [10.10.10.13]
>
> TASK [Yum update] ******************************************************************************************
> skipping: [10.10.10.12]
> skipping: [10.10.10.11]
> skipping: [10.10.10.13]
>
> TASK [Installs/Removes a package using yum] ****************************************************************
> changed: [10.10.10.12]
> changed: [10.10.10.13]
> changed: [10.10.10.11]
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Enable or disable services] **************************************************************************
> changed: [10.10.10.12] => (item=glusterd)
> changed: [10.10.10.11] => (item=glusterd)
> changed: [10.10.10.13] => (item=glusterd)
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [start/stop/restart/reload services] ******************************************************************
> changed: [10.10.10.11] => (item=glusterd)
> changed: [10.10.10.12] => (item=glusterd)
> changed: [10.10.10.13] => (item=glusterd)
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create the brick dirs, skips if present] *************************************************************
> changed: [10.10.10.11] => (item=/gluster/brick1/one)
> changed: [10.10.10.12] => (item=/gluster/brick1/one)
> changed: [10.10.10.13] => (item=/gluster/brick1/one)
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [master] **********************************************************************************************
>
> TASK [Creates a Trusted Storage Pool] **********************************************************************
> changed: [10.10.10.11]
>
> TASK [Pause for a few seconds] *****************************************************************************
> Pausing for 5 seconds
> (ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort)
> ok: [10.10.10.11]
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=2    changed=1    unreachable=0    failed=0
>
>
> PLAY [master] **********************************************************************************************
>
> TASK [Creates a volume] ************************************************************************************
> changed: [10.10.10.11]
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [master] **********************************************************************************************
>
> TASK [Starts a volume] *************************************************************************************
> changed: [10.10.10.11]
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [clients] *********************************************************************************************
>
> TASK [Create the dir to mount the volume, skips if present] ************************************************
> changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [clients] *********************************************************************************************
>
> TASK [Mount the volumes, if fstype is glusterfs] ***********************************************************
> changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [clients] *********************************************************************************************
>
> TASK [Gathering Facts] *************************************************************************************
> ok: [10.10.10.11]
> ok: [10.10.10.12]
> ok: [10.10.10.13]
>
> TASK [Uncomment STATD_PORT for rpc.statd to listen on] *****************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Uncomment LOCKD_TCPPORT for rpc.lockd to listen on] **************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Uncomment LOCKD_UDPPORT for rpc.lockd to listen on] **************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Uncomment MOUNTD_PORT for rpc.mountd to listen on] ***************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Restart nfs service (RHEL 6 only)] *******************************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Restart rpc-statd service] ***************************************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Restart nfs-config service] **************************************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Restart nfs-mountd service] **************************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Restart nfslock service (RHEL 6 & 7)] ****************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Mount the volumes if fstype is NFS] ******************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=0    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=0    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=0    unreachable=0    failed=0
>
>
> PLAY [clients] *********************************************************************************************
>
> TASK [Mount the volumes, if fstype is CIFS] ****************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=0    changed=0    unreachable=0    failed=0
> 10.10.10.12                : ok=0    changed=0    unreachable=0    failed=0
> 10.10.10.13                : ok=0    changed=0    unreachable=0    failed=0
>
>
> PLAY [master] **********************************************************************************************
>
> TASK [Delete lines from Ganesha export file] ***************************************************************
>
> TASK [Add lines to Ganesha export file] ********************************************************************
>
> TASK [Update lines to Ganesha export file] *****************************************************************
>
> TASK [Add a block to Ganesha export file] ******************************************************************
>
> TASK [Refresh NFS-Ganesha config] **************************************************************************
> fatal: [10.10.10.11]: FAILED! => {"changed": true, "cmd": "/usr/libexec/ganesha/ganesha-ha.sh --refresh-config \"/var/run/gluster/shared_storage/nfs-ganesha\" \"ganesha\"", "delta": "0:00:00.227074", "end": "2017-05-09 16:54:45.442219", "failed": true, "rc": 1, "start": "2017-05-09 16:54:45.215145", "stderr": "grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory\ngrep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory\nError: cluster is not currently running on this node\ngrep: /var/run/gluster/shared_storage/nfs-ganesha/exports/export.ganesha.conf: No such file or directory", "stderr_lines": ["grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory", "grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory", "Error: cluster is not currently running on this node", "grep: /var/run/gluster/shared_storage/nfs-ganesha/exports/expor
>  t.ganesha.conf: No such file or directory"], "stdout": "Error: refresh-config failed on localhost.", "stdout_lines": ["Error: refresh-config failed on localhost."]}
> 	to retry, use: --limit @/tmp/tmpacx8yi/ganesha-refresh-config.retry
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=0    changed=0    unreachable=0    failed=1
>
> Ignoring errors...
> You can view the generated configuration files inside /tmp/tmpacx8yi
> [root at linked-clone-of-centos-linux ~]#
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
>


More information about the Gluster-users mailing list