[Gluster-users] gdeploy not starting all the daemons for NFS-ganesha :(

Soumya Koduri skoduri at redhat.com
Wed May 10 05:15:57 UTC 2017


CCin Sac, Manisha, Arthy who could help with troubleshooting.

Thanks,
Soumya


On 05/09/2017 08:31 PM, hvjunk wrote:
> Hi there,
>
> Given the following config file, what am I doing wrong causing the error at the bottom  & no mounted /gluster_shared_storage?
>
> Hendrik
>
> [root at linked-clone-of-centos-linux ~]# cat t.conf
> [hosts]
> 10.10.10.11
> 10.10.10.12
> 10.10.10.13
>
> [backend-setup]
> devices=/dev/sdb
> mountpoints=/gluster/brick1
> brick_dirs=/gluster/brick1/one
> pools=pool1
>
> #Installing nfs-ganesha
> [yum]
> action=install
> repolist=
> gpgcheck=no
> update=no
> packages=glusterfs-ganesha
>
> [service]
> action=start
> service=glusterd
>
> [service]
> action=enable
> service=glusterd
>
> #This will create a volume. Skip this section if your volume already exists
> [volume]
> action=create
> volname=ganesha
> transport=tcp
> replica_count=3
> arbiter_count=1
> force=yes
>
> [clients]
> action=mount
> volname=ganesha
> hosts=10.10.10.11,10.10.10.12,10.10.10.13
> fstype=glusterfs
> client_mount_points=/mnt/ganesha_mnt
>
>
> #Creating a high availability cluster and exporting the volume
> [nfs-ganesha]
> action=create-cluster
> ha-name=ganesha-ha-360
> cluster-nodes=10.10.10.11,10.10.10.12
> vip=10.10.10.31,10.10.10.41
> volname=ganesha
>
>
> [nfs-ganesha]
> action=export-volume
> volname=ganesha
>
> [nfs-ganesha]
> action=refresh-config
> volname=ganesha
> [root at linked-clone-of-centos-linux ~]#
>
>
> [root at linked-clone-of-centos-linux ~]# gdeploy -c t.conf -k
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Clean up filesystem signature] ***********************************************************************
> skipping: [10.10.10.11] => (item=/dev/sdb)
> skipping: [10.10.10.12] => (item=/dev/sdb)
> skipping: [10.10.10.13] => (item=/dev/sdb)
>
> TASK [Create Physical Volume] ******************************************************************************
> changed: [10.10.10.13] => (item=/dev/sdb)
> changed: [10.10.10.11] => (item=/dev/sdb)
> changed: [10.10.10.12] => (item=/dev/sdb)
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create volume group on the disks] ********************************************************************
> changed: [10.10.10.12] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.13] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.11] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create logical volume named metadata] ****************************************************************
> changed: [10.10.10.12] => (item=GLUSTER_vg1)
> changed: [10.10.10.13] => (item=GLUSTER_vg1)
> changed: [10.10.10.11] => (item=GLUSTER_vg1)
>
> TASK [create data LV that has a size which is a multiple of stripe width] **********************************
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
>
> TASK [Convert the logical volume] **************************************************************************
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
>
> TASK [create stripe-aligned thin volume] *******************************************************************
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
>
> TASK [Change the attributes of the logical volume] *********************************************************
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=5    changed=5    unreachable=0    failed=0
> 10.10.10.12                : ok=5    changed=5    unreachable=0    failed=0
> 10.10.10.13                : ok=5    changed=5    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create an xfs filesystem] ****************************************************************************
> changed: [10.10.10.13] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
> changed: [10.10.10.12] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
> changed: [10.10.10.11] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create the backend disks, skips if present] **********************************************************
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
>
> TASK [Mount the volumes] ***********************************************************************************
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.12                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.13                : ok=2    changed=2    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Set SELinux labels on the bricks] ********************************************************************
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
>
> TASK [Restore the SELinux context] *************************************************************************
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.12                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.13                : ok=2    changed=2    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create/Enabling yum repos] ***************************************************************************
> skipping: [10.10.10.12] => (item=)
> skipping: [10.10.10.11] => (item=)
> skipping: [10.10.10.13] => (item=)
>
> TASK [Clean up the metadata] *******************************************************************************
> skipping: [10.10.10.12]
> skipping: [10.10.10.11]
> skipping: [10.10.10.13]
>
> TASK [Yum update] ******************************************************************************************
> skipping: [10.10.10.12]
> skipping: [10.10.10.11]
> skipping: [10.10.10.13]
>
> TASK [Installs/Removes a package using yum] ****************************************************************
> changed: [10.10.10.12]
> changed: [10.10.10.13]
> changed: [10.10.10.11]
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Enable or disable services] **************************************************************************
> changed: [10.10.10.12] => (item=glusterd)
> changed: [10.10.10.11] => (item=glusterd)
> changed: [10.10.10.13] => (item=glusterd)
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [start/stop/restart/reload services] ******************************************************************
> changed: [10.10.10.11] => (item=glusterd)
> changed: [10.10.10.12] => (item=glusterd)
> changed: [10.10.10.13] => (item=glusterd)
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] *************************************************************************************
>
> TASK [Create the brick dirs, skips if present] *************************************************************
> changed: [10.10.10.11] => (item=/gluster/brick1/one)
> changed: [10.10.10.12] => (item=/gluster/brick1/one)
> changed: [10.10.10.13] => (item=/gluster/brick1/one)
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [master] **********************************************************************************************
>
> TASK [Creates a Trusted Storage Pool] **********************************************************************
> changed: [10.10.10.11]
>
> TASK [Pause for a few seconds] *****************************************************************************
> Pausing for 5 seconds
> (ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort)
> ok: [10.10.10.11]
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=2    changed=1    unreachable=0    failed=0
>
>
> PLAY [master] **********************************************************************************************
>
> TASK [Creates a volume] ************************************************************************************
> changed: [10.10.10.11]
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [master] **********************************************************************************************
>
> TASK [Starts a volume] *************************************************************************************
> changed: [10.10.10.11]
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [clients] *********************************************************************************************
>
> TASK [Create the dir to mount the volume, skips if present] ************************************************
> changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [clients] *********************************************************************************************
>
> TASK [Mount the volumes, if fstype is glusterfs] ***********************************************************
> changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [clients] *********************************************************************************************
>
> TASK [Gathering Facts] *************************************************************************************
> ok: [10.10.10.11]
> ok: [10.10.10.12]
> ok: [10.10.10.13]
>
> TASK [Uncomment STATD_PORT for rpc.statd to listen on] *****************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Uncomment LOCKD_TCPPORT for rpc.lockd to listen on] **************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Uncomment LOCKD_UDPPORT for rpc.lockd to listen on] **************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Uncomment MOUNTD_PORT for rpc.mountd to listen on] ***************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Restart nfs service (RHEL 6 only)] *******************************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Restart rpc-statd service] ***************************************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Restart nfs-config service] **************************************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Restart nfs-mountd service] **************************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Restart nfslock service (RHEL 6 & 7)] ****************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> TASK [Mount the volumes if fstype is NFS] ******************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=1    changed=0    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=0    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=0    unreachable=0    failed=0
>
>
> PLAY [clients] *********************************************************************************************
>
> TASK [Mount the volumes, if fstype is CIFS] ****************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=0    changed=0    unreachable=0    failed=0
> 10.10.10.12                : ok=0    changed=0    unreachable=0    failed=0
> 10.10.10.13                : ok=0    changed=0    unreachable=0    failed=0
>
>
> PLAY [master] **********************************************************************************************
>
> TASK [Delete lines from Ganesha export file] ***************************************************************
>
> TASK [Add lines to Ganesha export file] ********************************************************************
>
> TASK [Update lines to Ganesha export file] *****************************************************************
>
> TASK [Add a block to Ganesha export file] ******************************************************************
>
> TASK [Refresh NFS-Ganesha config] **************************************************************************
> fatal: [10.10.10.11]: FAILED! => {"changed": true, "cmd": "/usr/libexec/ganesha/ganesha-ha.sh --refresh-config \"/var/run/gluster/shared_storage/nfs-ganesha\" \"ganesha\"", "delta": "0:00:00.227074", "end": "2017-05-09 16:54:45.442219", "failed": true, "rc": 1, "start": "2017-05-09 16:54:45.215145", "stderr": "grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory\ngrep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory\nError: cluster is not currently running on this node\ngrep: /var/run/gluster/shared_storage/nfs-ganesha/exports/export.ganesha.conf: No such file or directory", "stderr_lines": ["grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory", "grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory", "Error: cluster is not currently running on this node", "grep: /var/run/gluster/shared_storage/nfs-ganesha/exports/expor
>  t.ganesha.conf: No such file or directory"], "stdout": "Error: refresh-config failed on localhost.", "stdout_lines": ["Error: refresh-config failed on localhost."]}
> 	to retry, use: --limit @/tmp/tmpacx8yi/ganesha-refresh-config.retry
>
> PLAY RECAP *************************************************************************************************
> 10.10.10.11                : ok=0    changed=0    unreachable=0    failed=1
>
> Ignoring errors...
> You can view the generated configuration files inside /tmp/tmpacx8yi
> [root at linked-clone-of-centos-linux ~]#
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
>


More information about the Gluster-users mailing list