[Gluster-users] gdeploy not starting all the daemons for NFS-ganesha :(
hvjunk
hvjunk at gmail.com
Tue May 9 15:01:37 UTC 2017
Hi there,
Given the following config file, what am I doing wrong causing the error at the bottom & no mounted /gluster_shared_storage?
Hendrik
[root at linked-clone-of-centos-linux ~]# cat t.conf
[hosts]
10.10.10.11
10.10.10.12
10.10.10.13
[backend-setup]
devices=/dev/sdb
mountpoints=/gluster/brick1
brick_dirs=/gluster/brick1/one
pools=pool1
#Installing nfs-ganesha
[yum]
action=install
repolist=
gpgcheck=no
update=no
packages=glusterfs-ganesha
[service]
action=start
service=glusterd
[service]
action=enable
service=glusterd
#This will create a volume. Skip this section if your volume already exists
[volume]
action=create
volname=ganesha
transport=tcp
replica_count=3
arbiter_count=1
force=yes
[clients]
action=mount
volname=ganesha
hosts=10.10.10.11,10.10.10.12,10.10.10.13
fstype=glusterfs
client_mount_points=/mnt/ganesha_mnt
#Creating a high availability cluster and exporting the volume
[nfs-ganesha]
action=create-cluster
ha-name=ganesha-ha-360
cluster-nodes=10.10.10.11,10.10.10.12
vip=10.10.10.31,10.10.10.41
volname=ganesha
[nfs-ganesha]
action=export-volume
volname=ganesha
[nfs-ganesha]
action=refresh-config
volname=ganesha
[root at linked-clone-of-centos-linux ~]#
[root at linked-clone-of-centos-linux ~]# gdeploy -c t.conf -k
PLAY [gluster_servers] *************************************************************************************
TASK [Clean up filesystem signature] ***********************************************************************
skipping: [10.10.10.11] => (item=/dev/sdb)
skipping: [10.10.10.12] => (item=/dev/sdb)
skipping: [10.10.10.13] => (item=/dev/sdb)
TASK [Create Physical Volume] ******************************************************************************
changed: [10.10.10.13] => (item=/dev/sdb)
changed: [10.10.10.11] => (item=/dev/sdb)
changed: [10.10.10.12] => (item=/dev/sdb)
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [gluster_servers] *************************************************************************************
TASK [Create volume group on the disks] ********************************************************************
changed: [10.10.10.12] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'})
changed: [10.10.10.13] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'})
changed: [10.10.10.11] => (item={u'brick': u'/dev/sdb', u'vg': u'GLUSTER_vg1'})
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [gluster_servers] *************************************************************************************
TASK [Create logical volume named metadata] ****************************************************************
changed: [10.10.10.12] => (item=GLUSTER_vg1)
changed: [10.10.10.13] => (item=GLUSTER_vg1)
changed: [10.10.10.11] => (item=GLUSTER_vg1)
TASK [create data LV that has a size which is a multiple of stripe width] **********************************
changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
TASK [Convert the logical volume] **************************************************************************
changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
TASK [create stripe-aligned thin volume] *******************************************************************
changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
TASK [Change the attributes of the logical volume] *********************************************************
changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', u'vg': u'GLUSTER_vg1'})
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=5 changed=5 unreachable=0 failed=0
10.10.10.12 : ok=5 changed=5 unreachable=0 failed=0
10.10.10.13 : ok=5 changed=5 unreachable=0 failed=0
PLAY [gluster_servers] *************************************************************************************
TASK [Create an xfs filesystem] ****************************************************************************
changed: [10.10.10.13] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
changed: [10.10.10.12] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
changed: [10.10.10.11] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [gluster_servers] *************************************************************************************
TASK [Create the backend disks, skips if present] **********************************************************
changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
TASK [Mount the volumes] ***********************************************************************************
changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=2 changed=2 unreachable=0 failed=0
10.10.10.12 : ok=2 changed=2 unreachable=0 failed=0
10.10.10.13 : ok=2 changed=2 unreachable=0 failed=0
PLAY [gluster_servers] *************************************************************************************
TASK [Set SELinux labels on the bricks] ********************************************************************
changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
TASK [Restore the SELinux context] *************************************************************************
changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=2 changed=2 unreachable=0 failed=0
10.10.10.12 : ok=2 changed=2 unreachable=0 failed=0
10.10.10.13 : ok=2 changed=2 unreachable=0 failed=0
PLAY [gluster_servers] *************************************************************************************
TASK [Create/Enabling yum repos] ***************************************************************************
skipping: [10.10.10.12] => (item=)
skipping: [10.10.10.11] => (item=)
skipping: [10.10.10.13] => (item=)
TASK [Clean up the metadata] *******************************************************************************
skipping: [10.10.10.12]
skipping: [10.10.10.11]
skipping: [10.10.10.13]
TASK [Yum update] ******************************************************************************************
skipping: [10.10.10.12]
skipping: [10.10.10.11]
skipping: [10.10.10.13]
TASK [Installs/Removes a package using yum] ****************************************************************
changed: [10.10.10.12]
changed: [10.10.10.13]
changed: [10.10.10.11]
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [gluster_servers] *************************************************************************************
TASK [Enable or disable services] **************************************************************************
changed: [10.10.10.12] => (item=glusterd)
changed: [10.10.10.11] => (item=glusterd)
changed: [10.10.10.13] => (item=glusterd)
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [gluster_servers] *************************************************************************************
TASK [start/stop/restart/reload services] ******************************************************************
changed: [10.10.10.11] => (item=glusterd)
changed: [10.10.10.12] => (item=glusterd)
changed: [10.10.10.13] => (item=glusterd)
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [gluster_servers] *************************************************************************************
TASK [Create the brick dirs, skips if present] *************************************************************
changed: [10.10.10.11] => (item=/gluster/brick1/one)
changed: [10.10.10.12] => (item=/gluster/brick1/one)
changed: [10.10.10.13] => (item=/gluster/brick1/one)
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [master] **********************************************************************************************
TASK [Creates a Trusted Storage Pool] **********************************************************************
changed: [10.10.10.11]
TASK [Pause for a few seconds] *****************************************************************************
Pausing for 5 seconds
(ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort)
ok: [10.10.10.11]
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=2 changed=1 unreachable=0 failed=0
PLAY [master] **********************************************************************************************
TASK [Creates a volume] ************************************************************************************
changed: [10.10.10.11]
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
PLAY [master] **********************************************************************************************
TASK [Starts a volume] *************************************************************************************
changed: [10.10.10.11]
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
PLAY [clients] *********************************************************************************************
TASK [Create the dir to mount the volume, skips if present] ************************************************
changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [clients] *********************************************************************************************
TASK [Mount the volumes, if fstype is glusterfs] ***********************************************************
changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [clients] *********************************************************************************************
TASK [Gathering Facts] *************************************************************************************
ok: [10.10.10.11]
ok: [10.10.10.12]
ok: [10.10.10.13]
TASK [Uncomment STATD_PORT for rpc.statd to listen on] *****************************************************
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
TASK [Uncomment LOCKD_TCPPORT for rpc.lockd to listen on] **************************************************
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
TASK [Uncomment LOCKD_UDPPORT for rpc.lockd to listen on] **************************************************
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
TASK [Uncomment MOUNTD_PORT for rpc.mountd to listen on] ***************************************************
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
TASK [Restart nfs service (RHEL 6 only)] *******************************************************************
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
TASK [Restart rpc-statd service] ***************************************************************************
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
TASK [Restart nfs-config service] **************************************************************************
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
TASK [Restart nfs-mountd service] **************************************************************************
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
TASK [Restart nfslock service (RHEL 6 & 7)] ****************************************************************
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
TASK [Mount the volumes if fstype is NFS] ******************************************************************
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=1 changed=0 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=0 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=0 unreachable=0 failed=0
PLAY [clients] *********************************************************************************************
TASK [Mount the volumes, if fstype is CIFS] ****************************************************************
skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', u'fstype': u'fuse'})
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=0 changed=0 unreachable=0 failed=0
10.10.10.12 : ok=0 changed=0 unreachable=0 failed=0
10.10.10.13 : ok=0 changed=0 unreachable=0 failed=0
PLAY [master] **********************************************************************************************
TASK [Delete lines from Ganesha export file] ***************************************************************
TASK [Add lines to Ganesha export file] ********************************************************************
TASK [Update lines to Ganesha export file] *****************************************************************
TASK [Add a block to Ganesha export file] ******************************************************************
TASK [Refresh NFS-Ganesha config] **************************************************************************
fatal: [10.10.10.11]: FAILED! => {"changed": true, "cmd": "/usr/libexec/ganesha/ganesha-ha.sh --refresh-config \"/var/run/gluster/shared_storage/nfs-ganesha\" \"ganesha\"", "delta": "0:00:00.227074", "end": "2017-05-09 16:54:45.442219", "failed": true, "rc": 1, "start": "2017-05-09 16:54:45.215145", "stderr": "grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory\ngrep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory\nError: cluster is not currently running on this node\ngrep: /var/run/gluster/shared_storage/nfs-ganesha/exports/export.ganesha.conf: No such file or directory", "stderr_lines": ["grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory", "grep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or directory", "Error: cluster is not currently running on this node", "grep: /var/run/gluster/shared_storage/nfs-ganesha/exports/export.ganesha.conf: No such file or directory"], "stdout": "Error: refresh-config failed on localhost.", "stdout_lines": ["Error: refresh-config failed on localhost."]}
to retry, use: --limit @/tmp/tmpacx8yi/ganesha-refresh-config.retry
PLAY RECAP *************************************************************************************************
10.10.10.11 : ok=0 changed=0 unreachable=0 failed=1
Ignoring errors...
You can view the generated configuration files inside /tmp/tmpacx8yi
[root at linked-clone-of-centos-linux ~]#
More information about the Gluster-users
mailing list