[Gluster-users] gdeploy, Centos7 & Ansible 2.3
hvjunk
hvjunk at gmail.com
Mon May 8 08:35:59 UTC 2017
> On 08 May 2017, at 09:34 , knarra <knarra at redhat.com> wrote:
> Hi,
>
> There is a new version of gdeploy built where the above seen issues are fixed. Can you please update gdeploy to the version below [1] and run the test again ?
> [1] https://copr-be.cloud.fedoraproject.org/results/sac/gdeploy/epel-7-x86_64/00547404-gdeploy/gdeploy-2.0.2-6.noarch.rpm <https://copr-be.cloud.fedoraproject.org/results/sac/gdeploy/epel-7-x86_64/00547404-gdeploy/gdeploy-2.0.2-6.noarch.rpm>
> Thanks
>
> kasturi
Thank you Kasturi,
Please se the run with errors below, using this conf file:
===<snip>===
[hosts]
10.10.10.11
10.10.10.12
10.10.10.13
[backend-setup]
devices=/dev/sdb
mountpoints=/gluster/brick1
brick_dirs=/gluster/brick1/one
pools=pool1
#Installing nfs-ganesha
[yum]
action=install
repolist=
gpgcheck=no
update=no
packages=glusterfs-ganesha
#This will create a volume. Skip this section if your volume already exists
[volume]
action=create
volname=ganesha
transport=tcp
replica_count=3
arbiter=1
force=yes
#Creating a high availability cluster and exporting the volume
[nfs-ganesha]
action=create-cluster
ha-name=ganesha-ha-360
cluster-nodes=10.10.10.11,10.10.10.12
vip=10.10.10.31,10.10.10.41
volname=ganesha
===<snip>===
<quote>
[root at linked-clone-of-centos-linux ~]# gdeploy -c t.conf
ERROR! no action detected in task. This often indicates a misspelled module name, or incorrect module path.
The error appears to have been in '/tmp/tmpvhTM5i/pvcreate.yml': line 16, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
# Create pv on all the disks
- name: Create Physical Volume
^ here
The error appears to have been in '/tmp/tmpvhTM5i/pvcreate.yml': line 16, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
# Create pv on all the disks
- name: Create Physical Volume
^ here
Ignoring errors...
ERROR! no action detected in task. This often indicates a misspelled module name, or incorrect module path.
The error appears to have been in '/tmp/tmpvhTM5i/vgcreate.yml': line 8, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Create volume group on the disks
^ here
The error appears to have been in '/tmp/tmpvhTM5i/vgcreate.yml': line 8, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Create volume group on the disks
^ here
Ignoring errors...
ERROR! no action detected in task. This often indicates a misspelled module name, or incorrect module path.
The error appears to have been in '/tmp/tmpvhTM5i/auto_lvcreate_for_gluster.yml': line 7, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Create logical volume named metadata
^ here
The error appears to have been in '/tmp/tmpvhTM5i/auto_lvcreate_for_gluster.yml': line 7, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Create logical volume named metadata
^ here
Ignoring errors...
PLAY [gluster_servers] **********************************************************************************************************************************************************************
TASK [Create an xfs filesystem] *************************************************************************************************************************************************************
failed: [10.10.10.11] (item=/dev/GLUSTER_vg1/GLUSTER_lv1) => {"failed": true, "item": "/dev/GLUSTER_vg1/GLUSTER_lv1", "msg": "Device /dev/GLUSTER_vg1/GLUSTER_lv1 not found."}
failed: [10.10.10.13] (item=/dev/GLUSTER_vg1/GLUSTER_lv1) => {"failed": true, "item": "/dev/GLUSTER_vg1/GLUSTER_lv1", "msg": "Device /dev/GLUSTER_vg1/GLUSTER_lv1 not found."}
failed: [10.10.10.12] (item=/dev/GLUSTER_vg1/GLUSTER_lv1) => {"failed": true, "item": "/dev/GLUSTER_vg1/GLUSTER_lv1", "msg": "Device /dev/GLUSTER_vg1/GLUSTER_lv1 not found."}
to retry, use: --limit @/tmp/tmpvhTM5i/fscreate.retry
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=0 changed=0 unreachable=0 failed=1
10.10.10.12 : ok=0 changed=0 unreachable=0 failed=1
10.10.10.13 : ok=0 changed=0 unreachable=0 failed=1
Ignoring errors...
PLAY [gluster_servers] **********************************************************************************************************************************************************************
TASK [Create the backend disks, skips if present] *******************************************************************************************************************************************
ok: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
ok: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
ok: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
TASK [Mount the volumes] ********************************************************************************************************************************************************************
failed: [10.10.10.13] (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) => {"failed": true, "item": {"device": "/dev/GLUSTER_vg1/GLUSTER_lv1", "path": "/gluster/brick1"}, "msg": "Error mounting /gluster/brick1: mount: special device /dev/GLUSTER_vg1/GLUSTER_lv1 does not exist\n"}
failed: [10.10.10.11] (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) => {"failed": true, "item": {"device": "/dev/GLUSTER_vg1/GLUSTER_lv1", "path": "/gluster/brick1"}, "msg": "Error mounting /gluster/brick1: mount: special device /dev/GLUSTER_vg1/GLUSTER_lv1 does not exist\n"}
failed: [10.10.10.12] (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'}) => {"failed": true, "item": {"device": "/dev/GLUSTER_vg1/GLUSTER_lv1", "path": "/gluster/brick1"}, "msg": "Error mounting /gluster/brick1: mount: special device /dev/GLUSTER_vg1/GLUSTER_lv1 does not exist\n"}
to retry, use: --limit @/tmp/tmpvhTM5i/mount.retry
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=1 changed=0 unreachable=0 failed=1
10.10.10.12 : ok=1 changed=0 unreachable=0 failed=1
10.10.10.13 : ok=1 changed=0 unreachable=0 failed=1
Ignoring errors...
PLAY [gluster_servers] **********************************************************************************************************************************************************************
TASK [Set SELinux labels on the bricks] *****************************************************************************************************************************************************
changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
TASK [Restore the SELinux context] **********************************************************************************************************************************************************
changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', u'path': u'/gluster/brick1'})
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=2 changed=2 unreachable=0 failed=0
10.10.10.12 : ok=2 changed=2 unreachable=0 failed=0
10.10.10.13 : ok=2 changed=2 unreachable=0 failed=0
PLAY [gluster_servers] **********************************************************************************************************************************************************************
TASK [Create/Enabling yum repos] ************************************************************************************************************************************************************
skipping: [10.10.10.11] => (item=)
skipping: [10.10.10.12] => (item=)
skipping: [10.10.10.13] => (item=)
TASK [Clean up the metadata] ****************************************************************************************************************************************************************
skipping: [10.10.10.12]
skipping: [10.10.10.11]
skipping: [10.10.10.13]
TASK [Yum update] ***************************************************************************************************************************************************************************
skipping: [10.10.10.13]
skipping: [10.10.10.11]
skipping: [10.10.10.12]
TASK [Installs/Removes a package using yum] *************************************************************************************************************************************************
ok: [10.10.10.11]
ok: [10.10.10.13]
ok: [10.10.10.12]
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=1 changed=0 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=0 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=0 unreachable=0 failed=0
PLAY [gluster_servers] **********************************************************************************************************************************************************************
TASK [start/stop/restart/reload services] ***************************************************************************************************************************************************
ok: [10.10.10.11] => (item=glusterd)
ok: [10.10.10.13] => (item=glusterd)
ok: [10.10.10.12] => (item=glusterd)
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=1 changed=0 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=0 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=0 unreachable=0 failed=0
PLAY [gluster_servers] **********************************************************************************************************************************************************************
TASK [Create the brick dirs, skips if present] **********************************************************************************************************************************************
ok: [10.10.10.11] => (item=/gluster/brick1/one)
ok: [10.10.10.12] => (item=/gluster/brick1/one)
ok: [10.10.10.13] => (item=/gluster/brick1/one)
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=1 changed=0 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=0 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=0 unreachable=0 failed=0
ERROR! no action detected in task. This often indicates a misspelled module name, or incorrect module path.
The error appears to have been in '/tmp/tmpvhTM5i/gluster-peer-probe.yml': line 7, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Creates a Trusted Storage Pool
^ here
The error appears to have been in '/tmp/tmpvhTM5i/gluster-peer-probe.yml': line 7, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Creates a Trusted Storage Pool
^ here
Ignoring errors...
ERROR! no action detected in task. This often indicates a misspelled module name, or incorrect module path.
The error appears to have been in '/tmp/tmpvhTM5i/gluster-volume-create.yml': line 7, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Creates a volume
^ here
The error appears to have been in '/tmp/tmpvhTM5i/gluster-volume-create.yml': line 7, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Creates a volume
^ here
Ignoring errors...
ERROR! no action detected in task. This often indicates a misspelled module name, or incorrect module path.
The error appears to have been in '/tmp/tmpvhTM5i/gluster-volume-start.yml': line 7, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Starts a volume
^ here
The error appears to have been in '/tmp/tmpvhTM5i/gluster-volume-start.yml': line 7, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Starts a volume
^ here
Ignoring errors...
PLAY [cluster_nodes] ************************************************************************************************************************************************************************
TASK [Check if nfs-ganesha is installed] ****************************************************************************************************************************************************
[WARNING]: Consider using yum, dnf or zypper module rather than running rpm
changed: [10.10.10.12]
changed: [10.10.10.11]
TASK [fail] *********************************************************************************************************************************************************************************
skipping: [10.10.10.11]
skipping: [10.10.10.12]
TASK [Check if corosync is installed] *******************************************************************************************************************************************************
changed: [10.10.10.12]
changed: [10.10.10.11]
TASK [fail] *********************************************************************************************************************************************************************************
skipping: [10.10.10.11]
skipping: [10.10.10.12]
TASK [Check if pacemaker is installed] ******************************************************************************************************************************************************
changed: [10.10.10.12]
changed: [10.10.10.11]
TASK [fail] *********************************************************************************************************************************************************************************
skipping: [10.10.10.11]
skipping: [10.10.10.12]
TASK [Check if libntirpc is installed] ******************************************************************************************************************************************************
changed: [10.10.10.11]
changed: [10.10.10.12]
TASK [fail] *********************************************************************************************************************************************************************************
skipping: [10.10.10.11]
skipping: [10.10.10.12]
TASK [Check if pcs is installed] ************************************************************************************************************************************************************
changed: [10.10.10.12]
changed: [10.10.10.11]
TASK [fail] *********************************************************************************************************************************************************************************
skipping: [10.10.10.11]
skipping: [10.10.10.12]
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=5 changed=5 unreachable=0 failed=0
10.10.10.12 : ok=5 changed=5 unreachable=0 failed=0
PLAY [cluster_nodes] ************************************************************************************************************************************************************************
TASK [Stop kernel NFS] **********************************************************************************************************************************************************************
ok: [10.10.10.12]
ok: [10.10.10.11]
TASK [Stop network manager service] *********************************************************************************************************************************************************
changed: [10.10.10.11]
changed: [10.10.10.12]
TASK [Disable network manager service] ******************************************************************************************************************************************************
changed: [10.10.10.11]
changed: [10.10.10.12]
TASK [Start network service] ****************************************************************************************************************************************************************
ok: [10.10.10.12]
ok: [10.10.10.11]
TASK [Enable network service] ***************************************************************************************************************************************************************
ok: [10.10.10.11]
ok: [10.10.10.12]
TASK [Start pcsd service] *******************************************************************************************************************************************************************
changed: [10.10.10.11]
changed: [10.10.10.12]
TASK [Enable pcsd service] ******************************************************************************************************************************************************************
changed: [10.10.10.12]
changed: [10.10.10.11]
TASK [Enable pacemaker service] *************************************************************************************************************************************************************
changed: [10.10.10.12]
changed: [10.10.10.11]
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=8 changed=5 unreachable=0 failed=0
10.10.10.12 : ok=8 changed=5 unreachable=0 failed=0
PLAY [master_node] **************************************************************************************************************************************************************************
TASK [Generate ssh key in one of the nodes in HA cluster] ***********************************************************************************************************************************
changed: [10.10.10.11]
TASK [Copy the public key to the local] *****************************************************************************************************************************************************
changed: [10.10.10.11]
TASK [Copy the private key to the local] ****************************************************************************************************************************************************
changed: [10.10.10.11]
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=3 changed=3 unreachable=0 failed=0
PLAY [gluster_servers] **********************************************************************************************************************************************************************
TASK [Copy the public key to remote nodes] **************************************************************************************************************************************************
changed: [10.10.10.11] => (item=10.10.10.11)
changed: [10.10.10.13] => (item=10.10.10.11)
changed: [10.10.10.12] => (item=10.10.10.11)
TASK [Copy the private key to remote node] **************************************************************************************************************************************************
changed: [10.10.10.11] => (item=10.10.10.11)
changed: [10.10.10.12] => (item=10.10.10.11)
changed: [10.10.10.13] => (item=10.10.10.11)
TASK [Deploy the pubkey ~/root/.ssh/authorized_keys on all nodes] ***************************************************************************************************************************
changed: [10.10.10.13]
changed: [10.10.10.12]
changed: [10.10.10.11]
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=3 changed=3 unreachable=0 failed=0
10.10.10.12 : ok=3 changed=3 unreachable=0 failed=0
10.10.10.13 : ok=3 changed=3 unreachable=0 failed=0
ERROR! no action detected in task. This often indicates a misspelled module name, or incorrect module path.
The error appears to have been in '/tmp/tmpvhTM5i/gluster-shared-volume-mount.yml': line 7, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Make the volume a gluster shared volume
^ here
The error appears to have been in '/tmp/tmpvhTM5i/gluster-shared-volume-mount.yml': line 7, column 5, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Make the volume a gluster shared volume
^ here
Ignoring errors...
PLAY [gluster_servers] **********************************************************************************************************************************************************************
TASK [Create a user hacluster in all the cluster nodes if not already present] **************************************************************************************************************
changed: [10.10.10.11]
changed: [10.10.10.13]
changed: [10.10.10.12]
TASK [Set the hacluster user the same password on all machines] *****************************************************************************************************************************
changed: [10.10.10.12]
changed: [10.10.10.13]
changed: [10.10.10.11]
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=2 changed=2 unreachable=0 failed=0
10.10.10.12 : ok=2 changed=2 unreachable=0 failed=0
10.10.10.13 : ok=2 changed=2 unreachable=0 failed=0
PLAY [gluster_servers] **********************************************************************************************************************************************************************
TASK [Pcs cluster authenticate the hacluster users in all the nodes] ************************************************************************************************************************
changed: [10.10.10.13] => (item=10.10.10.11)
changed: [10.10.10.11] => (item=10.10.10.11)
changed: [10.10.10.12] => (item=10.10.10.11)
changed: [10.10.10.13] => (item=10.10.10.12)
changed: [10.10.10.11] => (item=10.10.10.12)
changed: [10.10.10.12] => (item=10.10.10.12)
TASK [Pause for a few seconds] **************************************************************************************************************************************************************
Pausing for 3 seconds
(ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort)
ok: [10.10.10.11]
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=2 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [master_node] **************************************************************************************************************************************************************************
TASK [Create configuration directory for nfs-ganesha] ***************************************************************************************************************************************
changed: [10.10.10.11]
TASK [Copy ganesha.conf to config directory on shared volume] *******************************************************************************************************************************
changed: [10.10.10.11]
TASK [Creating ganesha-ha.conf file] ********************************************************************************************************************************************************
changed: [10.10.10.11]
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=3 changed=3 unreachable=0 failed=0
PLAY [gluster_servers] **********************************************************************************************************************************************************************
TASK [Define service port] ******************************************************************************************************************************************************************
changed: [10.10.10.11]
changed: [10.10.10.12]
changed: [10.10.10.13]
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0
10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0
PLAY [master_node] **************************************************************************************************************************************************************************
TASK [Enable nfs-ganesha] *******************************************************************************************************************************************************************
changed: [10.10.10.11]
TASK [Pause for 30 seconds (takes a while to enable NFS Ganesha)] ***************************************************************************************************************************
Pausing for 30 seconds
(ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort)
ok: [10.10.10.11]
TASK [Check NFS Ganesha status] *************************************************************************************************************************************************************
changed: [10.10.10.11]
TASK [Report NFS Ganesha status] ************************************************************************************************************************************************************
ok: [10.10.10.11] => {
"changed": false,
"msg": "Cluster HA Status: FAILOVER"
}
TASK [Report NFS Ganesha status (If any errors)] ********************************************************************************************************************************************
skipping: [10.10.10.11]
PLAY RECAP **********************************************************************************************************************************************************************************
10.10.10.11 : ok=4 changed=2 unreachable=0 failed=0
ERROR! no action detected in task. This often indicates a misspelled module name, or incorrect module path.
The error appears to have been in '/tmp/tmpvhTM5i/gluster-volume-export-ganesha.yml': line 7, column 7, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Exports/Unexports volume through NFS-Ganesha
^ here
The error appears to have been in '/tmp/tmpvhTM5i/gluster-volume-export-ganesha.yml': line 7, column 7, but may
be elsewhere in the file depending on the exact syntax problem.
The offending line appears to be:
tasks:
- name: Exports/Unexports volume through NFS-Ganesha
^ here
Ignoring errors...
[root at linked-clone-of-centos-linux ~]#
</quote>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170508/bc1656c9/attachment.html>
More information about the Gluster-users
mailing list