[Gluster-users] Gluster volume not mounted

Atin Mukherjee amukherj at redhat.com
Wed Jun 28 04:08:22 UTC 2017


The mount log file of the volume would help in debugging the actual cause.

On Tue, Jun 27, 2017 at 6:33 PM, Joel Diaz <mrjoeldiaz at gmail.com> wrote:

> Good morning Gluster users,
>
> I'm very new to the Gluster file system. My apologies if this is not the
> correct way to seek assistance. However, I would appreciate some insight
> into understanding the issue I have.
>
> I have three nodes running two volumes, engine and data. The third node is
> the arbiter on both volumes. Both volumes were operation fine but one of
> the volumes, data,  no longer mounts.
>
> Please see below:
>
> gluster volume info all
>
> Volume Name: data
> Type: Replicate
> Volume ID: 1d6bb110-9be4-4630-ae91-36ec1cf6cc02
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1 x (2 + 1) = 3
> Transport-type: tcp
> Bricks:
> Brick1: 192.168.170.141:/gluster_bricks/data/data
> Brick2: 192.168.170.143:/gluster_bricks/data/data
> Brick3: 192.168.170.147:/gluster_bricks/data/data (arbiter)
> Options Reconfigured:
> nfs.disable: on
> performance.readdir-ahead: on
> transport.address-family: inet
> performance.quick-read: off
> performance.read-ahead: off
> performance.io-cache: off
> performance.stat-prefetch: off
> performance.low-prio-threads: 32
> network.remote-dio: off
> cluster.eager-lock: enable
> cluster.quorum-type: auto
> cluster.server-quorum-type: server
> cluster.data-self-heal-algorithm: full
> cluster.locking-scheme: granular
> cluster.shd-max-threads: 8
> cluster.shd-wait-qlength: 10000
> features.shard: on
> user.cifs: off
> storage.owner-uid: 36
> storage.owner-gid: 36
> network.ping-timeout: 30
> performance.strict-o-direct: on
> cluster.granular-entry-heal: enable
>
> Volume Name: engine
> Type: Replicate
> Volume ID: b160f0b2-8bd3-4ff2-a07c-134cab1519dd
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1 x (2 + 1) = 3
> Transport-type: tcp
> Bricks:
> Brick1: 192.168.170.141:/gluster_bricks/engine/engine
> Brick2: 192.168.170.143:/gluster_bricks/engine/engine
> Brick3: 192.168.170.147:/gluster_bricks/engine/engine (arbiter)
> Options Reconfigured:
> nfs.disable: on
> performance.readdir-ahead: on
> transport.address-family: inet
> performance.quick-read: off
> performance.read-ahead: off
> performance.io-cache: off
> performance.stat-prefetch: off
> performance.low-prio-threads: 32
> network.remote-dio: off
> cluster.eager-lock: enable
> cluster.quorum-type: auto
> cluster.server-quorum-type: server
> cluster.data-self-heal-algorithm: full
> cluster.locking-scheme: granular
> cluster.shd-max-threads: 8
> cluster.shd-wait-qlength: 10000
> features.shard: on
> user.cifs: off
> storage.owner-uid: 36
> storage.owner-gid: 36
> network.ping-timeout: 30
> performance.strict-o-direct: on
> cluster.granular-entry-heal: enable
>
> df -h
> Filesystem                                    Size  Used Avail Use%
> Mounted on
> /dev/mapper/centos_ovirt--hyp--01-root         50G  3.9G   47G   8% /
> devtmpfs                                      7.7G     0  7.7G   0% /dev
> tmpfs                                         7.8G     0  7.8G   0%
> /dev/shm
> tmpfs                                         7.8G  8.7M  7.7G   1% /run
> tmpfs                                         7.8G     0  7.8G   0%
> /sys/fs/cgroup
> /dev/mapper/centos_ovirt--hyp--01-home         61G   33M   61G   1% /home
> /dev/mapper/gluster_vg_sdb-gluster_lv_engine   50G  8.1G   42G  17%
> /gluster_bricks/engine
> /dev/sda1                                     497M  173M  325M  35% /boot
> /dev/mapper/gluster_vg_sdb-gluster_lv_data    730G  157G  574G  22%
> /gluster_bricks/data
> tmpfs                                         1.6G     0  1.6G   0%
> /run/user/0
> ovirt-hyp-01.reis.com:engine                   50G  8.1G   42G  17%
> /rhev/data-center/mnt/glusterSD/ovirt-hyp-01.reis.com:engine
>
> gluster volume status data
> Status of volume: data
> Gluster process                             TCP Port  RDMA Port  Online
>  Pid
> ------------------------------------------------------------
> ------------------
> Brick 192.168.170.141:/gluster_bricks/data/
> data                                        49157     0          Y
> 11967
> Brick 192.168.170.143:/gluster_bricks/data/
> data                                        49157     0          Y
> 2901
> Brick 192.168.170.147:/gluster_bricks/data/
> data                                        49158     0          Y
> 2626
> Self-heal Daemon on localhost               N/A       N/A        Y
> 16211
> Self-heal Daemon on 192.168.170.147         N/A       N/A        Y
> 3402
> Self-heal Daemon on 192.168.170.143         N/A       N/A        Y
> 20254
>
> Task Status of Volume data
> ------------------------------------------------------------
> ------------------
> There are no active volume tasks
>
> gluster peer status
> Number of Peers: 2
>
> Hostname: 192.168.170.143
> Uuid: b2b30d05-cf91-4567-92fd-022575e082f5
> State: Peer in Cluster (Connected)
> Other names:
> 10.0.0.2
>
> Hostname: 192.168.170.147
> Uuid: 4e50acc4-f3cb-422d-b499-fb5796a53529
> State: Peer in Cluster (Connected)
> Other names:
> 10.0.0.3
>
> Any assistance in understanding how and why the volume no longer mounts
> and a possible resolution would be greatly appreciated.
>
> Thank you,
>
> Joel
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170628/3bd19ebd/attachment.html>


More information about the Gluster-users mailing list