[Gluster-users] No possible to mount a gluster volume via /etc/fstab?
Sherry Reese
s.reese4u at gmail.com
Fri Jan 24 01:37:13 UTC 2020
Hello everyone,
I am using the following entry on a CentOS server.
gluster01.home:/videos /data2/plex/videos glusterfs _netdev 0 0
gluster01.home:/photos /data2/plex/photos glusterfs _netdev 0 0
I am able to use sudo mount -a to mount the volumes without any problems.
When I reboot my server, nothing is mounted.
I can see errors in /var/log/glusterfs/data2-plex-photos.log:
...
[2020-01-24 01:24:18.302191] I [glusterfsd.c:2594:daemonize] 0-glusterfs:
Pid of current running process is 3679
[2020-01-24 01:24:18.310017] E [MSGID: 101075]
[common-utils.c:505:gf_resolve_ip6] 0-resolver: getaddrinfo failed
(family:2) (Name or service not known)
[2020-01-24 01:24:18.310046] E
[name.c:266:af_inet_client_get_remote_sockaddr] 0-glusterfs: DNS resolution
failed on host gluster01.home
[2020-01-24 01:24:18.310187] I [MSGID: 101190]
[event-epoll.c:682:event_dispatch_epoll_worker] 0-epoll: Started thread
with index 0
...
I am able to to nslookup on gluster01 and gluster01.home without problems,
so "DNS resolution failed" is confusing to me. What happens here?
Output of my volumes.
sudo gluster volume status
Status of volume: documents
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick gluster01.home:/data/documents 49152 0 Y
5658
Brick gluster02.home:/data/documents 49152 0 Y
5340
Brick gluster03.home:/data/documents 49152 0 Y
5305
Self-heal Daemon on localhost N/A N/A Y
5679
Self-heal Daemon on gluster03.home N/A N/A Y
5326
Self-heal Daemon on gluster02.home N/A N/A Y
5361
Task Status of Volume documents
------------------------------------------------------------------------------
There are no active volume tasks
Status of volume: photos
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick gluster01.home:/data/photos 49153 0 Y
5779
Brick gluster02.home:/data/photos 49153 0 Y
5401
Brick gluster03.home:/data/photos 49153 0 Y
5366
Self-heal Daemon on localhost N/A N/A Y
5679
Self-heal Daemon on gluster03.home N/A N/A Y
5326
Self-heal Daemon on gluster02.home N/A N/A Y
5361
Task Status of Volume photos
------------------------------------------------------------------------------
There are no active volume tasks
Status of volume: videos
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick gluster01.home:/data/videos 49154 0 Y
5883
Brick gluster02.home:/data/videos 49154 0 Y
5452
Brick gluster03.home:/data/videos 49154 0 Y
5416
Self-heal Daemon on localhost N/A N/A Y
5679
Self-heal Daemon on gluster03.home N/A N/A Y
5326
Self-heal Daemon on gluster02.home N/A N/A Y
5361
Task Status of Volume videos
------------------------------------------------------------------------------
There are no active volume tasks
On the server (Ubuntu) following versions are installed.
glusterfs-client/bionic,now 7.2-ubuntu1~bionic1 armhf [installed,automatic]
glusterfs-common/bionic,now 7.2-ubuntu1~bionic1 armhf [installed,automatic]
glusterfs-server/bionic,now 7.2-ubuntu1~bionic1 armhf [installed]
On the client (CentOS) following versions are installed.
sudo rpm -qa | grep gluster
glusterfs-client-xlators-7.2-1.el7.x86_64
glusterfs-cli-7.2-1.el7.x86_64
glusterfs-libs-7.2-1.el7.x86_64
glusterfs-7.2-1.el7.x86_64
glusterfs-api-7.2-1.el7.x86_64
libvirt-daemon-driver-storage-gluster-4.5.0-23.el7_7.3.x86_64
centos-release-gluster7-1.0-1.el7.centos.noarch
glusterfs-fuse-7.2-1.el7.x86_64
I tried to disable IPv6 on the client voa sysctl with following parameters.
net.ipv6.conf.all.disable_ipv6 = 1
net.ipv6.conf.default.disable_ipv6 = 1
That did not help.
Volumes are configured with inet.
sudo gluster volume info videos
Volume Name: videos
Type: Replicate
Volume ID: 8fddde82-66b3-447f-8860-ed3768c51876
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: gluster01.home:/data/videos
Brick2: gluster02.home:/data/videos
Brick3: gluster03.home:/data/videos
Options Reconfigured:
features.ctime: on
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off
I tried turning off ctime but that did not work either.
Any ideas? How do I do this correctly?
Cheers
Sherry
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200124/cc8a2773/attachment.html>
More information about the Gluster-users
mailing list