[Gluster-users] Can't mount using volfile
Niels de Vos
ndevos at redhat.com
Fri Oct 3 14:23:51 UTC 2014
On Fri, Oct 03, 2014 at 03:26:04PM +0200, Peter Haraldson wrote:
> Hi all!
Hi Peter!
> I'm rather new to glusterfs, trying it out for redundant storage for my very
> small company.
> I have a minimal setup of glusterfs, 2 servers (storage1 & storage2) with
> one brick each, both added to volume "testvol1". I then mount "testvol1" on
> a third server (app1). This is fine as long as I use any one servers ip and
> filesystem glusterfs, but when trying to mount the volume using a file
> "/owndata/conf/glusterfs/storage.vol", it does not work. Also I can't use
> nfs - don't need it but maybe it's related.
It is not recommended to use the volume-file for mounting. Current
versions of Gluster manage the volume-file for you, and there is no need
to make changes there, or use it directly.
Mounting with "-t glusterfs ..." is the recommended way. If you would
like to fall-back on the 2nd server while mounting, you can use the
"backupvolfile-server=storage2" mount option.
I can not say why mounting over NFS fails to work. The output of
"gluster volume status" below shows that the NFS-server is running and
listening on port 2049. You can find the logs for the nfs-server in
/var/log/gluster/nfs.log, combine that with the output of
# mount -vvv -t nfs storage1:/testvol1 /mnt/tmp
to get some ideas on what might go wrong.
HTH,
Niels
>
> So:
> "mount -t glusterfs 192.168.160.21:/testvol1 /mnt/tmp/" works. I write a
> file to /mnt/tmp/filename, then mount 192.168.12.210:/testvol1 and the newly
> created file is there.
>
> Trying to mount using config file storage.vol:
>
> mount -t glusterfs /owndata/conf/glusterfs/storage.vol /mnt/tmp
> Mount failed. Please check the log file for more details.
>
> The main error in the log is
>
> /E [client-handshake.c:1778:client_query_portmap_cbk] 0-remote1:
> failed to get the port number for remote subvolume. Please run
> 'gluster volume status' on server to see if brick process is running/.
>
> There are lots & lots of pages on the 'net about this error message, none of
> the solutions I've found has worked.
>
> CentOS 6.5 on all servers, they are all kvm hosts under oVirt (this is just
> the testing stage, will be on real iron in production).
> No firewall anywhere, selinux is permissive.
>
> *File storage.vol:*
>
> volume remote1
> type protocol/client
> option transport-type tcp
> option remote-host 192.168.12.210
> option remote-subvolume testvol1
> end-volume
>
> volume remote2
> type protocol/client
> option transport-type tcp
> option remote-host 192.168.160.21
> option remote-subvolume testvol1
> end-volume
>
> volume replicate
> type cluster/replicate
> subvolumes remote1 remote2
> end-volume
>
> volume writebehind
> type performance/write-behind
> option window-size 1MB
> subvolumes replicate
> end-volume
>
> volume cache
> type performance/io-cache
> option cache-size 256MB
> subvolumes writebehind
> end-volume
>
> *# gluster volume info*
>
> Volume Name: testvol1
> Type: Replicate
> Volume ID: bcca4aa2-46c0-44a2-8175-1305faa8b4f9
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: 192.168.12.210:/export/vdb/brick1
> Brick2: 192.168.160.21:/export/vdb/brick1
>
> *# gluster volume status*
>
> Status of volume: testvol1
> Gluster process Port Online Pid
> -------------------------------------------------------------------------
> Brick 192.168.12.210:/export/vdb/brick1 49152
> Y 1656
> Brick 192.168.160.21:/export/vdb/brick1 49152
> Y 139090
> NFS Server on localhost 2049 Y
> 1670
> Self-heal Daemon on localhost N/A Y
> 1674
> NFS Server on 192.168.160.21 2049 Y
> 1481
> Self-heal Daemon on 192.168.160.21 N/A Y
> 139105
>
> Task Status of Volume testvol1
> ------------------------------------------------------------------------------
> There are no active volume tasks
>
> *Complete log after fail:
> *
>
> [2014-10-02 14:38:22.252235] I [glusterfsd.c:2026:main]
> 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version
> 3.4.0.57rhs (/usr/sbin/glusterfs
> --fuse-mountopts=allow_other,default_permissions,max_read=131072
> --volfile=/owndata/conf/glusterfs/storage.vol
> --fuse-mountopts=allow_other,default_permissions,max_read=131072
> /mnt/glust)
> [2014-10-02 14:38:22.284438] W [options.c:848:xl_opt_validate]
> 0-writebehind: option 'window-size' is deprecated, preferred is
> 'cache-size', continuing with correction
> [2014-10-02 14:38:22.284476] W [io-cache.c:1672:init] 0-cache:
> dangling volume. check volfile
> [2014-10-02 14:38:22.294306] I [socket.c:3505:socket_init]
> 0-remote2: SSL support is NOT enabled
> [2014-10-02 14:38:22.294339] I [socket.c:3520:socket_init]
> 0-remote2: using system polling thread
> [2014-10-02 14:38:22.294832] I [socket.c:3505:socket_init]
> 0-remote1: SSL support is NOT enabled
> [2014-10-02 14:38:22.294848] I [socket.c:3520:socket_init]
> 0-remote1: using system polling thread
> [2014-10-02 14:38:22.294870] I [client.c:2171:notify] 0-remote1:
> parent translators are ready, attempting connect on transport
> [2014-10-02 14:38:22.306697] I [client.c:2171:notify] 0-remote2:
> parent translators are ready, attempting connect on transport
> Final graph:
> +------------------------------------------------------------------------------+
> 1: volume remote1
> 2: type protocol/client
> 3: option remote-subvolume testvol1
> 4: option remote-host 192.168.12.210
> 5: option transport-type socket
> 6: end-volume
> 7:
> 8: volume remote2
> 9: type protocol/client
> 10: option remote-subvolume testvol1
> 11: option remote-host 192.168.160.21
> 12: option transport-type socket
> 13: end-volume
> 14:
> 15: volume replicate
> 16: type cluster/replicate
> 17: subvolumes remote1 remote2
> 18: end-volume
> 19:
> 20: volume writebehind
> 21: type performance/write-behind
> 22: option cache-size 1MB
> 23: subvolumes replicate
> 24: end-volume
> 25:
> 26: volume cache
> 27: type performance/io-cache
> 28: option cache-size 256MB
> 29: subvolumes writebehind
> 30: end-volume
> 31:
> +------------------------------------------------------------------------------+
> [2014-10-02 14:38:22.310830] E
> [client-handshake.c:1778:client_query_portmap_cbk] 0-remote1: failed
> to get the port number for remote subvolume. Please run 'gluster
> volume status' on server to see if brick process is running.
> [2014-10-02 14:38:22.310887] I [client.c:2103:client_rpc_notify]
> 0-remote1: disconnected from 192.168.12.210:24007. Client process
> will keep trying to connect to glusterd until brick's port is
> available.
> [2014-10-02 14:38:22.311031] E
> [client-handshake.c:1778:client_query_portmap_cbk] 0-remote2: failed
> to get the port number for remote subvolume. Please run 'gluster
> volume status' on server to see if brick process is running.
> [2014-10-02 14:38:22.311059] I [client.c:2103:client_rpc_notify]
> 0-remote2: disconnected from 192.168.160.21:24007. Client process
> will keep trying to connect to glusterd until brick's port is
> available.
> [2014-10-02 14:38:22.311070] E [afr-common.c:4025:afr_notify]
> 0-replicate: All subvolumes are down. Going offline until atleast
> one of them comes back up.
> [2014-10-02 14:38:22.314827] I [fuse-bridge.c:5874:fuse_graph_setup]
> 0-fuse: switched to graph 0
> [2014-10-02 14:38:22.316140] I [fuse-bridge.c:4811:fuse_init]
> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.13
> kernel 7.13
> [2014-10-02 14:38:22.321404] W [fuse-bridge.c:1134:fuse_attr_cbk]
> 0-glusterfs-fuse: 2: LOOKUP() / => -1 (Transport endpoint is not
> connected)
> [2014-10-02 14:38:22.324731] I [fuse-bridge.c:5715:fuse_thread_proc]
> 0-fuse: unmounting /mnt/glust
> [2014-10-02 14:38:22.324931] W [glusterfsd.c:1099:cleanup_and_exit]
> (-->/lib64/libc.so.6(clone+0x6d) [0x7f6e2ec5e86d]
> (-->/lib64/libpthread.so.0(+0x79d1) [0x7f6e2f2f19d1]
> (-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xcd) [0x4052ad]))) 0-:
> received signum (15), shutting down
> [2014-10-02 14:38:22.324946] I [fuse-bridge.c:6412:fini] 0-fuse:
> Unmounting '/mnt/glust'.
>
>
> Regards
> Peter H
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
More information about the Gluster-users
mailing list