[Gluster-users] Can't mount using volfile

Peter Haraldson registry at tuxit.se
Fri Oct 3 13:26:04 UTC 2014


Hi all!
I'm rather new to glusterfs, trying it out for redundant storage for my 
very small company.
I have a minimal setup of glusterfs, 2 servers (storage1 & storage2) 
with one brick each, both added to volume "testvol1". I then  mount 
"testvol1" on a third server (app1). This is fine as long as I use any 
one servers ip and filesystem glusterfs, but when trying to mount the 
volume using a file "/owndata/conf/glusterfs/storage.vol", it does not 
work. Also I can't use nfs - don't need it but maybe it's related.

So:
"mount -t glusterfs  192.168.160.21:/testvol1 /mnt/tmp/" works. I write 
a file to /mnt/tmp/filename, then mount 192.168.12.210:/testvol1 and the 
newly created file is there.

Trying to mount using config file storage.vol:

    mount -t glusterfs /owndata/conf/glusterfs/storage.vol /mnt/tmp
    Mount failed. Please check the log file for more details.

The main error in the log is

    /E [client-handshake.c:1778:client_query_portmap_cbk] 0-remote1:
    failed to get the port number for remote subvolume. Please run
    'gluster volume status' on server to see if brick process is running/.

There are lots & lots of pages on the 'net about this error message, 
none of the solutions I've found has worked.

CentOS 6.5 on all servers, they are all kvm hosts under oVirt (this is 
just the testing stage, will be on real iron in production).
No firewall anywhere, selinux is permissive.

*File storage.vol:*

    volume remote1
       type protocol/client
       option transport-type tcp
       option remote-host 192.168.12.210
       option remote-subvolume testvol1
    end-volume

    volume remote2
       type protocol/client
       option transport-type tcp
       option remote-host 192.168.160.21
       option remote-subvolume testvol1
    end-volume

    volume replicate
       type cluster/replicate
       subvolumes remote1 remote2
    end-volume

    volume writebehind
       type performance/write-behind
       option window-size 1MB
       subvolumes replicate
    end-volume

    volume cache
       type performance/io-cache
       option cache-size 256MB
       subvolumes writebehind
    end-volume

*# gluster volume info*

    Volume Name: testvol1
    Type: Replicate
    Volume ID: bcca4aa2-46c0-44a2-8175-1305faa8b4f9
    Status: Started
    Number of Bricks: 1 x 2 = 2
    Transport-type: tcp
    Bricks:
    Brick1: 192.168.12.210:/export/vdb/brick1
    Brick2: 192.168.160.21:/export/vdb/brick1

*# gluster volume status*

    Status of volume: testvol1
    Gluster process                                         Port Online  Pid
    -------------------------------------------------------------------------
    Brick 192.168.12.210:/export/vdb/brick1                 49152
    Y       1656
    Brick 192.168.160.21:/export/vdb/brick1                 49152
    Y       139090
    NFS Server on localhost                                 2049   
    Y       1670
    Self-heal Daemon on localhost                           N/A    
    Y       1674
    NFS Server on 192.168.160.21                            2049    Y
           1481
    Self-heal Daemon on 192.168.160.21                      N/A    
    Y       139105

    Task Status of Volume testvol1
    ------------------------------------------------------------------------------
    There are no active volume tasks

*Complete log after fail:
*

    [2014-10-02 14:38:22.252235] I [glusterfsd.c:2026:main]
    0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version
    3.4.0.57rhs (/usr/sbin/glusterfs
    --fuse-mountopts=allow_other,default_permissions,max_read=131072
    --volfile=/owndata/conf/glusterfs/storage.vol
    --fuse-mountopts=allow_other,default_permissions,max_read=131072
    /mnt/glust)
    [2014-10-02 14:38:22.284438] W [options.c:848:xl_opt_validate]
    0-writebehind: option 'window-size' is deprecated, preferred is
    'cache-size', continuing with correction
    [2014-10-02 14:38:22.284476] W [io-cache.c:1672:init] 0-cache:
    dangling volume. check volfile
    [2014-10-02 14:38:22.294306] I [socket.c:3505:socket_init]
    0-remote2: SSL support is NOT enabled
    [2014-10-02 14:38:22.294339] I [socket.c:3520:socket_init]
    0-remote2: using system polling thread
    [2014-10-02 14:38:22.294832] I [socket.c:3505:socket_init]
    0-remote1: SSL support is NOT enabled
    [2014-10-02 14:38:22.294848] I [socket.c:3520:socket_init]
    0-remote1: using system polling thread
    [2014-10-02 14:38:22.294870] I [client.c:2171:notify] 0-remote1:
    parent translators are ready, attempting connect on transport
    [2014-10-02 14:38:22.306697] I [client.c:2171:notify] 0-remote2:
    parent translators are ready, attempting connect on transport
    Final graph:
    +------------------------------------------------------------------------------+
       1: volume remote1
       2:     type protocol/client
       3:     option remote-subvolume testvol1
       4:     option remote-host 192.168.12.210
       5:     option transport-type socket
       6: end-volume
       7:
       8: volume remote2
       9:     type protocol/client
      10:     option remote-subvolume testvol1
      11:     option remote-host 192.168.160.21
      12:     option transport-type socket
      13: end-volume
      14:
      15: volume replicate
      16:     type cluster/replicate
      17:     subvolumes remote1 remote2
      18: end-volume
      19:
      20: volume writebehind
      21:     type performance/write-behind
      22:     option cache-size 1MB
      23:     subvolumes replicate
      24: end-volume
      25:
      26: volume cache
      27:     type performance/io-cache
      28:     option cache-size 256MB
      29:     subvolumes writebehind
      30: end-volume
      31:
    +------------------------------------------------------------------------------+
    [2014-10-02 14:38:22.310830] E
    [client-handshake.c:1778:client_query_portmap_cbk] 0-remote1: failed
    to get the port number for remote subvolume. Please run 'gluster
    volume status' on server to see if brick process is running.
    [2014-10-02 14:38:22.310887] I [client.c:2103:client_rpc_notify]
    0-remote1: disconnected from 192.168.12.210:24007. Client process
    will keep trying to connect to glusterd until brick's port is
    available.
    [2014-10-02 14:38:22.311031] E
    [client-handshake.c:1778:client_query_portmap_cbk] 0-remote2: failed
    to get the port number for remote subvolume. Please run 'gluster
    volume status' on server to see if brick process is running.
    [2014-10-02 14:38:22.311059] I [client.c:2103:client_rpc_notify]
    0-remote2: disconnected from 192.168.160.21:24007. Client process
    will keep trying to connect to glusterd until brick's port is
    available.
    [2014-10-02 14:38:22.311070] E [afr-common.c:4025:afr_notify]
    0-replicate: All subvolumes are down. Going offline until atleast
    one of them comes back up.
    [2014-10-02 14:38:22.314827] I [fuse-bridge.c:5874:fuse_graph_setup]
    0-fuse: switched to graph 0
    [2014-10-02 14:38:22.316140] I [fuse-bridge.c:4811:fuse_init]
    0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.13
    kernel 7.13
    [2014-10-02 14:38:22.321404] W [fuse-bridge.c:1134:fuse_attr_cbk]
    0-glusterfs-fuse: 2: LOOKUP() / => -1 (Transport endpoint is not
    connected)
    [2014-10-02 14:38:22.324731] I [fuse-bridge.c:5715:fuse_thread_proc]
    0-fuse: unmounting /mnt/glust
    [2014-10-02 14:38:22.324931] W [glusterfsd.c:1099:cleanup_and_exit]
    (-->/lib64/libc.so.6(clone+0x6d) [0x7f6e2ec5e86d]
    (-->/lib64/libpthread.so.0(+0x79d1) [0x7f6e2f2f19d1]
    (-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xcd) [0x4052ad]))) 0-:
    received signum (15), shutting down
    [2014-10-02 14:38:22.324946] I [fuse-bridge.c:6412:fini] 0-fuse:
    Unmounting '/mnt/glust'.


Regards
Peter H

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20141003/4fa2b1fd/attachment.html>


More information about the Gluster-users mailing list