[Gluster-users] Unable to mount gfs gv0 volume Enterprise Linux Enterprise Linux Server release 5.6 (Carthage)

Lalatendu Mohanty lmohanty at redhat.com
Thu Jan 16 05:36:57 UTC 2014


On 01/16/2014 03:58 AM, Jeffrey Brewster wrote:
>
> I'm not sure why the mount is failing. I followed the quick start guide...
>
>
> Data:
> --------------
>
>
>
> 1. Info check looks good
> gluster volume info
>   
> Volume Name: gv2
> Type: Replicate
> Volume ID: ca9f2409-3004-4287-af6f-1b455048710e
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: gcvs0139:/data/gv0/brick1/app
> Brick2: gcvs4056:/data/gv0/brick2/app1
>
>
> 2. Status looks good
>
> gluster volume status
>
> Status of volume: gv2
> Gluster process                                         Port    Online  Pid
> ------------------------------------------------------------------------------
> Brick gcvs0139:/data/gv0/brick1/app                     49152   Y       7648
> Brick gcvs4056:/data/gv0/brick2/app1                    49152   Y       12005
> NFS Server on localhost                                 2049    Y
>   12017
> Self-heal Daemon on localhost                           N/A     Y       12021
> NFS Server on gcvs0139                                  2049    Y       7660
> Self-heal Daemon on gcvs0139                            N/A     Y       7664
>   
> There are no active volume tasks
>
>
> 3. peer check looks good
>
> [root at gcvs4056 /]#  gluster peer probe gcvs0139
> peer probe: success: host gcvs0139 port 24007 already in peer list
> [root at gcvs4056 /]#
>
>
>
> 4. mount fails
>
> [root at gcvs4056 /]#   mount -t glusterfs gcvs4056:/gv2 /mnt
> Mount failed. Please check the log file for more details.
> [root at gcvs4056 /]#
>
>
> 5. mount log
>
>  From the mnt.log:
> -------------
>
>
> [2014-01-15 22:19:57.751543] I [afr-common.c:3698:afr_notify] 0-gv2-replicate-0: Subvolume 'gv2-client-1' came back up; going online.
> [2014-01-15 22:19:57.751614] I [rpc-clnt.c:1676:rpc_clnt_reconfig] 0-gv2-client-0: changing port to
>   49152 (from 0)
> [2014-01-15 22:19:57.751675] I [client-handshake.c:450:client_set_lk_version_cbk] 0-gv2-client-1: Server lk version = 1
> [2014-01-15 22:19:57.751712] W [socket.c:514:__socket_rwv] 0-gv2-client-0: readv failed (No data available)
> [2014-01-15 22:19:57.759041] W [common-utils.c:2247:gf_get_reserved_ports] 0-glusterfs: could not open the file /proc/sys/net/ipv4/ip_local_reserved_ports for getting reserved ports info (No such file or directory)
> [2014-01-15 22:19:57.759080] W [common-utils.c:2280:gf_process_reserved_ports] 0-glusterfs: Not able to get reserved ports, hence there is a possibility that glusterfs may consume reserved port
> [2014-01-15 22:19:57.762259] I [client-handshake.c:1659:select_server_supported_programs] 0-gv2-client-0: Using Program GlusterFS 3.3, Num (1298437), Version (330)
> [2014-01-15 22:19:57.762974] I [client-handshake.c:1456:client_setvolume_cbk] 0-gv2-client-0: Connected to 10.131.83.139:49152,
>   attached to remote volume '/data/gv0/brick1/app'.
> [2014-01-15 22:19:57.763008] I [client-handshake.c:1468:client_setvolume_cbk] 0-gv2-client-0: Server and Client lk-version numbers are not same, reopening the fds
> [2014-01-15 22:19:57.775406] I [fuse-bridge.c:4769:fuse_graph_setup] 0-fuse: switched to graph 0
> [2014-01-15 22:19:57.775695] I [client-handshake.c:450:client_set_lk_version_cbk] 0-gv2-client-0: Server lk version = 1
> [2014-01-15 22:19:57.779538] I [fuse-bridge.c:4628:fuse_thread_proc] 0-fuse: unmounting /mnt
> [2014-01-15 22:19:57.780102] W [glusterfsd.c:1002:cleanup_and_exit] (-->/lib64/libc.so.6(clone+0x6d) [0x31f6ad40cd] (-->/lib64/libpthread.so.0 [0x31f7e0673d] (-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0x138) [0x405328]))) 0-: received signum (15), shutting down
> [2014-01-15 22:19:57.780206] I [fuse-bridge.c:5260:fini] 0-fuse: Unmounting '/mnt'.
>
>
>
>
> On Tuesday, January 14, 2014 5:12 PM, Jeffrey Brewster 
> <jab2805 at yahoo.com> wrote:
>
> Hi Ben,
>
> 1. Port 24007 is open and all iptable rules have been flushed:
>
> [root at gcvs4056 run]# telnet gcvs0139 24007
> Trying 10.131.83.139...
> Connected to gcvs0139.
> Escape character is '^]'.
>
>
> 2. gluster peer status looks good from both boxes:
>
> box1
> [root at gcvs4056 run]# gluster peer status
> Number of Peers: 1
>
> Hostname: gcvs0139
> Uuid: d40ba14d-cbb4-40e7-86a2-62afaa99af4d
> State: Peer in Cluster (Connected)
>
> box 2:
>
>
> # gluster peer status
> Number of Peers: 1
>
> Hostname: gcvs4056
> Port: 24007
> Uuid: b1aae40a-78be-4303-bf48-49fb41d6bb30
> State: Peer in Cluster (Connected)
>
> 3. selinux is disabled on both boxes.
>
>  grep dis /etc/sysconfig/selinux
> #       disabled - No SELinux policy is loaded.
> SELINUX=disabled
>
>

I can see that selinux is disabled in the config file, but this does not 
come in to effect unless you reboot the server. Check the current status 
of selinux i.e. run "getenforce" .

Also what gluster packages are installed on the client side?

>
> Thanks for your help!
>
>
>
> On Tuesday, January 14, 2014 4:54 PM, Ben Turner <bturner at redhat.com> 
> wrote:
> ----- Original Message -----
> > From: "Jeffrey Brewster" <jab2805 at yahoo.com <mailto:jab2805 at yahoo.com>>
> > To: "Ben Turner" <bturner at redhat.com <mailto:bturner at redhat.com>>
> > Cc: gluster-users at gluster.org <mailto:gluster-users at gluster.org>
> > Sent: Tuesday, January 14, 2014 4:35:30 PM
> > Subject: Re: [Gluster-users] Unable to mount gfs gv0 volume 
> Enterprise Linux    Enterprise Linux Server release 5.6
> > (Carthage)
> >
> > Hi Ben,
> >
> >
> >
> >
> > I don't have any "E" (error I assume) lines in the mnt.log file. I 
> check all
> > the log files in the /var/log/glusterfs/ dir. I restarted glusterd  
> to see
> > if I could see any errors.
> >
>
> Make sure SELinux is disabled and your firewall is open to allow 
> gluster traffic.  Have a look at:
>
> http://www.gluster.org/community/documentation/index.php/Basic_Gluster_Troubleshooting
>
> For what ports you need open. As a test I would just try disabling 
> iptables and adding in the rules after you confirm it is working.
>
> -b
>
>
> >
> >
> >
> >
> > Data:
> >
> >
> >
> > Warnings from mount log:
> > -------------
> >
> > # grep W  mnt.log  | cat -n
> >
> >
> >      1  [2014-01-14
> > 19:32:22.920069] W [common-utils.c:2247:gf_get_reserved_ports]
> > 0-glusterfs: could not open the file
> > /proc/sys/net/ipv4/ip_local_reserved_ports for getting reserv
> >
> >  2  [2014-01-14 19:32:22.920108] W
> > [common-utils.c:2280:gf_process_reserved_ports] 0-glusterfs: Not able to
> >  get reserved ports, hence there is a possibility that glusterfs may c
> >
> >  3  [2014-01-14 19:32:22.935611] W
> > [common-utils.c:2247:gf_get_reserved_ports] 0-glusterfs: could not open
> > the file /proc/sys/net/ipv4/ip_local_reserved_ports for getting reserv
> >
> >  4  [2014-01-14 19:32:22.935646] W
> > [common-utils.c:2280:gf_process_reserved_ports] 0-glusterfs: Not able to
> >  get reserved ports, hence there is a possibility that glusterfs may c
> >
> >  5  [2014-01-14 19:32:22.938783] W
> > [common-utils.c:2247:gf_get_reserved_ports] 0-glusterfs: could not open
> >  the file /proc/sys/net/ipv4/ip_local_reserved_ports for getting reserv
> >
> >  6  [2014-01-14 19:32:22.938826] W
> > [common-utils.c:2280:gf_process_reserved_ports] 0-glusterfs: Not able to
> >  get reserved ports, hence there is a possibility that glusterfs may c
> >      7  [2014-01-14 19:32:22.941076] W [socket.c:514:__socket_rwv]
> >      0-gv0-client-1: readv failed (No data available)
> >
> >  8  [2014-01-14 19:32:22.945278] W
> > [common-utils.c:2247:gf_get_reserved_ports] 0-glusterfs: could not open
> > the file /proc/sys/net/ipv4/ip_local_reserved_ports for getting reserv
> >
> >  9  [2014-01-14 19:32:22.945312] W
> > [common-utils.c:2280:gf_process_reserved_ports] 0-glusterfs: Not able to
> >  get reserved ports, hence there is a possibility that glusterfs may c
> >     10  [2014-01-14
> >  19:32:22.946921] W [socket.c:514:__socket_rwv] 0-gv0-client-0: 
> readv failed
> >  (No data available)
> >
> >  11  [2014-01-14 19:32:22.953383] W
> > [common-utils.c:2247:gf_get_reserved_ports] 0-glusterfs: could not open
> > the file /proc/sys/net/ipv4/ip_local_reserved_ports for getting reserv
> >
> >  12  [2014-01-14 19:32:22.953423] W
> > [common-utils.c:2280:gf_process_reserved_ports] 0-glusterfs: Not able to
> >  get reserved ports, hence there is a possibility that glusterfs may c
> >
> >  13  [2014-01-14 19:32:22.976633] W [glusterfsd.c:1002:cleanup_and_exit]
> > (-->/lib64/libc.so.6(clone+0x6d) [0x31f6ad40cd]
> > (-->/lib64/libpthread.so.0 [0x
> >
> >
> >
> >
> >
> >
> >
> >
> >
> > After restarting glusterd:
> > -----------------------------
> >
> >
> > # grep  E *  | grep 21:25| cat -n
> >
> >
> >      1 etc-glusterfs-glusterd.vol.log:[2014-01-14 21:25:47.637082] E
> > [rpc-transport.c:253:rpc_transport_load] 0-rpc-transport:
> > /usr/lib64/glusterfs/3.4.2/rpc-transport/rdma.so: cannot open shared
> >      object file: No such file or directory
> >      2 etc-glusterfs-glusterd.vol.log:[2014-01-14 21:25:49.940650] E
> > [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key:
> >      brick-0
> >      3 etc-glusterfs-glusterd.vol.log:[2014-01-14 21:25:49.940698] E
> > [glusterd-store.c:1858:glusterd_store_retrieve_volume] 0-: Unknown key:
> >      brick-1
> >      4 etc-glusterfs-glusterd.vol.log:[2014-01-14 21:25:52.075563] E
> > [glusterd-utils.c:3801:glusterd_nodesvc_unlink_socket_file]
> >      0-management: Failed to remove
> > /var/run/3096dde11d292c28c8c2f97101c272e8.socket error: Resource
> >      temporarily unavailable
> >      5 etc-glusterfs-glusterd.vol.log:[2014-01-14 21:25:53.084722] E
> > [glusterd-utils.c:3801:glusterd_nodesvc_unlink_socket_file]
> >      0-management: Failed to remove
> > /var/run/15f2dcd004edbff6ab31364853d6b6b0.socket error: No such file or
> >      directory
> >      6 glustershd.log:[2014-01-14 21:25:42.392401] W
> > [socket.c:1962:__socket_proto_state_machine] 0-glusterfs: reading from
> >      socket failed. Error (No data available), peer (127.0.0.1:24007)
> >      7 glustershd.log:[2014-01-14 21:25:53.476026] E
> > [afr-self-heald.c:1067:afr_find_child_position] 0-gv0-replicate-0:
> >      getxattr failed on gv0-client-0 - (Transport endpoint is not 
> connected)
> >      8 nfs.log:[2014-01-14 21:25:42.391560] W
> > [socket.c:1962:__socket_proto_state_machine] 0-glusterfs: reading from
> >      socket failed. Error (No data available), peer (127.0.0.1:24007)
> >
> >
> >
> >
> > Procs After restrt:
> >
> >
> >  ps -ef | grep gluster
> > root      6345     1  0 18:35 ?        00:00:00 /usr/sbin/glusterfsd -s
> > gcvs4056 --volfile-id gv0.gcvs4056.data-gv0-brick1-app -p
> > /var/lib/glusterd/vols/gv0/run/gcvs4056-data-gv0-brick1-app.pid -S
> > /var/run/f2339d9fa145fd28662d8b970fbd4aab.socket --brick-name
> > /data/gv0/brick1/app -l 
> /var/log/glusterfs/bricks/data-gv0-brick1-app.log
> > --xlator-option 
> *-posix.glusterd-uuid=b1aae40a-78be-4303-bf48-49fb41d6bb30
> > --brick-port 49153 --xlator-option gv0-server.listen-port=49153
> > root      7240     1  0 21:25 ?        00:00:00 /usr/sbin/glusterd
> > --pid-file=/var/run/glusterd.pid
> > root      7266     1  0 21:25 ?        00:00:00 /usr/sbin/glusterfs -s
> > localhost --volfile-id gluster/nfs -p 
> /var/lib/glusterd/nfs/run/nfs.pid -l
> > /var/log/glusterfs/nfs.log -S
> > /var/run/3096dde11d292c28c8c2f97101c272e8.socket
> > root      7273     1  0 21:25 ?        00:00:00 /usr/sbin/glusterfs -s
> > localhost --volfile-id gluster/glustershd -p
> > /var/lib/glusterd/glustershd/run/glustershd.pid -l
> > /var/log/glusterfs/glustershd.log -S
> > /var/run/15f2dcd004edbff6ab31364853d6b6b0.socket --xlator-option
> > *replicate*.node-uuid=b1aae40a-78be-4303-bf48-49fb41d6bb30
> > root      7331  5375  0 21:34 pts/1    00:00:00 grep gluster
> >
> >
> >
> >
> >
> >
> >
> > On Tuesday, January 14, 2014 4:08 PM, Ben Turner <bturner at redhat.com 
> <mailto:bturner at redhat.com>> wrote:
> >
> > ----- Original Message -----
> > > From: "Jeffrey Brewster" <jab2805 at yahoo.com 
> <mailto:jab2805 at yahoo.com>>
> > > To: "Ben Turner" <bturner at redhat.com <mailto:bturner at redhat.com>>
> > > Cc: gluster-users at gluster.org <mailto:gluster-users at gluster.org>
> > > Sent: Tuesday, January 14, 2014 3:57:24 PM
> > > Subject: Re: [Gluster-users] Unable to mount gfs gv0 volume Enterprise
> > > Linux    Enterprise Linux Server release 5.6
> > > (Carthage)
> > >
> > > Thanks Ben,
> > >
> > >
> > >
> > > I tried that  it still failed.
> >
> > As Vijay suggested have a look at /var/log/glusterfs, there should 
> be a log
> > there with the mountpoint name that should give us a clue as to what is
> > going on.  To note if there is a problem with FUSE not being loaded 
> you will
> > see something like:
> >
> > [2013-01-12 01:58:22.213417] I [glusterfsd.c:1759:main]
> > 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version
> > 3.3.0.5rhs
> > [2013-01-12 01:58:22.213831] E [mount.c:596:gf_fuse_mount] 
> 0-glusterfs-fuse:
> > cannot open /dev/fuse (No such file or directory)
> > [2013-01-12 01:58:22.213856] E [xlator.c:385:xlator_init] 0-fuse:
> > Initialization of volume 'fuse' failed, review your volfile again
> >
> > If you can't tell the problem from the log shoot out the relevant 
> line and
> > I'll have a look.
> >
> > -b
> >
> >
> > >
> > >
> > >
> > > On Tuesday, January 14, 2014 3:22 PM, Ben Turner 
> <bturner at redhat.com <mailto:bturner at redhat.com>>
> > > wrote:
> > >
> > > ----- Original Message -----
> > > > From: "Jeffrey Brewster" <jab2805 at yahoo.com 
> <mailto:jab2805 at yahoo.com>>
> > > > To: gluster-users at gluster.org <mailto:gluster-users at gluster.org>
> > > > Sent: Tuesday, January 14, 2014 1:47:55 PM
> > > > Subject: [Gluster-users] Unable to mount gfs gv0 volume 
> Enterprise Linux
> > > >     Enterprise Linux Server release 5.6
> > > > (Carthage)
> > > >
> > > >
> > > >
> > > > Hi all,
> > > >
> > > > I have been following the quick start guide as part of a POC. I 
> created a
> > > > 10GB brick to be mounted. I'm unable to mount the volume. I 
> don't see any
> > > > thing in the logs. has anyone had the same issues? I was 
> thinking I need
> > > > to
> > > > install gluster-client but I don't see in the latest release rpms.
> > > >
> > > > Data:
> > > > ===========
> > > >
> > > > OS Version:
> > > > ------------
> > > >
> > > > Description: Enterprise Linux Enterprise Linux Server release 5.6
> > > > (Carthage
> > > >
> > > >
> > > > Installed packages on both servers
> > > > ------------
> > > >
> > > > # rpm -qa | grep gluster | cat -n
> > > > 1 glusterfs-libs-3.4.2-1.el5
> > > > 2 glusterfs-3.4.2-1.el5
> > > > 3 glusterfs-cli-3.4.2-1.el5
> > > > 4 glusterfs-geo-replication-3.4.2-1.el5
> > > > 5 glusterfs-fuse-3.4.2-1.el5
> > > > 6 glusterfs-server-3.4.2-1.el5
> > > >
> > > >
> > > > gluster peer probe successful:
> > > > -----------
> > > > peer probe: success: host gcvs0139 port 24007 already in peer list
> > > >
> > > > Gluster info:
> > > > ---------
> > > > gluster volume info | cat -n
> > > > 1
> > > > 2 Volume Name: gv0
> > > > 3 Type: Replicate
> > > > 4 Volume ID: 30a27041-ba1b-456f-b0bc-d8cdd2376c2f
> > > > 5 Status: Started
> > > > 6 Number of Bricks: 1 x 2 = 2
> > > > 7 Transport-type: tcp
> > > > 8 Bricks:
> > > > 9 Brick1: gcvs0139:/data/gv0/brick1/app
> > > > 10 Brick2: gcvs4056:/data/gv0/brick1/app
> > > >
> > > >
> > > > Mount Failure:
> > > > ----------
> > > >
> > > >
> > > > [root at gcvs4056 <mailto:root at gcvs4056> jbrewster]# mount -t 
> glusterfs gcvs4056:/gv0 /mnt
> > > > Mount failed. Please check the log file for more details.
> > > >
> > >
> > > I bet you need to modprobe the fuse module, in el5 its not loaded by
> > > default.
> > >
> > >
> > > -b
> > >
> > > >
> > > >
> > > > _______________________________________________
> > > > Gluster-users mailing list
> > > > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
> > > > http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140116/3506d827/attachment.html>


More information about the Gluster-users mailing list