[Gluster-users] NFS not start on localhost

Demeter Tibor tdemeter at itsmart.hu
Mon Oct 20 10:53:46 UTC 2014


Hi,

Thank you for you reply.

I did your recommendations, but there are no changes.

In the nfs.log there are no new things.


[root at node0 glusterfs]# reboot
Connection to 172.16.0.10 closed by remote host.
Connection to 172.16.0.10 closed.
[tdemeter at sirius-31 ~]$ ssh root at 172.16.0.10
root at 172.16.0.10's password: 
Last login: Mon Oct 20 11:02:13 2014 from 192.168.133.106
[root at node0 ~]# systemctl status nfs.target 
nfs.target - Network File System Server
   Loaded: loaded (/usr/lib/systemd/system/nfs.target; disabled)
   Active: inactive (dead)

[root at node0 ~]# gluster volume status engine
Status of volume: engine
Gluster process						Port	Online	Pid
------------------------------------------------------------------------------
Brick gs00.itsmart.cloud:/gluster/engine0		50160	Y	3271
Brick gs01.itsmart.cloud:/gluster/engine1		50160	Y	595
NFS Server on localhost					N/A	N	N/A
Self-heal Daemon on localhost				N/A	Y	3286
NFS Server on gs01.itsmart.cloud			2049	Y	6951
Self-heal Daemon on gs01.itsmart.cloud			N/A	Y	6958
 
Task Status of Volume engine
------------------------------------------------------------------------------
There are no active volume tasks
 
[root at node0 ~]# systemctl status 
Display all 262 possibilities? (y or n)
[root at node0 ~]# systemctl status nfs-lock
nfs-lock.service - NFS file locking service.
   Loaded: loaded (/usr/lib/systemd/system/nfs-lock.service; enabled)
   Active: inactive (dead)

[root at node0 ~]# systemctl stop nfs-lock
[root at node0 ~]# systemctl restart gluster
glusterd.service    glusterfsd.service  gluster.mount       
[root at node0 ~]# systemctl restart gluster
glusterd.service    glusterfsd.service  gluster.mount       
[root at node0 ~]# systemctl restart glusterfsd.service 
[root at node0 ~]# systemctl restart glusterd.service 
[root at node0 ~]# gluster volume status engine
Status of volume: engine
Gluster process						Port	Online	Pid
------------------------------------------------------------------------------
Brick gs00.itsmart.cloud:/gluster/engine0		50160	Y	5140
Brick gs01.itsmart.cloud:/gluster/engine1		50160	Y	2037
NFS Server on localhost					N/A	N	N/A
Self-heal Daemon on localhost				N/A	N	N/A
NFS Server on gs01.itsmart.cloud			2049	Y	6951
Self-heal Daemon on gs01.itsmart.cloud			N/A	Y	6958
 

Any other idea?

Tibor








----- Eredeti üzenet -----
> On Mon, Oct 20, 2014 at 09:04:2.8AM +0200, Demeter Tibor wrote:
> > Hi,
> > 
> > This is the full nfs.log after delete & reboot.
> > It is refers to portmap registering problem.
> > 
> > [root at node0 glusterfs]# cat nfs.log
> > [2014-10-20 06:48:43.221136] I [glusterfsd.c:1959:main]
> > 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.5.2
> > (/usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
> > /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log -S
> > /var/run/567e0bba7ad7102eae3049e2ad6c3ed7.socket)
> > [2014-10-20 06:48:43.224444] I [socket.c:3561:socket_init]
> > 0-socket.glusterfsd: SSL support is NOT enabled
> > [2014-10-20 06:48:43.224475] I [socket.c:3576:socket_init]
> > 0-socket.glusterfsd: using system polling thread
> > [2014-10-20 06:48:43.224654] I [socket.c:3561:socket_init] 0-glusterfs: SSL
> > support is NOT enabled
> > [2014-10-20 06:48:43.224667] I [socket.c:3576:socket_init] 0-glusterfs:
> > using system polling thread
> > [2014-10-20 06:48:43.235876] I
> > [rpcsvc.c:2127:rpcsvc_set_outstanding_rpc_limit] 0-rpc-service: Configured
> > rpc.outstanding-rpc-limit with value 16
> > [2014-10-20 06:48:43.254087] I [socket.c:3561:socket_init]
> > 0-socket.nfs-server: SSL support is NOT enabled
> > [2014-10-20 06:48:43.254116] I [socket.c:3576:socket_init]
> > 0-socket.nfs-server: using system polling thread
> > [2014-10-20 06:48:43.255241] I [socket.c:3561:socket_init]
> > 0-socket.nfs-server: SSL support is NOT enabled
> > [2014-10-20 06:48:43.255264] I [socket.c:3576:socket_init]
> > 0-socket.nfs-server: using system polling thread
> > [2014-10-20 06:48:43.257279] I [socket.c:3561:socket_init]
> > 0-socket.nfs-server: SSL support is NOT enabled
> > [2014-10-20 06:48:43.257315] I [socket.c:3576:socket_init]
> > 0-socket.nfs-server: using system polling thread
> > [2014-10-20 06:48:43.258135] I [socket.c:3561:socket_init] 0-socket.NLM:
> > SSL support is NOT enabled
> > [2014-10-20 06:48:43.258157] I [socket.c:3576:socket_init] 0-socket.NLM:
> > using system polling thread
> > [2014-10-20 06:48:43.293724] E
> > [rpcsvc.c:1314:rpcsvc_program_register_portmap] 0-rpc-service: Could not
> > register with portmap
> > [2014-10-20 06:48:43.293760] E [nfs.c:332:nfs_init_versions] 0-nfs: Program
> > NLM4 registration failed
> 
> The above line suggests that there already is a service registered at
> portmapper for the NLM4 program/service. This happens when the kernel
> module 'lockd' is loaded. The kernel NFS-client and NFS-server depend on
> this, but unfortunately it conflicts with the Gluster/nfs server.
> 
> Could you verify that the module is loaded?
>  - use 'lsmod | grep lockd' to check the modules
>  - use 'rpcinfo | grep nlockmgr' to check the rpcbind registrations
> 
> Make sure that you do not mount any NFS exports on the Gluster server.
> Unmount all NFS mounts.
> 
> You mentioned you are running CentOS-7, which is systemd based. You
> should be able to stop any conflicting NFS services like this:
> 
>  # systemctl stop nfs-lock.service
>  # systemctl stop nfs.target
>  # systemctl disable nfs.target
> 
> If all these services cleanup themselves, you should be able to start
> the Gluster/nfs service:
> 
>   # systemctl restart glusterd.service
> 
> In case some bits are still lingering around, it might be easier to
> reboot after disabling the 'nfs.target'.
> 
> > [2014-10-20 06:48:43.293771] E [nfs.c:1312:init] 0-nfs: Failed to
> > initialize protocols
> > [2014-10-20 06:48:43.293777] E [xlator.c:403:xlator_init] 0-nfs-server:
> > Initialization of volume 'nfs-server' failed, review your volfile again
> > [2014-10-20 06:48:43.293783] E [graph.c:307:glusterfs_graph_init]
> > 0-nfs-server: initializing translator failed
> > [2014-10-20 06:48:43.293789] E [graph.c:502:glusterfs_graph_activate]
> > 0-graph: init failed
> > pending frames:
> > frame : type(0) op(0)
> > 
> > patchset: git://git.gluster.com/glusterfs.git
> > signal received: 11
> > time of crash: 2014-10-20 06:48:43configuration details:
> > argp 1
> > backtrace 1
> > dlfcn 1
> > fdatasync 1
> > libpthread 1
> > llistxattr 1
> > setfsid 1
> > spinlock 1
> > epoll.h 1
> > xattr.h 1
> > st_atim.tv_nsec 1
> > package-string: glusterfs 3.5.2
> > [root at node0 glusterfs]# systemctl status portma
> > portma.service
> >    Loaded: not-found (Reason: No such file or directory)
> >    Active: inactive (dead)
> > 
> > 
> > 
> > Also I have checked the rpcbind service.
> > 
> > [root at node0 glusterfs]# systemctl status rpcbind.service
> > rpcbind.service - RPC bind service
> >    Loaded: loaded (/usr/lib/systemd/system/rpcbind.service; enabled)
> >    Active: active (running) since h 2014-10-20 08:48:39 CEST; 2min 52s ago
> >   Process: 1940 ExecStart=/sbin/rpcbind -w ${RPCBIND_ARGS} (code=exited,
> >   status=0/SUCCESS)
> >  Main PID: 1946 (rpcbind)
> >    CGroup: /system.slice/rpcbind.service
> >            └─1946 /sbin/rpcbind -w
> > 
> > okt 20 08:48:39 node0.itsmart.cloud systemd[1]: Starting RPC bind
> > service...
> > okt 20 08:48:39 node0.itsmart.cloud systemd[1]: Started RPC bind service.
> > 
> > The restart does not solve this problem.
> > 
> > 
> > I think this is the problem. Why are "exited" the portmap status?
> 
> The 'portmap' service has been replaced with 'rpcbind' since RHEL-6.
> They have the same functionality, 'rpcbind' just happens to be the newer
> version.
> 
> Did you file a bug for this already? As Vijay mentions, this crash seems
> to happen because the Gluster/nfs service fails to initialize correctly
> and then fails to cleanup correctly. The cleanup should get fixed, and
> we should also give an easier to understand error message.
> 
> Thanks,
> Niels
> 
> > 
> > 
> > On node1 is ok:
> > 
> > [root at node1 ~]# systemctl status rpcbind.service
> > rpcbind.service - RPC bind service
> >    Loaded: loaded (/usr/lib/systemd/system/rpcbind.service; enabled)
> >    Active: active (running) since p 2014-10-17 19:15:21 CEST; 2 days ago
> >  Main PID: 1963 (rpcbind)
> >    CGroup: /system.slice/rpcbind.service
> >            └─1963 /sbin/rpcbind -w
> > 
> > okt 17 19:15:21 node1.itsmart.cloud systemd[1]: Starting RPC bind
> > service...
> > okt 17 19:15:21 node1.itsmart.cloud systemd[1]: Started RPC bind service.
> > 
> > 
> > 
> > Thanks in advance
> > 
> > Tibor
> > 
> > 
> > 
> > ----- Eredeti üzenet -----
> > > On 10/19/2014 06:56 PM, Niels de Vos wrote:
> > > > On Sat, Oct 18, 2014 at 01:24:12PM +0200, Demeter Tibor wrote:
> > > >> Hi,
> > > >>
> > > >> [root at node0 ~]# tail -n 20 /var/log/glusterfs/nfs.log
> > > >> [2014-10-18 07:41:06.136035] E [graph.c:307:glusterfs_graph_init]
> > > >> 0-nfs-server: initializing translator failed
> > > >> [2014-10-18 07:41:06.136040] E [graph.c:502:glusterfs_graph_activate]
> > > >> 0-graph: init failed
> > > >> pending frames:
> > > >> frame : type(0) op(0)
> > > >>
> > > >> patchset: git://git.gluster.com/glusterfs.git
> > > >> signal received: 11
> > > >> time of crash: 2014-10-18 07:41:06configuration details:
> > > >> argp 1
> > > >> backtrace 1
> > > >> dlfcn 1
> > > >> fdatasync 1
> > > >> libpthread 1
> > > >> llistxattr 1
> > > >> setfsid 1
> > > >> spinlock 1
> > > >> epoll.h 1
> > > >> xattr.h 1
> > > >> st_atim.tv_nsec 1
> > > >> package-string: glusterfs 3.5.2
> > > >
> > > > This definitely is a gluster/nfs issue. For whatever reasone, the
> > > > gluster/nfs server crashes :-/ The log does not show enough details,
> > > > some more lines before this are needed.
> > > >
> > > 
> > > I wonder if the crash is due to a cleanup after the translator
> > > initialization failure. The complete logs might help in understanding
> > > why the initialization failed.
> > > 
> > > -Vijay
> > > 
> > > 
>


More information about the Gluster-users mailing list