[Gluster-users] recovering gluster volume || startup failure

srinivas jonn jmsrinivas at yahoo.com
Mon Jun 3 09:49:55 UTC 2013


Krish,
 
this is giving general volume information , can the state of volume known from any specific logs?
 
 
#gluster volume info gvol1
Volume Name: gvol1
Type: Distribute
Volume ID: aa25aa58-d191-432a-a84b-325051347af6
Status: Stopped
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: 10.0.0.30:/export/brick1
Options Reconfigured:
nfs.addr-namelookup: off
nfs.port: 2049

  

________________________________
 From: Krishnan Parthasarathi <kparthas at redhat.com>
To: srinivas jonn <jmsrinivas at yahoo.com> 
Cc: gluster-users at gluster.org 
Sent: Monday, 3 June 2013 3:14 PM
Subject: Re: [Gluster-users] recovering gluster volume || startup failure
  

Srinivas,

Could you paste the output of "gluster volume info gvol1"?
This should give us an idea as to what was the state of the volume
before the power loss.

thanks,
krish

----- Original Message -----
> Hello Gluster users:
> sorry for long post, I have run out of ideas here, kindly let me know if i am
> looking at right places for logs and any suggested actions.....thanks
> a sudden power loss casued hard reboot - now the volume does not start
> Glusterfs- 3.3.1 on Centos 6.1 transport: TCP
> sharing volume over NFS for VM storage - VHD Files
> Type: distributed - only 1 node (brick)
> XFS (LVM)
> mount /dev/datastore1/mylv1 /export/brick1 - mounts VHD files.......is there
> a way to recover these files?
> cat export-brick1.log
> [2013-06-02 09:29:00.832914] I [glusterfsd.c:1666:main]
> 0-/usr/sbin/glusterfsd: Started running /usr/sbin/glusterfsd version 3.3.1
> [2013-06-02 09:29:00.845515] I [graph.c:241:gf_add_cmdline_options]
> 0-gvol1-server: adding option 'listen-port' for volume 'gvol1-server' with
> value '24009'
> [2013-06-02 09:29:00.845558] I [graph.c:241:gf_add_cmdline_options]
> 0-gvol1-posix: adding option 'glusterd-uuid' for volume 'gvol1-posix' with
> value '16ee7a4e-ee9b-4543-bd61-9b444100693d'
> [2013-06-02 09:29:00.846654] W [options.c:782:xl_opt_validate]
> 0-gvol1-server: option 'listen-port' is deprecated, preferred is
> 'transport.socket.listen-port', continuing with correction
> Given volfile:
> +------------------------------------------------------------------------------+
> 1: volume gvol1-posix
> 2: type storage/posix
> 3: option directory /export/brick1
> 4: option volume-id aa25aa58-d191-432a-a84b-325051347af6
> 5: end-volume
> 6:
> 7: volume gvol1-access-control
> 8: type features/access-control
> 9: subvolumes gvol1-posix
> 10: end-volume
> 11:
> 12: volume gvol1-locks
> 13: type features/locks
> 14: subvolumes gvol1-access-control
> ----------
> -----------------
> 
> 46: option transport-type tcp
> 47: option auth.login./export/brick1.allow
> 6c4653bb-b708-46e8-b3f9-177b4cdbbf28
> 48: option auth.login.6c4653bb-b708-46e8-b3f9-177b4cdbbf28.password
> 091ae3b1-40c2-4d48-8870-6ad7884457ac
> 49: option auth.addr./export/brick1.allow *
> 50: subvolumes /export/brick1
> 51: end-volume
> +------------------------------------------------------------------------------+
> [2013-06-02 09:29:03.963001] W [socket.c:410:__socket_keepalive] 0-socket:
> failed to set keep idle on socket 8
> [2013-06-02 09:29:03.963046] W [socket.c:1876:socket_server_event_handler]
> 0-socket.glusterfsd: Failed to set keep-alive: Operation not supported
> [2013-06-02 09:29:04.850120] I [server-handshake.c:571:server_setvolume]
> 0-gvol1-server: accepted client from
> iiclab-oel1-9347-2013/06/02-09:29:00:835397-gvol1-client-0-0 (version:
> 3.3.1)
> [2013-06-02 09:32:16.973786] W [glusterfsd.c:831:cleanup_and_exit]
> (-->/usr/lib64/libgfrpc.so.0(rpcsvc_notify+0x93) [0x30cac0a5b3]
> (-->/usr/lib64/libgfrpc.so.0(rpcsvc_handle_rpc_call+0x293) [0x30cac0a443]
> (-->/usr/sbin/glusterfsd(glusterfs_handle_terminate+0x15) [0x40a955]))) 0-:
> received signum (15), shutting down
> [2013-06-02 09:32:16.973895] W [glusterfsd.c:831:cleanup_and_exit]
> (-->/lib64/libc.so.6(clone+0x6d) [0x3ef56e68ed] (-->/lib64/libpthread.so.0()
> [0x3ef5a077e1] (-->/usr/sbin/glusterfsd(glusterfs_sigwaiter+0xdd)
> [0x405d4d]))) 0-: received signum (15), shutting down
> NFS LOG
> [2013-06-02 09:29:00.918906] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
> 0-gvol1-client-0: changing port to 24009 (from 0)
> [2013-06-02 09:29:03.963023] W [socket.c:410:__socket_keepalive] 0-socket:
> failed to set keep idle on socket 8
> [2013-06-02 09:29:03.963062] W [socket.c:1876:socket_server_event_handler]
> 0-socket.glusterfsd: Failed to set keep-alive: Operation not supported
> [2013-06-02 09:29:04.849941] I
> [client-handshake.c:1636:select_server_supported_programs] 0-gvol1-client-0:
> Using Program GlusterFS 3.3.1, Num (1298437), Version (330)
> [2013-06-02 09:29:04.853016] I [client-handshake.c:1433:client_setvolume_cbk]
> 0-gvol1-client-0: Connected to 10.0.0.30:24009, attached to remote volume
> '/export/brick1'.
> [2013-06-02 09:29:04.853048] I [client-handshake.c:1445:client_setvolume_cbk]
> 0-gvol1-client-0: Server and Client lk-version numbers are not same,
> reopening the fds
> [2013-06-02 09:29:04.853262] I
> [client-handshake.c:453:client_set_lk_version_cbk] 0-gvol1-client-0: Server
> lk version = 1
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130603/9e5ce6b0/attachment.html>


More information about the Gluster-users mailing list