[Gluster-users] Gluster mount troubleshooting

Rob Leitman robl at voicebox.com
Thu Apr 17 21:35:22 UTC 2014


>From looking at the source, it appears that fuse_mount_sys failed.  There's no logging from there, unfortunately.

Is there a way to get logging from FUSE, or other layers, to debug this?

Thanks,

Rob

From: Rob Leitman
Sent: Wednesday, April 16, 2014 2:03 PM
To: 'gluster-users at gluster.org'
Subject: Gluster mount troubleshooting

I'm running glusterfs 3.4.3 built on Apr 14 2014 22:05:07

I have a two-brick volume created, but when I try to mount from the client machine I get an error:

# mount -t glusterfs -o backupvolfile-server=server2,volfile-max-fetch-attempts=2,log-level=WARNING,log-file=/var/log/gluster.log server1:/webs /mnt/glusterfs/
/usr/bin/fusermount-glusterfs: mount failed: Invalid argument
/usr/bin/fusermount-glusterfs: mount failed: Invalid argument
Mount failed. Please check the log file for more details.

Can I get troubleshooting advice, please?

The log file shows:

[2014-04-16 20:46:25.081640] E [mount.c:298:gf_fuse_mount] 0-glusterfs-fuse: mount of server1:/webs to /mnt/glusterfs/ (default_permissions,volfile-max-fetch-attempts=2,allow_other,max_read=131072) failed
[2014-04-16 20:46:25.085507] E [glusterfsd.c:1744:daemonize] 0-daemonize: mount failed
Given volfile:
+------------------------------------------------------------------------------+
  1: volume webs-client-0
  2:     type protocol/client
  3:     option transport-type tcp
  4:     option remote-subvolume /export/data
  5:     option remote-host server1
  6: end-volume
  7:
  8: volume webs-client-1
  9:     type protocol/client
10:     option transport-type tcp
11:     option remote-subvolume /export/data
12:     option remote-host server2
13: end-volume
14:
 15: volume webs-replicate-0
16:     type cluster/replicate
17:     subvolumes webs-client-0 webs-client-1
18: end-volume
19:
 20: volume webs-dht
21:     type cluster/distribute
22:     subvolumes webs-replicate-0
23: end-volume
24:
 25: volume webs-write-behind
26:     type performance/write-behind
27:     subvolumes webs-dht
28: end-volume
29:
 30: volume webs-read-ahead
31:     type performance/read-ahead
32:     subvolumes webs-write-behind
33: end-volume
34:
 35: volume webs-io-cache
36:     type performance/io-cache
37:     option cache-size 256MB
38:     subvolumes webs-read-ahead
39: end-volume
40:
 41: volume webs-quick-read
42:     type performance/quick-read
43:     option cache-size 256MB
44:     subvolumes webs-io-cache
45: end-volume
46:
 47: volume webs-open-behind
48:     type performance/open-behind
49:     subvolumes webs-quick-read
50: end-volume
51:
 52: volume webs-md-cache
53:     type performance/md-cache
54:     subvolumes webs-open-behind
55: end-volume
56:
 57: volume webs
58:     type debug/io-stats
59:     option count-fop-hits off
60:     option latency-measurement off
61:     subvolumes webs-md-cache
62: end-volume

+------------------------------------------------------------------------------+
[2014-04-16 20:46:25.284586] W [socket.c:514:__socket_rwv] 0-webs-client-0: readv failed (No data available)
[2014-04-16 20:46:25.296724] W [socket.c:514:__socket_rwv] 0-webs-client-1: readv failed (No data available)
[2014-04-16 20:46:25.304491] E [mount.c:298:gf_fuse_mount] 0-glusterfs-fuse: mount of server1:/webs to /mnt/glusterfs/ (default_permissions,volfile-max-fetch-attempts=2,allow_other,max_read=131072) failed
[2014-04-16 20:46:25.305488] E [glusterfsd.c:1744:daemonize] 0-daemonize: mount failed
Given volfile:
+------------------------------------------------------------------------------+
  1: volume webs-client-0
  2:     type protocol/client
  3:     option transport-type tcp
  4:     option remote-subvolume /export/data
  5:     option remote-host server1
  6: end-volume
  7:
  8: volume webs-client-1
  9:     type protocol/client
10:     option transport-type tcp
11:     option remote-subvolume /export/data
12:     option remote-host server2
13: end-volume
14:
15: volume webs-replicate-0
16:     type cluster/replicate
17:     subvolumes webs-client-0 webs-client-1
18: end-volume
19:
 20: volume webs-dht
21:     type cluster/distribute
22:     subvolumes webs-replicate-0
23: end-volume
24:
 25: volume webs-write-behind
26:     type performance/write-behind
27:     subvolumes webs-dht
28: end-volume
29:
 30: volume webs-read-ahead
31:     type performance/read-ahead
32:     subvolumes webs-write-behind
33: end-volume
34:
 35: volume webs-io-cache
36:     type performance/io-cache
37:     option cache-size 256MB
38:     subvolumes webs-read-ahead
39: end-volume
40:
 41: volume webs-quick-read
42:     type performance/quick-read
43:     option cache-size 256MB
44:     subvolumes webs-io-cache
45: end-volume
46:
 47: volume webs-open-behind
48:     type performance/open-behind
49:     subvolumes webs-quick-read
50: end-volume
51:
 52: volume webs-md-cache
53:     type performance/md-cache
54:     subvolumes webs-open-behind
55: end-volume
56:
 57: volume webs
58:     type debug/io-stats
59:     option count-fop-hits off
60:     option latency-measurement off
61:     subvolumes webs-md-cache
62: end-volume

+------------------------------------------------------------------------------+
[2014-04-16 20:46:25.391399] W [socket.c:514:__socket_rwv] 0-webs-client-0: readv failed (No data available)
[2014-04-16 20:46:25.397691] W [socket.c:514:__socket_rwv] 0-webs-client-1: readv failed (No data available)
[2014-04-16 20:46:25.436985] W [glusterfsd.c:1002:cleanup_and_exit] (-->/lib64/libc.so.6(clone+0x6d) [0x7f6bf417190d] (-->/lib64/libpthread.so.0(+0x7851) [0x7f6bf4803851] (-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xcd) [0x40533d]))) 0-: received signum (15), shutting down
[2014-04-16 20:46:25.456624] W [glusterfsd.c:1002:cleanup_and_exit] (-->/lib64/libc.so.6(clone+0x6d) [0x7f5a8696990d] (-->/lib64/libpthread.so.0(+0x7851) [0x7f5a86ffb851] (-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xcd) [0x40533d]))) 0-: received signum (15), shutting down

On both servers, I see good status:

# gluster volume status
Status of volume: webs
Gluster process                                                                                                Port       Online   Pid
------------------------------------------------------------------------------
Brick server1:/export/data                                                          49152    Y              6549
Brick server2:/export/data                                                          49152    Y              6536
NFS Server on localhost                                                                                2049       Y              6791
Self-heal Daemon on localhost                                                  N/A        Y              6798
NFS Server on server1                                                                   2049       Y              6724
Self-heal Daemon on server2                                                     N/A        Y              6731

There are no active volume tasks
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140417/f5d416c9/attachment.html>


More information about the Gluster-users mailing list