[Gluster-users] Trouble with gluster 3.5.3-1 on OEL 6.6
Duwaik, Hani
hduwaik at netsuite.com
Sat Feb 28 21:53:40 UTC 2015
Did some more investigating and research and come up a solution.
Short version, Œmount.glusterfs¹ does a Œstat -c %i <mount point>¹ to test
the mount. Unfortunately, the gluster mount takes a few seconds to mount
fully, but the script doesn¹t wait and instead the Œstat¹ fails so the
script assumes a failed mount and unmounts and exists:
[gfs1::root:bkp]# sh -x mount.glusterfs gfs2:data /mnt
+ _init gfs2:data /mnt
+ LOG_NONE=NONE
+ LOG_CRITICAL=CRITICAL
+ LOG_ERROR=ERROR
+ LOG_WARNING=WARNING
+ LOG_INFO=INFO
+ LOG_DEBUG=DEBUG
+ LOG_TRACE=TRACE
+ HOST_NAME_MAX=64
+ prefix=/usr
+ exec_prefix=/usr
++ echo /usr/sbin/glusterfs
+ cmd_line=/usr/sbin/glusterfs
+ case `uname -s` in
++ uname -s
+ getinode='stat -c %i '
+ getdev='stat -c %d '
+ lgetinode='stat -c %i -L'
+ lgetdev='stat -c %d -L'
+ mounttab=/etc/mtab
+ UPDATEDBCONF=/etc/updatedb.conf
+ main gfs2:data /mnt
++ sed -n 's/.*\--[ ]*\([^ ]*\).*/\1/p'
++ echo gfs2:data /mnt
+ helper=
+ in_opt=no
+ pos_args=0
+ for opt in '"$@"'
+ '[' no = yes ']'
+ '[' gfs2:data = -o ']'
+ case $pos_args in
+ volfile_loc=gfs2:data
+ pos_args=1
+ for opt in '"$@"'
+ '[' no = yes ']'
+ '[' /mnt = -o ']'
+ case $pos_args in
+ mount_point=/mnt
+ pos_args=2
+ '[' no = yes -o 2 -lt 2 ']'
+ '[' -r gfs2:data ']'
++ sed -n 's/\([a-zA-Z0-9:.\-]*\):.*/\1/p'
++ echo gfs2:data
+ server_ip=gfs2
++ sed -n 's/.*:\([^ ]*\).*/\1/p'
++ echo gfs2:data
+ test_str=data
+ '[' -n data ']'
+ volume_id=data
+ volfile_loc=
+ '[' -n '' ']'
+ '[' -z /mnt -o '!' -d /mnt ']'
+ grep -q '[[:space:]+]/mnt[[:space:]+]fuse' /etc/mtab
+ check_recursive_mount /mnt
+ '[' /mnt = / ']'
+ mnt_dir=/mnt
+ export PATH
+ which getfattr
+ '[' 0 -ne 0 ']'
+ grep -iq trusted.gfid=
+ getfattr -n trusted.gfid /mnt
+ '[' 1 -eq 0 ']'
+ GLUSTERD_WORKDIR=/var/lib/glusterd
+ ls -L /var/lib/glusterd/vols/data/bricks/gfs1:-gfs-brick0-data
/var/lib/glusterd/vols/data/bricks/gfs2:-gfs-brick0-data
+ '[' 0 -ne 0 ']'
++ cut -d = -f 2
++ grep '^path' /var/lib/glusterd/vols/data/bricks/gfs1:-gfs-brick0-data
/var/lib/glusterd/vols/data/bricks/gfs2:-gfs-brick0-data
+ brick_path='/gfs/brick0/data
/gfs/brick0/data'
++ stat -c %i -L /
+ root_inode=2
++ stat -c %d -L /
+ root_dev=64512
++ stat -c %i -L /mnt
+ mnt_inode=387856
++ stat -c %d -L /mnt
+ mnt_dev=64512
+ for brick in '"$brick_path"'
+ ls /gfs/brick0/data /gfs/brick0/data
+ '[' 0 -ne 0 ']'
+ grep -iq trusted.gfid=
+ getfattr -n trusted.gfid '/gfs/brick0/data
/gfs/brick0/data'
+ '[' 1 -ne 0 ']'
+ continue
+ test -f /etc/updatedb.conf
+ grep -q glusterfs /etc/updatedb.conf
+ start_glusterfs
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ volume_id_rdma=
+ '[' -z '' ']'
+ '[' -n gfs2 ']'
++ echo '/usr/sbin/glusterfs --volfile-server=gfs2'
+ cmd_line='/usr/sbin/glusterfs --volfile-server=gfs2'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n data ']'
+ '[' -n '' ']'
++ echo '/usr/sbin/glusterfs --volfile-server=gfs2 --volfile-id=data'
+ cmd_line='/usr/sbin/glusterfs --volfile-server=gfs2 --volfile-id=data'
+ '[' -n '' ']'
++ echo '/usr/sbin/glusterfs --volfile-server=gfs2 --volfile-id=data /mnt'
+ cmd_line='/usr/sbin/glusterfs --volfile-server=gfs2 --volfile-id=data
/mnt'
+ err=0
+ /usr/sbin/glusterfs --volfile-server=gfs2 --volfile-id=data /mnt
++ stat -c %i /mnt
+ inode=
+ '[' -z '' ']'
+ inode=0
+ '[' 0 -ne 1 ']'
+ err=1
+ '[' 1 -eq 1 ']'
+ echo 'Mount failed. Please check the log file for more details.'
Mount failed. Please check the log file for more details.
+ umount /mnt
+ exit 1
Adding a simple Œsleep 5¹ after the Œglusterfs¹ cmd seems to fix this:
[gfs1::root:bkp]# sh -x mount.glusterfs gfs2:data /mnt
+ _init gfs2:data /mnt
+ LOG_NONE=NONE
+ LOG_CRITICAL=CRITICAL
+ LOG_ERROR=ERROR
+ LOG_WARNING=WARNING
+ LOG_INFO=INFO
+ LOG_DEBUG=DEBUG
+ LOG_TRACE=TRACE
+ HOST_NAME_MAX=64
+ prefix=/usr
+ exec_prefix=/usr
++ echo /usr/sbin/glusterfs
+ cmd_line=/usr/sbin/glusterfs
+ case `uname -s` in
++ uname -s
+ getinode='stat -c %i '
+ getdev='stat -c %d '
+ lgetinode='stat -c %i -L'
+ lgetdev='stat -c %d -L'
+ mounttab=/etc/mtab
+ UPDATEDBCONF=/etc/updatedb.conf
+ main gfs2:data /mnt
++ sed -n 's/.*\--[ ]*\([^ ]*\).*/\1/p'
++ echo gfs2:data /mnt
+ helper=
+ in_opt=no
+ pos_args=0
+ for opt in '"$@"'
+ '[' no = yes ']'
+ '[' gfs2:data = -o ']'
+ case $pos_args in
+ volfile_loc=gfs2:data
+ pos_args=1
+ for opt in '"$@"'
+ '[' no = yes ']'
+ '[' /mnt = -o ']'
+ case $pos_args in
+ mount_point=/mnt
+ pos_args=2
+ '[' no = yes -o 2 -lt 2 ']'
+ '[' -r gfs2:data ']'
++ sed -n 's/\([a-zA-Z0-9:.\-]*\):.*/\1/p'
++ echo gfs2:data
+ server_ip=gfs2
++ sed -n 's/.*:\([^ ]*\).*/\1/p'
++ echo gfs2:data
+ test_str=data
+ '[' -n data ']'
+ volume_id=data
+ volfile_loc=
+ '[' -n '' ']'
+ '[' -z /mnt -o '!' -d /mnt ']'
+ grep -q '[[:space:]+]/mnt[[:space:]+]fuse' /etc/mtab
+ check_recursive_mount /mnt
+ '[' /mnt = / ']'
+ mnt_dir=/mnt
+ export PATH
+ which getfattr
+ '[' 0 -ne 0 ']'
+ grep -iq trusted.gfid=
+ getfattr -n trusted.gfid /mnt
+ '[' 1 -eq 0 ']'
+ GLUSTERD_WORKDIR=/var/lib/glusterd
+ ls -L /var/lib/glusterd/vols/data/bricks/gfs1:-gfs-brick0-data
/var/lib/glusterd/vols/data/bricks/gfs2:-gfs-brick0-data
+ '[' 0 -ne 0 ']'
++ cut -d = -f 2
++ grep '^path' /var/lib/glusterd/vols/data/bricks/gfs1:-gfs-brick0-data
/var/lib/glusterd/vols/data/bricks/gfs2:-gfs-brick0-data
+ brick_path='/gfs/brick0/data
/gfs/brick0/data'
++ stat -c %i -L /
+ root_inode=2
++ stat -c %d -L /
+ root_dev=64512
++ stat -c %i -L /mnt
+ mnt_inode=387856
++ stat -c %d -L /mnt
+ mnt_dev=64512
+ for brick in '"$brick_path"'
+ ls /gfs/brick0/data /gfs/brick0/data
+ '[' 0 -ne 0 ']'
+ grep -iq trusted.gfid=
+ getfattr -n trusted.gfid '/gfs/brick0/data
/gfs/brick0/data'
+ '[' 1 -ne 0 ']'
+ continue
+ test -f /etc/updatedb.conf
+ grep -q glusterfs /etc/updatedb.conf
+ start_glusterfs
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ volume_id_rdma=
+ '[' -z '' ']'
+ '[' -n gfs2 ']'
++ echo '/usr/sbin/glusterfs --volfile-server=gfs2'
+ cmd_line='/usr/sbin/glusterfs --volfile-server=gfs2'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n '' ']'
+ '[' -n data ']'
+ '[' -n '' ']'
++ echo '/usr/sbin/glusterfs --volfile-server=gfs2 --volfile-id=data'
+ cmd_line='/usr/sbin/glusterfs --volfile-server=gfs2 --volfile-id=data'
+ '[' -n '' ']'
++ echo '/usr/sbin/glusterfs --volfile-server=gfs2 --volfile-id=data /mnt'
+ cmd_line='/usr/sbin/glusterfs --volfile-server=gfs2 --volfile-id=data
/mnt'
+ err=0
+ /usr/sbin/glusterfs --volfile-server=gfs2 --volfile-id=data /mnt
+ sleep 5
++ stat -c %i /mnt
+ inode=1
+ '[' -z 1 ']'
+ '[' 1 -ne 1 ']'
+ '[' 0 -eq 1 ']'
It is not elegant, but it seems to fix the issue.
If anyone knows of a better fix, let me know.
Thanks,
---
Hani Duwaik | Mgr, Unix System Administration
650-430-5061 | hduwaik at netsuite.com | @NetSuite
<https://twitter.com/netsuite>
NetSuite <http://www.netsuite.com/>: Where Business is Going
On 2/28/15, 12:38 PM, "Niels de Vos" <ndevos at redhat.com> wrote:
>On Sat, Feb 28, 2015 at 06:49:29PM +0000, Duwaik, Hani wrote:
>> Having an issue trying to set up gluster 3.5.3-1 on OEL 6.6.
>>
>> After installing and setting up gluster, I am able to NFS mount the
>>volume without any issues. However, I am unable to complete a
>>Œglusterfs¹ mount.
>
>Unfortunately, the /var/log/glusterfs/data.log does not have an
>indication on why it failed. Could you mount with more verbose logging
>enabled?
>
> # mount -t glusterfs -o log-level=DEBUG server01x:/data /data
>
>If the log would still not explain whats wrong, you could replace DEBUG
>with TRACE to get even more messages.
>
>HTH,
>Niels
>
>
>>
>> Systems are OEL6.6 x86_64 with following packages installed:
>>
>> glusterfs-3.5.3-1.el6.x86_64
>> glusterfs-api-3.5.3-1.el6.x86_64
>> glusterfs-cli-3.5.3-1.el6.x86_64
>> glusterfs-fuse-3.5.3-1.el6.x86_64
>> glusterfs-libs-3.5.3-1.el6.x86_64
>> glusterfs-server-3.5.3-1.el6.x86_64
>>
>> Volume info:
>> ===========
>>
>> Volume Name: data
>>
>> Type: Distributed-Replicate
>>
>> Volume ID: 6e206cc1-d037-41a3-b614-869162ab3ece
>>
>> Status: Started
>>
>> Number of Bricks: 3 x 2 = 6
>>
>> Transport-type: tcp
>>
>> Bricks:
>>
>> Brick1: server01x:/gfs/brick0/data
>>
>> Brick2: server02x:/gfs/brick0/data
>>
>> Brick3: server03x:/gfs/brick0/data
>>
>> Brick4: server04x:/gfs/brick0/data
>>
>> Brick5: server05x:/gfs/brick0/data
>>
>> Brick6: server06x:/gfs/brick0/data
>>
>> Options Reconfigured:
>>
>> cluster.background-self-heal-count: 2
>>
>> cluster.data-self-heal-algorithm: full
>>
>> performance.client-io-threads: on
>>
>> performance.cache-size: 128MB
>>
>> performance.io-thread-count: 64
>>
>> When I try to mount via gluster, I get the generic failure:
>>
>> server02> mount -t glusterfs server01x:/data /data
>> Mount failed. Please check the log file for more details.
>>
>> The logs only show:
>>
>> /var/log/glusterfs/bricks/gfs-brick0-data.log:
>> [2015-02-28 18:11:54.837363] I
>>[server-handshake.c:575:server_setvolume] 0-data-server: accepted client
>>from server02-24970-2015/02/28-18:11:53:352986-data-client-0-0-0
>>(version: 3.5.3)
>> [2015-02-28 18:11:54.853135] I [server.c:520:server_rpc_notify]
>>0-data-server: disconnecting connection from
>>server02-24970-2015/02/28-18:11:53:352986-data-client-0-0-0
>> [2015-02-28 18:11:54.853226] I [client_t.c:417:gf_client_unref]
>>0-data-server: Shutting down connection
>>server02-24970-2015/02/28-18:11:53:352986-data-client-0-0-0
>>
>> And
>>
>> /var/log/glusterfs/data.log
>>
>> [2015-02-28 18:37:17.600728] I [glusterfsd.c:1959:main]
>>0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.5.3
>>(/usr/sbin/glusterfs --volfile-server=server01x.svale.netledger.com
>>--volfile-id=/data /data)
>>
>> [2015-02-28 18:37:17.605373] I [socket.c:3645:socket_init] 0-glusterfs:
>>SSL support is NOT enabled
>>
>> [2015-02-28 18:37:17.605404] I [socket.c:3660:socket_init] 0-glusterfs:
>>using system polling thread
>>
>> [2015-02-28 18:37:18.726440] I [dht-shared.c:314:dht_init_regex]
>>0-data-dht: using regex rsync-hash-regex = ^\.(.+)\.[^.]+$
>>
>> [2015-02-28 18:37:18.733444] I [socket.c:3645:socket_init]
>>0-data-client-5: SSL support is NOT enabled
>>
>> [2015-02-28 18:37:18.733473] I [socket.c:3660:socket_init]
>>0-data-client-5: using system polling thread
>>
>> [2015-02-28 18:37:18.733923] I [socket.c:3645:socket_init]
>>0-data-client-4: SSL support is NOT enabled
>>
>> [2015-02-28 18:37:18.733943] I [socket.c:3660:socket_init]
>>0-data-client-4: using system polling thread
>>
>> [2015-02-28 18:37:18.734404] I [socket.c:3645:socket_init]
>>0-data-client-3: SSL support is NOT enabled
>>
>> [2015-02-28 18:37:18.734426] I [socket.c:3660:socket_init]
>>0-data-client-3: using system polling thread
>>
>> [2015-02-28 18:37:18.734859] I [socket.c:3645:socket_init]
>>0-data-client-2: SSL support is NOT enabled
>>
>> [2015-02-28 18:37:18.734906] I [socket.c:3660:socket_init]
>>0-data-client-2: using system polling thread
>>
>> [2015-02-28 18:37:18.735364] I [socket.c:3645:socket_init]
>>0-data-client-1: SSL support is NOT enabled
>>
>> [2015-02-28 18:37:18.735386] I [socket.c:3660:socket_init]
>>0-data-client-1: using system polling thread
>>
>> [2015-02-28 18:37:18.735823] I [socket.c:3645:socket_init]
>>0-data-client-0: SSL support is NOT enabled
>>
>> [2015-02-28 18:37:18.735843] I [socket.c:3660:socket_init]
>>0-data-client-0: using system polling thread
>>
>> [2015-02-28 18:37:18.735903] I [client.c:2294:notify] 0-data-client-0:
>>parent translators are ready, attempting connect on transport
>>
>> [2015-02-28 18:37:18.744117] I [client.c:2294:notify] 0-data-client-1:
>>parent translators are ready, attempting connect on transport
>>
>> [2015-02-28 18:37:18.759585] I [client.c:2294:notify] 0-data-client-2:
>>parent translators are ready, attempting connect on transport
>>
>> [2015-02-28 18:37:18.767297] I [client.c:2294:notify] 0-data-client-3:
>>parent translators are ready, attempting connect on transport
>>
>> [2015-02-28 18:37:18.784718] I [client.c:2294:notify] 0-data-client-4:
>>parent translators are ready, attempting connect on transport
>>
>> [2015-02-28 18:37:18.793474] I [client.c:2294:notify] 0-data-client-5:
>>parent translators are ready, attempting connect on transport
>>
>> Final graph:
>>
>>
>>+------------------------------------------------------------------------
>>------+
>>
>> 1: volume data-client-0
>>
>> 2: type protocol/client
>>
>> 3: option remote-host server01x
>>
>> 4: option remote-subvolume /gfs/brick0/data
>>
>> 5: option transport-type socket
>>
>> 6: option send-gids true
>>
>> 7: end-volume
>>
>> 8:
>>
>> 9: volume data-client-1
>>
>> 10: type protocol/client
>>
>> 11: option remote-host server02x
>>
>> 12: option remote-subvolume /gfs/brick0/data
>>
>> 13: option transport-type socket
>>
>> 14: option send-gids true
>>
>> 15: end-volume
>>
>> 16:
>>
>> 17: volume data-replicate-0
>>
>> 18: type cluster/replicate
>>
>> 19: option background-self-heal-count 2
>>
>> 20: option data-self-heal-algorithm full
>>
>> 21: subvolumes data-client-0 data-client-1
>>
>> 22: end-volume
>>
>> 23:
>>
>> 24: volume data-client-2
>>
>> 25: type protocol/client
>>
>> 26: option remote-host server03x
>>
>> 27: option remote-subvolume /gfs/brick0/data
>>
>> 28: option transport-type socket
>>
>> 29: option send-gids true
>>
>> 30: end-volume
>>
>> 31:
>>
>> 32: volume data-client-3
>>
>> 33: type protocol/client
>>
>> 34: option remote-host server04x
>>
>> 35: option remote-subvolume /gfs/brick0/data
>>
>> 36: option transport-type socket
>>
>> 37: option send-gids true
>>
>> 38: end-volume
>>
>> 39:
>>
>> 40: volume data-replicate-1
>>
>> 41: type cluster/replicate
>>
>> 42: option background-self-heal-count 2
>>
>> 43: option data-self-heal-algorithm full
>>
>> 44: subvolumes data-client-2 data-client-3
>>
>> 45: end-volume
>>
>> 46:
>>
>> 47: volume data-client-4
>>
>> 48: type protocol/client
>>
>> 49: option remote-host server05x
>>
>> 50: option remote-subvolume /gfs/brick0/data
>>
>> 51: option transport-type socket
>>
>> 52: option send-gids true
>>
>> 53: end-volume
>>
>> 54:
>>
>> 55: volume data-client-5
>>
>> 56: type protocol/client
>>
>> 57: option remote-host server06x
>>
>> 58: option remote-subvolume /gfs/brick0/data
>>
>> 59: option transport-type socket
>>
>> 60: option send-gids true
>>
>> 61: end-volume
>>
>> 62:
>>
>> 63: volume data-replicate-2
>>
>> 64: type cluster/replicate
>>
>> 65: option background-self-heal-count 2
>>
>> 66: option data-self-heal-algorithm full
>>
>> 67: subvolumes data-client-4 data-client-5
>>
>> 68: end-volume
>>
>> 69:
>>
>> 70: volume data-dht
>>
>> 71: type cluster/distribute
>>
>> 72: subvolumes data-replicate-0 data-replicate-1 data-replicate-2
>>
>> 73: end-volume
>>
>> 74:
>>
>> 75: volume data-write-behind
>>
>> 76: type performance/write-behind
>>
>> 77: subvolumes data-dht
>>
>> 78: end-volume
>>
>> 79:
>>
>> 80: volume data-read-ahead
>>
>> 81: type performance/read-ahead
>>
>> 82: subvolumes data-write-behind
>>
>> 83: end-volume
>>
>> 84:
>>
>> 85: volume data-io-cache
>>
>> 86: type performance/io-cache
>>
>> 87: option cache-size 128MB
>>
>> 88: subvolumes data-read-ahead
>>
>> 89: end-volume
>>
>> 90:
>>
>> 91: volume data-quick-read
>>
>> 92: type performance/quick-read
>>
>> 93: option cache-size 128MB
>>
>> 94: subvolumes data-io-cache
>>
>> 95: end-volume
>>
>> 96:
>>
>> 97: volume data-open-behind
>>
>> 98: type performance/open-behind
>>
>> 99: subvolumes data-quick-read
>>
>> 100: end-volume
>>
>> 101:
>>
>> 102: volume data-md-cache
>>
>> 103: type performance/md-cache
>>
>> 104: subvolumes data-open-behind
>>
>> 105: end-volume
>>
>> 106:
>>
>> 107: volume data-io-threads
>>
>> 108: type performance/io-threads
>>
>> 109: option thread-count 64
>>
>> 110: subvolumes data-md-cache
>>
>> 111: end-volume
>>
>> 112:
>>
>> 113: volume data
>>
>> 114: type debug/io-stats
>>
>> 115: option latency-measurement off
>>
>> 116: option count-fop-hits off
>>
>> 117: subvolumes data-io-threads
>>
>> 118: end-volume
>>
>> 119:
>>
>>
>>+------------------------------------------------------------------------
>>------+
>>
>> [2015-02-28 18:37:18.803366] I [rpc-clnt.c:1729:rpc_clnt_reconfig]
>>0-data-client-1: changing port to 49154 (from 0)
>>
>> [2015-02-28 18:37:18.812481] I [rpc-clnt.c:1729:rpc_clnt_reconfig]
>>0-data-client-0: changing port to 49154 (from 0)
>>
>> [2015-02-28 18:37:18.821302] I [rpc-clnt.c:1729:rpc_clnt_reconfig]
>>0-data-client-2: changing port to 49154 (from 0)
>>
>> [2015-02-28 18:37:18.821368] I [rpc-clnt.c:1729:rpc_clnt_reconfig]
>>0-data-client-3: changing port to 49154 (from 0)
>>
>> [2015-02-28 18:37:18.821411] I
>>[client-handshake.c:1677:select_server_supported_programs]
>>0-data-client-1: Using Program GlusterFS 3.3, Num (1298437), Version
>>(330)
>>
>> [2015-02-28 18:37:18.821525] I [rpc-clnt.c:1729:rpc_clnt_reconfig]
>>0-data-client-5: changing port to 49154 (from 0)
>>
>> [2015-02-28 18:37:18.821569] I [rpc-clnt.c:1729:rpc_clnt_reconfig]
>>0-data-client-4: changing port to 49154 (from 0)
>>
>> [2015-02-28 18:37:18.850560] I
>>[client-handshake.c:1462:client_setvolume_cbk] 0-data-client-1:
>>Connected to 10.1.1.148:49154, attached to remote volume
>>'/gfs/brick0/data'.
>>
>> [2015-02-28 18:37:18.850597] I
>>[client-handshake.c:1474:client_setvolume_cbk] 0-data-client-1: Server
>>and Client lk-version numbers are not same, reopening the fds
>>
>> [2015-02-28 18:37:18.850695] I [afr-common.c:4267:afr_notify]
>>0-data-replicate-0: Subvolume 'data-client-1' came back up; going online.
>>
>> [2015-02-28 18:37:18.850863] I
>>[client-handshake.c:450:client_set_lk_version_cbk] 0-data-client-1:
>>Server lk version = 1
>>
>> [2015-02-28 18:37:18.850945] I
>>[client-handshake.c:1677:select_server_supported_programs]
>>0-data-client-0: Using Program GlusterFS 3.3, Num (1298437), Version
>>(330)
>>
>> [2015-02-28 18:37:18.851117] I
>>[client-handshake.c:1677:select_server_supported_programs]
>>0-data-client-2: Using Program GlusterFS 3.3, Num (1298437), Version
>>(330)
>>
>> [2015-02-28 18:37:18.851199] I
>>[client-handshake.c:1677:select_server_supported_programs]
>>0-data-client-5: Using Program GlusterFS 3.3, Num (1298437), Version
>>(330)
>>
>> [2015-02-28 18:37:18.851277] I
>>[client-handshake.c:1677:select_server_supported_programs]
>>0-data-client-3: Using Program GlusterFS 3.3, Num (1298437), Version
>>(330)
>>
>> [2015-02-28 18:37:18.851356] I
>>[client-handshake.c:1677:select_server_supported_programs]
>>0-data-client-4: Using Program GlusterFS 3.3, Num (1298437), Version
>>(330)
>>
>> [2015-02-28 18:37:18.851550] I
>>[client-handshake.c:1462:client_setvolume_cbk] 0-data-client-0:
>>Connected to 10.1.1.147:49154, attached to remote volume
>>'/gfs/brick0/data'.
>>
>> [2015-02-28 18:37:18.851592] I
>>[client-handshake.c:1474:client_setvolume_cbk] 0-data-client-0: Server
>>and Client lk-version numbers are not same, reopening the fds
>>
>> [2015-02-28 18:37:18.851836] I
>>[client-handshake.c:1462:client_setvolume_cbk] 0-data-client-2:
>>Connected to 10.1.1.149:49154, attached to remote volume
>>'/gfs/brick0/data'.
>>
>> [2015-02-28 18:37:18.851878] I
>>[client-handshake.c:1474:client_setvolume_cbk] 0-data-client-2: Server
>>and Client lk-version numbers are not same, reopening the fds
>>
>> [2015-02-28 18:37:18.851941] I [afr-common.c:4267:afr_notify]
>>0-data-replicate-1: Subvolume 'data-client-2' came back up; going online.
>>
>> [2015-02-28 18:37:18.852002] I
>>[client-handshake.c:450:client_set_lk_version_cbk] 0-data-client-0:
>>Server lk version = 1
>>
>> [2015-02-28 18:37:18.852039] I
>>[client-handshake.c:1462:client_setvolume_cbk] 0-data-client-5:
>>Connected to 10.1.1.152:49154, attached to remote volume
>>'/gfs/brick0/data'.
>>
>> [2015-02-28 18:37:18.852076] I
>>[client-handshake.c:1474:client_setvolume_cbk] 0-data-client-5: Server
>>and Client lk-version numbers are not same, reopening the fds
>>
>> [2015-02-28 18:37:18.852189] I [afr-common.c:4267:afr_notify]
>>0-data-replicate-2: Subvolume 'data-client-5' came back up; going online.
>>
>> [2015-02-28 18:37:18.852239] I
>>[client-handshake.c:1462:client_setvolume_cbk] 0-data-client-3:
>>Connected to 10.1.1.150:49154, attached to remote volume
>>'/gfs/brick0/data'.
>>
>> [2015-02-28 18:37:18.852256] I
>>[client-handshake.c:1474:client_setvolume_cbk] 0-data-client-3: Server
>>and Client lk-version numbers are not same, reopening the fds
>>
>> [2015-02-28 18:37:18.852846] I
>>[client-handshake.c:1462:client_setvolume_cbk] 0-data-client-4:
>>Connected to 10.1.1.151:49154, attached to remote volume
>>'/gfs/brick0/data'.
>>
>> [2015-02-28 18:37:18.852871] I
>>[client-handshake.c:1474:client_setvolume_cbk] 0-data-client-4: Server
>>and Client lk-version numbers are not same, reopening the fds
>>
>> [2015-02-28 18:37:18.861168] I [fuse-bridge.c:5016:fuse_graph_setup]
>>0-fuse: switched to graph 0
>>
>> [2015-02-28 18:37:18.861290] I
>>[client-handshake.c:450:client_set_lk_version_cbk] 0-data-client-5:
>>Server lk version = 1
>>
>> [2015-02-28 18:37:18.861323] I
>>[client-handshake.c:450:client_set_lk_version_cbk] 0-data-client-3:
>>Server lk version = 1
>>
>> [2015-02-28 18:37:18.861347] I
>>[client-handshake.c:450:client_set_lk_version_cbk] 0-data-client-4:
>>Server lk version = 1
>>
>> [2015-02-28 18:37:18.861435] I
>>[client-handshake.c:450:client_set_lk_version_cbk] 0-data-client-2:
>>Server lk version = 1
>>
>> [2015-02-28 18:37:18.861440] I [fuse-bridge.c:4857:fuse_thread_proc]
>>0-fuse: unmounting /data
>>
>> [2015-02-28 18:37:18.861920] W [glusterfsd.c:1095:cleanup_and_exit]
>>(-->/lib64/libc.so.6(clone+0x6d) [0x7f86e52939dd]
>>(-->/lib64/libpthread.so.0(+0x79d1) [0x7f86e59299d1]
>>(-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xd5) [0x4053e5]))) 0-:
>>received signum (15), shutting down
>>
>> [2015-02-28 18:37:18.861974] I [fuse-bridge.c:5514:fini] 0-fuse:
>>Unmounting '/data'.
>>
>>
>> Any assistance will be greatly appreciated.
>>
>> Thanks,
>>
>> -Hani
>>
>> To learn more about SuiteWorld, visit
>> www.netsuite.com/suiteworld.
>>
>>
>> NOTICE: This email and any attachments may contain confidential and
>>proprietary information of NetSuite Inc. and is for the sole use of the
>>intended recipient for the stated purpose. Any improper use or
>>distribution is prohibited. If you are not the intended recipient,
>>please notify the sender; do not review, copy or distribute; and
>>promptly delete or destroy all transmitted information. Please note that
>>all communications and information transmitted through this email system
>>may be monitored and retained by NetSuite or its agents and that all
>>incoming email is automatically scanned by a third party spam and
>>filtering service which may result in deletion of a legitimate e-mail
>>before it is read by the intended recipient.
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-users
>
More information about the Gluster-users
mailing list