[Gluster-users] Problem with add-brick

Atin Mukherjee amukherj at redhat.com
Thu Sep 29 14:19:05 UTC 2016


Dennis,

Thanks for sharing the logs.

It seems like a volume configured created with tcp,rdma transport fails to
start (atleast in my local set up). The issue here is although the brick
process comes up, but glusterd receives a non zero ret code from the runner
interface which spawns the brick process(es).

Raghavendra Talur/Rafi,

Is this an intended behaviour if rdma device is not configured? Please
chime in with your thoughts


On Wed, Sep 28, 2016 at 10:22 AM, Atin Mukherjee <amukherj at redhat.com>
wrote:

> Dennis,
>
> It seems like that add-brick has definitely failed and the entry is not
> committed into glusterd store. volume status and volume info commands are
> referring the in-memory data for fs4 (which exist) but post a restart they
> are no longer available. Could you run glusterd with debug log enabled
> (systemctl stop glusterd; glusterd -LDEBUG) and provide us cmd_history.log,
> glusterd log along with fs4 brick log files to further analyze the issue?
> Regarding the missing RDMA ports for fs2, fs3 brick can you cross check if
> glusterfs-rdma package is installed on both the nodes?
>
> On Wed, Sep 28, 2016 at 7:14 AM, Ravishankar N <ravishankar at redhat.com>
> wrote:
>
>> On 09/27/2016 10:29 PM, Dennis Michael wrote:
>>
>>
>>
>> [root at fs4 bricks]# gluster volume info
>>
>> Volume Name: cees-data
>> Type: Distribute
>> Volume ID: 27d2a59c-bdac-4f66-bcd8-e6124e53a4a2
>> Status: Started
>> Number of Bricks: 4
>> Transport-type: tcp,rdma
>> Bricks:
>> Brick1: fs1:/data/brick
>> Brick2: fs2:/data/brick
>> Brick3: fs3:/data/brick
>> Brick4: fs4:/data/brick
>> Options Reconfigured:
>> features.quota-deem-statfs: on
>> features.inode-quota: on
>> features.quota: on
>> performance.readdir-ahead: on
>> [root at fs4 bricks]# gluster volume status
>> Status of volume: cees-data
>> Gluster process                             TCP Port  RDMA Port  Online
>>  Pid
>> ------------------------------------------------------------
>> ------------------
>> Brick fs1:/data/brick                       49152     49153      Y
>> 1878
>> Brick fs2:/data/brick                       49152     0          Y
>> 1707
>> Brick fs3:/data/brick                       49152     0          Y
>> 4696
>> Brick fs4:/data/brick                       N/A       N/A        N
>> N/A
>> NFS Server on localhost                     2049      0          Y
>> 13808
>> Quota Daemon on localhost                   N/A       N/A        Y
>> 13813
>> NFS Server on fs1                           2049      0          Y
>> 6722
>> Quota Daemon on fs1                         N/A       N/A        Y
>> 6730
>> NFS Server on fs3                           2049      0          Y
>> 12553
>> Quota Daemon on fs3                         N/A       N/A        Y
>> 12561
>> NFS Server on fs2                           2049      0          Y
>> 11702
>> Quota Daemon on fs2                         N/A       N/A        Y
>> 11710
>>
>> Task Status of Volume cees-data
>> ------------------------------------------------------------
>> ------------------
>> There are no active volume tasks
>>
>> [root at fs4 bricks]# ps auxww | grep gluster
>> root     13791  0.0  0.0 701472 19768 ?        Ssl  09:06   0:00
>> /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level INFO
>> root     13808  0.0  0.0 560236 41420 ?        Ssl  09:07   0:00
>> /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
>> /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log -S
>> /var/run/gluster/01c61523374369658a62b75c582b5ac2.socket
>> root     13813  0.0  0.0 443164 17908 ?        Ssl  09:07   0:00
>> /usr/sbin/glusterfs -s localhost --volfile-id gluster/quotad -p
>> /var/lib/glusterd/quotad/run/quotad.pid -l /var/log/glusterfs/quotad.log
>> -S /var/run/gluster/3753def90f5c34f656513dba6a544f7d.socket
>> --xlator-option *replicate*.data-self-heal=off --xlator-option
>> *replicate*.metadata-self-heal=off --xlator-option
>> *replicate*.entry-self-heal=off
>> root     13874  0.0  0.0 1200472 31700 ?       Ssl  09:16   0:00
>> /usr/sbin/glusterfsd -s fs4 --volfile-id cees-data.fs4.data-brick -p
>> /var/lib/glusterd/vols/cees-data/run/fs4-data-brick.pid -S
>> /var/run/gluster/5203ab38be21e1d37c04f6bdfee77d4a.socket --brick-name
>> /data/brick -l /var/log/glusterfs/bricks/data-brick.log --xlator-option
>> *-posix.glusterd-uuid=f04b231e-63f8-4374-91ae-17c0c623f165 --brick-port
>> 49152 49153 --xlator-option cees-data-server.transport.rdma.listen-port=49153
>> --xlator-option cees-data-server.listen-port=49152
>> --volfile-server-transport=socket,rdma
>> root     13941  0.0  0.0 112648   976 pts/0    S+   09:50   0:00 grep
>> --color=auto gluster
>>
>> [root at fs4 bricks]# systemctl restart glusterfsd glusterd
>>
>> [root at fs4 bricks]# ps auxww | grep gluster
>> root     13808  0.0  0.0 560236 41420 ?        Ssl  09:07   0:00
>> /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
>> /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log -S
>> /var/run/gluster/01c61523374369658a62b75c582b5ac2.socket
>> root     13813  0.0  0.0 443164 17908 ?        Ssl  09:07   0:00
>> /usr/sbin/glusterfs -s localhost --volfile-id gluster/quotad -p
>> /var/lib/glusterd/quotad/run/quotad.pid -l /var/log/glusterfs/quotad.log
>> -S /var/run/gluster/3753def90f5c34f656513dba6a544f7d.socket
>> --xlator-option *replicate*.data-self-heal=off --xlator-option
>> *replicate*.metadata-self-heal=off --xlator-option
>> *replicate*.entry-self-heal=off
>> root     13953  0.1  0.0 570740 14988 ?        Ssl  09:51   0:00
>> /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level INFO
>> root     13965  0.0  0.0 112648   976 pts/0    S+   09:51   0:00 grep
>> --color=auto gluster
>>
>> [root at fs4 bricks]# gluster volume info
>>
>> Volume Name: cees-data
>> Type: Distribute
>> Volume ID: 27d2a59c-bdac-4f66-bcd8-e6124e53a4a2
>> Status: Started
>> Number of Bricks: 3
>> Transport-type: tcp,rdma
>> Bricks:
>> Brick1: fs1:/data/brick
>> Brick2: fs2:/data/brick
>> Brick3: fs3:/data/brick
>> Options Reconfigured:
>> performance.readdir-ahead: on
>> features.quota: on
>> features.inode-quota: on
>> features.quota-deem-statfs: on
>>
>>
>>
>> I'm not sure what's going on here. Restarting glusterd seems to change
>> the output of gluster volume info?  I also see you are using RDMA. Not sure
>> why the RDMA ports for fs2 and fs3 are not shown in the volume status
>> output. CC'ing some glusterd/rdma devs for pointers.
>>
>> -Ravi
>>
>>
>>
>> [root at fs4 bricks]# gluster volume status
>> Status of volume: cees-data
>> Gluster process                             TCP Port  RDMA Port  Online
>>  Pid
>> ------------------------------------------------------------
>> ------------------
>> Brick fs1:/data/brick                       49152     49153      Y
>> 1878
>> Brick fs2:/data/brick                       49152     0          Y
>> 1707
>> Brick fs3:/data/brick                       49152     0          Y
>> 4696
>> NFS Server on localhost                     2049      0          Y
>> 13968
>> Quota Daemon on localhost                   N/A       N/A        Y
>> 13976
>> NFS Server on fs2                           2049      0          Y
>> 11702
>> Quota Daemon on fs2                         N/A       N/A        Y
>> 11710
>> NFS Server on fs3                           2049      0          Y
>> 12553
>> Quota Daemon on fs3                         N/A       N/A        Y
>> 12561
>> NFS Server on fs1                           2049      0          Y
>> 6722
>>
>> Task Status of Volume cees-data
>> ------------------------------------------------------------
>> ------------------
>> There are no active volume tasks
>>
>> [root at fs4 bricks]# gluster peer status
>> Number of Peers: 3
>>
>> Hostname: fs1
>> Uuid: ddc0a23e-05e5-48f7-993e-a37e43b21605
>> State: Peer in Cluster (Connected)
>>
>> Hostname: fs2
>> Uuid: e37108f8-d2f1-4f28-adc8-0b3d3401df29
>> State: Peer in Cluster (Connected)
>>
>> Hostname: fs3
>> Uuid: 19a42201-c932-44db-b1a7-8b5b1af32a36
>> State: Peer in Cluster (Connected)
>>
>> Dennis
>>
>>
>> On Tue, Sep 27, 2016 at 9:40 AM, Ravishankar N <ravishankar at redhat.com>
>> wrote:
>>
>>> On 09/27/2016 09:53 PM, Dennis Michael wrote:
>>>
>>> Yes, you are right.  I mixed up the logs.  I just ran the add-brick
>>> command again after cleaning up fs4 and re-installing gluster.  This is the
>>> complete fs4 data-brick.log.
>>>
>>> [root at fs1 ~]# gluster volume add-brick cees-data fs4:/data/brick
>>> volume add-brick: failed: Commit failed on fs4. Please check log file
>>> for details.
>>>
>>> [root at fs4 bricks]# pwd
>>> /var/log/glusterfs/bricks
>>> [root at fs4 bricks]# cat data-brick.log
>>> [2016-09-27 16:16:28.095661] I [MSGID: 100030] [glusterfsd.c:2338:main]
>>> 0-/usr/sbin/glusterfsd: Started running /usr/sbin/glusterfsd version 3.7.14
>>> (args: /usr/sbin/glusterfsd -s fs4 --volfile-id cees-data.fs4.data-brick -p
>>> /var/lib/glusterd/vols/cees-data/run/fs4-data-brick.pid -S
>>> /var/run/gluster/5203ab38be21e1d37c04f6bdfee77d4a.socket --brick-name
>>> /data/brick -l /var/log/glusterfs/bricks/data-brick.log --xlator-option
>>> *-posix.glusterd-uuid=f04b231e-63f8-4374-91ae-17c0c623f165 --brick-port
>>> 49152 --xlator-option cees-data-server.transport.rdma.listen-port=49153
>>> --xlator-option cees-data-server.listen-port=49152
>>> --volfile-server-transport=socket,rdma)
>>> [2016-09-27 16:16:28.101547] I [MSGID: 101190]
>>> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread
>>> with index 1
>>> [2016-09-27 16:16:28.104637] I [graph.c:269:gf_add_cmdline_options]
>>> 0-cees-data-server: adding option 'listen-port' for volume
>>> 'cees-data-server' with value '49152'
>>> [2016-09-27 16:16:28.104646] I [graph.c:269:gf_add_cmdline_options]
>>> 0-cees-data-server: adding option 'transport.rdma.listen-port' for volume
>>> 'cees-data-server' with value '49153'
>>> [2016-09-27 16:16:28.104662] I [graph.c:269:gf_add_cmdline_options]
>>> 0-cees-data-posix: adding option 'glusterd-uuid' for volume
>>> 'cees-data-posix' with value 'f04b231e-63f8-4374-91ae-17c0c623f165'
>>> [2016-09-27 16:16:28.104808] I [MSGID: 115034]
>>> [server.c:403:_check_for_auth_option] 0-/data/brick: skip format check
>>> for non-addr auth option auth.login./data/brick.allow
>>> [2016-09-27 16:16:28.104814] I [MSGID: 115034]
>>> [server.c:403:_check_for_auth_option] 0-/data/brick: skip format check
>>> for non-addr auth option auth.login.18ddaf4c-ad98-4155-
>>> 9372-717eae718b4c.password
>>> [2016-09-27 16:16:28.104883] I [MSGID: 101190]
>>> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread
>>> with index 2
>>> [2016-09-27 16:16:28.105479] I [rpcsvc.c:2196:rpcsvc_set_outstanding_rpc_limit]
>>> 0-rpc-service: Configured rpc.outstanding-rpc-limit with value 64
>>> [2016-09-27 16:16:28.105532] W [MSGID: 101002]
>>> [options.c:957:xl_opt_validate] 0-cees-data-server: option
>>> 'listen-port' is deprecated, preferred is 'transport.socket.listen-port',
>>> continuing with correction
>>> [2016-09-27 16:16:28.109456] W [socket.c:3665:reconfigure]
>>> 0-cees-data-quota: NBIO on -1 failed (Bad file descriptor)
>>> [2016-09-27 16:16:28.489255] I [MSGID: 121050]
>>> [ctr-helper.c:259:extract_ctr_options] 0-gfdbdatastore: CTR Xlator is
>>> disabled.
>>> [2016-09-27 16:16:28.489272] W [MSGID: 101105]
>>> [gfdb_sqlite3.h:239:gfdb_set_sql_params] 0-cees-data-changetimerecorder:
>>> Failed to retrieve sql-db-pagesize from params.Assigning default value: 4096
>>> [2016-09-27 16:16:28.489278] W [MSGID: 101105]
>>> [gfdb_sqlite3.h:239:gfdb_set_sql_params] 0-cees-data-changetimerecorder:
>>> Failed to retrieve sql-db-journalmode from params.Assigning default value:
>>> wal
>>> [2016-09-27 16:16:28.489284] W [MSGID: 101105]
>>> [gfdb_sqlite3.h:239:gfdb_set_sql_params] 0-cees-data-changetimerecorder:
>>> Failed to retrieve sql-db-sync from params.Assigning default value: off
>>> [2016-09-27 16:16:28.489288] W [MSGID: 101105]
>>> [gfdb_sqlite3.h:239:gfdb_set_sql_params] 0-cees-data-changetimerecorder:
>>> Failed to retrieve sql-db-autovacuum from params.Assigning default value:
>>> none
>>> [2016-09-27 16:16:28.490431] I [trash.c:2412:init] 0-cees-data-trash: no
>>> option specified for 'eliminate', using NULL
>>> [2016-09-27 16:16:28.672814] W [graph.c:357:_log_if_unknown_option]
>>> 0-cees-data-server: option 'rpc-auth.auth-glusterfs' is not recognized
>>> [2016-09-27 16:16:28.672854] W [graph.c:357:_log_if_unknown_option]
>>> 0-cees-data-server: option 'rpc-auth.auth-unix' is not recognized
>>> [2016-09-27 16:16:28.672872] W [graph.c:357:_log_if_unknown_option]
>>> 0-cees-data-server: option 'rpc-auth.auth-null' is not recognized
>>> [2016-09-27 16:16:28.672924] W [graph.c:357:_log_if_unknown_option]
>>> 0-cees-data-quota: option 'timeout' is not recognized
>>> [2016-09-27 16:16:28.672955] W [graph.c:357:_log_if_unknown_option]
>>> 0-cees-data-trash: option 'brick-path' is not recognized
>>> Final graph:
>>> +-----------------------------------------------------------
>>> -------------------+
>>>   1: volume cees-data-posix
>>>   2:     type storage/posix
>>>   3:     option glusterd-uuid f04b231e-63f8-4374-91ae-17c0c623f165
>>>   4:     option directory /data/brick
>>>   5:     option volume-id 27d2a59c-bdac-4f66-bcd8-e6124e53a4a2
>>>   6:     option update-link-count-parent on
>>>   7: end-volume
>>>   8:
>>>   9: volume cees-data-trash
>>>  10:     type features/trash
>>>  11:     option trash-dir .trashcan
>>>  12:     option brick-path /data/brick
>>>  13:     option trash-internal-op off
>>>  14:     subvolumes cees-data-posix
>>>  15: end-volume
>>>  16:
>>>  17: volume cees-data-changetimerecorder
>>>  18:     type features/changetimerecorder
>>>  19:     option db-type sqlite3
>>>  20:     option hot-brick off
>>>  21:     option db-name brick.db
>>>  22:     option db-path /data/brick/.glusterfs/
>>>  23:     option record-exit off
>>>  24:     option ctr_link_consistency off
>>>  25:     option ctr_lookupheal_link_timeout 300
>>>  26:     option ctr_lookupheal_inode_timeout 300
>>>  27:     option record-entry on
>>>  28:     option ctr-enabled off
>>>  29:     option record-counters off
>>>  30:     option ctr-record-metadata-heat off
>>>  31:     option sql-db-cachesize 1000
>>>  32:     option sql-db-wal-autocheckpoint 1000
>>>  33:     subvolumes cees-data-trash
>>>  34: end-volume
>>>  35:
>>>  36: volume cees-data-changelog
>>>  37:     type features/changelog
>>>  38:     option changelog-brick /data/brick
>>>  39:     option changelog-dir /data/brick/.glusterfs/changelogs
>>>  40:     option changelog-barrier-timeout 120
>>>  41:     subvolumes cees-data-changetimerecorder
>>>  42: end-volume
>>>  43:
>>>  44: volume cees-data-bitrot-stub
>>>  45:     type features/bitrot-stub
>>>  46:     option export /data/brick
>>>  47:     subvolumes cees-data-changelog
>>>  48: end-volume
>>>  49:
>>>  50: volume cees-data-access-control
>>>  51:     type features/access-control
>>>  52:     subvolumes cees-data-bitrot-stub
>>>  53: end-volume
>>>  54:
>>>  55: volume cees-data-locks
>>>  56:     type features/locks
>>>  57:     subvolumes cees-data-access-control
>>>  58: end-volume
>>>  59:
>>>  60: volume cees-data-upcall
>>>  61:     type features/upcall
>>>  62:     option cache-invalidation off
>>>  63:     subvolumes cees-data-locks
>>>  64: end-volume
>>>  65:
>>>  66: volume cees-data-io-threads
>>>  67:     type performance/io-threads
>>>  68:     subvolumes cees-data-upcall
>>>  69: end-volume
>>>  70:
>>>  71: volume cees-data-marker
>>>  72:     type features/marker
>>>  73:     option volume-uuid 27d2a59c-bdac-4f66-bcd8-e6124e53a4a2
>>>  74:     option timestamp-file /var/lib/glusterd/vols/cees-da
>>> ta/marker.tstamp
>>>  75:     option quota-version 1
>>>  76:     option xtime off
>>>  77:     option gsync-force-xtime off
>>>  78:     option quota on
>>>  79:     option inode-quota on
>>>  80:     subvolumes cees-data-io-threads
>>>  81: end-volume
>>>  82:
>>>  83: volume cees-data-barrier
>>>  84:     type features/barrier
>>>  85:     option barrier disable
>>>  86:     option barrier-timeout 120
>>>  87:     subvolumes cees-data-marker
>>>  88: end-volume
>>>  89:
>>>  90: volume cees-data-index
>>>  91:     type features/index
>>>  92:     option index-base /data/brick/.glusterfs/indices
>>>  93:     subvolumes cees-data-barrier
>>>  94: end-volume
>>>  95:
>>>  96: volume cees-data-quota
>>>  97:     type features/quota
>>>  98:     option transport.socket.connect-path
>>> /var/run/gluster/quotad.socket
>>>  99:     option transport-type socket
>>> 100:     option transport.address-family unix
>>> 101:     option volume-uuid cees-data
>>> 102:     option server-quota on
>>> 103:     option timeout 0
>>> 104:     option deem-statfs on
>>> 105:     subvolumes cees-data-index
>>> 106: end-volume
>>> 107:
>>> 108: volume cees-data-worm
>>> 109:     type features/worm
>>> 110:     option worm off
>>> 111:     subvolumes cees-data-quota
>>> 112: end-volume
>>> 113:
>>> 114: volume cees-data-read-only
>>> 115:     type features/read-only
>>> 116:     option read-only off
>>> 117:     subvolumes cees-data-worm
>>> 118: end-volume
>>> 119:
>>> 120: volume /data/brick
>>> 121:     type debug/io-stats
>>> 122:     option log-level INFO
>>> 123:     option latency-measurement off
>>> 124:     option count-fop-hits off
>>> 125:     subvolumes cees-data-read-only
>>> 126: end-volume
>>> 127:
>>> 128: volume cees-data-server
>>> 129:     type protocol/server
>>> 130:     option transport.socket.listen-port 49152
>>> 131:     option rpc-auth.auth-glusterfs on
>>> 132:     option rpc-auth.auth-unix on
>>> 133:     option rpc-auth.auth-null on
>>> 134:     option rpc-auth-allow-insecure on
>>> 135:     option transport.rdma.listen-port 49153
>>> 136:     option transport-type tcp,rdma
>>> 137:     option auth.login./data/brick.allow
>>> 18ddaf4c-ad98-4155-9372-717eae718b4c
>>> 138:     option auth.login.18ddaf4c-ad98-4155-9372-717eae718b4c.password
>>> 9e913e92-7de0-47f9-94ed-d08cbb130d23
>>> 139:     option auth.addr./data/brick.allow *
>>> 140:     subvolumes /data/brick
>>> 141: end-volume
>>> 142:
>>> +-----------------------------------------------------------
>>> -------------------+
>>> [2016-09-27 16:16:30.079541] I [login.c:81:gf_auth] 0-auth/login:
>>> allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c
>>> [2016-09-27 16:16:30.079567] I [MSGID: 115029]
>>> [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted
>>> client from fs3-12560-2016/09/27-16:16:30:47674-cees-data-client-3-0-0
>>> (version: 3.7.14)
>>> [2016-09-27 16:16:30.081487] I [login.c:81:gf_auth] 0-auth/login:
>>> allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c
>>> [2016-09-27 16:16:30.081505] I [MSGID: 115029]
>>> [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted
>>> client from fs2-11709-2016/09/27-16:16:30:50047-cees-data-client-3-0-0
>>> (version: 3.7.14)
>>> [2016-09-27 16:16:30.111091] I [login.c:81:gf_auth] 0-auth/login:
>>> allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c
>>> [2016-09-27 16:16:30.111113] I [MSGID: 115029]
>>> [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted
>>> client from fs2-11701-2016/09/27-16:16:29:24060-cees-data-client-3-0-0
>>> (version: 3.7.14)
>>> [2016-09-27 16:16:30.112822] I [login.c:81:gf_auth] 0-auth/login:
>>> allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c
>>> [2016-09-27 16:16:30.112836] I [MSGID: 115029]
>>> [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted
>>> client from fs3-12552-2016/09/27-16:16:29:23041-cees-data-client-3-0-0
>>> (version: 3.7.14)
>>> [2016-09-27 16:16:31.950978] I [login.c:81:gf_auth] 0-auth/login:
>>> allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c
>>> [2016-09-27 16:16:31.950998] I [MSGID: 115029]
>>> [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted
>>> client from fs1-6721-2016/09/27-16:16:26:939991-cees-data-client-3-0-0
>>> (version: 3.7.14)
>>> [2016-09-27 16:16:31.981977] I [login.c:81:gf_auth] 0-auth/login:
>>> allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c
>>> [2016-09-27 16:16:31.981994] I [MSGID: 115029]
>>> [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted
>>> client from fs1-6729-2016/09/27-16:16:27:971228-cees-data-client-3-0-0
>>> (version: 3.7.14)
>>>
>>>
>>> Hmm, this shows the brick has started.
>>> Does gluster volume info on fs4 shows all 4 bricks? (I guess it does
>>> based on your first email).
>>> Does gluster volume status on fs4  (or ps aux|grep glusterfsd) show the
>>> brick as running?
>>> Does gluster peer status on all nodes list the other 3 nodes as
>>> connected?
>>>
>>> If yes, you could try `service glusterd restart` on fs4 and see if if
>>> brings up the brick? I'm just shooting in the dark here for possible clues.
>>> -Ravi
>>>
>>> On Tue, Sep 27, 2016 at 8:46 AM, Ravishankar N <ravishankar at redhat.com>
>>> wrote:
>>>
>>>> On 09/27/2016 09:06 PM, Dennis Michael wrote:
>>>>
>>>> Yes, the brick log /var/log/glusterfs/bricks/data-brick.log is created
>>>> on fs4, and the snippets showing the errors were from that log.
>>>>
>>>> Unless I'm missing something, the snippet below is from glusterd's log
>>>> and not the brick's as is evident from the function names.
>>>> -Ravi
>>>>
>>>> Dennis
>>>>
>>>> On Mon, Sep 26, 2016 at 5:58 PM, Ravishankar N <ravishankar at redhat.com>
>>>> wrote:
>>>>
>>>>> On 09/27/2016 05:25 AM, Dennis Michael wrote:
>>>>>
>>>>>> [2016-09-26 22:44:39.254921] E [MSGID: 106005]
>>>>>> [glusterd-utils.c:4771:glusterd_brick_start] 0-management: Unable to
>>>>>> start brick fs4:/data/brick
>>>>>> [2016-09-26 22:44:39.254949] E [MSGID: 106074]
>>>>>> [glusterd-brick-ops.c:2372:glusterd_op_add_brick] 0-glusterd: Unable
>>>>>> to add bricks
>>>>>>
>>>>>
>>>>> Is the brick log created on fs4? Does it contain warnings/errors?
>>>>>
>>>>> -Ravi
>>>>>
>>>>>
>>>>
>>>>
>>>
>>>
>>
>>
>
>
> --
>
> --Atin
>



-- 

--Atin
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160929/65490330/attachment.html>


More information about the Gluster-users mailing list