[Gluster-devel] [Gluster-users] 3.6.6 issues
David Robinson
david.robinson at corvidtec.com
Tue Oct 20 15:16:13 UTC 2015
Ben,
It is not still in this state. It seems to have settled out over the
weekend and is behaving normally. I am guessing here, but I think the
issue might have been a problem with 3.6.3 clients connecting to the
3.6.6 server. We have upgraded all of our clients to 3.6.6 and are no
longer having the problem.
Is there a way to check the client version of all clients connected to a
gluster server?
David
------ Original Message ------
From: "Ben Turner" <bturner at redhat.com>
To: "David Robinson" <drobinson at corvidtec.com>
Cc: gluster-users at gluster.org; "Gluster Devel"
<gluster-devel at gluster.org>
Sent: 10/19/2015 8:20:10 PM
Subject: Re: [Gluster-users] 3.6.6 issues
>Hi David. Is the cluster still in this state? If so can you grab a
>couple stack traces from the offending brick (gfs01a) process with
>gstack? Make sure that its the brick process spinning your CPUs with
>top or something, we want to be sure the stack traces are from the
>offending process. That will give us an idea of what it is chewing on.
> Other than that maybe you could take a couple sosreports on the
>servers and open a BZ. It may be a good idea to roll back versions
>until we can get this sorted, I don't know how long you can have the
>cluster in this state. Once you get a bugzilla open I'll try to repro
>what you are seeing to see if this is reproducible.
>
>-b
>
>----- Original Message -----
>> From: "David Robinson" <david.robinson at corvidtec.com>
>> To: gluster-users at gluster.org, "Gluster Devel"
>><gluster-devel at gluster.org>
>> Sent: Saturday, October 17, 2015 12:19:36 PM
>> Subject: [Gluster-users] 3.6.6 issues
>>
>> I upgraded my storage server from 3.6.3 to 3.6.6 and am now having
>>issues. My
>> setup (4x2) is shown below. One of the bricks (gfs01a) has a very
>>high
>> cpu-load even though the load on the other 3-bricks (gfs01b, gfs02a,
>>gfs02b)
>> is almost zero. The FUSE mounted partition is extremely slow and
>>basically
>> unuseable since the upgrade. I am getting a lot of the messages shown
>>below
>> in the logs on gfs01a and gfs01b. Nothing out of the ordinary is
>>showing up
>> on the gfs02a/gfs02b bricks.
>> Can someone help?
>> [root at gfs01b glusterfs]# gluster volume info homegfs
>>
>> Volume Name: homegfs
>> Type: Distributed-Replicate
>> Volume ID: 1e32672a-f1b7-4b58-ba94-58c085e59071
>> Status: Started
>> Number of Bricks: 4 x 2 = 8
>> Transport-type: tcp
>> Bricks:
>> Brick1: gfsib01a.corvidtec.com:/data/brick01a/homegfs
>> Brick2: gfsib01b.corvidtec.com:/data/brick01b/homegfs
>> Brick3: gfsib01a.corvidtec.com:/data/brick02a/homegfs
>> Brick4: gfsib01b.corvidtec.com:/data/brick02b/homegfs
>> Brick5: gfsib02a.corvidtec.com:/data/brick01a/homegfs
>> Brick6: gfsib02b.corvidtec.com:/data/brick01b/homegfs
>> Brick7: gfsib02a.corvidtec.com:/data/brick02a/homegfs
>> Brick8: gfsib02b.corvidtec.com:/data/brick02b/homegfs
>> Options Reconfigured:
>> changelog.rollover-time: 15
>> changelog.fsync-interval: 3
>> changelog.changelog: on
>> geo-replication.ignore-pid-check: on
>> geo-replication.indexing: off
>> storage.owner-gid: 100
>> network.ping-timeout: 10
>> server.allow-insecure: on
>> performance.write-behind-window-size: 128MB
>> performance.cache-size: 128MB
>> performance.io-thread-count: 32
>> server.manage-gids: on
>> [root@ gfs01a glusterfs]# tail -f cli.log
>> [2015-10-17 16:05:44.299933] I [socket.c:2353:socket_event_handler]
>> 0-transport: disconnecting now
>> [2015-10-17 16:05:44.331233] I [input.c:36:cli_batch] 0-: Exiting
>>with: 0
>> [2015-10-17 16:06:33.397631] I [socket.c:2353:socket_event_handler]
>> 0-transport: disconnecting now
>> [2015-10-17 16:06:33.432970] I [input.c:36:cli_batch] 0-: Exiting
>>with: 0
>> [2015-10-17 16:11:22.441290] I [socket.c:2353:socket_event_handler]
>> 0-transport: disconnecting now
>> [2015-10-17 16:11:22.472227] I [input.c:36:cli_batch] 0-: Exiting
>>with: 0
>> [2015-10-17 16:15:44.176391] I [socket.c:2353:socket_event_handler]
>> 0-transport: disconnecting now
>> [2015-10-17 16:15:44.205064] I [input.c:36:cli_batch] 0-: Exiting
>>with: 0
>> [2015-10-17 16:16:33.366424] I [socket.c:2353:socket_event_handler]
>> 0-transport: disconnecting now
>> [2015-10-17 16:16:33.377160] I [input.c:36:cli_batch] 0-: Exiting
>>with: 0
>> [root@ gfs01a glusterfs]# tail etc-glusterfs-glusterd.vol.log
>> [2015-10-17 15:56:33.177207] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume Source
>> [2015-10-17 16:01:22.303635] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume Software
>> [2015-10-17 16:05:44.320555] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume homegfs
>> [2015-10-17 16:06:17.204783] W [rpcsvc.c:254:rpcsvc_program_actor]
>> 0-rpc-service: RPC program not available (req 1298437 330)
>> [2015-10-17 16:06:17.204811] E
>>[rpcsvc.c:544:rpcsvc_check_and_reply_error]
>> 0-rpcsvc: rpc actor failed to complete successfully
>> [2015-10-17 16:06:33.408695] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume Source
>> [2015-10-17 16:11:22.462374] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume Software
>> [2015-10-17 16:12:30.608092] E
>>[glusterd-op-sm.c:207:glusterd_get_txn_opinfo]
>> 0-: Unable to get transaction opinfo for transaction ID :
>> d143b66b-2ac9-4fd9-8635-fe1eed41d56b
>> [2015-10-17 16:15:44.198292] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume homegfs
>> [2015-10-17 16:16:33.368170] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume Source
>> [root@ gfs01b glusterfs]# tail -f glustershd.log
>> [2015-10-17 16:11:45.996447] I
>> [afr-self-heal-metadata.c:54:__afr_selfheal_metadata_do]
>> 0-homegfs-replicate-1: performing metadata selfheal on
>> 0a65d73a-a416-418e-92f0-5cec7d240433
>> [2015-10-17 16:11:46.030947] I
>>[afr-self-heal-common.c:476:afr_log_selfheal]
>> 0-homegfs-replicate-1: Completed metadata selfheal on
>> 0a65d73a-a416-418e-92f0-5cec7d240433. source=1 sinks=0
>> [2015-10-17 16:11:46.031241] W
>>[client-rpc-fops.c:2772:client3_3_lookup_cbk]
>> 0-homegfs-client-3: remote operation failed: No such file or
>>directory.
>> Path: <gfid:d2714957-0c83-4ab2-8cfc-1931c8e9d0bf>
>> (d2714957-0c83-4ab2-8cfc-1931c8e9d0bf)
>> [2015-10-17 16:11:46.031633] W
>>[client-rpc-fops.c:2772:client3_3_lookup_cbk]
>> 0-homegfs-client-3: remote operation failed: No such file or
>>directory.
>> Path: <gfid:87c5f875-c3e7-4b14-807a-4e6d940750fc>
>> (87c5f875-c3e7-4b14-807a-4e6d940750fc)
>> [2015-10-17 16:11:47.043367] W
>>[client-rpc-fops.c:2772:client3_3_lookup_cbk]
>> 0-homegfs-client-3: remote operation failed: No such file or
>>directory.
>> Path: <gfid:d2714957-0c83-4ab2-8cfc-1931c8e9d0bf>
>> (d2714957-0c83-4ab2-8cfc-1931c8e9d0bf)
>> [2015-10-17 16:11:47.054199] W
>>[client-rpc-fops.c:2772:client3_3_lookup_cbk]
>> 0-homegfs-client-3: remote operation failed: No such file or
>>directory.
>> Path: <gfid:87c5f875-c3e7-4b14-807a-4e6d940750fc>
>> (87c5f875-c3e7-4b14-807a-4e6d940750fc)
>> [2015-10-17 16:12:48.001869] W
>>[client-rpc-fops.c:2772:client3_3_lookup_cbk]
>> 0-homegfs-client-3: remote operation failed: No such file or
>>directory.
>> Path: <gfid:d2714957-0c83-4ab2-8cfc-1931c8e9d0bf>
>> (d2714957-0c83-4ab2-8cfc-1931c8e9d0bf)
>> [2015-10-17 16:12:48.012671] W
>>[client-rpc-fops.c:2772:client3_3_lookup_cbk]
>> 0-homegfs-client-3: remote operation failed: No such file or
>>directory.
>> Path: <gfid:87c5f875-c3e7-4b14-807a-4e6d940750fc>
>> (87c5f875-c3e7-4b14-807a-4e6d940750fc)
>> [2015-10-17 16:13:49.011591] W
>>[client-rpc-fops.c:2772:client3_3_lookup_cbk]
>> 0-homegfs-client-3: remote operation failed: No such file or
>>directory.
>> Path: <gfid:d2714957-0c83-4ab2-8cfc-1931c8e9d0bf>
>> (d2714957-0c83-4ab2-8cfc-1931c8e9d0bf)
>> [2015-10-17 16:13:49.018600] W
>>[client-rpc-fops.c:2772:client3_3_lookup_cbk]
>> 0-homegfs-client-3: remote operation failed: No such file or
>>directory.
>> Path: <gfid:87c5f875-c3e7-4b14-807a-4e6d940750fc>
>> (87c5f875-c3e7-4b14-807a-4e6d940750fc)
>> [root@ gfs01b glusterfs]# tail cli.log
>> [2015-10-16 10:52:16.002922] I [input.c:36:cli_batch] 0-: Exiting
>>with: 0
>> [2015-10-16 10:52:16.167432] I [socket.c:2353:socket_event_handler]
>> 0-transport: disconnecting now
>> [2015-10-16 10:52:18.248024] I [input.c:36:cli_batch] 0-: Exiting
>>with: 0
>> [2015-10-17 16:12:30.607603] I [socket.c:2353:socket_event_handler]
>> 0-transport: disconnecting now
>> [2015-10-17 16:12:30.628810] I [input.c:36:cli_batch] 0-: Exiting
>>with: 0
>> [2015-10-17 16:12:33.992818] I [socket.c:2353:socket_event_handler]
>> 0-transport: disconnecting now
>> [2015-10-17 16:12:33.998944] I [input.c:36:cli_batch] 0-: Exiting
>>with: 0
>> [2015-10-17 16:12:38.604461] I [socket.c:2353:socket_event_handler]
>> 0-transport: disconnecting now
>> [2015-10-17 16:12:38.605532] I
>>[cli-rpc-ops.c:588:gf_cli_get_volume_cbk]
>> 0-cli: Received resp to get vol: 0
>> [2015-10-17 16:12:38.605659] I [input.c:36:cli_batch] 0-: Exiting
>>with: 0
>> [root@ gfs01b glusterfs]# tail etc-glusterfs-glusterd.vol.log
>> [2015-10-16 14:29:56.495120] E [rpcsvc.c:617:rpcsvc_handle_rpc_call]
>> 0-rpc-service: Request received from non-privileged port. Failing
>>request
>> [2015-10-16 14:29:59.369109] E [rpcsvc.c:617:rpcsvc_handle_rpc_call]
>> 0-rpc-service: Request received from non-privileged port. Failing
>>request
>> [2015-10-16 14:29:59.512093] E [rpcsvc.c:617:rpcsvc_handle_rpc_call]
>> 0-rpc-service: Request received from non-privileged port. Failing
>>request
>> [2015-10-16 14:30:02.383574] E [rpcsvc.c:617:rpcsvc_handle_rpc_call]
>> 0-rpc-service: Request received from non-privileged port. Failing
>>request
>> [2015-10-16 14:30:02.529206] E [rpcsvc.c:617:rpcsvc_handle_rpc_call]
>> 0-rpc-service: Request received from non-privileged port. Failing
>>request
>> [2015-10-16 16:01:20.389100] E [rpcsvc.c:617:rpcsvc_handle_rpc_call]
>> 0-rpc-service: Request received from non-privileged port. Failing
>>request
>> [2015-10-17 16:12:30.611161] W
>> [glusterd-op-sm.c:4066:glusterd_op_modify_op_ctx] 0-management:
>>op_ctx
>> modification failed
>> [2015-10-17 16:12:30.612433] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume Software
>> [2015-10-17 16:12:30.618444] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume Source
>> [2015-10-17 16:12:30.624005] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume homegfs
>> [2015-10-17 16:12:33.993869] I
>> [glusterd-handler.c:3836:__glusterd_handle_status_volume]
>>0-management:
>> Received status volume req for volume homegfs
>> [2015-10-17 16:12:38.605389] I
>> [glusterd-handler.c:1296:__glusterd_handle_cli_get_volume]
>>0-glusterd:
>> Received get vol req
>> [root at gfs01b glusterfs]# gluster volume status homegfs
>> Status of volume: homegfs
>> Gluster process Port Online Pid
>>
>>------------------------------------------------------------------------------
>> Brick gfsib01a.corvidtec.com:/data/brick01a/homegfs 49152 Y 3820
>> Brick gfsib01b.corvidtec.com:/data/brick01b/homegfs 49152 Y 3808
>> Brick gfsib01a.corvidtec.com:/data/brick02a/homegfs 49153 Y 3825
>> Brick gfsib01b.corvidtec.com:/data/brick02b/homegfs 49153 Y 3813
>> Brick gfsib02a.corvidtec.com:/data/brick01a/homegfs 49152 Y 3967
>> Brick gfsib02b.corvidtec.com:/data/brick01b/homegfs 49152 Y 3952
>> Brick gfsib02a.corvidtec.com:/data/brick02a/homegfs 49153 Y 3972
>> Brick gfsib02b.corvidtec.com:/data/brick02b/homegfs 49153 Y 3957
>> NFS Server on localhost 2049 Y 3822
>> Self-heal Daemon on localhost N/A Y 3827
>> NFS Server on 10.200.70.1 2049 Y 3834
>> Self-heal Daemon on 10.200.70.1 N/A Y 3839
>> NFS Server on gfsib02a.corvidtec.com 2049 Y 3981
>> Self-heal Daemon on gfsib02a.corvidtec.com N/A Y 3986
>> NFS Server on gfsib02b.corvidtec.com 2049 Y 3966
>> Self-heal Daemon on gfsib02b.corvidtec.com N/A Y 3971
>>
>> Task Status of Volume homegfs
>>
>>------------------------------------------------------------------------------
>> Task : Rebalance
>> ID : 58b6cc76-c29c-4695-93fe-c42b1112e171
>> Status : completed
>>
>>
>>
>> ========================
>>
>>
>>
>> David F. Robinson, Ph.D.
>>
>> President - Corvid Technologies
>>
>> 145 Overhill Drive
>>
>> Mooresville, NC 28117
>>
>> 704.799.6944 x101 [Office]
>>
>> 704.252.1310 [Cell]
>>
>> 704.799.7974 [Fax]
>>
>> d avid.robinson at corvidtec.com
>>
>> http://www.corvidtec.com
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-users
More information about the Gluster-devel
mailing list