[Gluster-users] Extremely high CPU load on one server

Dan Bretherton d.a.bretherton at reading.ac.uk
Thu Jan 31 18:46:36 UTC 2013


On 01/31/2013 05:37 PM, gluster-users-request at gluster.org wrote:
> Date: Thu, 31 Jan 2013 15:28:25 +0000 (UTC)
> From: Adri? Garc?a-Alz?rriz 	<adria.garcia-alzorriz at adam.es>
> To: gluster-users at gluster.org
> Subject: Re: [Gluster-users] Extremely high CPU load on one server
> Message-ID: <kee2ip$osk$1 at ger.gmane.org>
> Content-Type: text/plain; charset="us-ascii"
>
> El dia Thu, 31 Jan 2013 15:02:25 +0000, en/na Dan Bretherton va escriure:
>> Can anyone suggest a way to troubleshoot this problem?  The rebalance
>> logs don't show anything unusual but glustershd.log has a lot of
>> metadata split-brain warnings.   The brick logs are full of scary
>> looking warnings but none flagged 'E' or 'C'.  The trouble is that I see
>> messages like these on all the servers, and I can find nothing unusual
>> about the server with a CPU load of 70.  Users are complaining about
>> very poor performance, which has been going on or several weeks, so I
>> must at least find a work-around that allows people to work normally.
>>
>> -Dan.
> What's the output for your gluster volume info command?
> What kind of disks are you using for?
> How is the wait parameter in top?
Hello- Here is the information you requested; thanks.

> What's the output for your gluster volume info command?
There are ten volumes in the cluster, and the one I'm having the most 
trouble with is shown below.

[root at bdan11 ~]# gluster volume info atmos

Volume Name: atmos
Type: Distributed-Replicate
Volume ID: a4a7774e-cf4d-4a3e-9477-5c3d1b0efd93
Status: Started
Number of Bricks: 16 x 2 = 32
Transport-type: tcp
Bricks:
Brick1: bdan0.nerc-essc.ac.uk:/local/glusterfs
Brick2: bdan1.nerc-essc.ac.uk:/local/glusterfs
Brick3: bdan2.nerc-essc.ac.uk:/local/glusterfs
Brick4: bdan3.nerc-essc.ac.uk:/local/glusterfs
Brick5: bdan4.nerc-essc.ac.uk:/atmos/glusterfs
Brick6: bdan5.nerc-essc.ac.uk:/atmos/glusterfs
Brick7: bdan6.nerc-essc.ac.uk:/local/glusterfs
Brick8: bdan7.nerc-essc.ac.uk:/local/glusterfs
Brick9: bdan8.nerc-essc.ac.uk:/atmos/glusterfs
Brick10: bdan9.nerc-essc.ac.uk:/atmos/glusterfs
Brick11: bdan10.nerc-essc.ac.uk:/atmos/glusterfs
Brick12: bdan11.nerc-essc.ac.uk:/atmos/glusterfs
Brick13: bdan12.nerc-essc.ac.uk:/atmos/glusterfs
Brick14: bdan13.nerc-essc.ac.uk:/atmos/glusterfs
Brick15: bdan10.nerc-essc.ac.uk:/atmos2/glusterfs
Brick16: bdan11.nerc-essc.ac.uk:/atmos2/glusterfs
Brick17: bdan12.nerc-essc.ac.uk:/atmos2/glusterfs
Brick18: bdan13.nerc-essc.ac.uk:/atmos2/glusterfs
Brick19: bdan4.nerc-essc.ac.uk:/atmos2/glusterfs
Brick20: bdan5.nerc-essc.ac.uk:/atmos2/glusterfs
Brick21: bdan12.nerc-essc.ac.uk:/atmos3/glusterfs
Brick22: bdan13.nerc-essc.ac.uk:/atmos3/glusterfs
Brick23: bdan12.nerc-essc.ac.uk:/atmos4/glusterfs
Brick24: bdan13.nerc-essc.ac.uk:/atmos4/glusterfs
Brick25: bdan12.nerc-essc.ac.uk:/atmos5/glusterfs
Brick26: bdan13.nerc-essc.ac.uk:/atmos5/glusterfs
Brick27: pegasus.nerc-essc.ac.uk:/atmos/glusterfs
Brick28: bdan14.nerc-essc.ac.uk:/atmos/glusterfs
Brick29: bdan15.nerc-essc.ac.uk:/atmos/glusterfs
Brick30: bdan16.nerc-essc.ac.uk:/atmos/glusterfs
Brick31: bdan15.nerc-essc.ac.uk:/atmos2/glusterfs
Brick32: bdan16.nerc-essc.ac.uk:/atmos2/glusterfs
Options Reconfigured:
nfs.enable-ino32: on
server.allow-insecure: on
performance.stat-prefetch: off
performance.quick-read: off
cluster.min-free-disk: 338GB
nfs.rpc-auth-allow: 192.171.166.*,134.225.100.*
features.quota: off

> What kind of disks are you using for?
They are Hitachi Deskstar 7200 RPM 2TB SATA drives, attached to an Areca 
ARC1880 RAID controller.  I'm aware that it's an unusual choice of drive 
but the server vendor swears by them.  I have another pair of servers 
from the same vendor with 1TB Hitachi Deskstars and they have not given 
any GlusterFS related trouble.

The bricks are logical volumes on top of RAID-6 provided by the Areca.  
On this particular server they are formatted ext4, but I recently 
started using XFS for more recently added bricks on newer servers.  The 
bricks vary in size from 3.3TB to 500GB, depending on which volume they 
belong to.  I use smaller bricks for smaller volumes so that the brick 
size is an appropriate increment of volume growth.

> How is the wait parameter in top?
There isn't much I/O wait as far as I can gather.  Here is a view 
showing the top few processes.

top - 18:22:58 up  1:59,  1 user,  load average: 64.25, 64.36, 64.14
Tasks: 218 total,   9 running, 209 sleeping,   0 stopped,   0 zombie
Cpu(s): 78.2%us, 20.7%sy,  0.0%ni,  0.9%id,  0.0%wa,  0.0%hi, 0.1%si,  
0.0%st
Mem:  12290244k total,  6639128k used,  5651116k free,  2752944k buffers
Swap: 14352376k total,        0k used, 14352376k free,  1441712k cached

   PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+ COMMAND
  3661 root      15   0  289m  42m 2168 R 85.0  0.4 101:54.33 glusterfsd
  3697 root      15   0  304m  37m 2112 S 83.1  0.3  75:53.43 glusterfsd
  3655 root      15   0  289m  42m 2160 S 79.4  0.4  97:11.52 glusterfsd
  3685 root      15   0  550m  62m 2100 S 77.5  0.5 109:07.53 glusterfsd
  3691 root      15   0  483m  40m 2100 S 75.6  0.3  76:10.93 glusterfsd
  3679 root      15   0  272m  20m 2124 R  1.9  0.2   4:41.69 glusterfsd
     1 root      15   0 10368  692  580 S  0.0  0.0   0:00.74 init

It's having a breather; the load has gone down to 64!  You can see more 
load information on my Ganglia web frontend, here:

http://lovejoy.nerc-essc.ac.uk/ganglia/?c=ESSC%20Storage%20Cluster

The five glusterfsd processes using the most resources, listed in top 
above, are for bricks belonging to two different volumes, so I can't 
blame just one user or group for the load; each research group in the 
department has their own volume. That probably disproves my theory that 
poorly distributed files is the cause of the problem. There are bricks 
belonging to six different volumes on this server.

One thing I do notice about these five glusterfsd processes is that they 
all belong to the two volumes that are having rebalance...fix-layout 
performed on them following the addition of new bricks recently.  The 
new bricks are on a different pair of servers.  This does seem to point 
the finger at the fix-layout as the cause of the high load, but why 
would only one server be affected in this way?

-Dan









More information about the Gluster-users mailing list