[Gluster-users] Gluster speed sooo slow
Ivan Dimitrov
dobber at amln.net
Fri Aug 10 11:06:08 UTC 2012
So I stopped a node to check the BIOS and after it went up, the
rebalance kicked in. I was looking for those kind of speeds on a normal
write. The rebalance is much faster than my rsync/cp.
https://dl.dropbox.com/u/282332/Screen%20Shot%202012-08-10%20at%202.04.09%20PM.png
Best Regards
Ivan Dimitrov
On 8/10/12 1:23 PM, Ivan Dimitrov wrote:
> Hello
> What am I doing wrong?!?
>
> I have a test setup with 4 identical servers with 2 disks each in
> distribute-replicate 2. All servers are connected to a GB switch.
>
> I am experiencing really slow speeds at anything I do. Slow write,
> slow read, not to mention random write/reads.
>
> Here is an example:
> random-files is a directory with 32768 files with average size 16kb.
> [root at gltclient]:~# rsync -a /root/speedtest/random-files/
> /home/gltvolume/
> ^^ This will take more than 3 hours.
>
> On any of the servers if I do "iostat" the disks are not loaded at all:
> https://dl.dropbox.com/u/282332/Screen%20Shot%202012-08-10%20at%201.08.54%20PM.png
>
>
> This is similar result for all servers.
>
> Here is an example of simple "ls" command on the content.
> [root at gltclient]:~# unalias ls
> [root at gltclient]:~# /usr/bin/time -f "%e seconds" ls /home/gltvolume/
> | wc -l
> 2.81 seconds
> 5393
>
> almost 3 seconds to display 5000 files?!?! When they are 32,000, the
> ls will take around 35-45 seconds.
>
> This directory is on local disk:
> [root at gltclient]:~# /usr/bin/time -f "%e seconds" ls
> /root/speedtest/random-files/ | wc -l
> 1.45 seconds
> 32768
>
> [root at gltclient]:~# /usr/bin/time -f "%e seconds" cat
> /home/gltvolume/* >/dev/null
> 190.50 seconds
>
> [root at gltclient]:~# /usr/bin/time -f "%e seconds" du -sh /home/gltvolume/
> 126M /home/gltvolume/
> 75.23 seconds
>
>
> Here is the volume information.
>
> [root at glt1]:~# gluster volume info
>
> Volume Name: gltvolume
> Type: Distributed-Replicate
> Volume ID: 16edd852-8d23-41da-924d-710b753bb374
> Status: Started
> Number of Bricks: 4 x 2 = 8
> Transport-type: tcp
> Bricks:
> Brick1: 1.1.74.246:/home/sda3
> Brick2: glt2.network.net:/home/sda3
> Brick3: 1.1.74.246:/home/sdb1
> Brick4: glt2.network.net:/home/sdb1
> Brick5: glt3.network.net:/home/sda3
> Brick6: gltclient.network.net:/home/sda3
> Brick7: glt3.network.net:/home/sdb1
> Brick8: gltclient.network.net:/home/sdb1
> Options Reconfigured:
> performance.io-thread-count: 32
> performance.cache-size: 256MB
> cluster.self-heal-daemon: on
>
>
> [root at glt1]:~# gluster volume status all detail
> Status of volume: gltvolume
> ------------------------------------------------------------------------------
>
> Brick : Brick 1.1.74.246:/home/sda3
> Port : 24009
> Online : Y
> Pid : 1479
> File System : ext4
> Device : /dev/sda3
> Mount Options : rw,noatime
> Inode Size : 256
> Disk Space Free : 179.3GB
> Total Disk Space : 179.7GB
> Inode Count : 11968512
> Free Inodes : 11901550
> ------------------------------------------------------------------------------
>
> Brick : Brick glt2.network.net:/home/sda3
> Port : 24009
> Online : Y
> Pid : 1589
> File System : ext4
> Device : /dev/sda3
> Mount Options : rw,noatime
> Inode Size : 256
> Disk Space Free : 179.3GB
> Total Disk Space : 179.7GB
> Inode Count : 11968512
> Free Inodes : 11901550
> ------------------------------------------------------------------------------
>
> Brick : Brick 1.1.74.246:/home/sdb1
> Port : 24010
> Online : Y
> Pid : 1485
> File System : ext4
> Device : /dev/sdb1
> Mount Options : rw,noatime
> Inode Size : 256
> Disk Space Free : 228.8GB
> Total Disk Space : 229.2GB
> Inode Count : 15269888
> Free Inodes : 15202933
> ------------------------------------------------------------------------------
>
> Brick : Brick glt2.network.net:/home/sdb1
> Port : 24010
> Online : Y
> Pid : 1595
> File System : ext4
> Device : /dev/sdb1
> Mount Options : rw,noatime
> Inode Size : 256
> Disk Space Free : 228.8GB
> Total Disk Space : 229.2GB
> Inode Count : 15269888
> Free Inodes : 15202933
> ------------------------------------------------------------------------------
>
> Brick : Brick glt3.network.net:/home/sda3
> Port : 24009
> Online : Y
> Pid : 28963
> File System : ext4
> Device : /dev/sda3
> Mount Options : rw,noatime
> Inode Size : 256
> Disk Space Free : 179.3GB
> Total Disk Space : 179.7GB
> Inode Count : 11968512
> Free Inodes : 11906058
> ------------------------------------------------------------------------------
>
> Brick : Brick gltclient.network.net:/home/sda3
> Port : 24009
> Online : Y
> Pid : 3145
> File System : ext4
> Device : /dev/sda3
> Mount Options : rw,noatime
> Inode Size : 256
> Disk Space Free : 179.3GB
> Total Disk Space : 179.7GB
> Inode Count : 11968512
> Free Inodes : 11906058
> ------------------------------------------------------------------------------
>
> Brick : Brick glt3.network.net:/home/sdb1
> Port : 24010
> Online : Y
> Pid : 28969
> File System : ext4
> Device : /dev/sdb1
> Mount Options : rw,noatime
> Inode Size : 256
> Disk Space Free : 228.8GB
> Total Disk Space : 229.2GB
> Inode Count : 15269888
> Free Inodes : 15207375
> ------------------------------------------------------------------------------
>
> Brick : Brick gltclient.network.net:/home/sdb1
> Port : 24010
> Online : Y
> Pid : 3151
> File System : ext4
> Device : /dev/sdb1
> Mount Options : rw,noatime
> Inode Size : 256
> Disk Space Free : 228.8GB
> Total Disk Space : 229.2GB
> Inode Count : 15269888
> Free Inodes : 15207375
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
More information about the Gluster-users
mailing list