[Gluster-users] Slow read performance
Joe Julian
joe at julianfamily.org
Mon Mar 11 17:27:55 UTC 2013
No, the context is quite appropriate. Linus has made it clear for a very
long time that the kernel is not to break userspace programs with no
exceptions.
That said, however, Avati and Ted have been working together to
implement a fix in the kernel. No idea on an eta.
On 03/11/2013 10:18 AM, Rodrigo Severo wrote:
> I know where this statement came from. I believe you are both:
>
> * trying to apply some statement on a context it's not pertinent to and
> * fouling yourself and/or others arguing that this issue will/should
> be fixed in the kernel.
>
> The ext4 hash size change was applied in the kernel an year ago. I
> don't believe it will be undone. Gluster developers could argue that
> this change was hard on them, and that it shouldn't be backported to
> Enterprise kernels but after one year not having fixed it is on
> Gluster developers. Arguing otherwise seems rather foolish to me.
>
>
> Regards,
>
> Rodrigo Severo
>
>
>
> On Mon, Mar 11, 2013 at 2:00 PM, Joe Julian <joe at julianfamily.org
> <mailto:joe at julianfamily.org>> wrote:
>
> This bug is in the kernel.
>
> "If a change results in user programs breaking, it's a bug in the
> kernel. We never EVER blame the user programs." - Linus Torvalds
>
> http://lkml.org/lkml/2012/12/23/75
>
> On 03/08/2013 12:42 PM, Stephan von Krawczynski wrote:
>> I really do wonder if this bug in _glusterfs_ is not fixed. It really makes no
>> sense to do an implementation that breaks on the most used fs on linux.
>> And just as you said: don't wait on btrfs, it will never be production-ready.
>> And xfs is no solution, it is just a bad work-around.
>>
>>
>> On Fri, 8 Mar 2013 10:43:41 -0800
>> Bryan Whitehead<driver at megahappy.net> <mailto:driver at megahappy.net> wrote:
>>
>>> Here are some details about ext4 changes in the kernel screwing up
>>> glusterfs:
>>> http://www.gluster.org/2012/08/glusterfs-bit-by-ext4-structure-change/
>>> https://bugzilla.redhat.com/show_bug.cgi?id=838784
>>>
>>> I thought I read there was a work-around in recent versions of gluster but
>>> I think it came at a cost somewhere. I'm not sure since I've been using xfs
>>> since the 1.x days of gluster and only see random ext3/4 problems bubble up
>>> on these maillinglist. In general, ext4 was just a stopgap for the wait on
>>> btrfs getting flushed out. That said, I don't see ext4 going away for a
>>> long long time. :-/
>>>
>>> NOTE: I don't even know if this is your problem. You might try updating 2
>>> bricks that are replica pairs to use xfs then do some performance tests on
>>> files living on them to confirm. Example, you have 20 some servers/bricks.
>>> If hostD and hostE are replica pairs for some subset of files, shutdown
>>> glusterd on HostD, change fs to xfs, fire glusterd back up - let it resync
>>> and recover all the files, do the same on hostE (once hostD is good), then
>>> see if there is a read speed improvement for files living on those two host
>>> pairs.
>>>
>>>
>>> On Fri, Mar 8, 2013 at 6:40 AM, Thomas Wakefield<twake at cola.iges.org> <mailto:twake at cola.iges.org>wrote:
>>>
>>>> I am still confused how ext4 is suddenly slow to read when it's behind
>>>> Gluster, but plenty fast stand alone reading?
>>>>
>>>> And it writes really fast from both the server and client.
>>>>
>>>>
>>>>
>>>> On Mar 8, 2013, at 4:07 AM, "Jon Tegner"<tegner at renget.se> <mailto:tegner at renget.se> wrote:
>>>>
>>>> We had issues with ext4 about a bit less than a year ago, at that time I
>>>> upgraded the servers to CentOS-6.2. But that gave us large problems (more
>>>> than slow reads). Since I didn't want to reformat the disks at that time
>>>> (and switch to XFS) I went back to CentOS-5.5 (which we had used before).
>>>> On some link (think it was
>>>> https://bugzilla.redhat.com/show_bug.cgi?id=713546 but can't seem to
>>>> reach that now) it was stated that the ext4-issue was present even on later
>>>> versions of CentOS-5 (I _think_ 5.8 was affected).
>>>>
>>>> Are there hope that the "ext4-issue" will be solved in later
>>>> kernels/versions of gluster? If not, it seems one is eventually forced to
>>>> switch to XFS.
>>>>
>>>> Regards,
>>>>
>>>> /jon
>>>>
>>>> On Mar 8, 2013 03:27 "Thomas Wakefield"<twake at iges.org> <mailto:twake at iges.org> <twake at iges.org> <mailto:twake at iges.org>wrote:
>>>>
>>>>
>>>> inode size is 256.
>>>>
>>>>
>>>>
>>>>
>>>> Pretty stuck with these settings and ext4. I missed the memo that
>>>> Gluster started to prefer xfs, back in the 2.x days xfs was not the
>>>> preferred filesystem. At this point it's a 340TB filesystem with 160TB
>>>> used. I just added more space, and was doing some followup testing and
>>>> wasn't impressed with the results. But I am sure I was happier before
>>>> with the performance.
>>>>
>>>>
>>>>
>>>>
>>>> Still running CentOS 5.8
>>>>
>>>>
>>>>
>>>>
>>>> Anything else I could look at?
>>>>
>>>>
>>>>
>>>>
>>>> Thanks, Tom
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Mar 7, 2013, at 5:04 PM, Bryan Whitehead <<driver at megahappy.net> <mailto:driver at megahappy.net>>
>>>> wrote:
>>>>
>>>>
>>>>> I'm sure you know, but xfs is the recommended filesystem for
>>>>> glusterfs. Ext4 has a number of issues. (Particularly on
>>>>> CentOS/Redhat6).
>>>>>
>>>>>
>>>>>
>>>>> The default inode size for ext4 (and xfs) is small for the number of
>>>>> extended attributes glusterfs uses. This causes a minor hit in
>>>>> performance on xfs if theextended attributes grow more than 265 (xfs
>>>>> default size). In xfs, this is fixed by setting the size of an inode
>>>>> to 512. How big the impact is on ext4 is something I don't know
>>>>> offhand. But looking at a couple of boxes I have it looks like some
>>>>> ext4 filesystems have 128 inode size and some have 256 inode size
>>>>> (both of which are too small for glusterfs). The performance hit is
>>>>> everytimeextended attributes need to be read several inodes need to be
>>>>> seeked and found.
>>>>>
>>>>>
>>>>>
>>>>> run "dumpe2fs -h <blockdevice> | grep size" on your ext4 mountpoints.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> If it is not too much of a bother - I'd try xfs as your filesystem for
>>>>> the bricks
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> mkfs.xfs -i size=512 <blockdevice>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> Please see this for more detailed info:
>>>>>
>>>>> <https://access.redhat.com/knowledge/docs/en-US/Red_Hat_Storage/2.0/ht
>>>>> ml-single/Administration_Guide/index.html#chap-User_Guide-Setting_Volu
>>>>> mes> <https://access.redhat.com/knowledge/docs/en-US/Red_Hat_Storage/2.0/html-single/Administration_Guide/index.html#chap-User_Guide-Setting_Volumes>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Thu, Mar 7, 2013 at 12:08 PM, Thomas Wakefield
>>>>> <<twake at cola.iges.org> <mailto:twake at cola.iges.org>> wrote:
>>>>>
>>>>>> Everything is built as ext4, no options other than
>>>>>> lazy_itable_init=1 when I built the filesystems.
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Server mount example:
>>>>>>
>>>>>> LABEL=disk2a /storage/disk2a ext4defaults 0 0
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Client mount:
>>>>>>
>>>>>> fs-disk2:/shared /shared glusterfs defaults 0 0
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Remember, the slow reads are only from gluster clients, the disks
>>>>>> are really fast when I am local on the server testing the disks.
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> -Tom
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Mar 7, 2013, at 1:09 PM, Bryan Whitehead <<driver at megahappy.net> <mailto:driver at megahappy.net>>
>>>>>> wrote:
>>>>>>
>>>>>>
>>>>>>> Was just thinking, what is your mount options for your bricks
>>>>>>> (using inode64?)? Also, you are using xfs... right?
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> When you created the filesystems did you allocate more inode
>>>>>>> space? -i size=512 ?
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Thu, Mar 7, 2013 at 5:49 AM, Thomas Wakefield
>>>>>>> <<twake at cola.iges.org> <mailto:twake at cola.iges.org>> wrote:
>>>>>>>
>>>>>>>> Still looking for help.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mar 4, 2013, at 7:43 AM, Thomas Wakefield <<twake at iges.org> <mailto:twake at iges.org>>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>>> Also, I tested an NFS mount over the same 10GB link, and was
>>>>>>>>> able to pull almost 200MB/s. But Gluster is still much slower.
>>>>>>>>> Also I tested running it for a longer test, 105GB of data, and
>>>>>>>>> still showed that writing is MUCH faster. Which makes no sense
>>>>>>>>> when the disks can read 2x as fast as they can write.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Any other thoughts?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> [root at cpu_crew1 ~]# dd if=/dev/zero
>>>>>>>>> of=/shared/working/benchmark/test.cpucrew1 bs=512k
>>>>>>>>> count=200000 ; dd if=/shared/working/benchmark/test.cpucrew1
>>>>>>>>> of=/dev/null bs=512k
>>>>>>>>>
>>>>>>>>> 200000+0 records in
>>>>>>>>>
>>>>>>>>> 200000+0 records out
>>>>>>>>>
>>>>>>>>> 104857600000 bytes (105 GB) copied, 159.135 seconds, 659 MB/s
>>>>>>>>>
>>>>>>>>> 200000+0 records in
>>>>>>>>>
>>>>>>>>> 200000+0 records out
>>>>>>>>>
>>>>>>>>> 104857600000 bytes (105 GB) copied, 1916.87 seconds, 54.7 MB/s
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Mar 1, 2013, at 9:58 AM, Thomas Wakefield
>>>>>>>>> <<twake at iges.org> <mailto:twake at iges.org>> wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> The max setting for performance.read-ahead-page-count is 16,
>>>>>>>>>> which I did just try. No significant change.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Any other setting options?
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Feb 28, 2013, at 10:18 PM, Anand Avati
>>>>>>>>>> <<anand.avati at gmail.com> <mailto:anand.avati at gmail.com>> wrote:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>> Can you try "gluster volume set <volname>
>>>>>>>>>>> performance.read-ahead-page-count 64" or some value higher
>>>>>>>>>>> or lower?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Avati
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Thu, Feb 28, 2013 at 7:15 PM, Thomas Wakefield
>>>>>>>>>>> <<twake at iges.org> <mailto:twake at iges.org>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Good point, forgot to set a blcoksize, here are the
>>>>>>>>>>>> redone dd tests:
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> [root at cpu_crew1 ~]# dd
>>>>>>>>>>>> if=/shared/working/benchmark/test.cpucrew1 of=/dev/null
>>>>>>>>>>>> bs=128k
>>>>>>>>>>>>
>>>>>>>>>>>> 40000+0 records in
>>>>>>>>>>>>
>>>>>>>>>>>> 40000+0 records out
>>>>>>>>>>>>
>>>>>>>>>>>> 5242880000 bytes (5.2 GB) copied, 65.4928 seconds, 80.1
>>>>>>>>>>>> MB/s
>>>>>>>>>>>>
>>>>>>>>>>>> [root at cpu_crew1 ~]# dd
>>>>>>>>>>>> if=/shared/working/benchmark/test.cpucrew1 of=/dev/null
>>>>>>>>>>>> bs=1M
>>>>>>>>>>>>
>>>>>>>>>>>> 5000+0 records in
>>>>>>>>>>>>
>>>>>>>>>>>> 5000+0 records out
>>>>>>>>>>>>
>>>>>>>>>>>> 5242880000 bytes (5.2 GB) copied, 49.0907 seconds, 107
>>>>>>>>>>>> MB/s
>>>>>>>>>>>>
>>>>>>>>>>>> [root at cpu_crew1 ~]# dd
>>>>>>>>>>>> if=/shared/working/benchmark/test.cpucrew1 of=/dev/null
>>>>>>>>>>>> bs=4M
>>>>>>>>>>>>
>>>>>>>>>>>> 1250+0 records in
>>>>>>>>>>>>
>>>>>>>>>>>> 1250+0 records out
>>>>>>>>>>>>
>>>>>>>>>>>> 5242880000 bytes (5.2 GB) copied, 44.5724 seconds, 118
>>>>>>>>>>>> MB/s
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Still not impressive.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> -Tom
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Feb 28, 2013, at 8:42 PM, Jeff Anderson-Lee
>>>>>>>>>>>> <<jonah at eecs.berkeley.edu> <mailto:jonah at eecs.berkeley.edu>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>> Thomas,
>>>>>>>>>>>>>
>>>>>>>>>>>>> You have not specified a block size, so you are doing
>>>>>>>>>>>>> a huge number of small(ish) reads with associated
>>>>>>>>>>>>> round trips. What happens with dd bs=128k ..?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Jeff Anderson-Lee
>>>>>>>>>>>>>
>>>>>>>>>>>>> On 2/28/2013 5:30 PM, Thomas Wakefield wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Did a fresh dd test just to confirm, same results:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [root at cpu_crew1 benchmark]# dd if=/dev/zero
>>>>>>>>>>>>>> of=/shared/working/benchmark/test.cpucrew1 bs=512k
>>>>>>>>>>>>>> count=10000
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 10000+0 records in
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 10000+0 records out
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 5242880000 bytes (5.2 GB) copied, 7.43695 seconds,
>>>>>>>>>>>>>> 705 MB/s
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [root at cpu_crew1 benchmark]# dd
>>>>>>>>>>>>>> if=/shared/working/benchmark/test.cpucrew1
>>>>>>>>>>>>>> of=/dev/null
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 552126+0 records in
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 552125+0 records out
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 282688000 bytes (283 MB) copied, 37.8514 seconds,
>>>>>>>>>>>>>> 7.5 MB/s
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Feb 28, 2013, at 8:14 PM, Bryan Whitehead
>>>>>>>>>>>>>> <<driver at megahappy.net> <mailto:driver at megahappy.net>> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> How are you doing the reading? Is this still an
>>>>>>>>>>>>>>> iozone benchmark?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> if you simply dd if=/glustermount/bigfile
>>>>>>>>>>>>>>> of=/dev/null, is the speed better?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Thu, Feb 28, 2013 at 5:05 PM, Thomas Wakefield
>>>>>>>>>>>>>>> <<twake at iges.org> <mailto:twake at iges.org>> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> I get great speed locally, it's only when I add
>>>>>>>>>>>>>>>> gluster in that it slows down. I get 2GB/s
>>>>>>>>>>>>>>>> locally to the exact same brick. It's gluster
>>>>>>>>>>>>>>>> that is having the read issue (80MB/s). But
>>>>>>>>>>>>>>>> Gluster can write just fine, 800MB/s.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> The blockdev idea is a good one, and I have
>>>>>>>>>>>>>>>> already done it. Thanks though.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> -Tom
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Feb 28, 2013, at 7:53 PM, Ling Ho
>>>>>>>>>>>>>>>> <<ling at slac.stanford.edu> <mailto:ling at slac.stanford.edu>> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Tom,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> What type of disks do you have? If they are
>>>>>>>>>>>>>>>>> raid 5 or 6, have you try setting the
>>>>>>>>>>>>>>>>> read-ahead size to 8192 or 16384 (blockdev
>>>>>>>>>>>>>>>>> --setra 8192 /dev/<sd?> ?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>> ling
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On 02/28/2013 04:23 PM, Thomas Wakefield
>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>> Did anyone else have any ideas on
>>>>>>>>>>>>>>>>>> performance tuning for reads?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Feb 27, 2013, at 9:29 PM, Thomas
>>>>>>>>>>>>>>>>>> Wakefield <<twake at iges.org> <mailto:twake at iges.org>> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Bryan-
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Yes I can write at 700-800MBytes/sec, but
>>>>>>>>>>>>>>>>>>> i can only read at 70-80 MBytes/sec. I
>>>>>>>>>>>>>>>>>>> would be very happy if I could get it to
>>>>>>>>>>>>>>>>>>> read at the same speed it can write at.
>>>>>>>>>>>>>>>>>>> And the 70-80 is sequential, not random
>>>>>>>>>>>>>>>>>>> for reads, same exact test commands on the
>>>>>>>>>>>>>>>>>>> disk server are in the 2+GB/s range, so I
>>>>>>>>>>>>>>>>>>> know the disk server can do it.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> -Tom
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On Feb 27, 2013, at 7:41 PM, Bryan
>>>>>>>>>>>>>>>>>>> Whitehead <<driver at megahappy.net> <mailto:driver at megahappy.net>> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Are your figures 700-800MByte/sec?
>>>>>>>>>>>>>>>>>>>> Because that is probably as fast as your
>>>>>>>>>>>>>>>>>>>> 10G nic cards are able to do. You can
>>>>>>>>>>>>>>>>>>>> test that by trying to push a large
>>>>>>>>>>>>>>>>>>>> amount of data over nc or ftp.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Might want to try Infiniband. 40G cards
>>>>>>>>>>>>>>>>>>>> are pretty routine.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On Wed, Feb 27, 2013 at 3:45 PM, Thomas
>>>>>>>>>>>>>>>>>>>> Wakefield <<twake at iges.org> <mailto:twake at iges.org>> wrote:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> I also get the same performance
>>>>>>>>>>>>>>>>>>>>> running iozone for large file
>>>>>>>>>>>>>>>>>>>>> sizes, iozone -u 1 -r 512k -s 2G -I
>>>>>>>>>>>>>>>>>>>>> -F.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Large file IO is what I need the
>>>>>>>>>>>>>>>>>>>>> system to do. I am just shocked at the
>>>>>>>>>>>>>>>>>>>>> huge difference between local IO and
>>>>>>>>>>>>>>>>>>>>> gluster client IO. I know there should
>>>>>>>>>>>>>>>>>>>>> be some difference, but 10x is
>>>>>>>>>>>>>>>>>>>>> unacceptable.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> -Tom
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> On Feb 27, 2013, at 5:31 PM, Bryan
>>>>>>>>>>>>>>>>>>>>> Whitehead <<driver at megahappy.net> <mailto:driver at megahappy.net>>
>>>>>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Every time you open/close a file or
>>>>>>>>>>>>>>>>>>>>>> a directory you will have to wait
>>>>>>>>>>>>>>>>>>>>>> for locks which take time. This is
>>>>>>>>>>>>>>>>>>>>>> totally expected.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Why don't you share what you want to
>>>>>>>>>>>>>>>>>>>>>> do? iozone benchmarks look like crap
>>>>>>>>>>>>>>>>>>>>>> but serving qcow2 files to qemu
>>>>>>>>>>>>>>>>>>>>>> works fantastic for me. What are you
>>>>>>>>>>>>>>>>>>>>>> doing? Make a benchmark that does
>>>>>>>>>>>>>>>>>>>>>> that. If you are going to have many
>>>>>>>>>>>>>>>>>>>>>> files with a wide variety of sizes
>>>>>>>>>>>>>>>>>>>>>> glusterfs/fuse might not be what you
>>>>>>>>>>>>>>>>>>>>>> are looking for.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> On Wed, Feb 27, 2013 at 12:56 PM,
>>>>>>>>>>>>>>>>>>>>>> Thomas Wakefield
>>>>>>>>>>>>>>>>>>>>>> <<twake at cola.iges.org> <mailto:twake at cola.iges.org>> wrote:
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> I have tested everything, small
>>>>>>>>>>>>>>>>>>>>>>> and large files. I have used file
>>>>>>>>>>>>>>>>>>>>>>> sizes ranging from 128k up to
>>>>>>>>>>>>>>>>>>>>>>> multiple GB files. All the reads
>>>>>>>>>>>>>>>>>>>>>>> are bad.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Here is a fairly exhaustive iozone
>>>>>>>>>>>>>>>>>>>>>>> auto test:
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> random random bkwd record stride
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> KB reclen write rewrite read
>>>>>>>>>>>>>>>>>>>>>>> reread read write read rewrite
>>>>>>>>>>>>>>>>>>>>>>> read fwrite frewrite fread freread
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 64 4 40222 63492 26868 30060 1620
>>>>>>>>>>>>>>>>>>>>>>> 71037 1572 70570 31294 77096 72475
>>>>>>>>>>>>>>>>>>>>>>> 14736 13928
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 64 8 99207 116366 13591 13513 3214
>>>>>>>>>>>>>>>>>>>>>>> 97690 3155 109978
>>>>>>>>>>>>>>>>>>>>>>> <tel:3155%20%C2%A0%20109978> 28920
>>>>>>>>>>>>>>>>>>>>>>> 152018 158480 18936 17625
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 64 16 230257 253766 25156 28713
>>>>>>>>>>>>>>>>>>>>>>> 10867 223732 8873 244297 54796
>>>>>>>>>>>>>>>>>>>>>>> 303383 312204 15062 13545
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 64 32 255943 234481 5735102
>>>>>>>>>>>>>>>>>>>>>>> 7100397 11897 318502 13681 347801
>>>>>>>>>>>>>>>>>>>>>>> 24214 695778 528618 25838 28094
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 64 64 214096 681644 6421025
>>>>>>>>>>>>>>>>>>>>>>> 7100397 27453 292156 28117 621657
>>>>>>>>>>>>>>>>>>>>>>> 27338 376062 512471 28569 32534
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 128 4 74329 75468 26428
>>>>>>>>>>>>>>>>>>>>>>> <tel:75468%20%C2%A0%20%C2%A026428>
>>>>>>>>>>>>>>>>>>>>>>> 41089 1131 72857 1118 66976 1597
>>>>>>>>>>>>>>>>>>>>>>> 73778 78343 13351 13026
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 128 8 100862 135170 24966
>>>>>>>>>>>>>>>>>>>>>>> <tel:135170%20%C2%A0%20%C2%A024966
>>>>>>>>>>>>>>>>>>>>>>>> 16734 2617 118966 2560 120406
>>>>>>>>>>>>>>>>>>>>>>>> 39156 125121 146613 16177 16180
>>>>>>>>>>>>>>>>>>>>>>> 128 16 115114 253983 28212 17854
>>>>>>>>>>>>>>>>>>>>>>> 5307 246180 5431 229843 47335
>>>>>>>>>>>>>>>>>>>>>>> 255920 271173 27256 24445
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 128 32 256042 391360 39848 64258
>>>>>>>>>>>>>>>>>>>>>>> 11329 290230 9905 429563 38176
>>>>>>>>>>>>>>>>>>>>>>> 490380 463696 20917 19219
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 128 64 248573 592699 4557257
>>>>>>>>>>>>>>>>>>>>>>> 6812590 19583 452366 29263 603357
>>>>>>>>>>>>>>>>>>>>>>> 42967 814915 692017 76327 37604
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 128 128 921183 526444 5603747
>>>>>>>>>>>>>>>>>>>>>>> 5379161 45614 390222 65441 826202
>>>>>>>>>>>>>>>>>>>>>>> 41384 662962 1040839 78526 39023
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 256 4 76212 77337 40295
>>>>>>>>>>>>>>>>>>>>>>> <tel:77337%20%C2%A0%20%C2%A040295>
>>>>>>>>>>>>>>>>>>>>>>> 32125 1289 71866 1261 64645 1436
>>>>>>>>>>>>>>>>>>>>>>> 57309 53048 23073 29550
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 256 8 126922 141976 26237
>>>>>>>>>>>>>>>>>>>>>>> <tel:141976%20%C2%A0%20%C2%A026237
>>>>>>>>>>>>>>>>>>>>>>>> 25130 2566 128058 2565 138981
>>>>>>>>>>>>>>>>>>>>>>>> <tel:2565%20%C2%A0%20138981>
>>>>>>>>>>>>>>>>>>>>>>>> 2985 125060 133603 22840 24955
>>>>>>>>>>>>>>>>>>>>>>> 256 16 242883 263636 41850 24371
>>>>>>>>>>>>>>>>>>>>>>> 4902 250009 5290 248792 89353
>>>>>>>>>>>>>>>>>>>>>>> 243821 247303 26965 26199
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 256 32 409074 439732 40101 39335
>>>>>>>>>>>>>>>>>>>>>>> 11953 436870 11209 430218 83743
>>>>>>>>>>>>>>>>>>>>>>> 409542 479390 30821 27750
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 256 64 259935 571502 64840 71847
>>>>>>>>>>>>>>>>>>>>>>> 22537 <tel:71847%20%C2%A0%2022537>
>>>>>>>>>>>>>>>>>>>>>>> 617161 23383 392047 91852 672010
>>>>>>>>>>>>>>>>>>>>>>> 802614 41673 53111
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 256 128 847597 812329 185517 83198
>>>>>>>>>>>>>>>>>>>>>>> 49383 <tel:83198%20%C2%A0%2049383>
>>>>>>>>>>>>>>>>>>>>>>> 708831 44668 794889 74267 1180188
>>>>>>>>>>>>>>>>>>>>>>> 1662639 54303 41018
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 256 256 481324 709299 5217259
>>>>>>>>>>>>>>>>>>>>>>> 5320671 44668 719277 40954 808050
>>>>>>>>>>>>>>>>>>>>>>> 41302 790209 771473 62224 35754
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 512 4 77667 75226 35102 29696 1337
>>>>>>>>>>>>>>>>>>>>>>> 66262 1451 67680 1413 69265 69142
>>>>>>>>>>>>>>>>>>>>>>> 42084 27897
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 512 8 134311 144341 30144
>>>>>>>>>>>>>>>>>>>>>>> <tel:144341%20%C2%A0%20%C2%A030144
>>>>>>>>>>>>>>>>>>>>>>>> 24646 2102 134143 2209 134699
>>>>>>>>>>>>>>>>>>>>>>>> 2296 108110 128616 25104 29123
>>>>>>>>>>>>>>>>>>>>>>> 512 16 200085 248787 30235 25697
>>>>>>>>>>>>>>>>>>>>>>> 4196 247240 4179 256116
>>>>>>>>>>>>>>>>>>>>>>> <tel:4179%20%C2%A0%20256116> 4768
>>>>>>>>>>>>>>>>>>>>>>> 250003 226436 32351 28455
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 512 32 330341 439805 26440 39284
>>>>>>>>>>>>>>>>>>>>>>> 8744 457611 8006 424168
>>>>>>>>>>>>>>>>>>>>>>> <tel:8006%20%C2%A0%20424168>
>>>>>>>>>>>>>>>>>>>>>>> 125953 425935 448813 27660 26951
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 512 64 483906 733729 48747 41121
>>>>>>>>>>>>>>>>>>>>>>> 16032 555938 17424 587256 187343
>>>>>>>>>>>>>>>>>>>>>>> 366977 735740 41700 41548
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 512 128 836636 907717 69359 94921
>>>>>>>>>>>>>>>>>>>>>>> 42443 <tel:94921%20%C2%A0%2042443>
>>>>>>>>>>>>>>>>>>>>>>> 761031 36828 964378 123165 651383
>>>>>>>>>>>>>>>>>>>>>>> 695697 58368 44459
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 512 256 520879 860437 145534
>>>>>>>>>>>>>>>>>>>>>>> 135523 40267 847532 31585 663252
>>>>>>>>>>>>>>>>>>>>>>> 69696 1270846 1492545 48822 48092
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 512 512 782951 973118 3099691
>>>>>>>>>>>>>>>>>>>>>>> 2942541 42328 871966 46218 911184
>>>>>>>>>>>>>>>>>>>>>>> 49791 953248 1036527 52723 48347
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 1024 4 76218 69362 36431 28711
>>>>>>>>>>>>>>>>>>>>>>> 1137 66171 1174 68938 1125 70566
>>>>>>>>>>>>>>>>>>>>>>> 70845 34942 28914
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 1024 8 126045 140524 37836
>>>>>>>>>>>>>>>>>>>>>>> <tel:140524%20%C2%A0%20%C2%A037836
>>>>>>>>>>>>>>>>>>>>>>>> 15664 2698 126000 2557 125566
>>>>>>>>>>>>>>>>>>>>>>>> 2567 110858 127255 26764 27945
>>>>>>>>>>>>>>>>>>>>>>> 1024 16 243398 261429 40238 23263
>>>>>>>>>>>>>>>>>>>>>>> 3987 246400 3882 260746 4093
>>>>>>>>>>>>>>>>>>>>>>> 236652 236874 31429 25076
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 1024 32 383109 422076 41731 41605
>>>>>>>>>>>>>>>>>>>>>>> 8277 473441 7775 415261 8588
>>>>>>>>>>>>>>>>>>>>>>> 394765 407306 40089 28537
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 1024 64 590145 619156 39623 53267
>>>>>>>>>>>>>>>>>>>>>>> 15051 722717 14624 753000 257294
>>>>>>>>>>>>>>>>>>>>>>> 597784 620946 38619 44073
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 1024 128 1077836 1124099 56192
>>>>>>>>>>>>>>>>>>>>>>> <tel:1124099%20%C2%A0%20%C2%A05619
>>>>>>>>>>>>>>>>>>>>>>> 2> 64916 36851 1102176 37198
>>>>>>>>>>>>>>>>>>>>>>> 1082454 281548 829175 792604 47975
>>>>>>>>>>>>>>>>>>>>>>> 51913
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 1024 256 941918 1074331 72783
>>>>>>>>>>>>>>>>>>>>>>> 81450 26778
>>>>>>>>>>>>>>>>>>>>>>> <tel:81450%20%C2%A0%2026778>
>>>>>>>>>>>>>>>>>>>>>>> 1099636 32395 1060013 183218
>>>>>>>>>>>>>>>>>>>>>>> 1024121 995171 44371 45448
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 1024 512 697483 1130312 100324
>>>>>>>>>>>>>>>>>>>>>>> 114682 48215 1041758 41480 1058967
>>>>>>>>>>>>>>>>>>>>>>> 90156 994020 1563622 56328 46370
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 1024 1024 931702 1087111 4609294
>>>>>>>>>>>>>>>>>>>>>>> 4199201 44191 949834 45594 970656
>>>>>>>>>>>>>>>>>>>>>>> 56674 933525 1075676 44876 46115
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2048 4 71438 67066 58319 38913
>>>>>>>>>>>>>>>>>>>>>>> 1147 44147 1043 42916 967 66416
>>>>>>>>>>>>>>>>>>>>>>> 67205 45953 96750
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2048 8 141926 134567 61101 55445
>>>>>>>>>>>>>>>>>>>>>>> 2596 77528 2564 80402 4258 124211
>>>>>>>>>>>>>>>>>>>>>>> 120747 53888 100337
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2048 16 254344 255585 71550 74500
>>>>>>>>>>>>>>>>>>>>>>> 5410 139365 5201 141484 5171
>>>>>>>>>>>>>>>>>>>>>>> 205521 213113 67048 57304
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2048 32 397833 411261 56676 80027
>>>>>>>>>>>>>>>>>>>>>>> 10440 <tel:80027%20%C2%A0%2010440>
>>>>>>>>>>>>>>>>>>>>>>> 260034 10126 230238 10814 391665
>>>>>>>>>>>>>>>>>>>>>>> 383379 79333 60877
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2048 64 595167 687205 64262 87327
>>>>>>>>>>>>>>>>>>>>>>> 20772 <tel:87327%20%C2%A0%2020772>
>>>>>>>>>>>>>>>>>>>>>>> 456430 19960 477064 23190 540220
>>>>>>>>>>>>>>>>>>>>>>> 563096 86812 92565
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2048 128 833585 933403 121926
>>>>>>>>>>>>>>>>>>>>>>> 118621 37700 690020 37575 733254
>>>>>>>>>>>>>>>>>>>>>>> 567449 712337 734006 92011 104934
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2048 256 799003 949499 143688
>>>>>>>>>>>>>>>>>>>>>>> 125659 40871
>>>>>>>>>>>>>>>>>>>>>>> <tel:125659%20%C2%A0%2040871>
>>>>>>>>>>>>>>>>>>>>>>> 892757 37977 880494 458281 836263
>>>>>>>>>>>>>>>>>>>>>>> 901375 131332 110237
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2048 512 979936 1040724 120896
>>>>>>>>>>>>>>>>>>>>>>> 138013 54381 859783 48721 780491
>>>>>>>>>>>>>>>>>>>>>>> 279203 1068824 1087085 97886 98078
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2048 1024 901754 987938 53352
>>>>>>>>>>>>>>>>>>>>>>> 53043 72727
>>>>>>>>>>>>>>>>>>>>>>> <tel:53043%20%C2%A0%2072727>
>>>>>>>>>>>>>>>>>>>>>>> 1054522 68269 992275 181253
>>>>>>>>>>>>>>>>>>>>>>> 1309480 1524983 121600 95585
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 2048 2048 831890 1021540 4257067
>>>>>>>>>>>>>>>>>>>>>>> 3302797 75672 984203 80181 826209
>>>>>>>>>>>>>>>>>>>>>>> 94278 966920 1027159 111832 105921
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 4 66195 67316 62171 74785
>>>>>>>>>>>>>>>>>>>>>>> 1328 28963 1329 26397 1223 71470
>>>>>>>>>>>>>>>>>>>>>>> 69317 55903 84915
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 8 122221 120057 90537 60958
>>>>>>>>>>>>>>>>>>>>>>> 2598 47312 2468 59783 2640 128674
>>>>>>>>>>>>>>>>>>>>>>> 127872 41285 40422
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 16 238321 239251 29336 32121
>>>>>>>>>>>>>>>>>>>>>>> 4153 89262 3986 96930 4608 229970
>>>>>>>>>>>>>>>>>>>>>>> 237108 55039 56983
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 32 417110 421356 30974 50000
>>>>>>>>>>>>>>>>>>>>>>> 8382 156676 7886 153841 7900
>>>>>>>>>>>>>>>>>>>>>>> 359585 367288 26611 25952
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 64 648008 668066 32193 29389
>>>>>>>>>>>>>>>>>>>>>>> 14830 273265 14822 282211 19653
>>>>>>>>>>>>>>>>>>>>>>> 581898 620798 51281 50218
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 128 779422 848564 55594 60253
>>>>>>>>>>>>>>>>>>>>>>> 37108 <tel:60253%20%C2%A0%2037108>
>>>>>>>>>>>>>>>>>>>>>>> 451296 35908 491361 37567 738163
>>>>>>>>>>>>>>>>>>>>>>> 728059 67681 66440
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 256 865623 886986 71368 63947
>>>>>>>>>>>>>>>>>>>>>>> 44255 <tel:63947%20%C2%A0%2044255>
>>>>>>>>>>>>>>>>>>>>>>> 645961 42689 719491 736707 819696
>>>>>>>>>>>>>>>>>>>>>>> 837641 57059 60347
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 512 852099 889650 68870 73891
>>>>>>>>>>>>>>>>>>>>>>> 31185 845224 30259 830153 392334
>>>>>>>>>>>>>>>>>>>>>>> 910442 961983 60083 55558
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 1024 710357 867810 29377
>>>>>>>>>>>>>>>>>>>>>>> 29522 49954 846640 43665 926298
>>>>>>>>>>>>>>>>>>>>>>> 213677 986226 1115445 55130 59205
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 2048 826479 908420 43191
>>>>>>>>>>>>>>>>>>>>>>> 42075 59684 904022 58601 855664
>>>>>>>>>>>>>>>>>>>>>>> 115105 1418322 1524415 60548 66066
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 4096 4096 793351 855111 3232454
>>>>>>>>>>>>>>>>>>>>>>> 3673419 66018 861413 48833 847852
>>>>>>>>>>>>>>>>>>>>>>> 45914 852268 842075 42980 48374
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 4 67340 69421 42198 31740 994
>>>>>>>>>>>>>>>>>>>>>>> 23251 1166 16813 837 73827 73126
>>>>>>>>>>>>>>>>>>>>>>> 25169 29610
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 8 137150 125622 29131 36439
>>>>>>>>>>>>>>>>>>>>>>> 2051 44342 1988 48930 2315 134183
>>>>>>>>>>>>>>>>>>>>>>> 135367 31080 33573
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 16 237366 220826 24810 26584
>>>>>>>>>>>>>>>>>>>>>>> 3576 88004 3769 78717 4289 233751
>>>>>>>>>>>>>>>>>>>>>>> 235355 23302 28742
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 32 457447 454404 31594 27750
>>>>>>>>>>>>>>>>>>>>>>> 8141 142022 7846 143984 9322
>>>>>>>>>>>>>>>>>>>>>>> 353147 396188 34203 33265
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 64 670645 655259 28630 23255
>>>>>>>>>>>>>>>>>>>>>>> 16669 237476 16965 244968 15607
>>>>>>>>>>>>>>>>>>>>>>> 590365 575320 49998 43305
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 128 658676 760982 44197 47802
>>>>>>>>>>>>>>>>>>>>>>> 28693 379523 26614 378328 27184
>>>>>>>>>>>>>>>>>>>>>>> 720997 702038 51707 49733
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 256 643370 698683 56233 63165
>>>>>>>>>>>>>>>>>>>>>>> 28846 <tel:63165%20%C2%A0%2028846>
>>>>>>>>>>>>>>>>>>>>>>> 543952 27745 576739 44014 701007
>>>>>>>>>>>>>>>>>>>>>>> 725534 59611 58985
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 512 696884 776793 67258 52705
>>>>>>>>>>>>>>>>>>>>>>> 18711 698854 21004 694124 621695
>>>>>>>>>>>>>>>>>>>>>>> 784812 773331 43101 47659
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 1024 729664 810451 15470
>>>>>>>>>>>>>>>>>>>>>>> 15875 31318 801490 38123 812944
>>>>>>>>>>>>>>>>>>>>>>> 301222 804323 832765 54308 53376
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 2048 749217 68757 21914 22667
>>>>>>>>>>>>>>>>>>>>>>> 48971 783309 48132 782738 172848
>>>>>>>>>>>>>>>>>>>>>>> 907408 929324 51156 50565
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 4096 707677 763960 32063
>>>>>>>>>>>>>>>>>>>>>>> 31928 47809
>>>>>>>>>>>>>>>>>>>>>>> <tel:31928%20%C2%A0%2047809>
>>>>>>>>>>>>>>>>>>>>>>> 751692 49560 786339 93445 1046761
>>>>>>>>>>>>>>>>>>>>>>> 1297876 48037 51680
>>>>>>>>>>>>>>>>>>>>>>> <tel:48037%20%C2%A0%20%C2%A051680>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 8192 8192 623817 746288 2815955
>>>>>>>>>>>>>>>>>>>>>>> 3137358 48722 741633 35428 753787
>>>>>>>>>>>>>>>>>>>>>>> 49626 803683 823800 48977 52895
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 4 72372 73651 34471 30788
>>>>>>>>>>>>>>>>>>>>>>> 960 23610 903 22316 891 71445
>>>>>>>>>>>>>>>>>>>>>>> 71138 56451 55129
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 8 137920 141704 50830 33857
>>>>>>>>>>>>>>>>>>>>>>> 1935 41934 2275 35588 3608 130757
>>>>>>>>>>>>>>>>>>>>>>> 137801 51621 48525
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 16 245369 242460 41808 29770
>>>>>>>>>>>>>>>>>>>>>>> 3605 75682 4355 75315 4767 241100
>>>>>>>>>>>>>>>>>>>>>>> 239693 53263 30785
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 32 448877 433956 31846 35010
>>>>>>>>>>>>>>>>>>>>>>> 7973 118181 8819 112703 8177
>>>>>>>>>>>>>>>>>>>>>>> 381734 391651 57749 63417
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 64 710831 700712 66792 68864
>>>>>>>>>>>>>>>>>>>>>>> 20176 209806 19034 207852
>>>>>>>>>>>>>>>>>>>>>>> <tel:19034%20%C2%A0%20207852>
>>>>>>>>>>>>>>>>>>>>>>> 21255 589503 601379 104567 105162
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 128 836901 860867 104226
>>>>>>>>>>>>>>>>>>>>>>> 100373 40899 358865 40946 360562
>>>>>>>>>>>>>>>>>>>>>>> 39415 675968 691538 96086 105695
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 256 798081 828146 107103
>>>>>>>>>>>>>>>>>>>>>>> 120433 39084
>>>>>>>>>>>>>>>>>>>>>>> <tel:120433%20%C2%A0%2039084>
>>>>>>>>>>>>>>>>>>>>>>> 595325 39050 593110 56925 763466
>>>>>>>>>>>>>>>>>>>>>>> 797859 109645 113414
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 512 810851 843931 113564
>>>>>>>>>>>>>>>>>>>>>>> 106202 35111 714831 46244 745947
>>>>>>>>>>>>>>>>>>>>>>> 53636 802902 760172 110492 100879
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 1024 726399 820219 22106
>>>>>>>>>>>>>>>>>>>>>>> 22987 53087
>>>>>>>>>>>>>>>>>>>>>>> <tel:22987%20%C2%A0%2053087>
>>>>>>>>>>>>>>>>>>>>>>> 749053 54781 777705 1075341 772686
>>>>>>>>>>>>>>>>>>>>>>> 809723 100349 96619
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 2048 807772 856458 23920
>>>>>>>>>>>>>>>>>>>>>>> 23617 66320 829576 72105 740848
>>>>>>>>>>>>>>>>>>>>>>> 656379 864539 835446 93499 101714
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> 16384 4096 797470 840596 27270
>>>>>>>>>>>>>>>>>>>>>>> <tel:28132%20%C2%A0%2088784>
>>>>>> ...
>>>>>>
>>>>>> [Message clipped]
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
> --
>
> *Rodrigo Severo* | DIRETOR DE TECNOLOGIA
> Tel. *+55 61 3030-1515*
> Siga a Fábrica no twitter:*@empautaclipping*
>
> fabricadeideias.com <http://www.fabricadeideias.com/>
> 12 ANOS DE TECNOLOGIA E COMUNICAÇÃO
> NUMA COMBINAÇÃO PERFEITA
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130311/93d19f3e/attachment.html>
More information about the Gluster-users
mailing list