[Gluster-users] Gluster-users Digest, Vol 9, Issue 66

Keith Freedman freedman at FreeFormIT.com
Fri Jan 23 19:12:18 UTC 2009


dont forget to set :
option read-subvolume  BRICKNAME

in your AFR config, that'll improve your read performance significantly.


At 11:06 AM 1/23/2009, Evan wrote:
>Ideal what I'm trying to accomplish is to have multiple samba 
>servers connected with a 1.544Mbps pipe stay in sync with each 
>other. So its important for me to be able to have near really disk 
>access speeds when reading and writing to the local gluster Node in 
>the AFR group but as far as getting the data propagated out to the 
>other nodes I know the 1.544Mbps pipe can't keep up so I'll take the 
>fastest I can get as long as it doesn't affect the local performance 
>(which is what I am seeing).
>
>
>
>On Fri, Jan 23, 2009 at 10:56 AM, Keith Freedman 
><<mailto:freedman at freeformit.com>freedman at freeformit.com> wrote:
>At 10:18 AM 1/23/2009, Evan wrote:
>I added the following to the bottom of my spec file:
>
>volume writebehind
>  type performance/write-behind
>  option aggregate-size 10MB # default is 0bytes
>  option flush-behind off    # default is 'off'
>  subvolumes afr
>end-volume
>
>which gives me the following results when making a 10MB file
># time dd if=/dev/zero of=/tmp/disktest count=10240 bs=1024
>10240+0 records in
>10240+0 records out
>10485760 bytes (10 MB) copied, 0.173179 s, 60.5 MB/s
>
>real    0m0.183s
>user    0m0.000s
>sys     0m0.204s
>
># time dd if=/dev/zero of=/mnt/gluster/disktest count=10240 bs=1024
>10240+0 records in
>10240+0 records out
>10485760 bytes (10 MB) copied, 5.50861 s, 1.9 MB/s
>
>real    0m5.720s
>user    0m0.000s
>sys     0m0.060s
>
>Although this is better than I had before is there anyway to have 
>gluster write the data to the localBrick and then sync/afr in the 
>background so I could expect to see something closer to the 60 MB/s 
>I see when writing to the local disk directly?
>
>
>what you really want is a delayed replication.  I've asked for this 
>in this mailing list recently, and was told that it's something 
>they've considered (more as a DR feature than an HA feature), but 
>it's not currently on the list of priorities.
>
>The issue, as I see it, is if it's an HA feature, then you really 
>need to insure that the data is replicated before you let your 
>application think the data is written.  If the replication was 
>delayed, and the server went down, the data is lost forever.  This 
>is bad for HA.
>if it's a DR feature, then you're probably ok, because usually 
>disaster recovery scenarios can probably withstand some data loss, 
>and you're more interested in a point-in-time snapshot of the data.
>
>FUSE is a problem, and TCP/IP is a problem with much overhead and 
>large blocksizes.
>
>Ideally, glusters write-behind would be smart enough to aggregate 
>smaller blocks of data into a large write.  I think this would solve 
>a lot of the problem you're having in your tests.
>
>Thanks
>
>
>aghavendra G 
><<mailto:raghavendra at zresearch.com><mailto:raghavendra at zresearch.com>raghavendra at zresearch.com> 
>wrote:
>above afr with afr as a subvolume
>
>On Fri, Jan 23, 2009 at 12:59 AM, Evan 
><_<mailto:Gluster at devnada.com><mailto:Gluster at devnada.com>Gluster at devnada.com> 
>wrote:
>Where should I put the write-behind translator?
>Just above afr with afr as a subvolume? Or should I put it just 
>above my localBrick volume and below afr?
>
>
>Here is the output using /dev/zero:
># time dd if=/dev/zero of=/mnt/gluster/disktest count=1024 bs=1024
>
>1024+0 records in
>1024+0 records out
>1048576 bytes (1.0 MB) copied, 1.90119 s, 552 kB/s
>
>real    0m2.098s
>user    0m0.000s
>sys     0m0.016s
>
># time dd if=/dev/zero of=/tmp/disktest count=1024 bs=1024
>
>1024+0 records in
>1024+0 records out
>1048576 bytes (1.0 MB) copied, 0.0195388 s, 53.7 MB/s
>
>real    0m0.026s
>user    0m0.000s
>sys     0m0.028s
>
>
>Thanks
>
>
>On Thu, Jan 22, 2009 at 12:52 PM, Anand Avati 
><<mailto:avati at zresearch.com><mailto:avati at zresearch.com>avati at zresearch.com> 
>wrote:
>Do you have write-behind loaded on the client side? For IO testing,
>use /dev/zero instead of /dev/urandom.
>
>avati
>
>On Fri, Jan 23, 2009 at 2:14 AM, Evan 
><_<mailto:Gluster at devnada.com><mailto:Gluster at devnada.com>Gluster at devnada.com> 
>wrote:
> > I have a 2 node single process AFR setup with 1.544Mbps bandwidth between
> > the 2 nodes. When I write a 1MB file to the gluster share it 
> seems to AFR to
> > the other node in real time killing my disk IO speeds on the gluster mount
> > point. Is there anyway to fix this? Ideally I would like to see near real
> > disk IO speeds from/to the local gluster mount point and let the afr play
> > catch up in the background as the bandwidth becomes available.
> >
> > Gluster Spec File (same on both nodes) 
> <<http://pastebin.com/m58dc49d4>http://pastebin.com/m58dc49d4>http://pastebin.com/m58dc49d4
>
> > IO speed tests:
> > # time dd if=/dev/urandom of=/mnt/gluster/disktest count=1024 bs=1024
> > 1024+0 records in
> > 1024+0 records out
> > 1048576 bytes (1.0 MB) copied, 8.34701 s, 126 kB/s
> >
> > real    0m8.547s
> > user    0m0.000s
> > sys     0m0.372s
> >
> > # time dd if=/dev/urandom of=/tmp/disktest count=1024 bs=1024
> > 1024+0 records in
> > 1024+0 records out
> > 1048576 bytes (1.0 MB) copied, 0.253865 s, 4.1 MB/s
> >
> > real    0m0.259s
> > user    0m0.000s
> > sys     0m0.284s
> >
> >
> > Thanks
> >
> > _______________________________________________
> > Gluster-users mailing list
> > 
> <mailto:Gluster-users at gluster.org><mailto:Gluster-users at gluster.org>Gluster-users at gluster.org
>
> > http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
> >
> >
>
>
>
>_______________________________________________
>Gluster-users mailing list
><mailto:Gluster-users at gluster.org><mailto:Gluster-users at gluster.org>Gluster-users at gluster.org
>
>http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
>
>
>
>
>--
>Raghavendra G
>
>
>_______________________________________________
>Gluster-users mailing list
><mailto:Gluster-users at gluster.org>Gluster-users at gluster.org
>http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
>
>





More information about the Gluster-users mailing list