[Gluster-users] CHANGELOGs and new geo-replica sync taking forever

Brian Ericson bericson at ptc.com
Wed Nov 4 17:30:55 UTC 2015


tl;dr -- geo-replication of ~200,000 CHANGELOG files is killing me... Help!

I have about 125G spread over just shy of 5000 files that I'm 
replicating with
geo-replication to nodes around the world.  The content is fairly stable and
probably hasn't changed at all since I initially established the GlusterFS
nodes/network, which looks as follows:
x -> xx -> [xxx, xxy] (x geo-replicates to xx, xx geo-replicates to xxx/xxy)

Latency & throughput are markedly different (x -> xx is the fastest, xx 
-> xxx
the slowest (at about 1G/hour)). That said, all nodes were synced with 5 
days
of setting up the network.

I have since added another node, xxz, which is also geo-replicated from 
xx (xx
-> xxz). Its latency/throughput is clearly better than xx -> xxx's, but 
over 5
days later, I'm still replicating CHANGELOGs and haven't gotten to any real
content (the replicated volumes' mounted filesystems are empty).

Starting with x, you can see I have a "reasonable" number of CHANGELOGs:
x # find /bricks/*/.glusterfs/changelogs -name CHANGELOG\* | wc -l
186

However, xxz's source is xx, and I've got a real problem with xx:
xx # find /bricks/*/.glusterfs/changelogs -name CHANGELOG\* | wc -l
193450

5+ days into this, and I've hardly managed to dent this on xxz:
xxz # find /bricks/*/.glusterfs/changelogs -name CHANGELOG\* | wc -l
43211

On top of that, xx is generating new CHANGELOGs at a rate of ~6/minute (two
volumes at ~3/minute each), so chasing CHANGELOGs is a (quickly) moving 
target.

And these files are small! The "I'm alive" file is 92 bytes long, I've also
seen them also average about 4k. Demonstrating latency/throughput, you 
can see
that small files (for me) are a real killer:
### x -> xx (fastest route)
# for i in 1 10 100 1000; do file="$( dd if=/dev/urandom bs=1024 
count=$((4000/i)) 2> /dev/null )"; echo "$i ($(( $( echo -n "$file" | wc 
-c )/1024 ))k): $( ( time for i in $( seq 1 $i ); do echo -n "$file" | 
ssh xx 'cat > /dev/null'; done ) |& awk '/^real/{ print $2 }' )"; done
1 $i ); do echo -n "$file" | ssh $location 'cat > /dev/null'; done ) |& 
awk '/^real/{ print $2 }' )"; done
1 (3984k): 0m4.777s
10 (398k): 0m10.737s
100 (39k): 0m53.286s
1000 (3k): 7m21.493s

### xx -> xxx (slowest route)
# for i in 1 10 100 1000; do file="$( dd if=/dev/urandom bs=1024 
count=$((4000/i)) 2> /dev/null )"; echo "$i ($(( $( echo -n "$file" | wc 
-c )/1024 ))k): $( ( time for i in $( seq 1 $i ); do echo -n "$file" | 
ssh xxx 'cat > /dev/null'; done ) |& awk '/^real/{ print $2 }' )"; done
1 (3984k): 0m11.065s
10 (398k): 0m41.007s
100 (39k): 4m52.814s
1000 (3k): 39m23.009s

### xx -> xxz (the route I've added and am trying to sync)
# for i in 1 10 100 1000; do file="$( dd if=/dev/urandom bs=1024 
count=$((4000/i)) 2> /dev/null )"; echo "$i ($(( $( echo -n "$file" | wc 
-c )/1024 ))k): $( ( time for i in $( seq 1 $i ); do echo -n "$file" | 
ssh xxz 'cat > /dev/null'; done ) |& awk '/^real/{ print $2 }' )"; done
1 (3984k): 0m2.673s
10 (398k): 0m16.333s
100 (39k): 2m0.676s
1000 (3k): 17m28.265s

What you're looking at is the cost of transferring a total of 4000k: 1 
transfer
at 4000k, 10 at 400k, 100 at 40k, and 1000 at 4k. With 1 transfer at under 3s and 
1000
transfers at nearly 17 1/2 minutes for xx -> xxz and for the same total
transfer size, it's really a killer to transfer CHANGELOGs, especially 
almost
200,000 of them.

And, 92 byte files doesn't improve this:
### x -> xx (fastest route)
# file="$( dd if=/dev/urandom bs=92 count=1 2> /dev/null )"; i=100; echo 
"$i ($(( $( echo -n "$file" | wc -c ) ))): $( ( time for i in $( seq 1 
$i ); do echo -n "$file" | ssh xx 'cat > /dev/null'; done ) |& awk 
'/^real/{ print $2 }' )"
100 (92): 0m34.164s

### xx -> xxx (slowest route)
# file="$( dd if=/dev/urandom bs=92 count=1 2> /dev/null )"; i=100; echo 
"$i ($(( $( echo -n "$file" | wc -c ) ))): $( ( time for i in $( seq 1 
$i ); do echo -n "$file" | ssh xxx 'cat > /dev/null'; done ) |& awk 
'/^real/{ print $2 }' )"
100 (92): 3m53.388s

### xx -> xxz (the route I've added and am trying to sync)
# file="$( dd if=/dev/urandom bs=92 count=1 2> /dev/null )"; i=100; echo 
"$i ($(( $( echo -n "$file" | wc -c ) ))): $( ( time for i in $( seq 1 
$i ); do echo -n "$file" | ssh xxz 'cat > /dev/null'; done ) |& awk 
'/^real/{ print $2 }' )"
100 (92): 1m43.389s

Questions...:
o Why so many CHANGELOGs?

o Why so slow (in 5 days, I've transferred 43211 CHANGELOGs, so 
43211/5/24/60=6
   implies a real transfer rate of about 6 CHANGELOG files per minute, which
   brings me back to xx's generating new ones at about that rate...)?

o What can I do to "fix" this?



More information about the Gluster-users mailing list