[Gluster-users] Disastrous performance with rsync to mounted Gluster volume.

Ben Turner bturner at redhat.com
Fri Apr 24 18:03:10 UTC 2015


----- Original Message -----
> From: "Ernie Dunbar" <maillist at lightspeed.ca>
> To: "Gluster Users" <gluster-users at gluster.org>
> Sent: Friday, April 24, 2015 1:15:32 PM
> Subject: Re: [Gluster-users] Disastrous performance with rsync to mounted Gluster volume.
> 
> On 2015-04-23 18:10, Joe Julian wrote:
> > On 04/23/2015 04:41 PM, Ernie Dunbar wrote:
> >> On 2015-04-23 12:58, Ben Turner wrote:
> >> 
> >>> 
> >>> +1, lets nuke everything and start from a known good.  Those error
> >>> messages make me think something is really wrong with how we are
> >>> copying the data.  Gluster does NFS by default so you shouldn't have
> >>> have to reconfigure anything after you recreate the volume.
> >>> 
> >> 
> >> Okay... this is a silly question. How do I do that? Deleting the
> >> volume doesn't affect the files in the underlying filesystem, and I
> >> get the impression that trying to delete the files in the underlying
> >> filesystem without shutting down or deleting the volume would result
> >> in Gluster trying to write the files back where they "belong".
> >> 
> >> Should I stop the volume, delete it, then delete the files and start
> >> from scratch, re-creating the volume?
> > 
> > That's what I would do.
> > 
> 
> Well, apparently removing the .glusterfs directory from the brick is an
> exceptionally bad thing, and breaks gluster completely, rendering it
> inoperable. I'm going to have to post another thread about how to fix
> this mess now.

You are correct and I would just start from scratch Ernie.  Creating a gluster cluster is only about 3-4 commands and should only take a minute or two.  Also with all the problems you are having I am not confident in your data integrity.  All you need to do to clear EVERYTHING out is:

service glusterd stop
killall glusterfsd
killall glusterfs
sleep 1
for file in /var/lib/glusterd/*; do if ! echo $file | grep 'hooks' >/dev/null 2>&1;then rm -rf $file; fi; done

>From there restart the gluster service and recreate everything:

service glusterd restart
<make a new filesystem on your bricks, mount>
gluster peer probe <my peer>
gluster v create <my vol>
gluster v start <my vol>
gluster v info 

>From there mount the new volume on your system with the data  you want to migrate:

mount -t nfs -o vers=3 <my vol> <my mount>
rsync <your rsync command>

This should get you where you need to be.  Before you start to migrate the data maybe do a couple DDs and send me the output so we can get an idea of how your cluster performs:

time `dd if=/dev/zero of=<gluster-mount>/myfile bs=1024k count=1000; sync`
echo 3 > /proc/sys/vm/drop_caches
dd if=<gluster mount> of=/dev/null bs=1024k count=1000

If you are using gigabit and glusterfs mounts with replica 2 you should get ~55 MB / sec writes and ~110 MB / sec reads.  With NFS you will take a bit of a hit since NFS doesnt know where files live like glusterfs does.

-b

> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
> 


More information about the Gluster-users mailing list