[Gluster-users] Migrating data from a failing filesystem
Ravishankar N
ravishankar at redhat.com
Wed Sep 24 16:10:19 UTC 2014
On 09/24/2014 07:35 PM, james.bellinger at icecube.wisc.edu wrote:
> Thanks for the info!
> I started the remove-brick start and, of course, the brick went read-only
> in less than an hour.
> This morning I checked the status a couple of minutes apart and found:
>
> Node Rebalanced-files size scanned failures
> status
> --------- ----------- -------- --------- -----------
> ------------
> gfs-node04 6634 590.7GB 81799 14868 in
> progress
> ...
> gfs-node04 6669 596.5GB 86584 15271 in
> progress
>
> I'm not sure exactly what it is doing here: 4785 files scanned, 403
> failures, and 35 rebalanced.
What it is supposed to be doing is to scan all the files in the volume,
and for the files present in itself, i.e.gfs-node04:/sdb, migrate
(rebalance) it into other bricks in the volume. Let it go to completion.
The rebalance log should give you an idea of the 403 failures.
> The used amount on the partition hasn't
> changed.
Probably because after copying the files to the other bricks, the
unlinks/rmdirs on itself are failing because of the FS being mounted
read-only.
> If anything, the _other_ brick on the server is shrinking!
Because the data is being copied into this brick as a part of migration?
> (Which is related to the question I had before that you mention below.)
>
> gluster volume remove-brick scratch gfs-node04:/sdb start
What is your original volume configuration? (gluster vol info scratch)?
> but...
> df /sda
> Filesystem 1K-blocks Used Available Use% Mounted on
> /dev/sda 12644872688 10672989432 1844930140 86% /sda
> ...
> /dev/sda 12644872688 10671453672 1846465900 86% /sda
>
> Have I shot myself in the other foot?
> jim
>
>
>
>
>
>> On 09/23/2014 08:56 PM, james.bellinger at icecube.wisc.edu wrote:
>>> I inherited a non-replicated gluster system based on antique hardware.
>>>
>>> One of the brick filesystems is flaking out, and remounts read-only. I
>>> repair it and remount it, but this is only postponing the inevitable.
>>>
>>> How can I migrate files off a failing brick that intermittently turns
>>> read-only? I have enough space, thanks to a catastrophic failure on
>>> another brick; I don't want to present people with another one. But if
>>> I
>>> understand migration correctly references have to be deleted, which
>>> isn't
>>> possible if the filesystem turns read-only.
>> What you could do is initiate the migration with `remove-brick start'
>> and monitor the progress with 'remove-brick status`. Irrespective of
>> whether the rebalance completes or fails (due to the brick turning
>> read-only), you could anyway update the volume configuration with
>> 'remove-brick commit`. Now if the brick still has files left, mount the
>> gluster volume on that node and copy the files from the brick to the
>> volume via the mount. You can then safely rebuild the array/ add a
>> different brick or whatever.
>>
>>> What I want to do is migrate the files off, remove it from gluster,
>>> rebuild the array, rebuild the filesystem, and then add it back as a
>>> brick. (Actually what I'd really like is to hear that the students are
>>> all done with the system and I can turn the whole thing off, but theses
>>> aren't complete yet.)
>>>
>>> Any advice or words of warning will be appreciated.
>> Looks like your bricks are in trouble for over a year now
>> (http://gluster.org/pipermail/gluster-users.old/2013-September/014319.html).
>> Better get them fixed sooner than later! :-)
> Oddly enough the old XRAID systems are holding up better than the VTRAK
> arrays. That doesn't help me much, though, since they're so small.
>
>> HTH,
>> Ravi
>>
>>> James Bellinger
>>>
>>>
>>>
>>>
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>
>
More information about the Gluster-users
mailing list