[Gluster-users] problem with missing bricks
Todd Pfaff
pfaff at mcmaster.ca
Mon Jan 2 03:43:04 UTC 2012
Yes, that's correct, the drives are mounted on /gluster/brick*.
df -h
Filesystem Size Used Avail Use% Mounted on
/dev/sda1 20G 8.7G 11G 47% /
/dev/sda3 209G 21G 186G 11% /1
/dev/md2 3.6T 2.0T 1.7T 55% /4
/dev/md3 7.2T 5.5T 1.8T 76% /5
/dev/md4 7.2T 5.4T 1.9T 75% /6
/dev/md5 7.2T 5.6T 1.6T 79% /7
/dev/md6 7.2T 5.5T 1.4T 81% /8
/dev/md7 3.6T 1.2T 2.5T 31% /9
/dev/sdb1 917G 337G 572G 38% /gluster/brick01
/dev/sdc1 917G 236G 672G 27% /gluster/brick02
/dev/sdd1 917G 215G 693G 24% /gluster/brick03
/dev/sde1 917G 365G 544G 41% /gluster/brick04
/dev/sdf1 917G 6.1G 902G 1% /gluster/brick05
/dev/sdh1 917G 427G 482G 47% /gluster/brick06
/dev/sdi1 917G 345G 564G 38% /gluster/brick07
/dev/sdj1 917G 285G 624G 32% /gluster/brick08
/dev/sdk1 917G 449G 460G 50% /gluster/brick09
/dev/sdg1 917G 3.5G 905G 1% /gluster/brick10
localhost:/volume1 4.5T 1.5T 3.0T 34% /gluster/volume1
Good idea, thanks. So, just to clarify, what I did is to create a
subdirectory at the top of each of the brick filesystems and then
use that subdirectory in the brick path when creating the volume.
When the brick drive is not mounted, the subdirectory will not exist
and so gluster should not see the path as part of the volume.
However, I still find it hard to believe that gluster doesn't somehow
signify that a brick is really a brick with some sort of data at the
top of the filesystem of each brick.
Thanks,
Todd
On Sun, 1 Jan 2012, Mohit Anchlia wrote:
> Can you post the output of df -h? It looks like you have mounted the
> drives as /gluster/brick[0-9]? If that is correct then you can
> mitigate that by create a sub directory on that mount points accross
> all the bricks, and do writes only inside that sub-dir.
>
> On Sat, Dec 31, 2011 at 8:50 AM, Todd Pfaff <pfaff at rhpcs.mcmaster.ca> wrote:
>> Gluster-user folks,
>>
>> I'm trying to use gluster in a way that may be a considered an unusual use
>> case for gluster. Feel free to let me know if you think what I'm doing
>> is dumb. It just feels very comfortable doing this with gluster.
>> I have been using gluster in other, more orthodox configurations, for
>> several years.
>>
>> I have a single system with 45 inexpensive sata drives - it's a self-built
>> backblaze similar to that documented at this url but with some upgrades
>> and substitutions:
>>
>> http://blog.backblaze.com/2009/09/01/petabytes-on-a-budget-how-to-build-cheap-cloud-storage/
>>
>> We use this system for disk-to-disk backups only, no primary storage,
>> nothing mission critical.
>>
>> For the past two years I have been using this system with linux software
>> raid, with the drives organized as multiple raid 5/6/10 sets of 5 drives
>> per set. This has worked ok but I have suffered enough multiple
>> simultaneous drive failures to prompt me to explore alternatives to raid.
>> Yes, I know, that's what I get for using cheap sata drives.
>>
>> What I'm experimenting with now is creating gluster distributed-replicated
>> volumes on some of these drives, and maybe all in the future if this works
>> reasonably well.
>>
>> At this point I am using 10 of the drives configured as shown here:
>>
>> Volume Name: volume1
>> Type: Distributed-Replicate
>> Status: Started
>> Number of Bricks: 5 x 2 = 10
>> Transport-type: tcp
>> Bricks:
>> Brick1: host:/gluster/brick01
>> Brick2: host:/gluster/brick06
>> Brick3: host:/gluster/brick02
>> Brick4: host:/gluster/brick07
>> Brick5: host:/gluster/brick03
>> Brick6: host:/gluster/brick08
>> Brick7: host:/gluster/brick04
>> Brick8: host:/gluster/brick09
>> Brick9: host:/gluster/brick05
>> Brick10: host:/gluster/brick10
>> Options Reconfigured:
>> auth.allow: 127.0.0.1,10.10.10.10
>>
>>
>> For the most part this is working fine so far. The problem I have run
>> into several times now is that when a drive fails and the system is
>> rebooted, the volume comes up without that brick. Gluster happily writes
>> to the missing brick's mount point, thereby eventually filling up the root
>> filesystem. Once the root filesystem is full and processes writing to
>> gluster space are hung, I can never recover from this state without
>> rebooting.
>>
>> Is there any way to avoid this problem of gluster writing to a brick
>> path that isn't really populated by the intended brick filesystem?
>> Does gluster not create any sort of signature or meta-data that
>> indicates whether or not a path is really a gluster brick?
>> How do others deal with missing bricks?
>>
>> I realize that ultimately I should get the bricks replaced as soon as
>> possible but there may be times when I want to continue running for some
>> time with a "degraded" volume if you will.
>>
>> Any and all ideas, suggestions, comments, criticisms are welcome.
>>
>> Cheers and Happy New Year,
>> Todd
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
>
More information about the Gluster-users
mailing list