[Gluster-users] ZFS setup question

Shain Miley smiley at npr.org
Fri Jan 7 00:25:19 UTC 2011

Well I setup my zfs filesystems like this:

zfs create pool1/glusterfs
zfs create pool1/glusterfs/audio
zfs create pool1/glusterfs/video
zfs create pool1/glusterfs/documents

it all went well, I restarted gluster...could read the files 
fine...however...I see these errors when trying to write (touch a file):

[2011-01-06 19:11:23] W [posix.c:331:posix_fstat_with_gen] posix1: 
Access to fd 13 (on dev 47775760) is crossing device (47775759)
[2011-01-06 19:11:23] E [posix.c:2267:posix_create] posix1: fstat on 13 
failed: Cross-device link
[2011-01-06 19:15:58] W [posix.c:331:posix_fstat_with_gen] posix1: 
Access to fd 13 (on dev 47775761) is crossing device (47775759)
[2011-01-06 19:15:58] E [posix.c:2267:posix_create] posix1: fstat on 13 
failed: Cross-device link

which makes me think I cannot put everything under pool1/glusterfs like 

So I thought I would do something like this:

zfs create pool1/glusterfs01/audio
zfs create pool1/glusterfs02/video
zfs create pool1/glusterfs03/documents

but the problem with that is then I cannot just share out 
pool1/glusterfs in the .vol file like I wanted to.

Do I have any choice...other then a longer .vol file...or a single 
gluster zfs filesystem named pool1/glusterfs?

At this point the whole cluster is down...so I have to make a choice 

Thanks in advance,


On 01/06/2011 05:12 PM, Jacob Shucart wrote:
> Shain,
> That's correct.  There really is no downside to doing separate ZFS
> filesystems unless you consider the process of creating them or managing
> them a downside.  ZFS is pretty easy to administer, so my overall
> recommendation would be scenario #2.
> -Jacob
> -----Original Message-----
> From: Shain Miley [mailto:smiley at npr.org]
> Sent: Thursday, January 06, 2011 2:10 PM
> To: Jacob Shucart
> Cc: 'Gluster General Discussion List'
> Subject: Re: [Gluster-users] ZFS setup question
> Jocab,
> Thanks for the input.  I did consider that along with having the ability
> to set different properties on each (compression, dedup, etc)  none of
> which I plan on using right now...however I would at least have the
> option in the future.
> The only other thing I was able to come up with was this:
> If one of the shares did get out of sync for example (or if you simply
> wanted to know the size the shares for that matter)...it might be easier
> to tell which one using 'zfs list' or something like that...rather then
> having to do a 'du' on a several TB folder.
> Shain
> On 01/06/2011 04:58 PM, Jacob Shucart wrote:
>> Shain,
>> If you are planning on taking snapshots of the underlying filesystems
> then
>> #2 would be better.  If you are not planning on taking snapshots then #1
>> and #2 are equal really and so I would say that #1 is fine because there
>> are less filesystems to manage.  I hope this clarifies things.  Since
>> snapshots are done at the filesystem level, if you wanted to take a
>> snapshot of just music then you could not do that unless music was on
> its
>> own ZFS filesystem.
>> -Jacob
>> -----Original Message-----
>> From: gluster-users-bounces at gluster.org
>> [mailto:gluster-users-bounces at gluster.org] On Behalf Of Shain Miley
>> Sent: Thursday, January 06, 2011 1:47 PM
>> To: Gluster General Discussion List
>> Subject: [Gluster-users] ZFS setup question
>> Hello,
>> I am in the process of setting up my Gluster shares and I am looking at
>> the following two setup options and I am wondering if anyone can speak
>> to the pros/cons of either:
>> 1) Create one large zfs filesystem for gluster.
>> eg:
>> zfs create pool1/glusterfs
>> and then create several folders with 'mkdir' inside '/pool1/glusterfs'
>> (music,videos,documents).
>> 2) Create 1 zfs filesystem per share.
>> eg:
>> zfs create pool1/glusterfs
>> zfs create pool1/glusterfs/music
>> zfs create pool1/glusterfs/videos
>> zfs create pool1/glusterfs/documents
>> I would then share /pool1/glusterfs out with gluster (I do not want to
>> have to have an overly complicated .vol file with each share having it's
>> own gluster volume).
>> Any thoughts would be great.
>> Thanks,
>> Shain
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

More information about the Gluster-users mailing list