[Gluster-users] Just started using it...
Jason Tolsma
jtolsma at web.turner.com
Tue Apr 19 15:02:04 UTC 2011
Hello all, I am very interested in using gluster and I like what i
have seen so far. I have a few questions though about some weird
behavior I am seeing. Overall goal is to setup a replicated file
service as fully automated as possible.
1. The access permissions seem to be a little strange. It appears if I
want to restrict to only certain IP's it turns off NFS. How does NFS
and the gluster client stuff work together? Also, it seems to me that
if I change the parameters of auth.allow or nfs.rpc-auth-allow they
don't take effect immediately unless I do it while the volume is
offline?
For example see my config below.
I can mount test1 from anywhere via glusterfs and nfs, as expected...
I can mount test2 from only 10.165.20. on glusterfs, as expected. But I
can mount nfs from anywhere, I should only be able to do it from
10.165.20.* right?
I can mount test3 from only the IP's specified as glusterfs, good! But
the NFS daemon isn't exporting test3 at all?!?
2. Is there a way to turn off NFS completely, to restrict only access
via the fuse client? Is there a way to restrict ro/rw from the 'server'
perspective, or only from the client mount options?
3. I would really like to do gluster as a pool, with directories
'exported' out, but it appears that everyone would see the full
filesystem size and quotas are not really there. Is that a true
assumption? If so I will continue down the path of small lvm chunks for
each gluster volume.
4. If I wanted to automate this easier by managing the files themselves
instead of using the gluster commands, what is the best way to get
gluster to pick up the changes to the files?
Thanks in advance!
gluster volume info
Volume Name: test2
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: test8guest1:/gluster/test2/exp
Brick2: test8guest2:/gluster/test2/exp
Options Reconfigured:
nfs.rpc-auth-allow: 10.165.20.*
auth.allow: 10.165.20.*
Volume Name: test1
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: test8guest1:/gluster/test1/exp
Brick2: test8guest2:/gluster/test1/exp
Volume Name: test3
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: test8guest1:/gluster/test3/exp
Brick2: test8guest2:/gluster/test3/exp
Options Reconfigured:
auth.allow: 10.165.20.245,10.165.20.246
nfs.rpc-auth-allow: 10.165.20.245,10.165.20.246
volume nfs-server
type nfs/server
option nfs.dynamic-volumes on
option rpc-auth.addr.test2.allow *
option nfs3.test2.volume-id 4e479c60-3ff5-46c8-a3d5-c9ac78e03b67
option rpc-auth.addr.test1.allow *
option nfs3.test1.volume-id d56ef908-d160-4176-aaef-ebc2ae1232f9
option rpc-auth.addr.test3.allow *
option nfs3.test3.volume-id 98a1ed18-27b6-4d06-a92d-83963e885f85
subvolumes test2 test1 test3
end-volume
--
Jason Tolsma
More information about the Gluster-users
mailing list