[Gluster-users] Can't write to volume using vim/nano

Lindolfo Meira meira at cesup.ufrgs.br
Wed Jan 23 23:39:00 UTC 2019


Just checked: when the write is >= 340 bytes, everything works as 
supposed. If the write is smaller, the error takes place. And when it 
does, nothing is logged on the server. The client, however, logs the 
following:

[2019-01-23 23:28:54.554664] W [MSGID: 103046] 
[rdma.c:3502:gf_rdma_decode_header] 0-rpc-transport/rdma: received a msg 
of type RDMA_ERROR

[2019-01-23 23:28:54.554728] W [MSGID: 103046] 
[rdma.c:3939:gf_rdma_process_recv] 0-rpc-transport/rdma: peer 
(172.24.1.6:49152), couldn't encode or decode the msg properly or write 
chunks were not provided for replies that were bigger than 
RDMA_INLINE_THRESHOLD (2048)

[2019-01-23 23:28:54.554765] W [MSGID: 114031] 
[client-rpc-fops_v2.c:680:client4_0_writev_cbk] 0-gfs-client-5: remote 
operation failed [Transport endpoint is not connected]

[2019-01-23 23:28:54.554850] W [fuse-bridge.c:1436:fuse_err_cbk] 
0-glusterfs-fuse: 1723199: FLUSH() ERR => -1 (Transport endpoint is not 
connected)



Lindolfo Meira, MSc
Diretor Geral, Centro Nacional de Supercomputação
Universidade Federal do Rio Grande do Sul
+55 (51) 3308-3139

On Wed, 23 Jan 2019, Lindolfo Meira wrote:

> Hi Jim. Thanks for taking the time.
> 
> Sorry I didn't express myself properly. It's not a simple matter of 
> permissions. Users can write to the volume alright. It's when vim and nano 
> are used, or when small file writes are performed (by cat or echo), that 
> it doesn't work. The file is updated with the write in the server, but it 
> shows up as empty in the client.
> 
> I guess it has something to do with the size of the write, because I ran a 
> test writing to a file one byte at a time, and it never showed up as 
> having any content in the client (although in the server it kept growing 
> accordingly).
> 
> I should point out that I'm using a sharded volume. But when I was testing 
> a striped volume, it also happened. Output of "gluster volume info" 
> follows bellow:
> 
> Volume Name: gfs
> Type: Distribute
> Volume ID: b5ef065f-1ba2-481f-8108-e8f6d2d3f036
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 6
> Transport-type: rdma
> Bricks:
> Brick1: pfs01-ib:/mnt/data
> Brick2: pfs02-ib:/mnt/data
> Brick3: pfs03-ib:/mnt/data
> Brick4: pfs04-ib:/mnt/data
> Brick5: pfs05-ib:/mnt/data
> Brick6: pfs06-ib:/mnt/data
> Options Reconfigured:
> nfs.disable: on
> features.shard: on
> 
> 
> 
> Lindolfo Meira, MSc
> Diretor Geral, Centro Nacional de Supercomputação
> Universidade Federal do Rio Grande do Sul
> +55 (51) 3308-3139
> 
> On Wed, 23 Jan 2019, Jim Kinney wrote:
> 
> > Check permissions on the mount. I have multiple dozens of systems
> > mounting 18 "exports" using fuse and it works for multiple user
> > read/write based on user access permissions to the mount point space.
> > /home is mounted for 150+ users plus another dozen+ lab storage spaces.
> > I do manage user access with freeIPA across all systems to keep things
> > consistent.
> > On Wed, 2019-01-23 at 19:31 -0200, Lindolfo Meira wrote:
> > > Am I missing something here? A mere write operation, using vim or
> > > nano, cannot be performed on a gluster volume mounted over fuse! What
> > > gives?
> > > Lindolfo Meira, MScDiretor Geral, Centro Nacional de
> > > SupercomputaçãoUniversidade Federal do Rio Grande do Sul+55 (51)
> > > 3308-3139_______________________________________________Gluster-users 
> > > mailing listGluster-users at gluster.org
> > > https://lists.gluster.org/mailman/listinfo/gluster-users
> > -- 
> > James P. Kinney III
> > 
> > Every time you stop a school, you will have to build a jail. What you
> > gain at one end you lose at the other. It's like feeding a dog on his
> > own tail. It won't fatten the dog.
> > - Speech 11/23/1900 Mark Twain
> > 
> > http://heretothereideas.blogspot.com/
> > 
> > 


More information about the Gluster-users mailing list