[Gluster-users] [Gluster-devel] [Gluster-user] Sybase backup server failed to write to Gluster NFS

Peter Auyeung pauyeung at connexity.com
Thu Jan 22 17:41:03 UTC 2015


Hi Soumya,

I was able to mount the same volume on other NFS client and do writes

got the following nfs.log entries when write




[2015-01-22 17:39:03.528405] I [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-sas02-replicate-1:  metadata self heal  is successfully completed,   metadata self heal from source sas02-client-2 to sas02-client-3,  metadata - Pending matrix:  [ [ 0 0 ] [ 0 0 ] ], on /RepDBSata02
[2015-01-22 17:39:03.529407] I [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-sas02-replicate-2:  metadata self heal  is successfully completed,   metadata self heal from source sas02-client-4 to sas02-client-5,  metadata - Pending matrix:  [ [ 0 0 ] [ 0 0 ] ], on /RepDBSata02


Thanks
Peter
________________________________________
From: Soumya Koduri [skoduri at redhat.com]
Sent: Wednesday, January 21, 2015 9:05 PM
To: Peter Auyeung; gluster-devel at gluster.org; gluster-users at gluster.org
Subject: Re: [Gluster-devel] [Gluster-user] Sybase backup server failed to write to Gluster NFS

Hi Peter,

Can you please try manually mounting those volumes using any/other nfs
client and check if you are able to perform write operations. Also
please collect the gluster nfs log while doing so.

Thanks,
Soumya

On 01/22/2015 08:18 AM, Peter Auyeung wrote:
> Hi,
>
> We have been having 5 sybase servers doing dump/export to Gluster NFS
> for couple months and yesterday it started to give us these error on not
> able to write files
>
> The gluster NFS export is not full and we can still move and write files
> as sybase unix user from the sybase servers.
>
> There are no error logs on gluster nfs nor the bricks and etc-glusterfs
> logs and no nfs client error on the sybase servers neither.
>
> The NFS export was a replica 2 volume (3x2)
>
> I created another NFS export from same gluster but a distributed only
> volume and still giving out the same error.
>
> Any Clue?
>
> Thanks
> Peter
>
> Jan 20 20:04:17 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on
> device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.e'
> cannot be opened for write access. Mount another volume.
> Jan 20 20:04:17 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged
>          @session_id = 87,
>          @devname =
> '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.e',
>          @action = { 'PROCEED' | 'RETRY' | 'ABORT' }
> Jan 20 20:04:26 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on
> device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.a'
> cannot be opened for write access. Mount another volume.
> Jan 20 20:04:26 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged
>          @session_id = 87,
>          @devname =
> '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.a',
>          @action = { 'PROCEED' | 'RETRY' | 'ABORT' }
> Jan 20 20:05:41 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on
> device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.d'
> cannot be opened for write access. Mount another volume.
> Jan 20 20:05:41 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged
>          @session_id = 87,
>          @devname =
> '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.d',
>          @action = { 'PROCEED' | 'RETRY' | 'ABORT' }
>
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-devel
>


More information about the Gluster-users mailing list