[Gluster-users] [Gluster-devel] [Gluster-user] Sybase backup server failed to write to Gluster NFS

Soumya Koduri skoduri at redhat.com
Fri Jan 23 13:14:23 UTC 2015


In that case, most likely it seems to be an issue with the backup 
servers you are using.

Maybe you can first try verifying the NFS client on that machine. Issue 
write fops directly on the NFS mount points used by those servers.

Enable rpcdebug -> "rpcdebug -m nfs all" and check "/var/log/messages" 
for any errors.

Thanks,
Soumya

On 01/22/2015 11:11 PM, Peter Auyeung wrote:
> Hi Soumya,
>
> I was able to mount the same volume on other NFS client and do writes
>
> got the following nfs.log entries when write
>
>
>
>
> [2015-01-22 17:39:03.528405] I [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-sas02-replicate-1:  metadata self heal  is successfully completed,   metadata self heal from source sas02-client-2 to sas02-client-3,  metadata - Pending matrix:  [ [ 0 0 ] [ 0 0 ] ], on /RepDBSata02
> [2015-01-22 17:39:03.529407] I [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-sas02-replicate-2:  metadata self heal  is successfully completed,   metadata self heal from source sas02-client-4 to sas02-client-5,  metadata - Pending matrix:  [ [ 0 0 ] [ 0 0 ] ], on /RepDBSata02
>
>
> Thanks
> Peter
> ________________________________________
> From: Soumya Koduri [skoduri at redhat.com]
> Sent: Wednesday, January 21, 2015 9:05 PM
> To: Peter Auyeung; gluster-devel at gluster.org; gluster-users at gluster.org
> Subject: Re: [Gluster-devel] [Gluster-user] Sybase backup server failed to write to Gluster NFS
>
> Hi Peter,
>
> Can you please try manually mounting those volumes using any/other nfs
> client and check if you are able to perform write operations. Also
> please collect the gluster nfs log while doing so.
>
> Thanks,
> Soumya
>
> On 01/22/2015 08:18 AM, Peter Auyeung wrote:
>> Hi,
>>
>> We have been having 5 sybase servers doing dump/export to Gluster NFS
>> for couple months and yesterday it started to give us these error on not
>> able to write files
>>
>> The gluster NFS export is not full and we can still move and write files
>> as sybase unix user from the sybase servers.
>>
>> There are no error logs on gluster nfs nor the bricks and etc-glusterfs
>> logs and no nfs client error on the sybase servers neither.
>>
>> The NFS export was a replica 2 volume (3x2)
>>
>> I created another NFS export from same gluster but a distributed only
>> volume and still giving out the same error.
>>
>> Any Clue?
>>
>> Thanks
>> Peter
>>
>> Jan 20 20:04:17 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on
>> device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.e'
>> cannot be opened for write access. Mount another volume.
>> Jan 20 20:04:17 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged
>>           @session_id = 87,
>>           @devname =
>> '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.e',
>>           @action = { 'PROCEED' | 'RETRY' | 'ABORT' }
>> Jan 20 20:04:26 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on
>> device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.a'
>> cannot be opened for write access. Mount another volume.
>> Jan 20 20:04:26 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged
>>           @session_id = 87,
>>           @devname =
>> '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.a',
>>           @action = { 'PROCEED' | 'RETRY' | 'ABORT' }
>> Jan 20 20:05:41 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on
>> device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.d'
>> cannot be opened for write access. Mount another volume.
>> Jan 20 20:05:41 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged
>>           @session_id = 87,
>>           @devname =
>> '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.d',
>>           @action = { 'PROCEED' | 'RETRY' | 'ABORT' }
>>
>>
>>
>> _______________________________________________
>> Gluster-devel mailing list
>> Gluster-devel at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-devel
>>


More information about the Gluster-users mailing list