Peter Auyeung
2015-Jan-22 17:41 UTC
[Gluster-users] [Gluster-devel] [Gluster-user] Sybase backup server failed to write to Gluster NFS
Hi Soumya, I was able to mount the same volume on other NFS client and do writes got the following nfs.log entries when write [2015-01-22 17:39:03.528405] I [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-sas02-replicate-1: metadata self heal is successfully completed, metadata self heal from source sas02-client-2 to sas02-client-3, metadata - Pending matrix: [ [ 0 0 ] [ 0 0 ] ], on /RepDBSata02 [2015-01-22 17:39:03.529407] I [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-sas02-replicate-2: metadata self heal is successfully completed, metadata self heal from source sas02-client-4 to sas02-client-5, metadata - Pending matrix: [ [ 0 0 ] [ 0 0 ] ], on /RepDBSata02 Thanks Peter ________________________________________ From: Soumya Koduri [skoduri at redhat.com] Sent: Wednesday, January 21, 2015 9:05 PM To: Peter Auyeung; gluster-devel at gluster.org; gluster-users at gluster.org Subject: Re: [Gluster-devel] [Gluster-user] Sybase backup server failed to write to Gluster NFS Hi Peter, Can you please try manually mounting those volumes using any/other nfs client and check if you are able to perform write operations. Also please collect the gluster nfs log while doing so. Thanks, Soumya On 01/22/2015 08:18 AM, Peter Auyeung wrote:> Hi, > > We have been having 5 sybase servers doing dump/export to Gluster NFS > for couple months and yesterday it started to give us these error on not > able to write files > > The gluster NFS export is not full and we can still move and write files > as sybase unix user from the sybase servers. > > There are no error logs on gluster nfs nor the bricks and etc-glusterfs > logs and no nfs client error on the sybase servers neither. > > The NFS export was a replica 2 volume (3x2) > > I created another NFS export from same gluster but a distributed only > volume and still giving out the same error. > > Any Clue? > > Thanks > Peter > > Jan 20 20:04:17 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on > device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.e' > cannot be opened for write access. Mount another volume. > Jan 20 20:04:17 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged > @session_id = 87, > @devname > '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.e', > @action = { 'PROCEED' | 'RETRY' | 'ABORT' } > Jan 20 20:04:26 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on > device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.a' > cannot be opened for write access. Mount another volume. > Jan 20 20:04:26 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged > @session_id = 87, > @devname > '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.a', > @action = { 'PROCEED' | 'RETRY' | 'ABORT' } > Jan 20 20:05:41 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on > device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.d' > cannot be opened for write access. Mount another volume. > Jan 20 20:05:41 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged > @session_id = 87, > @devname > '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.d', > @action = { 'PROCEED' | 'RETRY' | 'ABORT' } > > > > _______________________________________________ > Gluster-devel mailing list > Gluster-devel at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-devel >
Soumya Koduri
2015-Jan-23 13:14 UTC
[Gluster-users] [Gluster-devel] [Gluster-user] Sybase backup server failed to write to Gluster NFS
In that case, most likely it seems to be an issue with the backup servers you are using. Maybe you can first try verifying the NFS client on that machine. Issue write fops directly on the NFS mount points used by those servers. Enable rpcdebug -> "rpcdebug -m nfs all" and check "/var/log/messages" for any errors. Thanks, Soumya On 01/22/2015 11:11 PM, Peter Auyeung wrote:> Hi Soumya, > > I was able to mount the same volume on other NFS client and do writes > > got the following nfs.log entries when write > > > > > [2015-01-22 17:39:03.528405] I [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-sas02-replicate-1: metadata self heal is successfully completed, metadata self heal from source sas02-client-2 to sas02-client-3, metadata - Pending matrix: [ [ 0 0 ] [ 0 0 ] ], on /RepDBSata02 > [2015-01-22 17:39:03.529407] I [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-sas02-replicate-2: metadata self heal is successfully completed, metadata self heal from source sas02-client-4 to sas02-client-5, metadata - Pending matrix: [ [ 0 0 ] [ 0 0 ] ], on /RepDBSata02 > > > Thanks > Peter > ________________________________________ > From: Soumya Koduri [skoduri at redhat.com] > Sent: Wednesday, January 21, 2015 9:05 PM > To: Peter Auyeung; gluster-devel at gluster.org; gluster-users at gluster.org > Subject: Re: [Gluster-devel] [Gluster-user] Sybase backup server failed to write to Gluster NFS > > Hi Peter, > > Can you please try manually mounting those volumes using any/other nfs > client and check if you are able to perform write operations. Also > please collect the gluster nfs log while doing so. > > Thanks, > Soumya > > On 01/22/2015 08:18 AM, Peter Auyeung wrote: >> Hi, >> >> We have been having 5 sybase servers doing dump/export to Gluster NFS >> for couple months and yesterday it started to give us these error on not >> able to write files >> >> The gluster NFS export is not full and we can still move and write files >> as sybase unix user from the sybase servers. >> >> There are no error logs on gluster nfs nor the bricks and etc-glusterfs >> logs and no nfs client error on the sybase servers neither. >> >> The NFS export was a replica 2 volume (3x2) >> >> I created another NFS export from same gluster but a distributed only >> volume and still giving out the same error. >> >> Any Clue? >> >> Thanks >> Peter >> >> Jan 20 20:04:17 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on >> device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.e' >> cannot be opened for write access. Mount another volume. >> Jan 20 20:04:17 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged >> @session_id = 87, >> @devname >> '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.e', >> @action = { 'PROCEED' | 'RETRY' | 'ABORT' } >> Jan 20 20:04:26 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on >> device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.a' >> cannot be opened for write access. Mount another volume. >> Jan 20 20:04:26 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged >> @session_id = 87, >> @devname >> '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.a', >> @action = { 'PROCEED' | 'RETRY' | 'ABORT' } >> Jan 20 20:05:41 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on >> device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.d' >> cannot be opened for write access. Mount another volume. >> Jan 20 20:05:41 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged >> @session_id = 87, >> @devname >> '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.d', >> @action = { 'PROCEED' | 'RETRY' | 'ABORT' } >> >> >> >> _______________________________________________ >> Gluster-devel mailing list >> Gluster-devel at gluster.org >> http://www.gluster.org/mailman/listinfo/gluster-devel >>