I get this: [root at gluster03-mi glusterfs]# git fetch git://review.gluster.org/glusterfs refs/changes/28/9328/4 && git checkout FETCH_HEAD fatal: Couldn't find remote ref refs/changes/28/9328/4 What's wrong with that? A. -----Original Message----- From: Atin Mukherjee [mailto:amukherj at redhat.com] Sent: venerd? 20 febbraio 2015 12:54 To: RASTELLI Alessandro Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] GlusterD uses 50% of RAM>From the cmd log history I could see lots of volume status commands were triggered parallely. This is a known issue for 3.6 and it would cause a memory leak. http://review.gluster.org/#/c/9328/ should solve it.~Atin On 02/20/2015 04:36 PM, RASTELLI Alessandro wrote:> 10MB log > sorry :) > > -----Original Message----- > From: Atin Mukherjee [mailto:amukherj at redhat.com] > Sent: venerd? 20 febbraio 2015 10:49 > To: RASTELLI Alessandro; gluster-users at gluster.org > Subject: Re: [Gluster-users] GlusterD uses 50% of RAM > > Could you please share the cmd_history.log & glusterd log file to analyze this high memory usage. > > ~Atin > > On 02/20/2015 03:10 PM, RASTELLI Alessandro wrote: >> Hi, >> I've noticed that one of our 6 gluster 3.6.2 nodes has "glusterd" >> process using 50% of RAM, on the other nodes usage is about 5% This can be a bug? >> Should I restart glusterd daemon? >> Thank you >> A >> >> From: Volnei Puttini [mailto:volnei at vcplinux.com.br] >> Sent: luned? 9 febbraio 2015 18:06 >> To: RASTELLI Alessandro; gluster-users at gluster.org >> Subject: Re: [Gluster-users] cannot access to CIFS export >> >> Hi Alessandro, >> >> My system: >> >> CentOS 7 >> >> samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64 >> samba-winbind-4.1.1-37.el7_0.x86_64 >> samba-libs-4.1.1-37.el7_0.x86_64 >> samba-common-4.1.1-37.el7_0.x86_64 >> samba-winbind-modules-4.1.1-37.el7_0.x86_64 >> samba-winbind-clients-4.1.1-37.el7_0.x86_64 >> samba-4.1.1-37.el7_0.x86_64 >> samba-client-4.1.1-37.el7_0.x86_64 >> >> glusterfs 3.6.2 built on Jan 22 2015 12:59:57 >> >> Try this, work fine for me: >> >> [GFSVOL] >> browseable = No >> comment = Gluster share of volume gfsvol >> path = / >> read only = No >> guest ok = Yes >> kernel share modes = No >> posix locking = No >> vfs objects = glusterfs >> glusterfs:loglevel = 7 >> glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log >> glusterfs:volume = vgtest >> glusterfs:volfile_server = 192.168.2.21 >> >> On 09-02-2015 14:45, RASTELLI Alessandro wrote: >> Hi, >> I've created and started a new replica volume "downloadstat" with CIFS export enabled on GlusterFS 3.6.2. >> I can see the following piece has been added automatically to smb.conf: >> [gluster-downloadstat] >> comment = For samba share of volume downloadstat vfs objects = >> glusterfs glusterfs:volume = downloadstat glusterfs:logfile = >> /var/log/samba/glusterfs-downloadstat.%M.log >> glusterfs:loglevel = 7 >> path = / >> read only = no >> guest ok = yes >> >> I restarted smb service, without errors. >> When I try to access from Win7 client to "\\gluster01-mi\gluster-downloadstat<file:///\\gluster01-mi\gluster-downloadstat>" it asks me a login (which user do I need to put?) and then gives me error "The network path was not found" >> and on Gluster smb.log I see: >> [2015/02/09 17:21:13.111639, 0] smbd/vfs.c:173(vfs_init_custom) >> error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL >> [2015/02/09 17:21:13.111709, 0] smbd/vfs.c:315(smbd_vfs_init) >> smbd_vfs_init: vfs_init_custom failed for glusterfs >> [2015/02/09 17:21:13.111741, 0] smbd/service.c:902(make_connection_snum) >> vfs_init failed for service gluster-downloadstat >> >> Can you explain how to fix? >> Thanks >> >> Alessandro >> >> From: >> gluster-users-bounces at gluster.org<mailto:gluster-users-bounces at gluste >> r .org> [mailto:gluster-users-bounces at gluster.org] On Behalf Of David >> F. >> Robinson >> Sent: domenica 8 febbraio 2015 18:19 >> To: Gluster Devel; >> gluster-users at gluster.org<mailto:gluster-users at gluster.org> >> Subject: [Gluster-users] cannot delete non-empty directory >> >> I am seeing these messsages after I delete large amounts of data using gluster 3.6.2. >> cannot delete non-empty directory: >> old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_fin >> a >> l >> >> >From the FUSE mount (as root), the directory shows up as empty: >> >> # pwd >> /backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storag >> e >> /Jimmy_Old/src_vj1.5_final >> >> # ls -al >> total 5 >> d--------- 2 root root 4106 Feb 6 13:55 . >> drwxrws--- 3 601 dmiller 72 Feb 6 13:55 .. >> >> However, when you look at the bricks, the files are still there (none on brick01bkp, all files are on brick02bkp). All of the files are 0-length and have ------T permissions. >> Any suggestions on how to fix this and how to prevent it from happening? >> >> # ls -al >> /data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/Rav >> e nCFD/Storage/Jimmy_Old/src_vj1.5_final >> /data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: >> total 4 >> d---------+ 2 root root 10 Feb 6 13:55 . >> drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. >> >> /data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: >> total 8 >> d---------+ 3 root root 4096 Dec 31 1969 . >> drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. >> ---------T 5 601 raven 0 Nov 20 00:08 read_inset.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 readbc.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 readcn.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 readinp.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_2.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_3.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 rotatept.f.gz >> d---------+ 2 root root 118 Feb 6 13:54 save1 >> ---------T 5 601 raven 0 Nov 20 00:08 sepvec.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 shadow.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 snksrc.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 source.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 step.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 stoprog.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 summer6.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 totforc.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 tritet.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 wallrsd.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 wheat.f.gz >> ---------T 5 601 raven 0 Nov 20 00:08 write_inset.f.gz >> >> >> This is using gluster 3.6.2 on a distributed gluster volume that resides on a single machine. Both of the bricks are on one machine consisting of 2x RAID-6 arrays. >> >> df -h | grep brick >> /dev/mapper/vg01-lvol1 88T 22T 66T 25% /data/brick01bkp >> /dev/mapper/vg02-lvol1 88T 22T 66T 26% /data/brick02bkp >> >> # gluster volume info homegfs_bkp >> Volume Name: homegfs_bkp >> Type: Distribute >> Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294 >> Status: Started >> Number of Bricks: 2 >> Transport-type: tcp >> Bricks: >> Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp >> Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp >> Options Reconfigured: >> storage.owner-gid: 100 >> performance.io-thread-count: 32 >> server.allow-insecure: on >> network.ping-timeout: 10 >> performance.cache-size: 128MB >> performance.write-behind-window-size: 128MB >> server.manage-gids: on >> changelog.rollover-time: 15 >> changelog.fsync-interval: 3 >> >> >> >> ==============================>> David F. Robinson, Ph.D. >> President - Corvid Technologies >> 704.799.6944 x101 [office] >> 704.252.1310 [cell] >> 704.799.7974 [fax] >> David.Robinson at corvidtec.com<mailto:David.Robinson at corvidtec.com> >> http://www.corvidtechnologies.com >> >> >> >> >> >> >> _______________________________________________ >> >> Gluster-users mailing list >> >> Gluster-users at gluster.org<mailto:Gluster-users at gluster.org> >> >> http://www.gluster.org/mailman/listinfo/gluster-users >> >> >> >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://www.gluster.org/mailman/listinfo/gluster-users >> > > -- > ~Atin >-- ~Atin
On Fri, Feb 20, 2015 at 01:50:38PM +0000, RASTELLI Alessandro wrote:> I get this: > > [root at gluster03-mi glusterfs]# git fetch git://review.gluster.org/glusterfs refs/changes/28/9328/4 && git checkout FETCH_HEAD > fatal: Couldn't find remote ref refs/changes/28/9328/4 > > What's wrong with that?I think anonymous git does not (always) work. You could try fetching over HTTP: $ git fetch http://review.gluster.org/glusterfs refs/changes/28/9328/4 && git checkout FETCH_HEAD Niels> > A. > > -----Original Message----- > From: Atin Mukherjee [mailto:amukherj at redhat.com] > Sent: venerd? 20 febbraio 2015 12:54 > To: RASTELLI Alessandro > Cc: gluster-users at gluster.org > Subject: Re: [Gluster-users] GlusterD uses 50% of RAM > > From the cmd log history I could see lots of volume status commands were triggered parallely. This is a known issue for 3.6 and it would cause a memory leak. http://review.gluster.org/#/c/9328/ should solve it. > > ~Atin > > On 02/20/2015 04:36 PM, RASTELLI Alessandro wrote: > > 10MB log > > sorry :) > > > > -----Original Message----- > > From: Atin Mukherjee [mailto:amukherj at redhat.com] > > Sent: venerd? 20 febbraio 2015 10:49 > > To: RASTELLI Alessandro; gluster-users at gluster.org > > Subject: Re: [Gluster-users] GlusterD uses 50% of RAM > > > > Could you please share the cmd_history.log & glusterd log file to analyze this high memory usage. > > > > ~Atin > > > > On 02/20/2015 03:10 PM, RASTELLI Alessandro wrote: > >> Hi, > >> I've noticed that one of our 6 gluster 3.6.2 nodes has "glusterd" > >> process using 50% of RAM, on the other nodes usage is about 5% This can be a bug? > >> Should I restart glusterd daemon? > >> Thank you > >> A > >> > >> From: Volnei Puttini [mailto:volnei at vcplinux.com.br] > >> Sent: luned? 9 febbraio 2015 18:06 > >> To: RASTELLI Alessandro; gluster-users at gluster.org > >> Subject: Re: [Gluster-users] cannot access to CIFS export > >> > >> Hi Alessandro, > >> > >> My system: > >> > >> CentOS 7 > >> > >> samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64 > >> samba-winbind-4.1.1-37.el7_0.x86_64 > >> samba-libs-4.1.1-37.el7_0.x86_64 > >> samba-common-4.1.1-37.el7_0.x86_64 > >> samba-winbind-modules-4.1.1-37.el7_0.x86_64 > >> samba-winbind-clients-4.1.1-37.el7_0.x86_64 > >> samba-4.1.1-37.el7_0.x86_64 > >> samba-client-4.1.1-37.el7_0.x86_64 > >> > >> glusterfs 3.6.2 built on Jan 22 2015 12:59:57 > >> > >> Try this, work fine for me: > >> > >> [GFSVOL] > >> browseable = No > >> comment = Gluster share of volume gfsvol > >> path = / > >> read only = No > >> guest ok = Yes > >> kernel share modes = No > >> posix locking = No > >> vfs objects = glusterfs > >> glusterfs:loglevel = 7 > >> glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log > >> glusterfs:volume = vgtest > >> glusterfs:volfile_server = 192.168.2.21 > >> > >> On 09-02-2015 14:45, RASTELLI Alessandro wrote: > >> Hi, > >> I've created and started a new replica volume "downloadstat" with CIFS export enabled on GlusterFS 3.6.2. > >> I can see the following piece has been added automatically to smb.conf: > >> [gluster-downloadstat] > >> comment = For samba share of volume downloadstat vfs objects = > >> glusterfs glusterfs:volume = downloadstat glusterfs:logfile = > >> /var/log/samba/glusterfs-downloadstat.%M.log > >> glusterfs:loglevel = 7 > >> path = / > >> read only = no > >> guest ok = yes > >> > >> I restarted smb service, without errors. > >> When I try to access from Win7 client to "\\gluster01-mi\gluster-downloadstat<file:///\\gluster01-mi\gluster-downloadstat>" it asks me a login (which user do I need to put?) and then gives me error "The network path was not found" > >> and on Gluster smb.log I see: > >> [2015/02/09 17:21:13.111639, 0] smbd/vfs.c:173(vfs_init_custom) > >> error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL > >> [2015/02/09 17:21:13.111709, 0] smbd/vfs.c:315(smbd_vfs_init) > >> smbd_vfs_init: vfs_init_custom failed for glusterfs > >> [2015/02/09 17:21:13.111741, 0] smbd/service.c:902(make_connection_snum) > >> vfs_init failed for service gluster-downloadstat > >> > >> Can you explain how to fix? > >> Thanks > >> > >> Alessandro > >> > >> From: > >> gluster-users-bounces at gluster.org<mailto:gluster-users-bounces at gluste > >> r .org> [mailto:gluster-users-bounces at gluster.org] On Behalf Of David > >> F. > >> Robinson > >> Sent: domenica 8 febbraio 2015 18:19 > >> To: Gluster Devel; > >> gluster-users at gluster.org<mailto:gluster-users at gluster.org> > >> Subject: [Gluster-users] cannot delete non-empty directory > >> > >> I am seeing these messsages after I delete large amounts of data using gluster 3.6.2. > >> cannot delete non-empty directory: > >> old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_fin > >> a > >> l > >> > >> >From the FUSE mount (as root), the directory shows up as empty: > >> > >> # pwd > >> /backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storag > >> e > >> /Jimmy_Old/src_vj1.5_final > >> > >> # ls -al > >> total 5 > >> d--------- 2 root root 4106 Feb 6 13:55 . > >> drwxrws--- 3 601 dmiller 72 Feb 6 13:55 .. > >> > >> However, when you look at the bricks, the files are still there (none on brick01bkp, all files are on brick02bkp). All of the files are 0-length and have ------T permissions. > >> Any suggestions on how to fix this and how to prevent it from happening? > >> > >> # ls -al > >> /data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/Rav > >> e nCFD/Storage/Jimmy_Old/src_vj1.5_final > >> /data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: > >> total 4 > >> d---------+ 2 root root 10 Feb 6 13:55 . > >> drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. > >> > >> /data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: > >> total 8 > >> d---------+ 3 root root 4096 Dec 31 1969 . > >> drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. > >> ---------T 5 601 raven 0 Nov 20 00:08 read_inset.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 readbc.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 readcn.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 readinp.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_2.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_3.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 rotatept.f.gz > >> d---------+ 2 root root 118 Feb 6 13:54 save1 > >> ---------T 5 601 raven 0 Nov 20 00:08 sepvec.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 shadow.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 snksrc.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 source.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 step.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 stoprog.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 summer6.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 totforc.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 tritet.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 wallrsd.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 wheat.f.gz > >> ---------T 5 601 raven 0 Nov 20 00:08 write_inset.f.gz > >> > >> > >> This is using gluster 3.6.2 on a distributed gluster volume that resides on a single machine. Both of the bricks are on one machine consisting of 2x RAID-6 arrays. > >> > >> df -h | grep brick > >> /dev/mapper/vg01-lvol1 88T 22T 66T 25% /data/brick01bkp > >> /dev/mapper/vg02-lvol1 88T 22T 66T 26% /data/brick02bkp > >> > >> # gluster volume info homegfs_bkp > >> Volume Name: homegfs_bkp > >> Type: Distribute > >> Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294 > >> Status: Started > >> Number of Bricks: 2 > >> Transport-type: tcp > >> Bricks: > >> Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp > >> Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp > >> Options Reconfigured: > >> storage.owner-gid: 100 > >> performance.io-thread-count: 32 > >> server.allow-insecure: on > >> network.ping-timeout: 10 > >> performance.cache-size: 128MB > >> performance.write-behind-window-size: 128MB > >> server.manage-gids: on > >> changelog.rollover-time: 15 > >> changelog.fsync-interval: 3 > >> > >> > >> > >> ==============================> >> David F. Robinson, Ph.D. > >> President - Corvid Technologies > >> 704.799.6944 x101 [office] > >> 704.252.1310 [cell] > >> 704.799.7974 [fax] > >> David.Robinson at corvidtec.com<mailto:David.Robinson at corvidtec.com> > >> http://www.corvidtechnologies.com > >> > >> > >> > >> > >> > >> > >> _______________________________________________ > >> > >> Gluster-users mailing list > >> > >> Gluster-users at gluster.org<mailto:Gluster-users at gluster.org> > >> > >> http://www.gluster.org/mailman/listinfo/gluster-users > >> > >> > >> > >> > >> _______________________________________________ > >> Gluster-users mailing list > >> Gluster-users at gluster.org > >> http://www.gluster.org/mailman/listinfo/gluster-users > >> > > > > -- > > ~Atin > > > > -- > ~Atin > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users-------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 181 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150220/a905f9d8/attachment.sig>
On 02/20/2015 07:20 PM, RASTELLI Alessandro wrote:> I get this: > > [root at gluster03-mi glusterfs]# git fetch git://review.gluster.org/glusterfs refs/changes/28/9328/4 && git checkout FETCH_HEAD > fatal: Couldn't find remote ref refs/changes/28/9328/4 > > What's wrong with that?Is your current branch at 3.6 ?> > A. > > -----Original Message----- > From: Atin Mukherjee [mailto:amukherj at redhat.com] > Sent: venerd? 20 febbraio 2015 12:54 > To: RASTELLI Alessandro > Cc: gluster-users at gluster.org > Subject: Re: [Gluster-users] GlusterD uses 50% of RAM > >>From the cmd log history I could see lots of volume status commands were triggered parallely. This is a known issue for 3.6 and it would cause a memory leak. http://review.gluster.org/#/c/9328/ should solve it. > > ~Atin > > On 02/20/2015 04:36 PM, RASTELLI Alessandro wrote: >> 10MB log >> sorry :) >> >> -----Original Message----- >> From: Atin Mukherjee [mailto:amukherj at redhat.com] >> Sent: venerd? 20 febbraio 2015 10:49 >> To: RASTELLI Alessandro; gluster-users at gluster.org >> Subject: Re: [Gluster-users] GlusterD uses 50% of RAM >> >> Could you please share the cmd_history.log & glusterd log file to analyze this high memory usage. >> >> ~Atin >> >> On 02/20/2015 03:10 PM, RASTELLI Alessandro wrote: >>> Hi, >>> I've noticed that one of our 6 gluster 3.6.2 nodes has "glusterd" >>> process using 50% of RAM, on the other nodes usage is about 5% This can be a bug? >>> Should I restart glusterd daemon? >>> Thank you >>> A >>> >>> From: Volnei Puttini [mailto:volnei at vcplinux.com.br] >>> Sent: luned? 9 febbraio 2015 18:06 >>> To: RASTELLI Alessandro; gluster-users at gluster.org >>> Subject: Re: [Gluster-users] cannot access to CIFS export >>> >>> Hi Alessandro, >>> >>> My system: >>> >>> CentOS 7 >>> >>> samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64 >>> samba-winbind-4.1.1-37.el7_0.x86_64 >>> samba-libs-4.1.1-37.el7_0.x86_64 >>> samba-common-4.1.1-37.el7_0.x86_64 >>> samba-winbind-modules-4.1.1-37.el7_0.x86_64 >>> samba-winbind-clients-4.1.1-37.el7_0.x86_64 >>> samba-4.1.1-37.el7_0.x86_64 >>> samba-client-4.1.1-37.el7_0.x86_64 >>> >>> glusterfs 3.6.2 built on Jan 22 2015 12:59:57 >>> >>> Try this, work fine for me: >>> >>> [GFSVOL] >>> browseable = No >>> comment = Gluster share of volume gfsvol >>> path = / >>> read only = No >>> guest ok = Yes >>> kernel share modes = No >>> posix locking = No >>> vfs objects = glusterfs >>> glusterfs:loglevel = 7 >>> glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log >>> glusterfs:volume = vgtest >>> glusterfs:volfile_server = 192.168.2.21 >>> >>> On 09-02-2015 14:45, RASTELLI Alessandro wrote: >>> Hi, >>> I've created and started a new replica volume "downloadstat" with CIFS export enabled on GlusterFS 3.6.2. >>> I can see the following piece has been added automatically to smb.conf: >>> [gluster-downloadstat] >>> comment = For samba share of volume downloadstat vfs objects = >>> glusterfs glusterfs:volume = downloadstat glusterfs:logfile = >>> /var/log/samba/glusterfs-downloadstat.%M.log >>> glusterfs:loglevel = 7 >>> path = / >>> read only = no >>> guest ok = yes >>> >>> I restarted smb service, without errors. >>> When I try to access from Win7 client to "\\gluster01-mi\gluster-downloadstat<file:///\\gluster01-mi\gluster-downloadstat>" it asks me a login (which user do I need to put?) and then gives me error "The network path was not found" >>> and on Gluster smb.log I see: >>> [2015/02/09 17:21:13.111639, 0] smbd/vfs.c:173(vfs_init_custom) >>> error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL >>> [2015/02/09 17:21:13.111709, 0] smbd/vfs.c:315(smbd_vfs_init) >>> smbd_vfs_init: vfs_init_custom failed for glusterfs >>> [2015/02/09 17:21:13.111741, 0] smbd/service.c:902(make_connection_snum) >>> vfs_init failed for service gluster-downloadstat >>> >>> Can you explain how to fix? >>> Thanks >>> >>> Alessandro >>> >>> From: >>> gluster-users-bounces at gluster.org<mailto:gluster-users-bounces at gluste >>> r .org> [mailto:gluster-users-bounces at gluster.org] On Behalf Of David >>> F. >>> Robinson >>> Sent: domenica 8 febbraio 2015 18:19 >>> To: Gluster Devel; >>> gluster-users at gluster.org<mailto:gluster-users at gluster.org> >>> Subject: [Gluster-users] cannot delete non-empty directory >>> >>> I am seeing these messsages after I delete large amounts of data using gluster 3.6.2. >>> cannot delete non-empty directory: >>> old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_fin >>> a >>> l >>> >>> >From the FUSE mount (as root), the directory shows up as empty: >>> >>> # pwd >>> /backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storag >>> e >>> /Jimmy_Old/src_vj1.5_final >>> >>> # ls -al >>> total 5 >>> d--------- 2 root root 4106 Feb 6 13:55 . >>> drwxrws--- 3 601 dmiller 72 Feb 6 13:55 .. >>> >>> However, when you look at the bricks, the files are still there (none on brick01bkp, all files are on brick02bkp). All of the files are 0-length and have ------T permissions. >>> Any suggestions on how to fix this and how to prevent it from happening? >>> >>> # ls -al >>> /data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/Rav >>> e nCFD/Storage/Jimmy_Old/src_vj1.5_final >>> /data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: >>> total 4 >>> d---------+ 2 root root 10 Feb 6 13:55 . >>> drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. >>> >>> /data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: >>> total 8 >>> d---------+ 3 root root 4096 Dec 31 1969 . >>> drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. >>> ---------T 5 601 raven 0 Nov 20 00:08 read_inset.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 readbc.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 readcn.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 readinp.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_2.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_3.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 rotatept.f.gz >>> d---------+ 2 root root 118 Feb 6 13:54 save1 >>> ---------T 5 601 raven 0 Nov 20 00:08 sepvec.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 shadow.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 snksrc.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 source.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 step.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 stoprog.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 summer6.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 totforc.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 tritet.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 wallrsd.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 wheat.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 write_inset.f.gz >>> >>> >>> This is using gluster 3.6.2 on a distributed gluster volume that resides on a single machine. Both of the bricks are on one machine consisting of 2x RAID-6 arrays. >>> >>> df -h | grep brick >>> /dev/mapper/vg01-lvol1 88T 22T 66T 25% /data/brick01bkp >>> /dev/mapper/vg02-lvol1 88T 22T 66T 26% /data/brick02bkp >>> >>> # gluster volume info homegfs_bkp >>> Volume Name: homegfs_bkp >>> Type: Distribute >>> Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294 >>> Status: Started >>> Number of Bricks: 2 >>> Transport-type: tcp >>> Bricks: >>> Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp >>> Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp >>> Options Reconfigured: >>> storage.owner-gid: 100 >>> performance.io-thread-count: 32 >>> server.allow-insecure: on >>> network.ping-timeout: 10 >>> performance.cache-size: 128MB >>> performance.write-behind-window-size: 128MB >>> server.manage-gids: on >>> changelog.rollover-time: 15 >>> changelog.fsync-interval: 3 >>> >>> >>> >>> ==============================>>> David F. Robinson, Ph.D. >>> President - Corvid Technologies >>> 704.799.6944 x101 [office] >>> 704.252.1310 [cell] >>> 704.799.7974 [fax] >>> David.Robinson at corvidtec.com<mailto:David.Robinson at corvidtec.com> >>> http://www.corvidtechnologies.com >>> >>> >>> >>> >>> >>> >>> _______________________________________________ >>> >>> Gluster-users mailing list >>> >>> Gluster-users at gluster.org<mailto:Gluster-users at gluster.org> >>> >>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >>> >>> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >> >> -- >> ~Atin >> > > -- > ~Atin >-- ~Atin
Hi, today the issue happened once again. Glusterd process was using 80% of RAM and its log was filling up the /var/log. One month ago, when the issue last happened, you suggested to install a patch, so I did this: git fetch git://review.gluster.org/glusterfs refs/changes/28/9328/4 git fetch http://review.gluster.org/glusterfs refs/changes/28/9328/4 && git checkout FETCH_HEAD Is this enough to install the patch or I missed something? Thank you Alessandro -----Original Message----- From: RASTELLI Alessandro Sent: marted? 24 febbraio 2015 10:28 To: 'Atin Mukherjee' Cc: gluster-users at gluster.org Subject: RE: [Gluster-users] GlusterD uses 50% of RAM Hi Atin, I managed to install the patch, it fixed the issue Thank you A. -----Original Message----- From: Atin Mukherjee [mailto:amukherj at redhat.com] Sent: marted? 24 febbraio 2015 08:03 To: RASTELLI Alessandro Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] GlusterD uses 50% of RAM On 02/20/2015 07:20 PM, RASTELLI Alessandro wrote:> I get this: > > [root at gluster03-mi glusterfs]# git fetch > git://review.gluster.org/glusterfs refs/changes/28/9328/4 && git > checkout FETCH_HEAD > fatal: Couldn't find remote ref refs/changes/28/9328/4 > > What's wrong with that?Is your current branch at 3.6 ?> > A. > > -----Original Message----- > From: Atin Mukherjee [mailto:amukherj at redhat.com] > Sent: venerd? 20 febbraio 2015 12:54 > To: RASTELLI Alessandro > Cc: gluster-users at gluster.org > Subject: Re: [Gluster-users] GlusterD uses 50% of RAM > >>From the cmd log history I could see lots of volume status commands were triggered parallely. This is a known issue for 3.6 and it would cause a memory leak. http://review.gluster.org/#/c/9328/ should solve it. > > ~Atin > > On 02/20/2015 04:36 PM, RASTELLI Alessandro wrote: >> 10MB log >> sorry :) >> >> -----Original Message----- >> From: Atin Mukherjee [mailto:amukherj at redhat.com] >> Sent: venerd? 20 febbraio 2015 10:49 >> To: RASTELLI Alessandro; gluster-users at gluster.org >> Subject: Re: [Gluster-users] GlusterD uses 50% of RAM >> >> Could you please share the cmd_history.log & glusterd log file to analyze this high memory usage. >> >> ~Atin >> >> On 02/20/2015 03:10 PM, RASTELLI Alessandro wrote: >>> Hi, >>> I've noticed that one of our 6 gluster 3.6.2 nodes has "glusterd" >>> process using 50% of RAM, on the other nodes usage is about 5% This can be a bug? >>> Should I restart glusterd daemon? >>> Thank you >>> A >>> >>> From: Volnei Puttini [mailto:volnei at vcplinux.com.br] >>> Sent: luned? 9 febbraio 2015 18:06 >>> To: RASTELLI Alessandro; gluster-users at gluster.org >>> Subject: Re: [Gluster-users] cannot access to CIFS export >>> >>> Hi Alessandro, >>> >>> My system: >>> >>> CentOS 7 >>> >>> samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64 >>> samba-winbind-4.1.1-37.el7_0.x86_64 >>> samba-libs-4.1.1-37.el7_0.x86_64 >>> samba-common-4.1.1-37.el7_0.x86_64 >>> samba-winbind-modules-4.1.1-37.el7_0.x86_64 >>> samba-winbind-clients-4.1.1-37.el7_0.x86_64 >>> samba-4.1.1-37.el7_0.x86_64 >>> samba-client-4.1.1-37.el7_0.x86_64 >>> >>> glusterfs 3.6.2 built on Jan 22 2015 12:59:57 >>> >>> Try this, work fine for me: >>> >>> [GFSVOL] >>> browseable = No >>> comment = Gluster share of volume gfsvol >>> path = / >>> read only = No >>> guest ok = Yes >>> kernel share modes = No >>> posix locking = No >>> vfs objects = glusterfs >>> glusterfs:loglevel = 7 >>> glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log >>> glusterfs:volume = vgtest >>> glusterfs:volfile_server = 192.168.2.21 >>> >>> On 09-02-2015 14:45, RASTELLI Alessandro wrote: >>> Hi, >>> I've created and started a new replica volume "downloadstat" with CIFS export enabled on GlusterFS 3.6.2. >>> I can see the following piece has been added automatically to smb.conf: >>> [gluster-downloadstat] >>> comment = For samba share of volume downloadstat vfs objects = >>> glusterfs glusterfs:volume = downloadstat glusterfs:logfile = >>> /var/log/samba/glusterfs-downloadstat.%M.log >>> glusterfs:loglevel = 7 >>> path = / >>> read only = no >>> guest ok = yes >>> >>> I restarted smb service, without errors. >>> When I try to access from Win7 client to "\\gluster01-mi\gluster-downloadstat<file:///\\gluster01-mi\gluster-downloadstat>" it asks me a login (which user do I need to put?) and then gives me error "The network path was not found" >>> and on Gluster smb.log I see: >>> [2015/02/09 17:21:13.111639, 0] smbd/vfs.c:173(vfs_init_custom) >>> error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL >>> [2015/02/09 17:21:13.111709, 0] smbd/vfs.c:315(smbd_vfs_init) >>> smbd_vfs_init: vfs_init_custom failed for glusterfs >>> [2015/02/09 17:21:13.111741, 0] smbd/service.c:902(make_connection_snum) >>> vfs_init failed for service gluster-downloadstat >>> >>> Can you explain how to fix? >>> Thanks >>> >>> Alessandro >>> >>> From: >>> gluster-users-bounces at gluster.org<mailto:gluster-users-bounces at glust >>> e r .org> [mailto:gluster-users-bounces at gluster.org] On Behalf Of >>> David F. >>> Robinson >>> Sent: domenica 8 febbraio 2015 18:19 >>> To: Gluster Devel; >>> gluster-users at gluster.org<mailto:gluster-users at gluster.org> >>> Subject: [Gluster-users] cannot delete non-empty directory >>> >>> I am seeing these messsages after I delete large amounts of data using gluster 3.6.2. >>> cannot delete non-empty directory: >>> old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_fi >>> n >>> a >>> l >>> >>> >From the FUSE mount (as root), the directory shows up as empty: >>> >>> # pwd >>> /backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Stora >>> g >>> e >>> /Jimmy_Old/src_vj1.5_final >>> >>> # ls -al >>> total 5 >>> d--------- 2 root root 4106 Feb 6 13:55 . >>> drwxrws--- 3 601 dmiller 72 Feb 6 13:55 .. >>> >>> However, when you look at the bricks, the files are still there (none on brick01bkp, all files are on brick02bkp). All of the files are 0-length and have ------T permissions. >>> Any suggestions on how to fix this and how to prevent it from happening? >>> >>> # ls -al >>> /data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/Ra >>> v e nCFD/Storage/Jimmy_Old/src_vj1.5_final >>> /data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: >>> total 4 >>> d---------+ 2 root root 10 Feb 6 13:55 . >>> drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. >>> >>> /data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: >>> total 8 >>> d---------+ 3 root root 4096 Dec 31 1969 . >>> drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. >>> ---------T 5 601 raven 0 Nov 20 00:08 read_inset.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 readbc.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 readcn.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 readinp.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_2.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_3.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 rotatept.f.gz >>> d---------+ 2 root root 118 Feb 6 13:54 save1 >>> ---------T 5 601 raven 0 Nov 20 00:08 sepvec.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 shadow.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 snksrc.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 source.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 step.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 stoprog.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 summer6.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 totforc.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 tritet.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 wallrsd.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 wheat.f.gz >>> ---------T 5 601 raven 0 Nov 20 00:08 write_inset.f.gz >>> >>> >>> This is using gluster 3.6.2 on a distributed gluster volume that resides on a single machine. Both of the bricks are on one machine consisting of 2x RAID-6 arrays. >>> >>> df -h | grep brick >>> /dev/mapper/vg01-lvol1 88T 22T 66T 25% /data/brick01bkp >>> /dev/mapper/vg02-lvol1 88T 22T 66T 26% /data/brick02bkp >>> >>> # gluster volume info homegfs_bkp >>> Volume Name: homegfs_bkp >>> Type: Distribute >>> Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294 >>> Status: Started >>> Number of Bricks: 2 >>> Transport-type: tcp >>> Bricks: >>> Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp >>> Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp >>> Options Reconfigured: >>> storage.owner-gid: 100 >>> performance.io-thread-count: 32 >>> server.allow-insecure: on >>> network.ping-timeout: 10 >>> performance.cache-size: 128MB >>> performance.write-behind-window-size: 128MB >>> server.manage-gids: on >>> changelog.rollover-time: 15 >>> changelog.fsync-interval: 3 >>> >>> >>> >>> ==============================>>> David F. Robinson, Ph.D. >>> President - Corvid Technologies >>> 704.799.6944 x101 [office] >>> 704.252.1310 [cell] >>> 704.799.7974 [fax] >>> David.Robinson at corvidtec.com<mailto:David.Robinson at corvidtec.com> >>> http://www.corvidtechnologies.com >>> >>> >>> >>> >>> >>> >>> _______________________________________________ >>> >>> Gluster-users mailing list >>> >>> Gluster-users at gluster.org<mailto:Gluster-users at gluster.org> >>> >>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >>> >>> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >> >> -- >> ~Atin >> > > -- > ~Atin >-- ~Atin