Could you please share the cmd_history.log & glusterd log file to
analyze this high memory usage.
~Atin
On 02/20/2015 03:10 PM, RASTELLI Alessandro wrote:> Hi,
> I've noticed that one of our 6 gluster 3.6.2 nodes has
"glusterd" process using 50% of RAM, on the other nodes usage is about
5%
> This can be a bug?
> Should I restart glusterd daemon?
> Thank you
> A
>
> From: Volnei Puttini [mailto:volnei at vcplinux.com.br]
> Sent: luned? 9 febbraio 2015 18:06
> To: RASTELLI Alessandro; gluster-users at gluster.org
> Subject: Re: [Gluster-users] cannot access to CIFS export
>
> Hi Alessandro,
>
> My system:
>
> CentOS 7
>
> samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64
> samba-winbind-4.1.1-37.el7_0.x86_64
> samba-libs-4.1.1-37.el7_0.x86_64
> samba-common-4.1.1-37.el7_0.x86_64
> samba-winbind-modules-4.1.1-37.el7_0.x86_64
> samba-winbind-clients-4.1.1-37.el7_0.x86_64
> samba-4.1.1-37.el7_0.x86_64
> samba-client-4.1.1-37.el7_0.x86_64
>
> glusterfs 3.6.2 built on Jan 22 2015 12:59:57
>
> Try this, work fine for me:
>
> [GFSVOL]
> browseable = No
> comment = Gluster share of volume gfsvol
> path = /
> read only = No
> guest ok = Yes
> kernel share modes = No
> posix locking = No
> vfs objects = glusterfs
> glusterfs:loglevel = 7
> glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log
> glusterfs:volume = vgtest
> glusterfs:volfile_server = 192.168.2.21
>
> On 09-02-2015 14:45, RASTELLI Alessandro wrote:
> Hi,
> I've created and started a new replica volume "downloadstat"
with CIFS export enabled on GlusterFS 3.6.2.
> I can see the following piece has been added automatically to smb.conf:
> [gluster-downloadstat]
> comment = For samba share of volume downloadstat
> vfs objects = glusterfs
> glusterfs:volume = downloadstat
> glusterfs:logfile = /var/log/samba/glusterfs-downloadstat.%M.log
> glusterfs:loglevel = 7
> path = /
> read only = no
> guest ok = yes
>
> I restarted smb service, without errors.
> When I try to access from Win7 client to
"\\gluster01-mi\gluster-downloadstat<file:///\\gluster01-mi\gluster-downloadstat>"
it asks me a login (which user do I need to put?) and then gives me error
"The network path was not found"
> and on Gluster smb.log I see:
> [2015/02/09 17:21:13.111639, 0] smbd/vfs.c:173(vfs_init_custom)
> error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL
> [2015/02/09 17:21:13.111709, 0] smbd/vfs.c:315(smbd_vfs_init)
> smbd_vfs_init: vfs_init_custom failed for glusterfs
> [2015/02/09 17:21:13.111741, 0] smbd/service.c:902(make_connection_snum)
> vfs_init failed for service gluster-downloadstat
>
> Can you explain how to fix?
> Thanks
>
> Alessandro
>
> From: gluster-users-bounces at gluster.org<mailto:gluster-users-bounces
at gluster.org> [mailto:gluster-users-bounces at gluster.org] On Behalf Of
David F. Robinson
> Sent: domenica 8 febbraio 2015 18:19
> To: Gluster Devel; gluster-users at gluster.org<mailto:gluster-users at
gluster.org>
> Subject: [Gluster-users] cannot delete non-empty directory
>
> I am seeing these messsages after I delete large amounts of data using
gluster 3.6.2.
> cannot delete non-empty directory:
old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final
>
>>From the FUSE mount (as root), the directory shows up as empty:
>
> # pwd
>
/backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final
>
> # ls -al
> total 5
> d--------- 2 root root 4106 Feb 6 13:55 .
> drwxrws--- 3 601 dmiller 72 Feb 6 13:55 ..
>
> However, when you look at the bricks, the files are still there (none on
brick01bkp, all files are on brick02bkp). All of the files are 0-length and
have ------T permissions.
> Any suggestions on how to fix this and how to prevent it from happening?
>
> # ls -al
/data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final
>
/data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final:
> total 4
> d---------+ 2 root root 10 Feb 6 13:55 .
> drwxrws---+ 3 601 raven 36 Feb 6 13:55 ..
>
>
/data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final:
> total 8
> d---------+ 3 root root 4096 Dec 31 1969 .
> drwxrws---+ 3 601 raven 36 Feb 6 13:55 ..
> ---------T 5 601 raven 0 Nov 20 00:08 read_inset.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 readbc.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 readcn.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 readinp.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_2.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_3.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 rotatept.f.gz
> d---------+ 2 root root 118 Feb 6 13:54 save1
> ---------T 5 601 raven 0 Nov 20 00:08 sepvec.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 shadow.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 snksrc.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 source.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 step.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 stoprog.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 summer6.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 totforc.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 tritet.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 wallrsd.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 wheat.f.gz
> ---------T 5 601 raven 0 Nov 20 00:08 write_inset.f.gz
>
>
> This is using gluster 3.6.2 on a distributed gluster volume that resides on
a single machine. Both of the bricks are on one machine consisting of 2x RAID-6
arrays.
>
> df -h | grep brick
> /dev/mapper/vg01-lvol1 88T 22T 66T 25%
/data/brick01bkp
> /dev/mapper/vg02-lvol1 88T 22T 66T 26%
/data/brick02bkp
>
> # gluster volume info homegfs_bkp
> Volume Name: homegfs_bkp
> Type: Distribute
> Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294
> Status: Started
> Number of Bricks: 2
> Transport-type: tcp
> Bricks:
> Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp
> Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp
> Options Reconfigured:
> storage.owner-gid: 100
> performance.io-thread-count: 32
> server.allow-insecure: on
> network.ping-timeout: 10
> performance.cache-size: 128MB
> performance.write-behind-window-size: 128MB
> server.manage-gids: on
> changelog.rollover-time: 15
> changelog.fsync-interval: 3
>
>
>
> ==============================> David F. Robinson, Ph.D.
> President - Corvid Technologies
> 704.799.6944 x101 [office]
> 704.252.1310 [cell]
> 704.799.7974 [fax]
> David.Robinson at corvidtec.com<mailto:David.Robinson at
corvidtec.com>
> http://www.corvidtechnologies.com
>
>
>
>
>
>
> _______________________________________________
>
> Gluster-users mailing list
>
> Gluster-users at gluster.org<mailto:Gluster-users at gluster.org>
>
> http://www.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
--
~Atin