Re: GlusterD uses 50% of RAM

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



>From the cmd log history I could see lots of volume status commands were
triggered parallely. This is a known issue for 3.6 and it would cause a
memory leak. http://review.gluster.org/#/c/9328/ should solve it.

~Atin

On 02/20/2015 04:36 PM, RASTELLI Alessandro wrote:
> 10MB log
> sorry :)
> 
> -----Original Message-----
> From: Atin Mukherjee [mailto:amukherj@xxxxxxxxxx] 
> Sent: venerdì 20 febbraio 2015 10:49
> To: RASTELLI Alessandro; gluster-users@xxxxxxxxxxx
> Subject: Re:  GlusterD uses 50% of RAM
> 
> Could you please share the cmd_history.log & glusterd log file to analyze this high memory usage.
> 
> ~Atin
> 
> On 02/20/2015 03:10 PM, RASTELLI Alessandro wrote:
>> Hi,
>> I've noticed that one of our 6 gluster 3.6.2 nodes has "glusterd" 
>> process using 50% of RAM, on the other nodes usage is about 5% This can be a bug?
>> Should I restart glusterd daemon?
>> Thank you
>> A
>>
>> From: Volnei Puttini [mailto:volnei@xxxxxxxxxxxxxxx]
>> Sent: lunedì 9 febbraio 2015 18:06
>> To: RASTELLI Alessandro; gluster-users@xxxxxxxxxxx
>> Subject: Re:  cannot access to CIFS export
>>
>> Hi Alessandro,
>>
>> My system:
>>
>> CentOS 7
>>
>> samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64
>> samba-winbind-4.1.1-37.el7_0.x86_64
>> samba-libs-4.1.1-37.el7_0.x86_64
>> samba-common-4.1.1-37.el7_0.x86_64
>> samba-winbind-modules-4.1.1-37.el7_0.x86_64
>> samba-winbind-clients-4.1.1-37.el7_0.x86_64
>> samba-4.1.1-37.el7_0.x86_64
>> samba-client-4.1.1-37.el7_0.x86_64
>>
>> glusterfs 3.6.2 built on Jan 22 2015 12:59:57
>>
>> Try this, work fine for me:
>>
>> [GFSVOL]
>>     browseable = No
>>     comment = Gluster share of volume gfsvol
>>     path = /
>>     read only = No
>>     guest ok = Yes
>>     kernel share modes = No
>>     posix locking = No
>>     vfs objects = glusterfs
>>     glusterfs:loglevel = 7
>>     glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log
>>     glusterfs:volume = vgtest
>>     glusterfs:volfile_server = 192.168.2.21
>>
>> On 09-02-2015 14:45, RASTELLI Alessandro wrote:
>> Hi,
>> I've created and started a new replica volume "downloadstat" with CIFS export enabled on GlusterFS 3.6.2.
>> I can see the following piece has been added automatically to smb.conf:
>> [gluster-downloadstat]
>> comment = For samba share of volume downloadstat vfs objects = 
>> glusterfs glusterfs:volume = downloadstat glusterfs:logfile = 
>> /var/log/samba/glusterfs-downloadstat.%M.log
>> glusterfs:loglevel = 7
>> path = /
>> read only = no
>> guest ok = yes
>>
>> I restarted smb service, without errors.
>> When I try to access from Win7 client to "\\gluster01-mi\gluster-downloadstat<file:///\\gluster01-mi\gluster-downloadstat>" it asks me a login (which user do I need to put?) and then gives me error "The network path was not found"
>> and on Gluster smb.log I see:
>> [2015/02/09 17:21:13.111639,  0] smbd/vfs.c:173(vfs_init_custom)
>>   error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL
>> [2015/02/09 17:21:13.111709,  0] smbd/vfs.c:315(smbd_vfs_init)
>>   smbd_vfs_init: vfs_init_custom failed for glusterfs
>> [2015/02/09 17:21:13.111741,  0] smbd/service.c:902(make_connection_snum)
>>   vfs_init failed for service gluster-downloadstat
>>
>> Can you explain how to fix?
>> Thanks
>>
>> Alessandro
>>
>> From: 
>> gluster-users-bounces@xxxxxxxxxxx<mailto:gluster-users-bounces@gluster
>> .org> [mailto:gluster-users-bounces@xxxxxxxxxxx] On Behalf Of David F. 
>> Robinson
>> Sent: domenica 8 febbraio 2015 18:19
>> To: Gluster Devel; 
>> gluster-users@xxxxxxxxxxx<mailto:gluster-users@xxxxxxxxxxx>
>> Subject:  cannot delete non-empty directory
>>
>> I am seeing these messsages after I delete large amounts of data using gluster 3.6.2.
>> cannot delete non-empty directory: 
>> old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_fina
>> l
>>
>> >From the FUSE mount (as root), the directory shows up as empty:
>>
>> # pwd
>> /backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage
>> /Jimmy_Old/src_vj1.5_final
>>
>> # ls -al
>> total 5
>> d--------- 2 root root    4106 Feb  6 13:55 .
>> drwxrws--- 3  601 dmiller   72 Feb  6 13:55 ..
>>
>> However, when you look at the bricks, the files are still there (none on brick01bkp, all files are on brick02bkp).  All of the files are 0-length and have ------T permissions.
>> Any suggestions on how to fix this and how to prevent it from happening?
>>
>> #  ls -al 
>> /data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/Rave
>> nCFD/Storage/Jimmy_Old/src_vj1.5_final
>> /data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final:
>> total 4
>> d---------+ 2 root root  10 Feb  6 13:55 .
>> drwxrws---+ 3  601 raven 36 Feb  6 13:55 ..
>>
>> /data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final:
>> total 8
>> d---------+ 3 root root  4096 Dec 31  1969 .
>> drwxrws---+ 3  601 raven   36 Feb  6 13:55 ..
>> ---------T  5  601 raven    0 Nov 20 00:08 read_inset.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 readbc.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 readcn.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 readinp.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 readinp_v1_2.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 readinp_v1_3.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 rotatept.f.gz
>> d---------+ 2 root root   118 Feb  6 13:54 save1
>> ---------T  5  601 raven    0 Nov 20 00:08 sepvec.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 shadow.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 snksrc.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 source.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 step.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 stoprog.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 summer6.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 totforc.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 tritet.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 wallrsd.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 wheat.f.gz
>> ---------T  5  601 raven    0 Nov 20 00:08 write_inset.f.gz
>>
>>
>> This is using gluster 3.6.2 on a distributed gluster volume that resides on a single machine.  Both of the bricks are on one machine consisting of 2x RAID-6 arrays.
>>
>> df -h | grep brick
>> /dev/mapper/vg01-lvol1                       88T   22T   66T  25% /data/brick01bkp
>> /dev/mapper/vg02-lvol1                       88T   22T   66T  26% /data/brick02bkp
>>
>> # gluster volume info homegfs_bkp
>> Volume Name: homegfs_bkp
>> Type: Distribute
>> Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294
>> Status: Started
>> Number of Bricks: 2
>> Transport-type: tcp
>> Bricks:
>> Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp
>> Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp
>> Options Reconfigured:
>> storage.owner-gid: 100
>> performance.io-thread-count: 32
>> server.allow-insecure: on
>> network.ping-timeout: 10
>> performance.cache-size: 128MB
>> performance.write-behind-window-size: 128MB
>> server.manage-gids: on
>> changelog.rollover-time: 15
>> changelog.fsync-interval: 3
>>
>>
>>
>> ===============================
>> David F. Robinson, Ph.D.
>> President - Corvid Technologies
>> 704.799.6944 x101 [office]
>> 704.252.1310 [cell]
>> 704.799.7974 [fax]
>> David.Robinson@xxxxxxxxxxxxx<mailto:David.Robinson@xxxxxxxxxxxxx>
>> http://www.corvidtechnologies.com
>>
>>
>>
>>
>>
>>
>> _______________________________________________
>>
>> Gluster-users mailing list
>>
>> Gluster-users@xxxxxxxxxxx<mailto:Gluster-users@xxxxxxxxxxx>
>>
>> http://www.gluster.org/mailman/listinfo/gluster-users
>>
>>
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users@xxxxxxxxxxx
>> http://www.gluster.org/mailman/listinfo/gluster-users
>>
> 
> --
> ~Atin
> 

-- 
~Atin
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users





[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux