How to disable io-cache routine ? I will try it and report back :) thanks Matus 2011/6/8 Mohit Anchlia <mohitanchlia at gmail.com>: > On Wed, Jun 8, 2011 at 12:29 PM, Justice London <jlondon at lawinfo.com> wrote: >> Hopefully this will help some people... try disabling the io-cache routine >> in the fuse configurations for your share. Let me know if you need >> instruction on doing this. It solved all of the lockup issues I was >> experiencing. I believe there is some sort of as-yet-undetermined memory >> leak here. > > Was there a bug filed? If you think this is a bug it will help others as well. >> >> Justice London >> >> -----Original Message----- >> From: gluster-users-bounces at gluster.org >> [mailto:gluster-users-bounces at gluster.org] On Behalf Of bxmatus at gmail.com >> Sent: Wednesday, June 08, 2011 12:22 PM >> To: gluster-users at gluster.org >> Subject: Re: uninterruptible processes writing >> toglusterfsshare >> >> Im using kernel 2.6.34 + fuse 2.5.5 + gluster 3.2 from beginning and >> it happen again today ... >> php-fpm freeze and reboot was only solution. >> >> Matus >> >> >> 2011/6/7 Markus Fr?hlich <markus.froehlich at xidras.com>: >>> hi! >>> >>> there ist no relavant output from dmesg. >>> no entries in the server log - only the one line in the client-server log, >> I >>> already posted. >>> >>> the glusterfs version on the server had been updated to gfs 3.2.0 more >> than >>> a month ago. >>> because of the troubles on the backup server, I deleted the whole backup >>> share and started from scratch. >>> >>> >>> I looked for a update of "fuse" and upgraded from 2.7.2-61.18.1 to >>> 2.8.5-41.1 >>> maybe this helps. >>> >>> here is the changelog info: >>> >>> Authors: >>> -------- >>> ? ?Miklos Szeredi <miklos at szeredi.hu> >>> Distribution: systemsmanagement:baracus / SLE_11_SP1 >>> * Tue Mar 29 2011 dbahi at novell.com >>> - remove the --no-canonicalize usage for suse_version <= 11.3 >>> >>> * Mon Mar 21 2011 coolo at novell.com >>> - licenses package is about to die >>> >>> * Thu Feb 17 2011 mszeredi at suse.cz >>> - In case of failure to add to /etc/mtab don't umount. [bnc#668820] >>> ?[CVE-2011-0541] >>> >>> * Tue Nov 16 2010 mszeredi at suse.cz >>> - Fix symlink attack for mount and umount [bnc#651598] >>> >>> * Wed Oct 27 2010 mszeredi at suse.cz >>> - Remove /etc/init.d/boot.fuse [bnc#648843] >>> >>> * Tue Sep 28 2010 mszeredi at suse.cz >>> - update to 2.8.5 >>> ?* fix option escaping for fusermount [bnc#641480] >>> >>> * Wed Apr 28 2010 mszeredi at suse.cz >>> - keep examples and internal docs in devel package (from jnweiger) >>> >>> * Mon Apr 26 2010 mszeredi at suse.cz >>> - update to 2.8.4 >>> ?* fix checking for symlinks in umount from /tmp >>> ?* fix umounting if /tmp is a symlink >>> >>> >>> kind regards >>> markus froehlich >>> >>> Am 06.06.2011 21:19, schrieb Anthony J. Biacco: >>>> >>>> Could be fuse, check 'dmesg' for kernel module timeouts. >>>> >>>> In a similar vein, has anyone seen signifigant performance/reliability >>>> with diff fuse versions? say, latest source vs. Rhel distro rpms vers. >>>> >>>> -Tony >>>> >>>> >>>> >>>> -----Original Message----- >>>> From: Mohit Anchlia<mohitanchlia at gmail.com> >>>> Sent: June 06, 2011 1:14 PM >>>> To: Markus Fr?hlich<markus.froehlich at xidras.com> >>>> Cc: gluster-users at gluster.org<gluster-users at gluster.org> >>>> Subject: Re: uninterruptible processes writing to >>>> glusterfsshare >>>> >>>> Is there anything in the server logs? Does it follow any particular >>>> pattern before going in this mode? >>>> >>>> Did you upgrade Gluster or is this new install? >>>> >>>> 2011/6/6 Markus Fr?hlich<markus.froehlich at xidras.com>: >>>>> >>>>> hi! >>>>> >>>>> sometimes we've on some client-servers hanging uninterruptible processes >>>>> ("ps aux" stat is on "D" ) and on one the CPU wait I/O grows within some >>>>> minutes to 100%. >>>>> you are not able to kill such processes - also "kill -9" doesnt work - >>>>> when >>>>> you connect via "strace" to such an process, you wont see anything and >>>>> you >>>>> cannot detach it again. >>>>> >>>>> there are only two possibilities: >>>>> killing the glusterfs process (umount GFS share) or rebooting the >> server. >>>>> >>>>> the only log entry I found, was on one client - just a single line: >>>>> [2011-06-06 10:44:18.593211] I >>>>> [afr-common.c:581:afr_lookup_collect_xattr] >>>>> 0-office-data-replicate-0: data self-heal is pending for >>>>> >>>>> >> /pc-partnerbet-public/Promotionaktionen/Mailakquise_2009/Webmaster_2010/HTML >> /bilder/Thumbs.db. >>>>> >>>>> one of the client-servers is a samba-server, the other one a >>>>> backup-server >>>>> based on rsync with millions of small files. >>>>> >>>>> gfs-servers + gfs-clients: SLES11 x86_64, glusterfs V 3.2.0 >>>>> >>>>> and here are the configs from server and client: >>>>> server config >>>>> >>>>> >> "/etc/glusterd/vols/office-data/office-data.gfs-01-01.GFS-office-data02.vol" >> : >>>>> volume office-data-posix >>>>> ? ?type storage/posix >>>>> ? ?option directory /GFS/office-data02 >>>>> end-volume >>>>> >>>>> volume office-data-access-control >>>>> ? ?type features/access-control >>>>> ? ?subvolumes office-data-posix >>>>> end-volume >>>>> >>>>> volume office-data-locks >>>>> ? ?type features/locks >>>>> ? ?subvolumes office-data-access-control >>>>> end-volume >>>>> >>>>> volume office-data-io-threads >>>>> ? ?type performance/io-threads >>>>> ? ?subvolumes office-data-locks >>>>> end-volume >>>>> >>>>> volume office-data-marker >>>>> ? ?type features/marker >>>>> ? ?option volume-uuid 3c6e633d-a0bb-4c52-8f05-a2db9bc9c659 >>>>> ? ?option timestamp-file /etc/glusterd/vols/office-data/marker.tstamp >>>>> ? ?option xtime off >>>>> ? ?option quota off >>>>> ? ?subvolumes office-data-io-threads >>>>> end-volume >>>>> >>>>> volume /GFS/office-data02 >>>>> ? ?type debug/io-stats >>>>> ? ?option latency-measurement off >>>>> ? ?option count-fop-hits off >>>>> ? ?subvolumes office-data-marker >>>>> end-volume >>>>> >>>>> volume office-data-server >>>>> ? ?type protocol/server >>>>> ? ?option transport-type tcp >>>>> ? ?option auth.addr./GFS/office-data02.allow * >>>>> ? ?subvolumes /GFS/office-data02 >>>>> end-volume >>>>> >>>>> >>>>> -------------- >>>>> client config "/etc/glusterd/vols/office-data/office-data-fuse.vol": >>>>> volume office-data-client-0 >>>>> ? ?type protocol/client >>>>> ? ?option remote-host gfs-01-01 >>>>> ? ?option remote-subvolume /GFS/office-data02 >>>>> ? ?option transport-type tcp >>>>> end-volume >>>>> >>>>> volume office-data-replicate-0 >>>>> ? ?type cluster/replicate >>>>> ? ?subvolumes office-data-client-0 >>>>> end-volume >>>>> >>>>> volume office-data-write-behind >>>>> ? ?type performance/write-behind >>>>> ? ?subvolumes office-data-replicate-0 >>>>> end-volume >>>>> >>>>> volume office-data-read-ahead >>>>> ? ?type performance/read-ahead >>>>> ? ?subvolumes office-data-write-behind >>>>> end-volume >>>>> >>>>> volume office-data-io-cache >>>>> ? ?type performance/io-cache >>>>> ? ?subvolumes office-data-read-ahead >>>>> end-volume >>>>> >>>>> volume office-data-quick-read >>>>> ? ?type performance/quick-read >>>>> ? ?subvolumes office-data-io-cache >>>>> end-volume >>>>> >>>>> volume office-data-stat-prefetch >>>>> ? ?type performance/stat-prefetch >>>>> ? ?subvolumes office-data-quick-read >>>>> end-volume >>>>> >>>>> volume office-data >>>>> ? ?type debug/io-stats >>>>> ? ?option latency-measurement off >>>>> ? ?option count-fop-hits off >>>>> ? ?subvolumes office-data-stat-prefetch >>>>> end-volume >>>>> >>>>> >>>>> ?-- Mit freundlichen Gr?ssen >>>>> >>>>> Markus Fr?hlich >>>>> Techniker >>>>> >>>>> _______________________________________________ >>>>> Gluster-users mailing list >>>>> Gluster-users at gluster.org >>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>> >>> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >> >