Oops! I replied before doing a 'tla update' after getting up :O Thing is, this bug is fixed in archive 'glusterfs--mainline--2.5--patch-264' -Amar > > On Thu, May 8, 2008 at 9:03 AM, Dan Parsons <dparsons@xxxxxxxx> wrote: > >> Amar, any update on this issue? >> >> >> Dan Parsons >> >> >> >> On May 6, 2008, at 11:21 PM, Dan Parsons wrote: >> >> Amar, quick question. I've switched to readahead but really wish I could >>> use io-cache. How likely do you think changing block-size from 128kb to 1MB >>> (the same as what stripe uses, based on your advice) would fix the crash >>> issue? >>> >>> >>> Dan Parsons >>> >>> >>> On May 6, 2008, at 12:43 PM, Amar S. Tumballi wrote: >>> >>> Mostly give the same block-size (and page-size) in stripe and io-cache. >>>> Just >>>> for checking it. But currently you can fall back to read-ahead. >>>> >>>> Regards, >>>> Amar >>>> >>>> On Tue, May 6, 2008 at 12:38 PM, Dan Parsons <dparsons@xxxxxxxx> wrote: >>>> >>>> Ah, so it's not something I'm doing wrong? Do you think changing >>>>> cache-size back to 32MB will prevent the problem from happening? >>>>> >>>>> Perhaps I should switch to readahead until fix? >>>>> >>>>> >>>>> Dan Parsons >>>>> >>>>> >>>>> >>>>> On May 6, 2008, at 12:37 PM, Amar S. Tumballi wrote: >>>>> >>>>> Thanks for the bug report, We will get back to you in another 2-3 days >>>>> >>>>>> about it. Mostly with a fix :) >>>>>> >>>>>> Regards, >>>>>> Amar >>>>>> >>>>>> On Tue, May 6, 2008 at 10:14 AM, Dan Parsons <dparsons@xxxxxxxx> >>>>>> wrote: >>>>>> Oh, one more useful bit of information, I see lines like the below a >>>>>> lot >>>>>> in glusterfs log files, what do they mean? >>>>>> >>>>>> 2008-05-05 21:20:11 W [fuse-bridge.c:402:fuse_entry_cbk] >>>>>> glusterfs-fuse: >>>>>> 18054459: (34) >>>>>> >>>>>> /bio/data/fast-hmmsearch-all/tmpDCex3b_fast-hmmsearch-all_job/result.tigrfam.TIGR02736.hmmhits >>>>>> => 610503040 Rehashing because st_nlink less than dentry maps >>>>>> >>>>>> Dan Parsons >>>>>> >>>>>> >>>>>> >>>>>> On May 6, 2008, at 10:13 AM, Dan Parsons wrote: >>>>>> >>>>>> I'm experiencing a glusterfs client crash, signal 11, under the >>>>>> io-cache >>>>>> xlator. This is on our bioinformatics cluster- the crash happened on 2 >>>>>> out >>>>>> of 33 machines. I've verified the hardware stability of the machines. >>>>>> >>>>>> Running v1.3.8 built May 5th, 2008 from latest downloadable version. >>>>>> >>>>>> Here is the crash message: >>>>>> >>>>>> [0xffffe420] >>>>>> >>>>>> >>>>>> /usr/local/lib/glusterfs/1.3.8/xlator/performance/io-cache.so(ioc_page_wakeup+0x67)[0xb76c5f67] >>>>>> >>>>>> >>>>>> /usr/local/lib/glusterfs/1.3.8/xlator/performance/io-cache.so(ioc_inode_wakeup+0xb2)[0xb76c6902] >>>>>> >>>>>> >>>>>> /usr/local/lib/glusterfs/1.3.8/xlator/performance/io-cache.so(ioc_cache_validate_cbk+0xae)[0xb76c1e5e] >>>>>> >>>>>> >>>>>> /usr/local/lib/glusterfs/1.3.8/xlator/cluster/stripe.so(stripe_stack_unwind_buf_cbk+0x98)[0xb76cd038] >>>>>> >>>>>> >>>>>> /usr/local/lib/glusterfs/1.3.8/xlator/protocol/client.so(client_fstat_cbk+0xcc)[0xb76dd13c] >>>>>> >>>>>> >>>>>> /usr/local/lib/glusterfs/1.3.8/xlator/protocol/client.so(notify+0xa97)[0xb76db117] >>>>>> /usr/local/lib/libglusterfs.so.0(transport_notify+0x38)[0xb7efe978] >>>>>> /usr/local/lib/libglusterfs.so.0(sys_epoll_iteration+0xd6)[0xb7eff906] >>>>>> /usr/local/lib/libglusterfs.so.0(poll_iteration+0x98)[0xb7efeb28] >>>>>> [glusterfs](main+0x85e)[0x804a14e] >>>>>> /lib/libc.so.6(__libc_start_main+0xdc)[0x7b1dec] >>>>>> [glusterfs][0x8049391] >>>>>> >>>>>> And here is my config file. The only thing I can think of is maybe my >>>>>> cache-size is too big. I want a lot of cache, we have big files, and >>>>>> the >>>>>> boxes have the RAM. Anyway, below is the config. If you see any >>>>>> problems >>>>>> with it, please let me know. There are no errors on the glusterfsd >>>>>> servers, >>>>>> except for an EOF from the machines where glusterfs client segfaulted. >>>>>> >>>>>> volume fuse >>>>>> type mount/fuse >>>>>> option direct-io-mode 1 >>>>>> option entry-timeout 1 >>>>>> option attr-timeout 1 >>>>>> option mount-point /glusterfs >>>>>> subvolumes ioc >>>>>> end-volume >>>>>> >>>>>> volume ioc >>>>>> type performance/io-cache >>>>>> option priority *.psiblast:3,*.seq:2,*:1 >>>>>> option force-revalidate-timeout 5 >>>>>> option cache-size 1200MB >>>>>> option page-size 128KB >>>>>> subvolumes stripe0 >>>>>> end-volume >>>>>> >>>>>> volume stripe0 >>>>>> type cluster/stripe >>>>>> option alu.disk-usage.exit-threshold 100MB >>>>>> option alu.disk-usage.entry-threshold 2GB >>>>>> option alu.write-usage.exit-threshold 4% >>>>>> option alu.write-usage.entry-threshold 20% >>>>>> option alu.read-usage.exit-threshold 4% >>>>>> option alu.read-usage.entry-threshold 20% >>>>>> option alu.order read-usage:write-usage:disk-usage >>>>>> option scheduler alu >>>>>> option block-size *:1MB >>>>>> subvolumes distfs01 distfs02 distfs03 distfs04 >>>>>> end-volume >>>>>> >>>>>> volume distfs04 >>>>>> type protocol/client >>>>>> option remote-subvolume brick >>>>>> option remote-host 10.8.101.54 >>>>>> option transport-type tcp/client >>>>>> end-volume >>>>>> >>>>>> volume distfs03 >>>>>> type protocol/client >>>>>> option remote-subvolume brick >>>>>> option remote-host 10.8.101.53 >>>>>> option transport-type tcp/client >>>>>> end-volume >>>>>> >>>>>> volume distfs02 >>>>>> type protocol/client >>>>>> option remote-subvolume brick >>>>>> option remote-host 10.8.101.52 >>>>>> option transport-type tcp/client >>>>>> end-volume >>>>>> >>>>>> volume distfs01 >>>>>> type protocol/client >>>>>> option remote-subvolume brick >>>>>> option remote-host 10.8.101.51 >>>>>> option transport-type tcp/client >>>>>> end-volume >>>>>> >>>>>> >>>>>> Dan Parsons >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Gluster-devel mailing list >>>>>> Gluster-devel@xxxxxxxxxx >>>>>> http://lists.nongnu.org/mailman/listinfo/gluster-devel >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Gluster-devel mailing list >>>>>> Gluster-devel@xxxxxxxxxx >>>>>> http://lists.nongnu.org/mailman/listinfo/gluster-devel >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Amar Tumballi >>>>>> Gluster/GlusterFS Hacker >>>>>> [bulde on #gluster/irc.gnu.org] >>>>>> http://www.zresearch.com - Commoditizing Super Storage! >>>>>> >>>>>> >>>>> >>>>> >>>>> >>>> >>>> -- >>>> Amar Tumballi >>>> Gluster/GlusterFS Hacker >>>> [bulde on #gluster/irc.gnu.org] >>>> http://www.zresearch.com - Commoditizing Super Storage! >>>> >>> >>> >>> >>> _______________________________________________ >>> Gluster-devel mailing list >>> Gluster-devel@xxxxxxxxxx >>> http://lists.nongnu.org/mailman/listinfo/gluster-devel >>> >>> >> >> > > > -- > Amar Tumballi > Gluster/GlusterFS Hacker > [bulde on #gluster/irc.gnu.org] > http://www.zresearch.com - Commoditizing Super Storage! > -- Amar Tumballi Gluster/GlusterFS Hacker [bulde on #gluster/irc.gnu.org] http://www.zresearch.com - Commoditizing Super Storage!