Ananth,
Thank you for your
kindness.
I will try to use them as you
advise. For past days, I have research the ReiserFS, about which some papers
said it is no limitation on the number of the subdirectory, is it
right?
Thanks,
Yaomin
From: Ananth
Sent: Friday, January 09, 2009 2:47 PM
To: yaomin @ gmail
Subject: Re: Re: Gluster-devel Digest, Vol 41, Issue
13 >From what I remember both XFS and the rather nascent EXT4 have a higher subdirectory limit. You could probably check ZFS as well. You can convert your ext3 filesystem into ext4 : http://ext4.wiki.kernel.org/index.php/Ext4_Howto#Converting_an_ext3_filesystem_to_ext4 These are just pointers as to what you can do, hence please check if the change in filesystem is feasible / appropriate for you, and ensure your data is backed up before proceeding. Regards, Ananth Z Research -----Original Message----- From: yaomin @ gmail <yangyaomin@xxxxxxxxx> To: gluster-devel@xxxxxxxxxx Subject: Re: Gluster-devel Digest, Vol 41, Issue 13 Date: Wed, 7 Jan 2009 22:54:51 +0800 All, When I use the ext3 as the filesystem on the server, I have a new trouble that one directory at most have 31998 subdirectories. Do you have any advice for me? Thanks, Yaomin -------------------------------------------------- From: <gluster-devel-request@xxxxxxxxxx> Sent: Tuesday, January 06, 2009 8:21 PM To: <gluster-devel@xxxxxxxxxx> Subject: Gluster-devel Digest, Vol 41, Issue 13 > Send Gluster-devel mailing list submissions to > gluster-devel@xxxxxxxxxx > > To subscribe or unsubscribe via the World Wide Web, visit > http://lists.nongnu.org/mailman/listinfo/gluster-devel > or, via email, send a message with subject or body 'help' to > gluster-devel-request@xxxxxxxxxx > > You can reach the person managing the list at > gluster-devel-owner@xxxxxxxxxx > > When replying, please edit your Subject line so it is more specific > than "Re: Contents of Gluster-devel digest..." > > > Today's Topics: > > 1. Re: Cascading different translator doesn't work as > expectation (yaomin @ gmail) > 2. Re: Cascading different translator doesn't work as > expectation (Krishna Srinivas) > 3. Re: Cascading different translator doesn't work as > expectation (yaomin @ gmail) > > > ---------------------------------------------------------------------- > > Message: 1 > Date: Tue, 6 Jan 2009 17:13:49 +0800 > From: "yaomin @ gmail" <yangyaomin@xxxxxxxxx> > Subject: Re: Cascading different translator doesn't > work as expectation > To: "Krishna Srinivas" <krishna@xxxxxxxxxxxxx> > Cc: gluster-devel@xxxxxxxxxx > Message-ID: <D7CA065B4BF644348A6DE543D8213029@yangyaomin> > Content-Type: text/plain; charset="iso-8859-1" > > Krishna, > > Thank you for your kind help before. > > According to your advice, I confront a new error. The storage node has > no log information, and the client's log is like following: > > /lib64/libc.so.6[0x3fbb2300a0] > /usr/local/lib/glusterfs/1.3.9/xlator/cluster/afr.so(afr_setxattr+0x6a)[0x2aaaaaf0658a] > /usr/local/lib/glusterfs/1.3.9/xlator/cluster/stripe.so(notify+0x220)[0x2aaaab115c80] > /usr/local/lib/libglusterfs.so.0(default_notify+0x25)[0x2aaaaaab8f55] > /usr/local/lib/glusterfs/1.3.9/xlator/cluster/afr.so(notify+0x16d)[0x2aaaaaefc19d] > /usr/local/lib/glusterfs/1.3.9/xlator/protocol/client.so(notify+0x681)[0x2aaaaacebac1] > /usr/local/lib/libglusterfs.so.0(sys_epoll_iteration+0xbb)[0x2aaaaaabe14b] > /usr/local/lib/libglusterfs.so.0(poll_iteration+0x79)[0x2aaaaaabd509] > [glusterfs](main+0x66a)[0x4026aa] > /lib64/libc.so.6(__libc_start_main+0xf4)[0x3fbb21d8a4] > [glusterfs][0x401b69] > --------- > > > [root@IP6 ~]# df -h > Filesystem Size Used Avail Use% Mounted on > /dev/sda2 9.5G 6.8G 2.2G 76% / > /dev/sda1 190M 12M 169M 7% /boot > tmpfs 1006M 0 1006M 0% /dev/shm > /dev/sda4 447G 2.8G 422G 1% /locfs > /dev/sdb1 459G 199M 435G 1% /locfsb > df: `/mnt/new': Transport endpoint is not connected > > Thanks, > Yaomin > > -------------------------------------------------- > From: "Krishna Srinivas" <krishna@xxxxxxxxxxxxx> > Sent: Tuesday, January 06, 2009 1:09 PM > To: "yaomin @ gmail" <yangyaomin@xxxxxxxxx> > Cc: <gluster-devel@xxxxxxxxxx> > Subject: Re: Cascading different translator doesn't work > as expectation > >> Alfred, >> Your vol files are wrong. you need to remove all the volume >> definitions below "writeback" in the client vol file. For server vol >> file the definition of performance translators is not having any >> effect. Also you need to use "features/locks" translator above >> "storage/posix" >> Krishna >> >> On Tue, Jan 6, 2009 at 8:51 AM, yaomin @ gmail <yangyaomin@xxxxxxxxx> >> wrote: >>> All, >>> >>> It seems difficult for you. >>> >>> There is a new problem when I tested. >>> >>> When I kill all the storage nodes, the client still try to send >>> data, >>> and doesn't quit. >>> >>> Thanks, >>> Alfred >>> From: yaomin @ gmail >>> Sent: Monday, January 05, 2009 10:52 PM >>> To: Krishna Srinivas >>> Cc: gluster-devel@xxxxxxxxxx >>> Subject: Re: Cascading different translator doesn't work >>> as >>> expectation >>> Krishna, >>> Thank you for your quick response. >>> There are two log information in the client's log file when setting >>> up >>> the client. >>> 2009-01-05 18:44:59 W [fuse-bridge.c:389:fuse_entry_cbk] >>> glusterfs-fuse: >>> 2: (34) / => 1 Rehashing 0/0 >>> 2009-01-05 18:48:04 W [fuse-bridge.c:389:fuse_entry_cbk] >>> glusterfs-fuse: >>> 2: (34) / => 1 Rehashing 0/0 >>> >>> There is no any information in the storage node's log file. >>> >>> Although I changed the scheduler from ALU to RR, there only the >>> No.3(192.168.13.5) and No.4(192.168.13.7) storage nodes on working. >>> >>> Each machine has 2GB memory. >>> >>> Thanks, >>> Alfred >>> > -------------- next part -------------- > An HTML attachment was scrubbed... > URL: > http://lists.gnu.org/pipermail/gluster-devel/attachments/20090106/76074e85/attachment.html > > ------------------------------ > > Message: 2 > Date: Tue, 6 Jan 2009 15:06:42 +0530 > From: "Krishna Srinivas" <krishna@xxxxxxxxxxxxx> > Subject: Re: Cascading different translator doesn't > work as expectation > To: "yaomin @ gmail" <yangyaomin@xxxxxxxxx> > Cc: gluster-devel@xxxxxxxxxx > Message-ID: > <ad4bc5820901060136k2a3c0943nd89b0d4f41240e22@xxxxxxxxxxxxxx> > Content-Type: text/plain; charset=ISO-8859-1 > > Yaomin, > > Can you: > * mention what version you are using > * give the modified client and server vol file (to see if there are any > errors) > * give gdb backtrace from the core file? "gdb -c /core.pid glusterfs" > and then type "bt" > > Krishna > > On Tue, Jan 6, 2009 at 2:43 PM, yaomin @ gmail <yangyaomin@xxxxxxxxx> > wrote: >> Krishna, >> >> Thank you for your kind help before. >> >> According to your advice, I confront a new error. The storage node >> has >> no log information, and the client's log is like following: >> >> /lib64/libc.so.6[0x3fbb2300a0] >> /usr/local/lib/glusterfs/1.3.9/xlator/cluster/afr.so(afr_setxattr+0x6a)[0x2aaaaaf0658a] >> /usr/local/lib/glusterfs/1.3.9/xlator/cluster/stripe.so(notify+0x220)[0x2aaaab115c80] >> /usr/local/lib/libglusterfs.so.0(default_notify+0x25)[0x2aaaaaab8f55] >> /usr/local/lib/glusterfs/1.3.9/xlator/cluster/afr.so(notify+0x16d)[0x2aaaaaefc19d] >> /usr/local/lib/glusterfs/1.3.9/xlator/protocol/client.so(notify+0x681)[0x2aaaaacebac1] >> /usr/local/lib/libglusterfs.so.0(sys_epoll_iteration+0xbb)[0x2aaaaaabe14b] >> /usr/local/lib/libglusterfs.so.0(poll_iteration+0x79)[0x2aaaaaabd509] >> [glusterfs](main+0x66a)[0x4026aa] >> /lib64/libc.so.6(__libc_start_main+0xf4)[0x3fbb21d8a4] >> [glusterfs][0x401b69] >> --------- >> >> [root@IP6 ~]# df -h >> Filesystem Size Used Avail Use% Mounted on >> /dev/sda2 9.5G 6.8G 2.2G 76% / >> /dev/sda1 190M 12M 169M 7% /boot >> tmpfs 1006M 0 1006M 0% /dev/shm >> /dev/sda4 447G 2.8G 422G 1% /locfs >> /dev/sdb1 459G 199M 435G 1% /locfsb >> df: `/mnt/new': Transport endpoint is not connected >> >> Thanks, >> Yaomin >> -------------------------------------------------- >> From: "Krishna Srinivas" <krishna@xxxxxxxxxxxxx> >> Sent: Tuesday, January 06, 2009 1:09 PM >> To: "yaomin @ gmail" <yangyaomin@xxxxxxxxx> >> Cc: <gluster-devel@xxxxxxxxxx> >> Subject: Re: Cascading different translator doesn't work >> as >> expectation >> >>> Alfred, >>> Your vol files are wrong. you need to remove all the volume >>> definitions below "writeback" in the client vol file. For server vol >>> file the definition of performance translators is not having any >>> effect. Also you need to use "features/locks" translator above >>> "storage/posix" >>> Krishna >>> >>> On Tue, Jan 6, 2009 at 8:51 AM, yaomin @ gmail <yangyaomin@xxxxxxxxx> >>> wrote: >>>> All, >>>> >>>> It seems difficult for you. >>>> >>>> There is a new problem when I tested. >>>> >>>> When I kill all the storage nodes, the client still try to send >>>> data, >>>> and doesn't quit. >>>> >>>> Thanks, >>>> Alfred >>>> From: yaomin @ gmail >>>> Sent: Monday, January 05, 2009 10:52 PM >>>> To: Krishna Srinivas >>>> Cc: gluster-devel@xxxxxxxxxx >>>> Subject: Re: Cascading different translator doesn't >>>> work >>>> as >>>> expectation >>>> Krishna, >>>> Thank you for your quick response. >>>> There are two log information in the client's log file when setting >>>> up >>>> the client. >>>> 2009-01-05 18:44:59 W [fuse-bridge.c:389:fuse_entry_cbk] >>>> glusterfs-fuse: >>>> 2: (34) / => 1 Rehashing 0/0 >>>> 2009-01-05 18:48:04 W [fuse-bridge.c:389:fuse_entry_cbk] >>>> glusterfs-fuse: >>>> 2: (34) / => 1 Rehashing 0/0 >>>> >>>> There is no any information in the storage node's log file. >>>> >>>> Although I changed the scheduler from ALU to RR, there only the >>>> No.3(192.168.13.5) and No.4(192.168.13.7) storage nodes on working. >>>> >>>> Each machine has 2GB memory. >>>> >>>> Thanks, >>>> Alfred >>>> > > > > > ------------------------------ > > Message: 3 > Date: Tue, 6 Jan 2009 20:21:35 +0800 > From: "yaomin @ gmail" <yangyaomin@xxxxxxxxx> > Subject: Re: Cascading different translator doesn't > work as expectation > To: "Krishna Srinivas" <krishna@xxxxxxxxxxxxx> > Cc: gluster-devel@xxxxxxxxxx > Message-ID: <CA759DEFFF2C42AA877946E88BD53E42@yangyaomin> > Content-Type: text/plain; charset="iso-8859-1" > > Krishna, > > 1, The version is 1.3.9 > 2, the client and server vol files are in the attachments. > 3, The result is "No Stack" > > Thanks, > Yaomin > > -------------------------------------------------- > From: "Krishna Srinivas" <krishna@xxxxxxxxxxxxx> > Sent: Tuesday, January 06, 2009 5:36 PM > To: "yaomin @ gmail" <yangyaomin@xxxxxxxxx> > Cc: <gluster-devel@xxxxxxxxxx> > Subject: Re: Cascading different translator doesn't work > as > expectation > >> Yaomin, >> >> Can you: >> * mention what version you are using >> * give the modified client and server vol file (to see if there are any >> errors) >> * give gdb backtrace from the core file? "gdb -c /core.pid glusterfs" >> and then type "bt" >> >> Krishna >> >> On Tue, Jan 6, 2009 at 2:43 PM, yaomin @ gmail <yangyaomin@xxxxxxxxx> >> wrote: >>> Krishna, >>> >>> Thank you for your kind help before. >>> >>> According to your advice, I confront a new error. The storage node >>> has >>> no log information, and the client's log is like following: >>> >>> /lib64/libc.so.6[0x3fbb2300a0] >>> /usr/local/lib/glusterfs/1.3.9/xlator/cluster/afr.so(afr_setxattr+0x6a)[0x2aaaaaf0658a] >>> /usr/local/lib/glusterfs/1.3.9/xlator/cluster/stripe.so(notify+0x220)[0x2aaaab115c80] >>> /usr/local/lib/libglusterfs.so.0(default_notify+0x25)[0x2aaaaaab8f55] >>> /usr/local/lib/glusterfs/1.3.9/xlator/cluster/afr.so(notify+0x16d)[0x2aaaaaefc19d] >>> /usr/local/lib/glusterfs/1.3.9/xlator/protocol/client.so(notify+0x681)[0x2aaaaacebac1] >>> /usr/local/lib/libglusterfs.so.0(sys_epoll_iteration+0xbb)[0x2aaaaaabe14b] >>> /usr/local/lib/libglusterfs.so.0(poll_iteration+0x79)[0x2aaaaaabd509] >>> [glusterfs](main+0x66a)[0x4026aa] >>> /lib64/libc.so.6(__libc_start_main+0xf4)[0x3fbb21d8a4] >>> [glusterfs][0x401b69] >>> --------- >>> >>> [root@IP6 ~]# df -h >>> Filesystem Size Used Avail Use% Mounted on >>> /dev/sda2 9.5G 6.8G 2.2G 76% / >>> /dev/sda1 190M 12M 169M 7% /boot >>> tmpfs 1006M 0 1006M 0% /dev/shm >>> /dev/sda4 447G 2.8G 422G 1% /locfs >>> /dev/sdb1 459G 199M 435G 1% /locfsb >>> df: `/mnt/new': Transport endpoint is not connected >>> >>> Thanks, >>> Yaomin >>> -------------------------------------------------- >>> From: "Krishna Srinivas" <krishna@xxxxxxxxxxxxx> >>> Sent: Tuesday, January 06, 2009 1:09 PM >>> To: "yaomin @ gmail" <yangyaomin@xxxxxxxxx> >>> Cc: <gluster-devel@xxxxxxxxxx> >>> Subject: Re: Cascading different translator doesn't work >>> as >>> expectation >>> >>>> Alfred, >>>> Your vol files are wrong. you need to remove all the volume >>>> definitions below "writeback" in the client vol file. For server vol >>>> file the definition of performance translators is not having any >>>> effect. Also you need to use "features/locks" translator above >>>> "storage/posix" >>>> Krishna >>>> >>>> On Tue, Jan 6, 2009 at 8:51 AM, yaomin @ gmail <yangyaomin@xxxxxxxxx> >>>> wrote: >>>>> All, >>>>> >>>>> It seems difficult for you. >>>>> >>>>> There is a new problem when I tested. >>>>> >>>>> When I kill all the storage nodes, the client still try to send >>>>> data, >>>>> and doesn't quit. >>>>> >>>>> Thanks, >>>>> Alfred >>>>> From: yaomin @ gmail >>>>> Sent: Monday, January 05, 2009 10:52 PM >>>>> To: Krishna Srinivas >>>>> Cc: gluster-devel@xxxxxxxxxx >>>>> Subject: Re: Cascading different translator doesn't >>>>> work >>>>> as >>>>> expectation >>>>> Krishna, >>>>> Thank you for your quick response. >>>>> There are two log information in the client's log file when >>>>> setting >>>>> up >>>>> the client. >>>>> 2009-01-05 18:44:59 W [fuse-bridge.c:389:fuse_entry_cbk] >>>>> glusterfs-fuse: >>>>> 2: (34) / => 1 Rehashing 0/0 >>>>> 2009-01-05 18:48:04 W [fuse-bridge.c:389:fuse_entry_cbk] >>>>> glusterfs-fuse: >>>>> 2: (34) / => 1 Rehashing 0/0 >>>>> >>>>> There is no any information in the storage node's log file. >>>>> >>>>> Although I changed the scheduler from ALU to RR, there only the >>>>> No.3(192.168.13.5) and No.4(192.168.13.7) storage nodes on working. >>>>> >>>>> Each machine has 2GB memory. >>>>> >>>>> Thanks, >>>>> Alfred >>>>> > -------------- next part -------------- > volume client-ns > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 192.168.13.2 # IP address of the remote brick > # option remote-port 6996 # default server port is 6996 > # option transport-timeout 30 # seconds to wait for a response > # from server for each request > option remote-subvolume name_space # name of the remote volume > end-volume > > volume client11 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 192.168.13.2 # IP address of the remote brick > # option remote-port 6996 # default server port is 6996 > # option transport-timeout 30 # seconds to wait for a response > # from server for each request > option remote-subvolume brick1 # name of the remote volume > end-volume > > volume client12 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 192.168.13.2 # IP address of the remote brick > # option remote-port 6996 # default server port is 6996 > # option transport-timeout 30 # seconds to wait for a response > # from server for each request > option remote-subvolume brick2 # name of the remote volume > end-volume > > > volume client21 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 192.168.13.4 # IP address of the remote brick > # option remote-port 6996 # default server port is 6996 > # option transport-timeout 30 # seconds to wait for a response > # from server for each request > option remote-subvolume brick1 # name of the remote volume > end-volume > > volume client22 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 192.168.13.4 # IP address of the remote brick > # option remote-port 6996 # default server port is 6996 > # option transport-timeout 30 # seconds to wait for a response > # from server for each request > option remote-subvolume brick2 # name of the remote volume > end-volume > > volume client31 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 192.168.13.5 # IP address of the remote brick > # option remote-port 6996 # default server port is 6996 > # option transport-timeout 30 # seconds to wait for a response > # from server for each request > option remote-subvolume brick1 # name of the remote volume > end-volume > > volume client32 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 192.168.13.5 # IP address of the remote brick > # option remote-port 6996 # default server port is 6996 > # option transport-timeout 30 # seconds to wait for a response > # from server for each request > option remote-subvolume brick2 # name of the remote volume > end-volume > > volume client41 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 192.168.13.7 # IP address of the remote brick > # option remote-port 6996 # default server port is 6996 > # option transport-timeout 30 # seconds to wait for a response > # from server for each request > option remote-subvolume brick1 # name of the remote volume > end-volume > > volume client42 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 192.168.13.7 # IP address of the remote brick > # option remote-port 6996 # default server port is 6996 > # option transport-timeout 30 # seconds to wait for a response > # from server for each request > option remote-subvolume brick2 # name of the remote volume > end-volume > > volume afr1 > type cluster/afr > subvolumes client11 client21 > option debug off # turns on detailed debug messages > # in log by default is debugging off > option self-heal on # turn off self healing default is on > end-volume > > volume afr2 > type cluster/afr > subvolumes client31 client41 > option debug off # turns on detailed debug messages > # in log by default is debugging off > option self-heal on # turn off self healing default is on > end-volume > > volume afr3 > type cluster/afr > subvolumes client12 client22 > option debug off # turns on detailed debug messages > # in log by default is debugging off > option self-heal on # turn off self healing default is on > end-volume > > volume afr4 > type cluster/afr > subvolumes client32 client42 > option debug off # turns on detailed debug messages > # in log by default is debugging off > option self-heal on # turn off self healing default is on > end-volume > > volume stripe1 > type cluster/stripe > option block-size 1MB #default size is 128KB > subvolumes afr1 afr2 > end-volume > > volume stripe2 > type cluster/stripe > option block-size 1MB #default size is 128KB > subvolumes afr3 afr4 > end-volume > > > > volume bricks > type cluster/unify > subvolumes stripe1 stripe2 > option namespace client-ns > option scheduler rr > end-volume > > > ### Add io-threads feature > volume iot > type performance/io-threads > option thread-count 1 # deault is 1 > option cache-size 16MB #64MB > > subvolumes bricks #stripe #afr #bricks > end-volume > > ### Add readahead feature > volume readahead > type performance/read-ahead > option page-size 1MB # unit in bytes > option page-count 4 # cache per file = (page-count x page-size) > subvolumes iot > end-volume > > ### Add IO-Cache feature > volume iocache > type performance/io-cache > option page-size 256KB > option page-count 8 > subvolumes readahead > end-volume > > ### Add writeback feature > volume writeback > type performance/write-behind > option aggregate-size 1MB #option flush-behind off > option window-size 3MB # default is 0bytes > # option flush-behind on # default is 'off' > subvolumes iocache > end-volume > -------------- next part -------------- > volume name_space > type storage/posix > option directory /locfsb/name_space > end-volume > > volume brick_1 > type storage/posix # POSIX FS translator > option directory /locfs/brick # Export this directory > end-volume > > > volume brick1 > type features/posix-locks # POSIX FS translator > subvolumes brick_1 > end-volume > > volume brick_2 > type storage/posix # POSIX FS translator > option directory /locfsb/brick # Export this directory > end-volume > > > volume brick2 > type features/posix-locks # POSIX FS translator > subvolumes brick_2 > end-volume > > volume server > type protocol/server > option transport-type tcp/server # For TCP/IP transport > # option listen-port 6996 # Default is 6996 > # option client-volume-filename /etc/glusterfs/glusterfs-client.vol > subvolumes brick1 brick2 name_space > option auth.ip.brick1.allow 192.168.13.* # Allow access to "brick1" > volume > option auth.ip.brick2.allow 192.168.13.* # Allow access to "brick2" > volume > option auth.ip.name_space.allow 192.168.13.* # Allow access to > "name_space" volume > end-volume > > ------------------------------ > > _______________________________________________ > Gluster-devel mailing list > Gluster-devel@xxxxxxxxxx > http://lists.nongnu.org/mailman/listinfo/gluster-devel > > > End of Gluster-devel Digest, Vol 41, Issue 13 > ********************************************* _______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxx http://lists.nongnu.org/mailman/listinfo/gluster-devel |