Hi Andrey, I am looking into the issue. Will get back to you soon. Btw, is it very regular (ie, you mount glusterfs, and create a file, and it segfaults)? -Amar On 9/6/07, NovA <av.nova@xxxxxxxxx> wrote: > > Hello everybody! > > Today I've tried to switch from tla patch 184 to the 1.3.1 release. > But client crashes with core, when writing something to imported unify > volume. Here there are the backtrace and my specs: > > ------------ > Using host libthread_db library "/lib64/libthread_db.so.1". > Core was generated by `[glusterfs] > '. > Program terminated with signal 8, Arithmetic exception. > #0 0x00002aaaab569af2 in get_stats_free_disk (this=0x7fffa19e8420) at > alu.c:71 > in alu.c > #0 0x00002aaaab569af2 in get_stats_free_disk (this=0x7fffa19e8420) at > alu.c:71 > #1 0x00002aaaab56a21d in update_stat_array_cbk (frame=0x65d2c0, > cookie=0x6147e0, xl=<value optimized out>, op_ret=0, > op_errno=<value optimized out>, trav_stats=0x7fffa19e8420) at alu.c > :483 > #2 0x00002aaaaaaaf9d5 in client_stats_cbk (frame=0x6560e0, > args=<value optimized out>) at client-protocol.c:3974 > #3 0x00002aaaaaab2c33 in notify (this=0x614800, event=<value > optimized out>, data=0x62cd80) at client-protocol.c:4409 > #4 0x00002aac092eeb22 in sys_epoll_iteration (ctx=<value optimized > out>) at epoll.c:53 > #5 0x000000000040348b in main (argc=6, argv=0x7fffa19e8728) at > glusterfs.c:388 > ------------ > > > --------- server-head.vol ------ > # Namespace > volume brick-ns > type storage/posix # POSIX FS translator > option directory /mnt/glusterfs-ns # Export this directory > end-volume > > # Data > volume disk > type storage/posix # POSIX FS translator > option directory /mnt/hd # Export this directory > end-volume > > volume locks > type features/posix-locks > option mandatory on > subvolumes disk > end-volume > > volume brick #iothreads can give performance a boost > type performance/io-threads > option thread-count 8 > subvolumes locks > end-volume > > volume server > type protocol/server > option transport-type tcp/server # For TCP/IP transport > # option bind-address 192.168.1.10 # Default is to listen on all > interfaces > # option listen-port 6996 # Default is 6996 > # option client-volume-filename /etc/glusterfs/client.vol > subvolumes brick brick-ns > option auth.ip.brick.allow 10.1.0.* # Allow access to "brick" volume > option auth.ip.brick-ns.allow 10.1.0.* # Allow access to "brick-ns" > volume > end-volume > --------------------------------------- > > -------- server-node.vol ----------- > volume disk > type storage/posix > option directory /mnt/hd > end-volume > > volume locks > type features/posix-locks > option mandatory on > subvolumes disk > end-volume > > volume brick > type performance/io-threads > subvolumes locks > end-volume > > volume server > type protocol/server > option transport-type tcp/server # For TCP/IP transport > # option bind-address 192.168.1.10 # Default is to listen on all > interfaces > # option listen-port 6996 # Default is 6996 > option client-volume-filename /etc/glusterfs/client.vol > subvolumes brick > option auth.ip.brick.allow 10.1.0.* # Allow access to "brick" volume > end-volume > ----------------------------------- > > ------------ client.vol ----------- > ### Remote subvolumes > volume c0 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 10.1.0.1 # IP address of the remote brick > # option remote-port 6996 # default server port is 6996 > option remote-subvolume brick # name of the remote volume > end-volume > > volume c-ns # namespace > type protocol/client > option transport-type tcp/client > option remote-host 10.1.0.1 > option remote-subvolume brick-ns > end-volume > > volume c1 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 10.1.0.2 # IP address of the remote brick > option remote-subvolume brick # name of the remote volume > end-volume > > [ ...skipped... ] > > volume c10 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host 10.1.0.11 # IP address of the remote brick > option remote-subvolume brick # name of the remote volume > end-volume > > volume bricks > type cluster/unify > option namespace c-ns # this will not be storage child of unify. > subvolumes c0 c1 c2 c3 c4 c5 c6 c7 c8 c9 c10 > > option scheduler alu > option alu.limits.min-free-disk 5% #% > option alu.limits.max-open-files 10000 > option alu.order > disk-usage:read-usage:write-usage:open-files-usage:disk-speed-usage > option alu.disk-usage.entry-threshold 2GB > option alu.disk-usage.exit-threshold 128MB > option alu.open-files-usage.entry-threshold 1024 > option alu.open-files-usage.exit-threshold 32 > option alu.read-usage.entry-threshold 20 #% > option alu.read-usage.exit-threshold 4 #% > option alu.write-usage.entry-threshold 20 #% > option alu.write-usage.exit-threshold 4 #% > option alu.disk-speed-usage.entry-threshold 0 > option alu.disk-speed-usage.exit-threshold 0 > option alu.stat-refresh.interval 10sec > option alu.stat-refresh.num-file-create 10 > end-volume > > #volume debug > # type debug/trace > # subvolumes bricks > #end-volume > > volume threads > type performance/io-threads > option thread-count 11 > subvolumes bricks > end-volume > > volume wb > type performance/write-behind > option aggregate-size 1MB > subvolumes threads > end-volume > --------------------------------- > > With best regards, > Andrey > > > _______________________________________________ > Gluster-devel mailing list > Gluster-devel@xxxxxxxxxx > http://lists.nongnu.org/mailman/listinfo/gluster-devel > -- Amar Tumballi Engineer - Gluster Core Team [bulde on #gluster/irc.gnu.org] http://www.zresearch.com - Commoditizing Supercomputing and Superstorage!