Drishna,
Thanks. It seems OK.
Thanks,
Yaomin
--------------------------------------------------
From: "Krishna Srinivas" <krishna@xxxxxxxxxxxxx>
Sent: Tuesday, January 06, 2009 10:08 PM
To: "yaomin @ gmail" <yangyaomin@xxxxxxxxx>
Cc: <gluster-devel@xxxxxxxxxx>
Subject: Re: Cascading different translator doesn't work as
expectation
The vol files look fine.
1.3.9 is old and lot of fixes have gone in since then, can you
download the latest release here and try:
http://ftp.gluster.com/pub/gluster/glusterfs/1.4/
Please get back to us if you still face the problem.
Thanks
Krishna
On Tue, Jan 6, 2009 at 5:51 PM, yaomin @ gmail <yangyaomin@xxxxxxxxx>
wrote:
Krishna,
1, The version is 1.3.9
2, the client and server vol files are in the attachments.
3, The result is "No Stack"
Thanks,
Yaomin
--------------------------------------------------
From: "Krishna Srinivas" <krishna@xxxxxxxxxxxxx>
Sent: Tuesday, January 06, 2009 5:36 PM
To: "yaomin @ gmail" <yangyaomin@xxxxxxxxx>
Cc: <gluster-devel@xxxxxxxxxx>
Subject: Re: Cascading different translator doesn't work
as
expectation
Yaomin,
Can you:
* mention what version you are using
* give the modified client and server vol file (to see if there are any
errors)
* give gdb backtrace from the core file? "gdb -c /core.pid glusterfs"
and then type "bt"
Krishna
On Tue, Jan 6, 2009 at 2:43 PM, yaomin @ gmail <yangyaomin@xxxxxxxxx>
wrote:
Krishna,
Thank you for your kind help before.
According to your advice, I confront a new error. The storage node
has
no log information, and the client's log is like following:
/lib64/libc.so.6[0x3fbb2300a0]
/usr/local/lib/glusterfs/1.3.9/xlator/cluster/afr.so(afr_setxattr+0x6a)[0x2aaaaaf0658a]
/usr/local/lib/glusterfs/1.3.9/xlator/cluster/stripe.so(notify+0x220)[0x2aaaab115c80]
/usr/local/lib/libglusterfs.so.0(default_notify+0x25)[0x2aaaaaab8f55]
/usr/local/lib/glusterfs/1.3.9/xlator/cluster/afr.so(notify+0x16d)[0x2aaaaaefc19d]
/usr/local/lib/glusterfs/1.3.9/xlator/protocol/client.so(notify+0x681)[0x2aaaaacebac1]
/usr/local/lib/libglusterfs.so.0(sys_epoll_iteration+0xbb)[0x2aaaaaabe14b]
/usr/local/lib/libglusterfs.so.0(poll_iteration+0x79)[0x2aaaaaabd509]
[glusterfs](main+0x66a)[0x4026aa]
/lib64/libc.so.6(__libc_start_main+0xf4)[0x3fbb21d8a4]
[glusterfs][0x401b69]
---------
[root@IP6 ~]# df -h
Filesystem Size Used Avail Use% Mounted on
/dev/sda2 9.5G 6.8G 2.2G 76% /
/dev/sda1 190M 12M 169M 7% /boot
tmpfs 1006M 0 1006M 0% /dev/shm
/dev/sda4 447G 2.8G 422G 1% /locfs
/dev/sdb1 459G 199M 435G 1% /locfsb
df: `/mnt/new': Transport endpoint is not connected
Thanks,
Yaomin
--------------------------------------------------
From: "Krishna Srinivas" <krishna@xxxxxxxxxxxxx>
Sent: Tuesday, January 06, 2009 1:09 PM
To: "yaomin @ gmail" <yangyaomin@xxxxxxxxx>
Cc: <gluster-devel@xxxxxxxxxx>
Subject: Re: Cascading different translator doesn't
work
as
expectation
Alfred,
Your vol files are wrong. you need to remove all the volume
definitions below "writeback" in the client vol file. For server vol
file the definition of performance translators is not having any
effect. Also you need to use "features/locks" translator above
"storage/posix"
Krishna
On Tue, Jan 6, 2009 at 8:51 AM, yaomin @ gmail <yangyaomin@xxxxxxxxx>
wrote:
All,
It seems difficult for you.
There is a new problem when I tested.
When I kill all the storage nodes, the client still try to send
data,
and doesn't quit.
Thanks,
Alfred
From: yaomin @ gmail
Sent: Monday, January 05, 2009 10:52 PM
To: Krishna Srinivas
Cc: gluster-devel@xxxxxxxxxx
Subject: Re: Cascading different translator doesn't
work
as
expectation
Krishna,
Thank you for your quick response.
There are two log information in the client's log file when
setting
up
the client.
2009-01-05 18:44:59 W [fuse-bridge.c:389:fuse_entry_cbk]
glusterfs-fuse:
2: (34) / => 1 Rehashing 0/0
2009-01-05 18:48:04 W [fuse-bridge.c:389:fuse_entry_cbk]
glusterfs-fuse:
2: (34) / => 1 Rehashing 0/0
There is no any information in the storage node's log file.
Although I changed the scheduler from ALU to RR, there only the
No.3(192.168.13.5) and No.4(192.168.13.7) storage nodes on working.
Each machine has 2GB memory.
Thanks,
Alfred