Roland, please set 'option lookup-unhashed on' and stat those filenames once. Or you could upgrade to 2.0.6 which fixes this issue. Avati On 9/9/09, Roland Rabben <roland at jotta.no> wrote: > Anyone know if this problem has been addressed in v. 2.0.6? > > Roland > > 2009/9/9 Liam Slusser <lslusser at gmail.com>: > > > You should really upgrade to gluster 2.0.6, there has been many bug fixes. > > > > ls > > > > > > > > On Sep 9, 2009, at 4:36 AM, Roland Rabben <roland at jotta.no> wrote: > > > >> Hi > >> I am using GlusterFS 2.02 on Ubuntu 9.04 64 bit. I have 4 data-nodes and 3 > >> clients. Se my vol files at the end of this email. > >> > >> After adding more disks to my data-nodes for more capacity and > >> reconfiguring > >> GlusterFS to include those drives I am experiencing problems. > >> > >> I am getting "No such file or directory" if I try to copy a new file into > >> an > >> existing directory. > >> However, if I copy a new file into a new directory everyting works fine. > >> > >> It seems that if I create the folderstructure from the old data-nodes on > >> the > >> new disks, everything works fine. > >> > >> So my questions are? > >> > >> 1. Am I doing somthing wrong in the upgrade process? > >> 2. Do I need to manually create the existing folders on the new hard > >> drives? > >> 3. Self heal does not fix this. Shouldn't it? > >> 4. Is there a tool that will create the folderstructure on the new disks > >> for > >> me? > >> > >> > >> Client vol file example: > >> ================= > >> # DN-000 > >> volume dn-000-01 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-000 > >> option remote-subvolume brick-01 > >> end-volume > >> > >> volume dn-000-02 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-000 > >> option remote-subvolume brick-02 > >> end-volume > >> > >> volume dn-000-03 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-000 > >> option remote-subvolume brick-03 > >> end-volume > >> > >> volume dn-000-04 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-000 > >> option remote-subvolume brick-04 > >> end-volume > >> > >> > >> volume dn-000-ns > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-000 > >> option remote-subvolume brick-ns > >> end-volume > >> > >> # DN-001 > >> volume dn-001-01 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-001 > >> option remote-subvolume brick-01 > >> end-volume > >> > >> volume dn-001-02 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-001 > >> option remote-subvolume brick-02 > >> end-volume > >> > >> volume dn-001-03 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-001 > >> option remote-subvolume brick-03 > >> end-volume > >> > >> volume dn-001-04 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-001 > >> option remote-subvolume brick-04 > >> end-volume > >> > >> volume dn-001-ns > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-001 > >> option remote-subvolume brick-ns > >> end-volume > >> > >> # DN-002 > >> volume dn-002-01 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-002 > >> option remote-subvolume brick-01 > >> end-volume > >> > >> volume dn-002-02 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-002 > >> option remote-subvolume brick-02 > >> end-volume > >> > >> volume dn-002-03 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-002 > >> option remote-subvolume brick-03 > >> end-volume > >> > >> volume dn-002-04 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-002 > >> option remote-subvolume brick-04 > >> end-volume > >> > >> # DN-003 > >> volume dn-003-01 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-003 > >> option remote-subvolume brick-01 > >> end-volume > >> > >> volume dn-003-02 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-003 > >> option remote-subvolume brick-02 > >> end-volume > >> > >> volume dn-003-03 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-003 > >> option remote-subvolume brick-03 > >> end-volume > >> > >> volume dn-003-04 > >> type protocol/client > >> option transport-type tcp > >> option remote-host dn-003 > >> option remote-subvolume brick-04 > >> end-volume > >> > >> # Replicate data between the servers > >> # Use pairs, but swtich the order to distribute read load > >> volume repl-000-001-01 > >> type cluster/replicate > >> subvolumes dn-000-01 dn-001-01 > >> end-volume > >> > >> volume repl-000-001-02 > >> type cluster/replicate > >> subvolumes dn-001-02 dn-000-02 > >> end-volume > >> > >> volume repl-000-001-03 > >> type cluster/replicate > >> subvolumes dn-000-03 dn-001-03 > >> end-volume > >> > >> volume repl-000-001-04 > >> type cluster/replicate > >> subvolumes dn-001-04 dn-000-04 > >> end-volume > >> > >> > >> volume repl-002-003-01 > >> type cluster/replicate > >> subvolumes dn-002-01 dn-003-01 > >> end-volume > >> > >> volume repl-002-003-02 > >> type cluster/replicate > >> subvolumes dn-003-02 dn-002-02 > >> end-volume > >> > >> volume repl-002-003-03 > >> type cluster/replicate > >> subvolumes dn-002-03 dn-003-03 > >> end-volume > >> > >> volume repl-002-003-04 > >> type cluster/replicate > >> subvolumes dn-003-04 dn-002-04 > >> end-volume > >> > >> > >> # Also replicate the namespace > >> volume repl-ns > >> type cluster/replicate > >> subvolumes dn-000-ns dn-001-ns > >> end-volume > >> > >> # Distribute the data using the "adaptive least usage" scheduler > >> # We have a 5GB treshold for disk-usage first, then we look at > >> write-usage, > >> and finanly read-usage > >> volume dfs > >> type cluster/unify > >> option namespace repl-ns > >> option scheduler alu > >> option scheduler.limits.min-free-disk 5% > >> option scheduler.alu.order disk-usage:write-usage:read-usage > >> option scheduler.alu.disk-usage.entry-threshold 5GB > >> option scheduler.alu.disk-usage.exit-threshold 1GB > >> option scheduler.alu.write-usage.entry-threshold 25 > >> option scheduler.alu.write-usage.exit-threshold 5 > >> option scheduler.alu.read-usage.entry-threshold 25 > >> option scheduler.alu.read-usage.exit-threshold 5 > >> subvolumes repl-000-001-01 repl-000-001-02 repl-000-001-03 > >> repl-000-001-04 repl-002-003-01 repl-002-003-02 repl-002-003-03 > >> repl-002-003-04 > >> > >> end-volume > >> > >> # Enable write-behind to decrease write latency > >> volume wb > >> type performance/write-behind > >> option flush-behind off > >> option cache-size 128MB > >> subvolumes dfs > >> end-volume > >> > >> volume cache > >> type performance/io-cache > >> option cache-size 1024MB > >> subvolumes wb > >> end-volume > >> > >> > >> > >> > >> > >> Server vol file example: > >> ================== > >> # The posix volumes > >> volume posix-01 > >> type storage/posix > >> option directory /mnt/data01 > >> end-volume > >> > >> volume posix-02 > >> type storage/posix > >> option directory /mnt/data02 > >> end-volume > >> > >> volume posix-03 > >> type storage/posix > >> option directory /mnt/data03 > >> end-volume > >> > >> volume posix-04 > >> type storage/posix > >> option directory /mnt/data04 > >> end-volume > >> > >> > >> volume posix-ns > >> type storage/posix > >> option directory /var/lib/glusterfs/ns > >> end-volume > >> > >> # Add locking capabilities > >> volume locks-01 > >> type features/locks > >> subvolumes posix-01 > >> end-volume > >> > >> volume locks-02 > >> type features/locks > >> subvolumes posix-02 > >> end-volume > >> > >> volume locks-03 > >> type features/locks > >> subvolumes posix-03 > >> end-volume > >> > >> volume locks-04 > >> type features/locks > >> subvolumes posix-04 > >> end-volume > >> > >> > >> volume locks-ns > >> type features/locks > >> subvolumes posix-ns > >> end-volume > >> > >> # Finally add threads to the briks > >> volume brick-01 > >> type performance/io-threads > >> option thread-count 8 > >> subvolumes locks-01 > >> end-volume > >> > >> volume brick-02 > >> type performance/io-threads > >> option thread-count 8 > >> subvolumes locks-02 > >> end-volume > >> > >> volume brick-03 > >> type performance/io-threads > >> option thread-count 8 > >> subvolumes locks-03 > >> end-volume > >> > >> volume brick-04 > >> type performance/io-threads > >> option thread-count 8 > >> subvolumes locks-04 > >> end-volume > >> > >> > >> volume brick-ns > >> type performance/io-threads > >> option thread-count 8 > >> subvolumes locks-ns > >> end-volume > >> > >> # Mount the posix drives as a network drive > >> volume server > >> type protocol/server > >> option transport-type tcp > >> subvolumes brick-01 brick-02 brick-03 brick-04 brick-ns > >> option auth.addr.brick-01.allow 10.0.* > >> option auth.addr.brick-02.allow 10.0.* > >> option auth.addr.brick-03.allow 10.0.* > >> option auth.addr.brick-04.allow 10.0.* > >> option auth.addr.brick-ns.allow 10.0.* > >> end-volume > >> > >> > >> > >> Regards > >> > >> Roland Rabben > >> Founder & CEO Jotta AS > >> Cell: +47 90 85 85 39 > >> Phone: +47 21 04 29 00 > >> Email: roland at jotta.no > >> _______________________________________________ > >> Gluster-users mailing list > >> Gluster-users at gluster.org > >> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > > > > > > > -- > > Roland Rabben > Founder & CEO Jotta AS > Cell: +47 90 85 85 39 > Phone: +47 21 04 29 00 > Email: roland at jotta.no > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >