Can't start a volume. [root at sfsccl03 ~]# gluster volume start brick1 brick: sfsccl03:/data/brick-sdc2/glusterfs/dht, path creation failed, reason: No such file or directory But ... [root at sfsccl03 ~]# ls -alF /data/brick-sdc2/glusterfs total 0 drwxr-xr-x 4 root root 27 Dec 2 13:00 ./ drwxr-xr-x 4 root root 107 Jul 5 11:55 ../ drwxrwxrwt 7 root root 61 Sep 15 11:35 dht/ drwxr-xr-x 2 root root 6 Dec 2 13:00 dht2/ So it is there. [root at sfsccl03 ~]# ls -alF /data/brick-sdc2/glusterfs/dht total 128 drwxrwxrwt 7 root root 61 Sep 15 11:35 ./ drwxr-xr-x 4 root root 27 Dec 2 13:00 ../ drwxr-xr-x 1230 root root 65536 Oct 24 09:14 equity/ drwxr-xr-x 1740 oracle root 65536 Nov 30 23:33 opra/ drwxr-xr-x 35 oracle oinstall 501 Jul 9 17:07 tag/ drwxr-xr-x 11 root root 126 Jul 1 08:51 taq/ drwxr-xr-x 2 root root 34 Jul 11 19:44 test/ and it is readable. More info: [root at sfsccl03 ~]# gluster volume info brick1 Volume Name: brick1 Type: Distribute Status: Stopped Number of Bricks: 4 Transport-type: tcp Bricks: Brick1: sfsccl01:/data/glusterfs/dht Brick2: sfsccl02:/data/glusterfs/dht Brick3: sfsccl03:/data/brick-sdc2/glusterfs/dht Brick4: sfsccl03:/data/brick-sdd2/glusterfs/dht [root at sfsccl03 ~]# gluster peer status Number of Peers: 2 Hostname: sfsccl02 Uuid: 6e72d1a8-bdeb-4bfb-806c-7fa8b98cb697 State: Peer in Cluster (Connected) Hostname: sfsccl01 Uuid: 116197cd-5dfe-4881-85ad-5de2be484ba6 State: Peer in Cluster (Connected) a volume reset doesn't help. [root at sfsccl03 ~]# gluster volume reset brick1 reset volume successful [root at sfsccl03 ~]# gluster volume start brick1 brick: sfsccl03:/data/brick-sdc2/glusterfs/dht, path creation failed, reason: No such file or directory New volume creation also fails. [root at sfsccl03 ~]# gluster volume create brick2 transport tcp sfsccl01:/data/glusterfs/dht2 sfsccl03:/data/brick-sdc2/glusterfs/dht2 sfsccl02:/data/glusterfs/dht2 sfsccl03:/data/brick-sdd2/glusterfs/dht2 brick: sfsccl03:/data/brick-sdc2/glusterfs/dht2, path creation failed, reason: No such file or directory Not good. Taking out the 03 machine [root at sfsccl03 ~]# gluster volume create brick2 transport tcp sfsccl01:/data/glusterfs/dht2 sfsccl02:/data/glusterfs/dht2Creation of volume brick2 has been successful. Please start the volume to access data. I am wondering if I should remove the 03 machine from the volume, start it up with 01 and 02, and then add the 03 machine in, after forcing the volume back up. Any thoughts? -- Joseph Landman, Ph.D Founder and CEO Scalable Informatics Inc. email: landman at scalableinformatics.com web : http://scalableinformatics.com http://scalableinformatics.com/sicluster phone: +1 734 786 8423 x121 fax : +1 866 888 3112 cell : +1 734 612 4615