[SPAM?] STALE NFS Handle (Transport endpoint is not connected)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Heiko.

Since you have your bricks setup as Distribute, both bricks participate in the file system - if you look at the underlying directories that make up each brick, you will find some files on each server that appear to be under the same directory when mounted by a client.

You must have both bricks up to present a valid filesystem.

To get the scenario you are describing, you would have to delete and recreate your Gluster volume as Replicated-Distribute, using the arguments "replica 2" in your gluster volume create stanza.

James Burnash, Unix Engineering

-----Original Message-----
From: gluster-users-bounces at gluster.org [mailto:gluster-users-bounces at gluster.org] On Behalf Of Heiko Schr?ter
Sent: Wednesday, March 16, 2011 8:26 AM
To: gluster-users at gluster.org
Subject: [SPAM?] STALE NFS Handle (Transport endpoint is not connected)
Importance: Low

Hello,

the following produces STALE NFS Handle (Transport endpoint is not connected)
Scenario: 2 bricks, 1 client
Version: 3.1.2

1) Bring gluster system up
rd32 ~ # gluster volume info
Volume Name: test-volume
Type: Distribute
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: rd32:/mnt/gluster1
Brick2: rd33:/mnt/gluster1

1a) mount /mnt/gluster on client.

2) Bring bricks down

3) Try to access /mnt/gluster on client:
client# ls -la /mnt/gluster
ls: cannot access /mnt/gluster: Transport endpoint is not connected

3) Start glusterd on Brick1:
client# ls -la /mnt/gluster
ls: cannot access /mnt/gluster: Stale NFS file handle

4) Start glusterd on Brick2:
client# ls -la /mnt/gluster
total 102408
drwxr-xr-x 2 root root        58 Mar 14 15:48 .
drwxr-xr-x 7 root root      4096 Mar  2 10:21 ..
-rw-r--r-- 1 root root         1 Mar 14 14:36 .version
-rw-r--r-- 1 root root 104857600 Mar 14 15:48 willi.dat

STALE NFS Handle only vanishs when ALL bricks are up and running, when started as above.
This is reproducable.

Expected behaviour:
Client should never ever see a STALE NFS Handle.
i.e
If no brick is running -> /mnt/gluster should be just empty.
If at least one brick is running -> /mnt/gluster should show the files on that single brick.


Thanks and Regards
Heiko


_______________________________________________
Gluster-users mailing list
Gluster-users at gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


DISCLAIMER:
This e-mail, and any attachments thereto, is intended only for use by the addressee(s) named herein and may contain legally privileged and/or confidential information. If you are not the intended recipient of this e-mail, you are hereby notified that any dissemination, distribution or copying of this e-mail, and any attachments thereto, is strictly prohibited. If you have received this in error, please immediately notify me and permanently delete the original and any copy of any e-mail and any printout thereof. E-mail transmission cannot be guaranteed to be secure or error-free. The sender therefore does not accept liability for any errors or omissions in the contents of this message which arise as a result of e-mail transmission.
NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, at its discretion, monitor and review the content of all e-mail communications. http://www.knight.com


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux