[SPAM?] STALE NFS Handle (Transport endpoint is not connected)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Am Mittwoch 16 M?rz 2011, um 13:33:27 schrieben Sie:

Hi Burnash,

thanks for the info. Hm, yes i could do that but that is not the intention.
Even with a "distribute" setup the client should never hang when a (or more) server goes down. No matter what.

Taken the "redistribute" scenario into account, when will that going to hang ? When both bricks are down (replica 2) ?
Or what will happen if the cloud of the two bricks cannot be reached (network failure) ?

So to my understanding a network application should never just hang (or STALE), but give a proper error message or whatever is appropriate.
To me more appropriate, is that the "ressource" (brick) is not there, i.e directory empty or so.

Correct if i'am wrong but adding replicas is just a way of saying it is "more unlikely" to happen, but it still could ...

Regards
Heiko




> Hi Heiko.
> 
> Since you have your bricks setup as Distribute, both bricks participate in the file system - if you look at the underlying directories that make up each brick, you will find some files on each server that appear to be under the same directory when mounted by a client.
> 
> You must have both bricks up to present a valid filesystem.
> 
> To get the scenario you are describing, you would have to delete and recreate your Gluster volume as Replicated-Distribute, using the arguments "replica 2" in your gluster volume create stanza.
> 
> James Burnash, Unix Engineering
> 
> -----Original Message-----
> From: gluster-users-bounces at gluster.org [mailto:gluster-users-bounces at gluster.org] On Behalf Of Heiko Schr?ter
> Sent: Wednesday, March 16, 2011 8:26 AM
> To: gluster-users at gluster.org
> Subject: [SPAM?] STALE NFS Handle (Transport endpoint is not connected)
> Importance: Low
> 
> Hello,
> 
> the following produces STALE NFS Handle (Transport endpoint is not connected)
> Scenario: 2 bricks, 1 client
> Version: 3.1.2
> 
> 1) Bring gluster system up
> rd32 ~ # gluster volume info
> Volume Name: test-volume
> Type: Distribute
> Status: Started
> Number of Bricks: 2
> Transport-type: tcp
> Bricks:
> Brick1: rd32:/mnt/gluster1
> Brick2: rd33:/mnt/gluster1
> 
> 1a) mount /mnt/gluster on client.
> 
> 2) Bring bricks down
> 
> 3) Try to access /mnt/gluster on client:
> client# ls -la /mnt/gluster
> ls: cannot access /mnt/gluster: Transport endpoint is not connected
> 
> 3) Start glusterd on Brick1:
> client# ls -la /mnt/gluster
> ls: cannot access /mnt/gluster: Stale NFS file handle
> 
> 4) Start glusterd on Brick2:
> client# ls -la /mnt/gluster
> total 102408
> drwxr-xr-x 2 root root        58 Mar 14 15:48 .
> drwxr-xr-x 7 root root      4096 Mar  2 10:21 ..
> -rw-r--r-- 1 root root         1 Mar 14 14:36 .version
> -rw-r--r-- 1 root root 104857600 Mar 14 15:48 willi.dat
> 
> STALE NFS Handle only vanishs when ALL bricks are up and running, when started as above.
> This is reproducable.
> 
> Expected behaviour:
> Client should never ever see a STALE NFS Handle.
> i.e
> If no brick is running -> /mnt/gluster should be just empty.
> If at least one brick is running -> /mnt/gluster should show the files on that single brick.
> 
> 
> Thanks and Regards
> Heiko
> 
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
> 
> 
> DISCLAIMER:
> This e-mail, and any attachments thereto, is intended only for use by the addressee(s) named herein and may contain legally privileged and/or confidential information. If you are not the intended recipient of this e-mail, you are hereby notified that any dissemination, distribution or copying of this e-mail, and any attachments thereto, is strictly prohibited. If you have received this in error, please immediately notify me and permanently delete the original and any copy of any e-mail and any printout thereof. E-mail transmission cannot be guaranteed to be secure or error-free. The sender therefore does not accept liability for any errors or omissions in the contents of this message which arise as a result of e-mail transmission.
> NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, at its discretion, monitor and review the content of all e-mail communications. http://www.knight.com


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux