I/O error for one folder within the mountpoint

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello everyone,

first time on the ML so excuse me if I'm not following well the rules, I'll improve if I get comments.

We got one volume "applicatif" on three nodes (2 and 1 arbiter), each following command was made on node ipvr8.xxx:

# gluster volume info applicatif
 
Volume Name: applicatif
Type: Replicate
Volume ID: ac222863-9210-4354-9636-2c822b332504
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: ipvr7.xxx:/mnt/gluster-applicatif/brick
Brick2: ipvr8.xxx:/mnt/gluster-applicatif/brick
Brick3: ipvr9.xxx:/mnt/gluster-applicatif/brick (arbiter)
Options Reconfigured:
performance.read-ahead: on
performance.cache-size: 1024MB
performance.quick-read: off
performance.stat-prefetch: on
performance.io-cache: off
transport.address-family: inet
performance.readdir-ahead: on
nfs.disable: off

# gluster volume status applicatif
Status of volume: applicatif
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick ipvr7.xxx:/mnt/gluster-applicatif/
brick                                       49154     0          Y       2814
Brick ipvr8.xxx:/mnt/gluster-applicatif/
brick                                       49154     0          Y       2672
Brick ipvr9.xxx:/mnt/gluster-applicatif/
brick                                       49154     0          Y       3424
NFS Server on localhost                     2049      0          Y       26530
Self-heal Daemon on localhost               N/A       N/A        Y       26538
NFS Server on ipvr9.xxx                  2049      0          Y       12238
Self-heal Daemon on ipvr9.xxx            N/A       N/A        Y       12246
NFS Server on ipvr7.xxx                  2049      0          Y       2234
Self-heal Daemon on ipvr7.xxx            N/A       N/A        Y       2243
 
Task Status of Volume applicatif
------------------------------------------------------------------------------
There are no active volume tasks

The volume is mounted with autofs (nfs) in /home/applicatif and one folder is "broken":

l /home/applicatif/services/
ls: cannot access /home/applicatif/services/snooper: Input/output error
total 16
lrwxrwxrwx  1 applicatif applicatif    9 Apr  6 15:53 config -> ../config
lrwxrwxrwx  1 applicatif applicatif    7 Apr  6 15:54 .pwd -> ../.pwd
drwxr-xr-x  3 applicatif applicatif 4096 Apr 12 10:24 querybuilder
d?????????  ? ?          ?             ?            ? snooper
drwxr-xr-x  3 applicatif applicatif 4096 Jul  6 02:57 snooper_new
drwxr-xr-x 16 applicatif applicatif 4096 Jul  6 02:58 snooper_old
drwxr-xr-x  4 applicatif applicatif 4096 Jul  4 23:45 ssnooper

I checked wether there was a heal, and it seems so:

# gluster volume heal applicatif statistics heal-count
Gathering count of entries to be healed on volume applicatif has been successful

Brick ipvr7.xxx:/mnt/gluster-applicatif/brick
Number of entries: 8

Brick ipvr8.xxx:/mnt/gluster-applicatif/brick
Number of entries: 29

Brick ipvr9.xxx:/mnt/gluster-applicatif/brick
Number of entries: 8

But actually in the brick on each server the folder "snooper" is fine.

I tried rebooting the servers, restarting gluster after killing every process using it but it's not working.

Has anyone already experienced that ? Any help would be nice.

Thanks a lot !

--

Cordialement,

 

Florian LELEU
Responsable Hosting, Cognix Systems

Rennes | Brest | Saint-Malo | Paris
florian.leleu@xxxxxxxxxxxxxxxxxx

Tél. : 02 99 27 75 92

   
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux