do you have disabled fuse pagecache on your clients ceph.conf ? [client] fuse_disable_pagecache = true ----- Mail original ----- De: "Florent Bautista" <florent@xxxxxxxxxxx> À: "ceph-users" <ceph-users@xxxxxxxxxxxxxx> Envoyé: Vendredi 8 Décembre 2017 10:54:59 Objet: Re: Corrupted files on CephFS since Luminous upgrade On 08/12/2017 10:44, Wido den Hollander wrote: > > > On 12/08/2017 10:27 AM, Florent B wrote: >> Hi everyone, >> >> A few days ago I upgraded a cluster from Kraken to Luminous. >> >> I have a few mail servers on it, running Ceph-Fuse & Dovecot. >> >> And since the day of upgrade, Dovecot is reporting some corrupted files >> on a large account : >> >> doveadm(myuser@xxxxxxxxxxxx): Error: Corrupted dbox file >> /mnt/maildata1/mydomain.com/myuser//mdbox/storage/m.5808 (around >> offset=79178): purging found mismatched offsets (79148 vs 72948, >> 13/1313) >> doveadm(myuser@xxxxxxxxxxxx): Warning: fscking index file >> /mnt/maildata1/mydomain.com/myuser//mdbox/storage/dovecot.map.index >> doveadm(myuser@xxxxxxxxxxxx): Warning: mdbox >> /mnt/maildata1/mydomain.com/myuser//mdbox/storage: rebuilding indexes >> doveadm(myuser@xxxxxxxxxxxx): Warning: Transaction log file >> /mnt/maildata1/mydomain.com/myuser//mdbox/storage/dovecot.map.index.log >> was locked for 1249 seconds (mdbox storage rebuild) >> doveadm(myuser@xxxxxxxxxxxx): Error: Purging namespace '' failed: >> Corrupted dbox file >> /mnt/maildata1/mydomain.com/myuser//mdbox/storage/m.5808 (around >> offset=79178): purging found mismatched offsets (79148 vs 72948, >> 13/1313) >> >> Even if Dovecot fixes this problem, every day new files are corrupted. >> >> I never had this problem before ! And Ceph status is reporting some "MDS >> slow requests" ! >> >> Do you have an idea ? >> > > Not really, but could you share a bit more information: > > - Which version if Luminous? > - Running with BlueStore or FileStore? > - Replication? > - Cache tiering? > - Which kernel version do the clients use? > > Wido > Luminous 12.2.1 upgraded to 12.2.2 yesterday, and always the same problem today. FileStore only (xfs). Replication is 3 copies for these mail files. No Cache Tiering. Kernel on clients is default Debian Jessie (3.16.43-2+deb8u5) but I'm using ceph-fuse, not kernel client. _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com