On Fri, Dec 8, 2017 at 6:51 PM, Florent B <florent@xxxxxxxxxxx> wrote: > I don't know I didn't touched that setting. Which one is recommended ? > > If multiple dovecot instances are running at the same time and they all modify the same files. you need to set fuse_disable_pagecache to true. > On 08/12/2017 11:49, Alexandre DERUMIER wrote: >> do you have disabled fuse pagecache on your clients ceph.conf ? >> >> >> [client] >> fuse_disable_pagecache = true >> >> ----- Mail original ----- >> De: "Florent Bautista" <florent@xxxxxxxxxxx> >> À: "ceph-users" <ceph-users@xxxxxxxxxxxxxx> >> Envoyé: Vendredi 8 Décembre 2017 10:54:59 >> Objet: Re: Corrupted files on CephFS since Luminous upgrade >> >> On 08/12/2017 10:44, Wido den Hollander wrote: >>> >>> On 12/08/2017 10:27 AM, Florent B wrote: >>>> Hi everyone, >>>> >>>> A few days ago I upgraded a cluster from Kraken to Luminous. >>>> >>>> I have a few mail servers on it, running Ceph-Fuse & Dovecot. >>>> >>>> And since the day of upgrade, Dovecot is reporting some corrupted files >>>> on a large account : >>>> >>>> doveadm(myuser@xxxxxxxxxxxx): Error: Corrupted dbox file >>>> /mnt/maildata1/mydomain.com/myuser//mdbox/storage/m.5808 (around >>>> offset=79178): purging found mismatched offsets (79148 vs 72948, >>>> 13/1313) >>>> doveadm(myuser@xxxxxxxxxxxx): Warning: fscking index file >>>> /mnt/maildata1/mydomain.com/myuser//mdbox/storage/dovecot.map.index >>>> doveadm(myuser@xxxxxxxxxxxx): Warning: mdbox >>>> /mnt/maildata1/mydomain.com/myuser//mdbox/storage: rebuilding indexes >>>> doveadm(myuser@xxxxxxxxxxxx): Warning: Transaction log file >>>> /mnt/maildata1/mydomain.com/myuser//mdbox/storage/dovecot.map.index.log >>>> was locked for 1249 seconds (mdbox storage rebuild) >>>> doveadm(myuser@xxxxxxxxxxxx): Error: Purging namespace '' failed: >>>> Corrupted dbox file >>>> /mnt/maildata1/mydomain.com/myuser//mdbox/storage/m.5808 (around >>>> offset=79178): purging found mismatched offsets (79148 vs 72948, >>>> 13/1313) >>>> >>>> Even if Dovecot fixes this problem, every day new files are corrupted. >>>> >>>> I never had this problem before ! And Ceph status is reporting some "MDS >>>> slow requests" ! >>>> >>>> Do you have an idea ? >>>> >>> Not really, but could you share a bit more information: >>> >>> - Which version if Luminous? >>> - Running with BlueStore or FileStore? >>> - Replication? >>> - Cache tiering? >>> - Which kernel version do the clients use? >>> >>> Wido >>> >> Luminous 12.2.1 upgraded to 12.2.2 yesterday, and always the same >> problem today. >> >> FileStore only (xfs). >> >> Replication is 3 copies for these mail files. >> >> No Cache Tiering. >> >> Kernel on clients is default Debian Jessie (3.16.43-2+deb8u5) but I'm >> using ceph-fuse, not kernel client. >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com