On Tue, May 23, 2023 at 11:52 PM Dietmar Rieder <dietmar.rieder@xxxxxxxxxxx> wrote: > > On 5/23/23 15:58, Gregory Farnum wrote: > > On Tue, May 23, 2023 at 3:28 AM Dietmar Rieder > > <dietmar.rieder@xxxxxxxxxxx> wrote: > >> > >> Hi, > >> > >> can the cephfs "max_file_size" setting be changed at any point in the > >> lifetime of a cephfs? > >> Or is it critical for existing data if it is changed after some time? Is > >> there anything to consider when changing, let's say, from 1TB (default) > >> to 4TB ? > > > > Larger files take longer to delete (the MDS has to issue a delete op > > on every one of the objects that may exist), and longer to recover if > > their client crashes and the MDS has to probe all the objects looking > > for the actual size and mtime. > > This is all throttled so it shouldn't break anything, we just want to > > avoid the situation somebody ran into once where they accidentally > > created a 1 exabyte RBD on their little 3-node cluster and then had to > > suffer through "deleting" it. :D > > -Greg > > > > Thanks for your detailed explanation. > > Would it also be ok if we set the max to 5 TB create some big files > (>1TB) and then set the max back to 1 TB? Would the big files then still > be available and usable? It does look like that would work, but I wouldn't recommend it. Workarounds like that will always be finicky. -Greg > > Best > Dietmar > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx