Hi. On Thu, Aug 04, 2022 at 12:25:31PM +0200, Emmanouil Vamvakopoulos wrote: > hello Carlos and Dave > > thank you for the replies > > a) for the mismatch in alignment bewteen xfs and underlying raid volume I have to re-check > but from preliminary tests , when I mount the partition with a static allocsize ( e.g. allocsize=256k) > we have large file with large number of externs ( up to 40) but the sizes from du was comparable. allocsize mount option controls the EOF preallocation size, which, by default is dynamic, so, you just fixed it to a small size, and it might well be the reason why you ended up with so many extents, as the main goal of speculative preallocation is to try to reduce fragmentation by creating bigger extents, and as Dave mentioned, the extra space will be removed after file is closed. I'm not the best to explain details on speculative preallocation, but I suppose you're seeing a closer size report from du modes due the smaller preallocated space, even though you have more extents, the extra preallocated space is still very small. > > b) for the speculative preallocation beyond EOF of my files as I understood have to run xfs_fsr to get the space back. No, speculative preallocation is dynamically removed. > > but why the inodes of those files remains dirty at least for 300 sec after the closing of the file and lost the automatic removal of the preallocation ? > IIRC, speculative preallocated blocks can be kept around even after the file is closed, I believe append-only files are one example of that, where the speculative preallocated blocks will be kept after a file is closed. But I don't have a deep knowledge on the speculative prealloc algorithm to give more details. But I'm pretty sure it's tied up with the file's write patterns, maybe you can describe more how this file is written to? > we are runing on CentOS Stream release 8 with 4.18.0-383.el8.x86_64 > > but we never see something simliar on CentOS Linux release 7.9.2009 (Core) with 3.10.0-1160.45.1.el7.x86_64 > (for similar pattern of file sizes, but truly with different distributed strorage application) > That's a question more for the distribution not for the upstream project =/ unlikely somebody will remember what changed between 3.10 and 4.18 and also what the distribution backported (or not). > > > > ----- Original Message ----- > From: "Dave Chinner" <david@xxxxxxxxxxxxx> > To: "emmanouil vamvakopoulos" <emmanouil.vamvakopoulos@xxxxxxxxxxxxxxx> > Cc: "linux-xfs" <linux-xfs@xxxxxxxxxxxxxxx> > Sent: Wednesday, 3 August, 2022 23:59:09 > Subject: Re: s_bmap and flags explanation > > On Wed, Aug 03, 2022 at 04:56:43PM +0200, Emmanouil Vamvakopoulos wrote: > > > > > > Hello developers > > > > It is possible to explain the FLAGS field in xfs_bmap output of a file > > > > EXT: FILE-OFFSET BLOCK-RANGE AG AG-OFFSET TOTAL FLAGS > > 0: [0..7]: 49700520968..49700520975 30 (8..15) 8 001111 > > 1: [8..4175871]: 49708756480..49712932343 30 (8235520..12411383) 4175864 000111 > > 2: [4175872..19976191]: 49715788288..49731588607 30 (15267328..31067647) 15800320 000011 > > 3: [19976192..25153535]: 49731588608..49736765951 30 (31067648..36244991) 5177344 000011 > > 4: [25153536..41930743]: 49767625216..49784402423 30 (67104256..83881463) 16777208 000111 > > 5: [41930744..58707951]: 49784402424..49801179631 30 (83881464..100658671) 16777208 001111 > > 6: [58707952..58959935]: 49801179632..49801431615 30 (100658672..100910655) 251984 001111 > > 7: [58959936..75485159]: 49801431616..49817956839 30 (100910656..117435879) 16525224 001111 > > $ man xfs_bmap > ..... > -v Shows verbose information. When this flag is specified, > additional AG specific information is appended to each > line in the following form: > > agno (startagoffset..endagoffset) nblocks flags > > A second -v option will print out the flags legend. > ..... > > So: > > $ xfs_bmap -vvp foo > foo: > EXT: FILE-OFFSET BLOCK-RANGE AG AG-OFFSET TOTAL FLAGS > 0: [0..7]: 440138672..440138679 4 (687024..687031) 8 000000 > FLAG Values: > 0100000 Shared extent > 0010000 Unwritten preallocated extent > 0001000 Doesn't begin on stripe unit > 0000100 Doesn't end on stripe unit > 0000010 Doesn't begin on stripe width > 0000001 Doesn't end on stripe width > > And there's what the flags mean. > > > with > > > > [disk06]# du -sh ./00000869/014886f4 > > 36G ./00000869/014886f4 > > [disk06]# du -sh --apparent-size ./00000869/014886f4 > > 29G ./00000869/014886f4 > > > > I try to understand if this file contains unused externs > > and how those file are created like this (if we assume that the free space was not fragmented ) > > > > we are running CentOS Stream release 8 with 4.18.0-383.el8.x86_64 > > > > if I defrag the file above the difference bewteen apparent size and size with du disappered ! > > It will be a result of speculative preallocation beyond EOF as the > file is grown to ensure it doesn't get fragmented badly. Files in > the size range of tens of GB or larger will have preallocation > extend out to 8GB beyond EOF. It will get removed when the inode is > reclaimed from memory (i.e. no longer in active use). > > Cheers, > > Dave. > -- > Dave Chinner > david@xxxxxxxxxxxxx -- Carlos Maiolino