Hi Theodore, Thank you for your analysis. As you suggested, finally we decided to increase block size to 4K, and it seems that problem is resolved. By default, yocto/bitbake do not configure the block size, so that mkfs.ext4 used default value - 1K. We fixed it like here: (see the last line) https://git.yoctoproject.org/cgit/cgit.cgi/meta-qcom/tree/conf/machine/dragonboard-820c.conf?h=master#n26 BR, Vilo Confidentiality Notice: This e-mail is privileged and confidential and for the use of the addressee only. Should you have received this e-mail in error please notify us by replying directly to the sender or by sending a message to info@xxxxxxxxxxx. Unauthorised dissemination, disclosure or copying of the contents of this e-mail, or any similar action, is prohibited. -----Original Message----- From: Theodore Y. Ts'o <tytso@xxxxxxx> Sent: Friday, 6. December 2019 4:51 To: Lejcik Viliam <Viliam.Lejcik@xxxxxxxxxxx> Cc: linux-ext4@xxxxxxxxxxxxxxx Subject: Re: e2fsprogs: setting UUID with tune2fs corrupts an ext4 fs image On Thu, Dec 05, 2019 at 12:36:35PM +0000, Viliam Lejcik wrote: > > The problem for tune2fs is "Number of entries", when count==limit > (126). In this case it fails within the following 'if' statement: > https://git.kernel.org/pub/scm/fs/ext2/e2fsprogs.git/tree/misc/tune2fs.c#n544 > > Then it prints out error, sets 'not clean' fs state in superblock, > and exits. What fsck does, it recomputes checksums, sets 'clean' fs > state, and that's all. It doesn't change number of entries, > count+limit stays the same (126). So that's why rerunning tune2fs > corrupts the fs again. So what's going on is that the code in question was originally designed for file systems with a 4k block size, and when *adding* a checksum to a directory which didn't already have a checksum "tail" already reserved. In that case, you have to recreate the htree index for that directory. Since tune2fs didn't want to deal with that corner case, it would throw up its hands and tell the user to run e2fsck -fD. Since the UUID had alrady been changed, and the checksum seed was based on the checksum seed, e2fsck would report corruptions, but that was actually what was expected. Unfortunately the message printed by tune2fs is super-confusing, and logic for checking for this case is faulty in that (a) it doesn't take into account the block size, and (b) it doesn't take into account if there is a checksum "tail" already present for that particular htree directory. Most people don't see this because they are using file systems with 4k block sizes, and it's much less likely they will run into that situation, since the interior node fanout is significantly larger with 4k blocks. (Is there a reason you are using a 1k block size? This adds quite a bit of inefficiency to the file system, and while it does reduce internal fragmentation, bytes are quite cheap these days, and it's probably not worth it if you care about performance at all to use a 1k block size instead of a 4k block size.) The workaround I would suggest is assuming you are using a kernel which is 4.4 or newer (and in 2019, you really should), to turn on the metadata_csum_seed field, either when the file system is originally formatted, or using "tune2fs -O ^metadata_csum_seed". This allows you to change the UUID without needing to rewrite all of the metadata blocks, which is faster, works while the file system is mounted, and avoids the bug in tune2fs. So using the test file system you sent me, this works just fine: % tune2fs -O metadata_csum_seed -U random core-image.ext4 tune2fs 1.45.4 (23-Sep-2019) % e2fsck -fy !$ e2fsck -fy core-image.ext4 e2fsck 1.45.4 (23-Sep-2019) Pass 1: Checking inodes, blocks, and sizes Pass 2: Checking directory structure Pass 3: Checking directory connectivity Pass 4: Checking reference counts Pass 5: Checking group summary information core-image.ext4: 13237/89408 files (0.6% non-contiguous), 249888/357064 blocks cheers, - Ted