On 1/19/24 2:34 PM, Kent Overstreet wrote: > On Fri, Jan 19, 2024 at 02:22:04PM -0700, Jens Axboe wrote: >> On 1/19/24 5:25 AM, Mia Kanashi wrote: >>> This issue was originally reported here: https://github.com/koverstreet/bcachefs/issues/628 >>> >>> Transferring large amounts of files to the bcachefs from the btrfs >>> causes I/O timeouts and freezes the whole system. This doesn't seem to >>> be related to the btrfs, but rather to the heavy I/O on the drive, as >>> it happens without btrfs being mounted. Transferring the files to the >>> HDD, and then from it to the bcachefs on the NVME sometimes doesn't >>> make the problem occur. The problem only happens on the bcachefs, not >>> on btrfs or ext4. It doesn't happen on the HDD, I can't test with >>> other NVME drives sadly. The behaviour when it is frozen is like this: >>> all drive accesses can't process, when not cached in ram, so every app >>> that is loaded in the ram, continues to function, but at the moment it >>> tries to access the drive it freezes, until the drive is reset and >>> those abort status messages appear in the dmesg, after that system is >>> unfrozen for a moment, if you keep copying the files then the problem >>> reoccurs once again. >>> >>> This drive is known to have problems with the power management in the >>> past: >>> https://wiki.archlinux.org/title/Solid_state_drive/NVMe#Troubleshooting >>> But those problems where since fixed with kernel workarounds / >>> firmware updates. This issue is may be related, perhaps bcachefs does >>> something different from the other filesystems, and workarounds don't >>> apply, which causes the bug to occur only on it. It may be a problem >>> in the nvme subsystem, or just some edge case in the bcachefs too, who >>> knows. I tried to disable ASPM and setting latency to 0 like was >>> suggested, it didn't fix the problem, so I don't know. If this is >>> indeed related to that specific drive it would be hard to reproduce. >> >> From a quick look, looks like a broken drive/firmware. It is suspicious >> that all failed IO is 256 blocks. You could try and limit the transfer >> size and see if that helps: >> >> # echo 64 > /sys/block/nvme0n1/queue/max_sectors_kb >> >> Or maybe the transfer size is just a red herring, who knows. The error >> code seems wonky: > > Does nvme have a blacklist/quirks mechanism, if that ends up resolving > it? It does, in fact this drive is already marked as having broken suspend. So not too surprising if it's misbehaving in other ways too, I guess. -- Jens Axboe