On Thu, Mar 09, 2023 at 08:54:39AM +1100, Dave Chinner wrote: > On Wed, Mar 08, 2023 at 06:12:06PM +0100, Jan Kara wrote: > > Hi! > > > > I'm interested in this topic. Some comments below. > > > > On Tue 28-02-23 12:49:03, Darrick J. Wong wrote: > > > Five years ago[0], we started a conversation about cross-filesystem > > > userspace tooling for online fsck. I think enough time has passed for > > > us to have another one, since a few things have happened since then: > > > > > > 1. ext4 has gained the ability to send corruption reports to a userspace > > > monitoring program via fsnotify. Thanks, Collabora! > > > > > > 2. XFS now tracks successful scrubs and corruptions seen during runtime > > > and during scrubs. Userspace can query this information. > > > > > > 3. Directory parent pointers, which enable online repair of the > > > directory tree, is nearing completion. > > > > > > 4. Dave and I are working on merging online repair of space metadata for > > > XFS. Online repair of directory trees is feature complete, but we > > > still have one or two unresolved questions in the parent pointer > > > code. > > > > > > 5. I've gotten a bit better[1] at writing systemd service descriptions > > > for scheduling and performing background online fsck. > > > > > > Now that fsnotify_sb_error exists as a result of (1), I think we > > > should figure out how to plumb calls into the readahead and writeback > > > code so that IO failures can be reported to the fsnotify monitor. I > > > suspect there may be a few difficulties here since fsnotify (iirc) > > > allocates memory and takes locks. > > > > Well, if you want to generate fsnotify events from an interrupt handler, > > you're going to have a hard time, I don't have a good answer for that. > > I don't think we ever do that, or need to do that. IO completions > that can throw corruption errors are already running in workqueue > contexts in XFS. > > Worst case, we throw all bios that have IO errors flagged to the > same IO completion workqueues, and the problem of memory allocation, > locks, etc in interrupt context goes away entire. Indeed. For XFS I think the only time we might need to fsnotify about errors from interrupt context is writeback completions for a pure overwrite? We could punt those to a workqueue as Dave says. Or figure out a way for whoever's initiating writeback to send it for us? I think this is a general issue for the pagecache, not XFS. I'll brainstorm with willy the next time I encounter him. > > But > > offloading of error event generation to a workqueue should be doable (and > > event delivery is async anyway so from userspace POV there's no > > difference). > > Unless I'm misunderstanding you (possible!), that requires a memory > allocation to offload the error information to the work queue to > allow the fsnotify error message to be generated in an async manner. > That doesn't seem to solve anything. > > > Otherwise locking shouldn't be a problem AFAICT. WRT memory > > allocation, we currently preallocate the error events to avoid the loss of > > event due to ENOMEM. With current usecases (filesystem catastrophical error > > reporting) we have settled on a mempool with 32 preallocated events (note > > that preallocated event gets used only if normal kmalloc fails) for > > simplicity. If the error reporting mechanism is going to be used > > significantly more, we may need to reconsider this but it should be doable. > > And frankly if you have a storm of fs errors *and* the system is going > > ENOMEM at the same time, I have my doubts loosing some error report is > > going to do any more harm ;). > > Once the filesystem is shut down, it will need to turn off > individual sickness notifications because everything is sick at this > point. I was thinking that the existing fsnotify error set should adopt a 'YOUR FS IS DEAD' notification. Then when the fs goes down due to errors or the shutdown ioctl, we can broadcast that as the final last gasp of the filesystem. > > > As a result of (2), XFS now retains quite a bit of incore state about > > > its own health. The structure that fsnotify gives to userspace is very > > > generic (superblock, inode, errno, errno count). How might XFS export > > > a greater amount of information via this interface? We can provide > > > details at finer granularity -- for example, a specific data structure > > > under an allocation group or an inode, or specific quota records. > > > > Fsnotify (fanotify in fact) interface is fairly flexible in what can be > > passed through it. So if you need to pass some (reasonably short) binary > > blob to userspace which knows how to decode it, fanotify can handle that > > (with some wrapping). Obviously there's a tradeoff to make how much of the > > event is generic (as that is then easier to process by tools common for all > > filesystems) and how much is fs specific (which allows to pass more > > detailed information). But I guess we need to have concrete examples of > > events to discuss this. > > Fine grained health information will always be filesystem specific - > IMO it's not worth trying to make it generic when there is only one > filesystem that tracking and exporting fine-grained health > information. Once (if) we get multiple filesystems tracking fine > grained health information, then we'll have the information we need > to implement a useful generic set of notifications, but until then I > don't think we should try. Same here. XFS might want to send the generic notifications and follow them up with more specific information? > We should just export the notifications the filesystem utilities > need to do their work for the moment. When management applications > (e.g Stratis) get to the point where they can report/manage > filesystem health and need that information from multiple > filesystems types, then we can work out a useful common subset of > fine grained events across those filesystems that the applications > can listen for. If someone wants to write xfs_scrubd that listens for events and issues XFS_IOC_SCRUB_METADATA calls I'd be all ears. :) --D > Cheers, > > Dave. > -- > Dave Chinner > david@xxxxxxxxxxxxx