On Mon, Jun 21, 2010 at 10:09:24AM -0700, Michael Rubin wrote: > Thanks for looking at this. > > On Sun, Jun 20, 2010 at 4:10 PM, Dave Chinner <david@xxxxxxxxxxxxx> wrote: > >> Michael Rubin (3): > >> writeback: Creating /sys/kernel/mm/writeback/writeback > >> writeback: per bdi monitoring > >> writeback: tracking subsystems causing writeback > > > > I'm not sure we want to export statistics that represent internal > > implementation details into a fixed userspace API. Who, other than > > developers, are going to understand and be able to make use of this > > information? > > I think there are varying degrees of internal exposure on the patches. > > >> writeback: Creating /sys/kernel/mm/writeback/writeback > This one seems to not expose any new internals. We already expose the > concept of "dirty", "writeback" and thresholds in /proc/meminfo. I don't see any probems with these stats - no matter the implementation, they'll still be relevant. > >> writeback: per bdi monitoring > > Looking at it again. I think this one is somewhat of a mixed bag. > BDIReclaimable, BdiWriteback, and the dirty thresholds seems safe to > export.While I agree the rest should stay in debugfs. Would that be > amenable? I'd much prefer all the bdi stats in the one spot. It's hard enough to find what you're looking for without splitting them into multiple locations. The other thing to consider is that tracing requires debugfѕ to be mounted. Hence most kernels are going to have the debug stats available, anyway.... > >> writeback: tracking subsystems causing writeback > > I definitely agree that this one is too revealing and needs to be > redone. But I think we might want to add the details for concepts > which we already expose. > The idea of a "periodic writeback" is already exposed in /proc/sys/vm/ > and I don't see that changing in the kernel as a method to deal with > buffered IO. Neither will sync. I don't see much value in exposing this information outside of development environments. I think it's much better to add trace points for events like this so that we do fine-grained analysis of when the events occur during problematic workloads.... > The laptop stuff and the names of > "balance_dirty_pages" are bad, but maybe we can come up with something > more high level. Like "writeback due to low memory" > > > FWIW, I've got to resend the writeback tracing patches to Jens that I > > have that give better visibility into the writeback behaviour. > > Perhaps those tracing events are a better basis for tracking down > > writeback problems - the bugs I found with the tracing could not > > have been found with these statistics... > > Yeah I have been watching the tracing stuff you have posted and I > think it will help. There were some other trace points I wanted to add > to this patch but was waiting to learn from your submission on the > best way to integrate them. I've got more work to do on them first.... :/ > > That's really why I'm asking - if the stats are just there to help > > development and debugging, then I think that improving the writeback > > tracing is a better approach to improving visibility of writeback > > behaviour... > > Maybe I should not have put all these patches in one series. The first > one with the /sys/kernel/vm file is very useful for user space > developers. System Administrators who are trying to classify IO > problems often need to know if the disk is bad These stats aren't the place for observing that a disk is bad ;) > or if the buffered data > is not even being written to disk over time.. I think this can be obtained from the existing info in /proc/meminfo. I'm not saying the new stats aren't necessary, just that we already have the high level information available for this.... > Also at Google we tend > to run our jobs with very little unused RAM. Pushing things close to > their limits results in many surprises and writeback is often one of > them. Knowing the thresholds and rate of dirty and cleaning of pages > can help systems do the right thing. Yes, I hear this all the time from appliance developers that cache everything they need in userspace - they just want the kernel to stay out of the way and not use the unused RAM for caching stuff that doesn't matter to the application. Normally the issue is unbounded growth of the inode and dentry caches, but I can see how exceeding writeback limits can be just as much of a problem. Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html