On Fri, 2008-03-14 at 13:45 +0100, Thomas Graf wrote: > * Ian Kent <raven@xxxxxxxxxx> 2008-03-13 16:00 > > The function that is a problem is the sending of expire requests. In the > > current implementation this function is synchronous. An ioctl is used to > > ask the kernel module (autofs4) to check for mounts that can be expired > > and, if a candidate is found the module sends a request to the user space > > daemon asking it to try and umount the select mount after which the daemon > > sends a success or fail status back to the module which marks the > > completion of the original ioctl expire request. > > > > The Generic Netlink interface won't allow this because only one concurrent > > send request can be active for "all" Generic Netlink Families in use, > > since the socket receive function is bracketed by a single mutex. So I would > > need to use a workqueue to queue the request but that has it's own set of > > problems. > > > > The next issue is that in order to keep track of multiple in flight > > requests a separate Netlink socket would need to be opened for every > > expire request in order to ensure that the Netlink completion reply makes > > it back to the original requesting thread (Is that actually correct?). Not > > really such a big problem but it defeats another aim of the > > re-implementation, which is to reduce the selinux user space exposure to > > file descriptors that are open but don't yet have close-on-exec flag set > > when a mount or umount is spawned by the automount daemon. This can > > obviously be resolved by adding a mutex around the fork/exec code but > > isn't a popular idea due to added performance overhead. > > Netlink is a messaging protocol, synchronization is up to the user. Yes, I realize that, but what I'm curious about are the options that I have within the messaging system to control delivery of message replies, other than using separate sockets. Can this be achieved by using the pid set in the originating message? > > I suggest that you send a netlink notification to a multicast group for > every expire candiate when an expire request is received. Unmount > daemons simply subscribe to the group and wait for work to do. Put the > request onto a list including the netlink pid and sequence number so you > can address the orignal source of the request later on. Exit the netlink > receive function and wait for the userspace daemon to get back to you. I'll have to think about what you've said here to relate it to the situation I have. I don't actually have umount daemons, at the moment I request an expire and the daemon creates a thread to do the umount and sends a status message to the kernel module. But that may not matter, see below. > > The userspace daemon notifies you of successful or unsuccesful unmount > attempts by sending notifications. Update your list entry accordingly > and once the request is fullfilled send a notification to the original > source of the request by using the stored pid and sequence number. > > The userspace application requesting the expire can simply block on the > receival of this notification in order to make the whole operation > synchronous. Actually, I've progressed on this since posting. I've implemented the first steps toward using a workqueue to perform the expire and, to my surprise, my code worked for a simple test case. Basically, a thread in the daemon issues the expire, the kernel module queues the work and replies. The expire workqueue task does the expire check and if no candidates are found it sends an expire complete notification, or it sends a umount request to the daemon and waits for the status result, then returns that result as the expire compete notification. Seems to work quite well. I expect this is possibly the method you're suggesting above anyway. Unfortunately, having now stepped up intensity of the testing, I'm getting a hard hang on my system. I've setup to reduce the message functions used to only two simple notification messages to the kernel module to ensure it isn't the expire implementation causing the problem. It's hard to see where I could have messed these two functions as they are essentially re-entrant but there is fairly heavy mount activity of about 10-15 mounts a second. Such is life! Any ideas on what might be causing this? Ian -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html