On Mon, Dec 03, 2012 at 12:36:33PM +0400, Glauber Costa wrote: > On 11/30/2012 09:55 PM, Pavel Emelyanov wrote: > > Hello, > > > > This is an attempt to implement support for memory snapshot for the the > > checkpoint-restore project (http://criu.org). > > > > To create a dump of an application(s) we save all the information about it > > to files. No surprise, the biggest part of such dump is the contents of tasks' > > memory. However, in some usage scenarios it's not required to get _all_ the > > task memory while creating a dump. For example, when doing periodical dumps > > it's only required to take full memory dump only at the first step and then > > take incremental changes of memory. Another example is live migration. In the > > simplest form it looks like -- create dump, copy it on the remote node then > > restore tasks from dump files. While all this dump-copy-restore thing goes all > > the process must be stopped. However, if we can monitor how tasks change their > > memory, we can dump and copy it in smaller chunks, periodically updating it > > and thus freezing tasks only at the very end for the very short time to pick > > up the recent changes. > > > > That said, some help from kernel to watch how processes modify the contents of > > their memory is required. I'd like to propose one possible solution of this > > task -- with the help of page-faults and trace events. > > > > Briefly the approach is -- remap some memory regions as read-only, get the #pf > > on task's attempt to modify the memory and issue a trace event of that. Since > > we're only interested in parts of memory of some tasks, make it possible to mark > > the vmas we're interested in and issue events for them only. Also, to be aware > > of tasks unmapping the vma-s being watched, also issue an event when the marked > > vma is removed (and for symmetry -- an event when a vma is marked). > > > > What do you think about this approach? Is this way of supporting mem snapshot > > OK for you, or should we invent some better one? > > > > The page fault mechanism is pretty obvious - anything that deals with > dirty pages will end up having to do this. So there is nothing crazy > about this. > > What concerns me, however, is that should this go in, we'll have two > dirty mem loggers in the kernel: one to support CRIU, one to support > KVM. And the worst part: They have the exact the same purpose!! > > So to begin with, I think one thing to consider, would be to generalize > KVM's dirty memory notification so it can work on a normal process > memory region. KVM api requires a "memory slot" to be passed, something > we are unlikely to have. But KVM can easily keep its API and use an > alternate mechanics, that's trivial... > > Generally speaking, KVM will do polling with this ioctl. I prefer your > tracing mechanism better. The only difference, is that KVM tends to > transfer large chunks of memory in some loads - in the high gigs range. > So the proposal tracing API should be able to optionally batch requests > within a time frame. > > It would also be good to hear what does the KVM guys think of it as well There are significant differences. KVM's dirty logging works for guest translations (NPT/shadow) and is optimized for specific use cases. Above is about dirty logging of userspace memory areas. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html