Hi Pavel, On 15/08/2019 21:09, Pavel Tatashin wrote: >>> Also, I'd appreciate if anyone could test this series on vhe hardware >>> with vhe kernel, it does not look like QEMU can emulate it yet >> >> This locks up during resume from hibernate on my AMD Seattle, a regular v8.0 machine. > > Thanks for reporting a bug I will root cause and fix it. >> Please try and build the series to reduce review time. What you have here is an all-new >> page-table generation API, which you switch hibernate and kexec too. This is effectively a >> new implementation of hibernate and kexec. There are three things here that need review. >> >> You have a regression in your all-new implementation of hibernate. It took six months (and >> lots of review) to get the existing code right, please don't rip it out if there is >> nothing wrong with it. > >> Instead, please just move the hibernate copy_page_tables() code, and then wire kexec up. >> You shouldn't need to change anything in the copy_page_tables() code as the linear map is >> the same in both cases. > It is not really an all-new implementation of hibernate (for kexec it > is true though). I used the current implementation of hibernate as > bases, and simply generalized the functions by providing a flexible > interface. So what you are asking is actually exactly what I am doing. I disagree. The resume page-table code is the bulk of the complexity in hibernate.c. Your first patch dumps ~200 lines of differently-complex code, and your second switches hibernate over to it. Instead, please move that code, keeping it as it is. git will spot the move, and the generated diffstat should only reflect the build-system changes. You don't need to 'switch hibernate to transitional page tables.' Adding kexec will then show-up what needs changing, each change comes with a commit message explaining why. Having these as 'generalisations' in the first patch is a mess. There is existing code that we don't want to break. Any changes need to be done as a sequence of small incremental changes. It can't be reviewed any other way. > I realize, that I introduced a bug that I will fix. Done as a sequence of small incremental changes, I could bisect it to the patch that introduces the bug, and probably fix it from the description in the commit message. >> It looks like you are creating the page tables just after the kexec:segments have been >> loaded. This will go horribly wrong if anything changes between then and kexec time. (e.g. >> memory you've got mapped gets hot-removed). >> This needs to be done as late as possible, so we don't waste memory, and the world can't >> change around us. Reboot notifiers run before kexec, can't we do the memory-allocation there? > Kexec by design does not allow allocate during kexec time. This is > because we cannot fail during kexec syscall. This problem needs solving. | Reboot notifiers run before kexec, can't we do the memory-allocation there? > All allocations must be done during kexec load time. This increases the memory footprint. I don't think we should waste ~2MB per GB of kernel memory on this feature. (Assuming 4K pages and rodata_full) Another option is to allocate this memory at load time, but then free it so it can be used in the meantime. You can keep the list of allocated pfn, as we know they aren't in use by the running kernel, kexec metadata, loaded images etc. Memory hotplug would need handling carefully, as would anything that 'donates' memory to another agent. (I suspect the TEE stuff does this, I don't know how it interacts with kexec) > Kernel memory cannot be hot-removed, as > it is not part of ZONE_MOVABLE, and cannot be migrated. Today, yes. Tomorrow?, "arm64/mm: Enable memory hot remove": https://lore.kernel.org/r/1563171470-3117-1-git-send-email-anshuman.khandual@xxxxxxx >>>> Previously: >>>> kernel shutdown 0.022131328s >>>> relocation 0.440510736s >>>> kernel startup 0.294706768s >>>> >>>> Relocation was taking: 58.2% of reboot time >>>> >>>> Now: >>>> kernel shutdown 0.032066576s >>>> relocation 0.022158152s >>>> kernel startup 0.296055880s >>>> >>>> Now: Relocation takes 6.3% of reboot time >>>> >>>> Total reboot is x2.16 times faster. >> >> When I first saw these numbers they were ~'0.29s', which I wrongly assumed was 29 seconds. >> Savings in milliseconds, for _reboot_ is a hard sell. I'm hoping that on the machines that >> take minutes to kexec we'll get numbers that make this change more convincing. > Sure, this userland is very small kernel+userland is only 47M. Here is > another data point: fitImage: 380M, it contains a larger userland. > The numbers for kernel shutdown and startup are the same as this is > the same kernel, but relocation takes: 3.58s > shutdown: 0.02s > relocation: 3.58s > startup: 0.30s > > Relocation take 88% of reboot time. And, we must have it under one second. Where does this one second number come from? (was it ever a reasonable starting point?) Thanks, James