On 11/10/21 at 12:21pm, David Hildenbrand wrote: > On 10.11.21 12:11, Dave Young wrote: > > Hi David, > > On 11/10/21 at 09:10am, David Hildenbrand wrote: > >> On 10.11.21 08:22, Baoquan He wrote: > >>> On 11/08/21 at 06:31pm, Andrew Morton wrote: > >>>> From: David Hildenbrand <david@xxxxxxxxxx> > >>>> Subject: proc/vmcore: convert oldmem_pfn_is_ram callback to more generic vmcore callbacks > >>>> > >>>> Let's support multiple registered callbacks, making sure that registering > >>>> vmcore callbacks cannot fail. Make the callback return a bool instead of > >>>> an int, handling how to deal with errors internally. Drop unused > >>>> HAVE_OLDMEM_PFN_IS_RAM. > >>>> > >>>> We soon want to make use of this infrastructure from other drivers: > >>>> virtio-mem, registering one callback for each virtio-mem device, to > >>>> prevent reading unplugged virtio-mem memory. > >>>> > >>>> Handle it via a generic vmcore_cb structure, prepared for future > >>>> extensions: for example, once we support virtio-mem on s390x where the > >>>> vmcore is completely constructed in the second kernel, we want to detect > >>>> and add plugged virtio-mem memory ranges to the vmcore in order for them > >>>> to get dumped properly. > >>>> > >>>> Handle corner cases that are unexpected and shouldn't happen in sane > >>>> setups: registering a callback after the vmcore has already been opened > >>>> (warn only) and unregistering a callback after the vmcore has already been > >>>> opened (warn and essentially read only zeroes from that point on). > >>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > >>> > >>> I am fine with the whole patch except of one concern. As above sentence > >>> underscored states, if a callback is unregistered when vmcore has been > >>> opened, it will read out zeros from that point on. And it's done by > >>> judging global variable 'vmcore_cb_unstable' in pfn_is_ram(). This will > >>> cause vmcore dumping in makedumpfile only being able to read out zero > >>> page since then, and may cost long extra time to finish. > >>> > >>> Please see remap_oldmem_pfn_checked(). In makedumpfile, we default to > >>> mmap 4M memory region at one time, then copy out. With this patch, and if > >>> vmcore_cb_unstable is true, kernel will mmap page by page. The extra > >>> time could be huge, e.g on machine with TBs memory, and we only get a > >>> useless vmcore because of loss of core data with high probability. > >> > >> Thanks Baoquan for the quick review! > >> > >> This code is really just to handle the unlikely case of a driver getting > >> unbound from a device that has a callback registered (e.g., a > >> virtio-mem-pci device). Something like this will never happen in > >> practice in a *sane* environment. > >> > >> The only known way I know is if userspace manually unbinds the driver > >> from a virtio-mem-pci device -- which is possible but especially in a > >> kdump environment something without any sane use case. In that case, we'll > >> > >> pr_warn_once("Unexpected vmcore callback unregistration\n"); > >> > >> to let user space know that something weird/unsupported is going on. > >> > >> Long story short: if user space does something nasty, I don't see a > >> problem in some action taking a little longer. > >> > >> > >>> > >>> I am thinking if we can simply panic in the case, since the left dumping > >>> are all zeroed, very likely the vmcore is unavailable any more. > >> > >> IMHO panic() is a little bit too much. Instead of returning zeroes, we > >> could fail the read/mmap operation -- I considered that as an option > >> when I crafted/tested this patch, however, this approach here turned out > >> to be the easiest way to handle something that's really not > >> supported/advised and won't really happen in a sane environment. > > > > I would still say that the most important task for kdump is to save the > > vmcore successfully. Even the above issue is not a common case it could > > cause the vmcore to be useless. It is understandable if the zeroed part > > is only the virtio-mem part, but if all the remaining vmcore is zeroed > > that it is bad and not acceptable for kdump. > > Again, in a sane environment this will never happen. > > Why are we discussing on how to optimize a scenario where user space > does something that's clearly unsupported and will not happen in real life? > > My take is to warn and fail as simple as possible, without hacking > around the issue (like blocking driver unloading while user space has > /proc/vmcore opened. > > "remaining vmcore is zeroed that it is bad and not acceptable for kdump." > > Which scenario are you concerned about? User space plays stupid games > (unbining a driver from a virtio-mem device in a *kdump kernel* after > opening /proc/vmcore) and wins stupid prices (a warning and a vmcore > filled (partially) with zeroes). Why isn't a warning sufficient for > something like that? Hi David, Suppose we have the use case below: A user plays with the game (Probably in hypervisor part, but the user is not aware that the guest panicked and in a kdump kernel), then we get a zeroed vmcore. But the panic can not be easily reproduced any more, then the warning is not useful. But if you think user is playing the game in kdump kernel, eg. in guest os while kdump is saving vmcore then it is nearly not possible to happen I agree with you it is a very trival problem. Probably we have some misunderstanding, but it would be good to make it clear :) > > I appreciate all the feedback (even if it comes in late :) ), but I'm > missing why we are trying to optimize something here. > > I'm happy to send a patch that does whatever we decide to do, but I > really don't see the need for a change. Most probably I'm missing > something important? > > (the patch landed mainline in the meantime) > > -- > Thanks, > > David / dhildenb > Thanks Dave