On Tue 2020-06-09 17:56:03, John Ogness wrote: > On 2020-06-09, Petr Mladek <pmladek@xxxxxxxx> wrote: > >> --- /dev/null > >> +++ b/kernel/printk/printk_ringbuffer.c > >> +/* > >> + * Advance the desc ring tail. This function advances the tail by one > >> + * descriptor, thus invalidating the oldest descriptor. Before advancing > >> + * the tail, the tail descriptor is made reusable and all data blocks up to > >> + * and including the descriptor's data block are invalidated (i.e. the data > >> + * ring tail is pushed past the data block of the descriptor being made > >> + * reusable). > >> + */ > >> +static bool desc_push_tail(struct printk_ringbuffer *rb, > >> + unsigned long tail_id) > >> +{ > >> + struct prb_desc_ring *desc_ring = &rb->desc_ring; > >> + enum desc_state d_state; > >> + struct prb_desc desc; > >> + > >> + d_state = desc_read(desc_ring, tail_id, &desc); > >> + > >> + switch (d_state) { > >> + case desc_miss: > >> + /* > >> + * If the ID is exactly 1 wrap behind the expected, it is > >> + * in the process of being reserved by another writer and > >> + * must be considered reserved. > >> + */ > >> + if (DESC_ID(atomic_long_read(&desc.state_var)) == > >> + DESC_ID_PREV_WRAP(desc_ring, tail_id)) { > >> + return false; > >> + } > >> + > >> + /* > >> + * The ID has changed. Another writer must have pushed the > >> + * tail and recycled the descriptor already. Success is > >> + * returned because the caller is only interested in the > >> + * specified tail being pushed, which it was. > >> + */ > >> + return true; > >> + case desc_reserved: > >> + return false; > >> + case desc_committed: > >> + desc_make_reusable(desc_ring, tail_id); > >> + break; > >> + case desc_reusable: > >> + break; > >> + } > >> + > >> + /* > >> + * Data blocks must be invalidated before their associated > >> + * descriptor can be made available for recycling. Invalidating > >> + * them later is not possible because there is no way to trust > >> + * data blocks once their associated descriptor is gone. > >> + */ > >> + > >> + if (!data_push_tail(rb, &rb->text_data_ring, desc.text_blk_lpos.next)) > >> + return false; > >> + if (!data_push_tail(rb, &rb->dict_data_ring, desc.dict_blk_lpos.next)) > >> + return false; > >> + > >> + /* > >> + * Check the next descriptor after @tail_id before pushing the tail > >> + * to it because the tail must always be in a committed or reusable > >> + * state. The implementation of prb_first_seq() relies on this. > >> + * > >> + * A successful read implies that the next descriptor is less than or > >> + * equal to @head_id so there is no risk of pushing the tail past the > >> + * head. > >> + */ > >> + d_state = desc_read(desc_ring, DESC_ID(tail_id + 1), > >> + &desc); /* LMM(desc_push_tail:A) */ > >> + if (d_state == desc_committed || d_state == desc_reusable) { > >> + /* > >> + * Any CPU that loads the new tail ID, must see that the > >> + * descriptor at @tail_id is in the reusable state. See the > >> + * read memory barrier part of desc_reserve:D for details. > >> + */ > >> + atomic_long_cmpxchg_relaxed(&desc_ring->tail_id, tail_id, > >> + DESC_ID(tail_id + 1)); /* LMM(desc_push_tail:B) */ > > > > I was quite confused by the above comment. Does it mean that we need > > a barrier here? Or does it explain why the cmpxchg has its own > > LMM marker? > > This LMM marker is referenced quite often, but since it is a relaxed > cmpxchg(), its significance is not immediately clear. I was hoping to > add some hints as to why it is significant. The comment that it is > referring to is: > > /* > * Guarantee the tail ID is read before validating the > * recycled descriptor state. A read memory barrier is > * sufficient for this. This pairs with data_push_tail:C. > * > * Memory barrier involvement: > * > * If desc_reserve:C reads from desc_push_tail:B, then > * desc_reserve:F reads from desc_make_reusable:A. > * > * Relies on: > * > * MB from desc_make_reusable:A to desc_push_tail:B > * matching > * RMB from desc_reserve:C to desc_reserve:F > * > * Note: desc_make_reusable:A, desc_push_tail:B, and > * data_push_tail:C can all be different CPUs. However, > * the desc_push_tail:B CPU must have previously seen > * data_push_tail:D and the data_push_tail:D CPU (which > * performs the full memory barrier) must have > * previously seen desc_make_reusable:A. > */ > > English translation: > > In order to push the data tail, a CPU must first see that the associated > descriptor is in the reusable state. Since a full memory barrier is > performed after that sighting and before doing the data tail push, _any_ > CPU that sees the pushed data tail will be able to see that the > associated descriptor is in the reusable state. > > In order to push the descriptor tail, a CPU must first see that the > associated data tail has been pushed. Therefore, that CPU would also see > that the associated descriptor is in the reusable state. Thanks a lot for this detailed description. It helped a lot. Let me try another description from slightly different angle: All this relies on the fact the the full barrier is called in data_push_tail() and data_push_tail() is called right above. But there are two situations where the barrier is not called. It is when: 1. desc.text_blk_lpos.next already is behind data_ring->tail_lpos. This is safe. It might happen when there was a race in the past. CPU1 reserved a descriptor before CPU2 and CPU2 was able to allocate data block before CPU1. As a result, both descriptors and both data blocks were moved into reusable state when the earlier descriptor was reused. It is because it pointed to newer data block and the older data block must have been invalidated together with the newer descriptor. Now, the full barrier was called before tail_lpos was moved. Both descriptors must have been in the reusable state already. 2. desc.text_blk_lpos == INVALID_LPOS. It seems that this is not synchronized and other CPUs might see the old state. It happens for data blocks that do not have any data. So it probably does not cause real problems but ... > > I think that we actually need a full barrier here to make sure that > > all CPUs see the changes made by data_push_tail() before we > > allow to rewrite the descriptor. The changes in data_push_tail() might > > be done on different CPUs. > > How so? That memory barrier exists to make sure the reusable descriptor > state is stored before pushing the data tail. This is important for > readers (which start from the data tail) so they can notice if the > descriptor has since been invalidated (reusable state). > > But where is it important that the data tail change is seen before the > descriptor tail change? How are the data tail and descriptor tail > significantly related to each other? I have to admit that I did not think about it deeply enough. It was more about feeling and seeing similar pattern. You have a point. The state value of the descriptor is the central point. It is used to synchronize operations with both desc_ring and data_ring. The state is modified only once. So one (full) memory barrier should be enough to synchronize all CPUs. The state has to be modified before the data block could be reused. Therefore the full barrier has to be already in the data_ring code. The question is what to do with the empty data case. I see three possibilities: 1. Ignore the case with empty block because it (probably) does not cause real problems. 2. Call the full barrier in data_push_tail() even when the data block is empty. 3. Call the full barrier also in desc_push_tail() as I suggested. My opinion: I prefer 3rd solution. The barrier would be superfluous in the most common situation. But it would create more error-proof code. We could always optimize it when it causes problems. Anyway, I would feel very uneasy about the the 1st solution. And the 2nd solution is weird. It would be hard to explain. Best Regards, Petr PS: I start feeling more confident about the code. The more barriers the less possible races ;-) _______________________________________________ kexec mailing list kexec@xxxxxxxxxxxxxxxxxxx http://lists.infradead.org/mailman/listinfo/kexec