On Fri, Jun 17, 2016 at 03:55:56PM -0400, Chuck Lever wrote: > > > On Jun 17, 2016, at 5:20 AM, Leon Romanovsky <leon@xxxxxxxxxx> wrote: > > > > On Thu, Jun 16, 2016 at 05:58:29PM -0400, Chuck Lever wrote: > >> > >>> On Jun 16, 2016, at 5:10 PM, Sagi Grimberg <sagigrim@xxxxxxxxx> wrote: > >> > >>> First of all, IIRC the patch author was Christoph wasn't he. > >>> > >>> Plus, you do realize that this patch makes the pages allocation > >>> in granularity of pages. In systems with a large page size this > >>> is completely redundant, it might even be harmful as the storage > >>> ULPs need lots of MRs. > >> > >> I agree that the official fix should take a conservative > >> approach to allocating this resource; there will be lots > >> of MRs in an active system. This fix doesn't seem too > >> careful. > > > > In mlx5 system, we always added 2048 bytes to such allocations, for > > reasons unknown to me. And it doesn't seem as a conservative approach > > either. > > The mlx5 approach is much better than allocating a whole > page, when you consider platforms with 64KB pages. > > A 1MB payload (for NFS) on such a platform comprises just > 16 pages. So xprtrdma will allocate MRs with support for > 16 pages. That's a priv pages array of 128 bytes, and you > just put it in a 64KB page all by itself. > > So maybe adding 2048 bytes is not optimal either. But I > think sticking with kmalloc here is a more optimal choice. I agree with your's and Sagi's points, just preferred working solution over optimal. I'll send an optimal version. > > > >>> Also, I don't see how that solves the issue, I'm not sure I even > >>> understand the issue. Do you? Were you able to reproduce it? > >> > >> The issue is that dma_map_single() does not seem to DMA map > >> portions of a memory region that are past the end of the first > >> page of that region. Maybe that's a bug? > > > > No, I didn't find support for that. Function dma_map_single expects > > contiguous memory aligned to cache line, there is no limitation to be > > page bounded. > > There certainly isn't, but that doesn't mean there can't > be a bug somewhere ;-) and maybe not in dma_map_single. > It could be that the "array on one page only" limitation > is somewhere else in the mlx4 driver, or even in the HCA > firmware. We checked with HW/FW/arch teams prior to respond. > > > >> This patch works around that behavior by guaranteeing that > >> > >> a) the memory region starts at the beginning of a page, and > >> b) the memory region is never larger than a page > > > > b) the memory region ends on cache line. > > I think we demonstrated pretty clearly that the issue > occurs only when the end of the priv pages array crosses > into a new page. > > We didn't see any problem otherwise. SLUB debug do exactly one thing, change alignment and this is why no issue there observed before. > > >> This patch is not sufficient to repair mlx5, because b) > >> cannot be satisfied in that case; the array of __be64's can > >> be larger than 512 entries. > >> > >> > >>> IFF the pages buffer end not being aligned to a cacheline is problematic > >>> then why not extent it to end in a cacheline? Why in the next full page? > >> > >> I think the patch description justifies the choice of > >> solution, but does not describe the original issue at > >> all. The original issue had nothing to do with cacheline > >> alignment. > > > > I disagree, kmalloc with supplied flags will return contiguous memory > > which is enough for dma_map_single. It is cache line alignment. > > The reason I find this hard to believe is that there is > no end alignment guarantee at all in this code, but it > works without issue when SLUB debugging is not enabled. > > xprtrdma allocates 256 elements in this array on x86. > The code makes the array start on an 0x40 byte boundary. > I'm pretty sure that means the end of that array will > also be on at least an 0x40 byte boundary, and thus > aligned to the DMA cacheline, whether or not SLUB > debugging is enabled. > > Notice that in the current code, if the consumer requests > an odd number of SGs, that array can't possibly end on > an alignment boundary. But we've never had a complaint. > > SLUB debugging changes the alignment of lots of things, > but mlx4_alloc_priv_pages is the only breakage that has > been reported. I think it is related to custom logic which is in this function only. I posted grep output earlier to emphasize it. For example adds 2K region after the actual data and it will ensure alignment. > > DMA-API.txt says: > > > [T]he mapped region must begin exactly on a cache line > > boundary and end exactly on one (to prevent two separately > > mapped regions from sharing a single cache line) > > The way I read this, cacheline alignment shouldn't be > an issue at all, as long as DMA cachelines aren't > shared between mappings. > > If I simply increase the memory allocation size a little > and ensure the end of the mapping is aligned, that should > be enough to prevent DMA cacheline sharing with another > memory allocation on the same page. But I still see Local > Protection Errors when SLUB debugging is enabled, on my > system (with patches to allocate more pages per MR). > > I'm not convinced this has anything to do with DMA > cacheline alignment. The reason your patch fixes this > issue is because it keeps the entire array on one page. If you don't mind, we can do an experiment. Let's add padding which will prevent alignment issue and for sure will cross the page boundary. diff --git a/drivers/infiniband/hw/mlx4/mr.c b/drivers/infiniband/hw/mlx4/mr.c index 6312721..41e277e 100644 --- a/drivers/infiniband/hw/mlx4/mr.c +++ b/drivers/infiniband/hw/mlx4/mr.c @@ -280,8 +280,10 @@ mlx4_alloc_priv_pages(struct ib_device *device, int size = max_pages * sizeof(u64); int add_size; int ret; - +/* add_size = max_t(int, MLX4_MR_PAGES_ALIGN - ARCH_KMALLOC_MINALIGN, 0); +*/ + add_size = 2048; mr->pages_alloc = kzalloc(size + add_size, GFP_KERNEL); if (!mr->pages_alloc)
Attachment:
signature.asc
Description: Digital signature