On Wed, Nov 16, 2011 at 11:57:37AM -0500, j.glisse@xxxxxxxxx wrote: > From: Jerome Glisse <jglisse@xxxxxxxxxx> > > Move dma data to a superset ttm_dma_tt structure which herit > from ttm_tt. This allow driver that don't use dma functionalities > to not have to waste memory for it. > > V2 Rebase on top of no memory account changes (where/when is my > delorean when i need it ?) > V3 Make sure page list is initialized empty > V4 typo/syntax fixes > > Signed-off-by: Jerome Glisse <jglisse@xxxxxxxxxx> > Reviewed-by: Thomas Hellstrom <thellstrom@xxxxxxxxxx> Reviewed-by: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx> > --- > drivers/gpu/drm/nouveau/nouveau_bo.c | 18 +++-- > drivers/gpu/drm/nouveau/nouveau_sgdma.c | 22 ++++-- > drivers/gpu/drm/radeon/radeon_ttm.c | 43 ++++++------ > drivers/gpu/drm/ttm/ttm_page_alloc.c | 114 +++++++++++++++--------------- > drivers/gpu/drm/ttm/ttm_page_alloc_dma.c | 35 +++++---- > drivers/gpu/drm/ttm/ttm_tt.c | 60 +++++++++++++--- > drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c | 2 + > include/drm/ttm/ttm_bo_driver.h | 32 ++++++++- > include/drm/ttm/ttm_page_alloc.h | 33 +-------- > 9 files changed, 203 insertions(+), 156 deletions(-) > > diff --git a/drivers/gpu/drm/nouveau/nouveau_bo.c b/drivers/gpu/drm/nouveau/nouveau_bo.c > index e603909..4347776 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_bo.c > +++ b/drivers/gpu/drm/nouveau/nouveau_bo.c > @@ -1052,6 +1052,7 @@ nouveau_bo_fence(struct nouveau_bo *nvbo, struct nouveau_fence *fence) > static int > nouveau_ttm_tt_populate(struct ttm_tt *ttm) > { > + struct ttm_dma_tt *ttm_dma = (void *)ttm; > struct drm_nouveau_private *dev_priv; > struct drm_device *dev; > unsigned i; > @@ -1065,7 +1066,7 @@ nouveau_ttm_tt_populate(struct ttm_tt *ttm) > > #ifdef CONFIG_SWIOTLB > if (swiotlb_nr_tbl()) { > - return ttm_dma_populate(ttm, dev->dev); > + return ttm_dma_populate((void *)ttm, dev->dev); > } > #endif > > @@ -1075,14 +1076,14 @@ nouveau_ttm_tt_populate(struct ttm_tt *ttm) > } > > for (i = 0; i < ttm->num_pages; i++) { > - ttm->dma_address[i] = pci_map_page(dev->pdev, ttm->pages[i], > + ttm_dma->dma_address[i] = pci_map_page(dev->pdev, ttm->pages[i], > 0, PAGE_SIZE, > PCI_DMA_BIDIRECTIONAL); > - if (pci_dma_mapping_error(dev->pdev, ttm->dma_address[i])) { > + if (pci_dma_mapping_error(dev->pdev, ttm_dma->dma_address[i])) { > while (--i) { > - pci_unmap_page(dev->pdev, ttm->dma_address[i], > + pci_unmap_page(dev->pdev, ttm_dma->dma_address[i], > PAGE_SIZE, PCI_DMA_BIDIRECTIONAL); > - ttm->dma_address[i] = 0; > + ttm_dma->dma_address[i] = 0; > } > ttm_pool_unpopulate(ttm); > return -EFAULT; > @@ -1094,6 +1095,7 @@ nouveau_ttm_tt_populate(struct ttm_tt *ttm) > static void > nouveau_ttm_tt_unpopulate(struct ttm_tt *ttm) > { > + struct ttm_dma_tt *ttm_dma = (void *)ttm; > struct drm_nouveau_private *dev_priv; > struct drm_device *dev; > unsigned i; > @@ -1103,14 +1105,14 @@ nouveau_ttm_tt_unpopulate(struct ttm_tt *ttm) > > #ifdef CONFIG_SWIOTLB > if (swiotlb_nr_tbl()) { > - ttm_dma_unpopulate(ttm, dev->dev); > + ttm_dma_unpopulate((void *)ttm, dev->dev); > return; > } > #endif > > for (i = 0; i < ttm->num_pages; i++) { > - if (ttm->dma_address[i]) { > - pci_unmap_page(dev->pdev, ttm->dma_address[i], > + if (ttm_dma->dma_address[i]) { > + pci_unmap_page(dev->pdev, ttm_dma->dma_address[i], > PAGE_SIZE, PCI_DMA_BIDIRECTIONAL); > } > } > diff --git a/drivers/gpu/drm/nouveau/nouveau_sgdma.c b/drivers/gpu/drm/nouveau/nouveau_sgdma.c > index ee1eb7c..47f245e 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_sgdma.c > +++ b/drivers/gpu/drm/nouveau/nouveau_sgdma.c > @@ -8,7 +8,10 @@ > #define NV_CTXDMA_PAGE_MASK (NV_CTXDMA_PAGE_SIZE - 1) > > struct nouveau_sgdma_be { > - struct ttm_tt ttm; > + /* this has to be the first field so populate/unpopulated in > + * nouve_bo.c works properly, otherwise have to move them here > + */ > + struct ttm_dma_tt ttm; > struct drm_device *dev; > u64 offset; > }; > @@ -20,6 +23,7 @@ nouveau_sgdma_destroy(struct ttm_tt *ttm) > > if (ttm) { > NV_DEBUG(nvbe->dev, "\n"); > + ttm_dma_tt_fini(&nvbe->ttm); > kfree(nvbe); > } > } > @@ -38,7 +42,7 @@ nv04_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem) > nvbe->offset = mem->start << PAGE_SHIFT; > pte = (nvbe->offset >> NV_CTXDMA_PAGE_SHIFT) + 2; > for (i = 0; i < ttm->num_pages; i++) { > - dma_addr_t dma_offset = ttm->dma_address[i]; > + dma_addr_t dma_offset = nvbe->ttm.dma_address[i]; > uint32_t offset_l = lower_32_bits(dma_offset); > > for (j = 0; j < PAGE_SIZE / NV_CTXDMA_PAGE_SIZE; j++, pte++) { > @@ -97,7 +101,7 @@ nv41_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem) > struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm; > struct drm_nouveau_private *dev_priv = nvbe->dev->dev_private; > struct nouveau_gpuobj *pgt = dev_priv->gart_info.sg_ctxdma; > - dma_addr_t *list = ttm->dma_address; > + dma_addr_t *list = nvbe->ttm.dma_address; > u32 pte = mem->start << 2; > u32 cnt = ttm->num_pages; > > @@ -206,7 +210,7 @@ nv44_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem) > struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm; > struct drm_nouveau_private *dev_priv = nvbe->dev->dev_private; > struct nouveau_gpuobj *pgt = dev_priv->gart_info.sg_ctxdma; > - dma_addr_t *list = ttm->dma_address; > + dma_addr_t *list = nvbe->ttm.dma_address; > u32 pte = mem->start << 2, tmp[4]; > u32 cnt = ttm->num_pages; > int i; > @@ -282,10 +286,11 @@ static struct ttm_backend_func nv44_sgdma_backend = { > static int > nv50_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem) > { > + struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm; > struct nouveau_mem *node = mem->mm_node; > > /* noop: bound in move_notify() */ > - node->pages = ttm->dma_address; > + node->pages = nvbe->ttm.dma_address; > return 0; > } > > @@ -316,12 +321,13 @@ nouveau_sgdma_create_ttm(struct ttm_bo_device *bdev, > return NULL; > > nvbe->dev = dev; > - nvbe->ttm.func = dev_priv->gart_info.func; > + nvbe->ttm.ttm.func = dev_priv->gart_info.func; > > - if (ttm_tt_init(&nvbe->ttm, bdev, size, page_flags, dummy_read_page)) { > + if (ttm_dma_tt_init(&nvbe->ttm, bdev, size, page_flags, dummy_read_page)) { > + kfree(nvbe); > return NULL; > } > - return &nvbe->ttm; > + return &nvbe->ttm.ttm; > } > > int > diff --git a/drivers/gpu/drm/radeon/radeon_ttm.c b/drivers/gpu/drm/radeon/radeon_ttm.c > index f499b2c..e111a38 100644 > --- a/drivers/gpu/drm/radeon/radeon_ttm.c > +++ b/drivers/gpu/drm/radeon/radeon_ttm.c > @@ -501,7 +501,7 @@ static bool radeon_sync_obj_signaled(void *sync_obj, void *sync_arg) > * TTM backend functions. > */ > struct radeon_ttm_tt { > - struct ttm_tt ttm; > + struct ttm_dma_tt ttm; > struct radeon_device *rdev; > u64 offset; > }; > @@ -509,17 +509,16 @@ struct radeon_ttm_tt { > static int radeon_ttm_backend_bind(struct ttm_tt *ttm, > struct ttm_mem_reg *bo_mem) > { > - struct radeon_ttm_tt *gtt; > + struct radeon_ttm_tt *gtt = (void*)ttm; > int r; > > - gtt = container_of(ttm, struct radeon_ttm_tt, ttm); > gtt->offset = (unsigned long)(bo_mem->start << PAGE_SHIFT); > if (!ttm->num_pages) { > WARN(1, "nothing to bind %lu pages for mreg %p back %p!\n", > ttm->num_pages, bo_mem, ttm); > } > r = radeon_gart_bind(gtt->rdev, gtt->offset, > - ttm->num_pages, ttm->pages, ttm->dma_address); > + ttm->num_pages, ttm->pages, gtt->ttm.dma_address); > if (r) { > DRM_ERROR("failed to bind %lu pages at 0x%08X\n", > ttm->num_pages, (unsigned)gtt->offset); > @@ -530,18 +529,17 @@ static int radeon_ttm_backend_bind(struct ttm_tt *ttm, > > static int radeon_ttm_backend_unbind(struct ttm_tt *ttm) > { > - struct radeon_ttm_tt *gtt; > + struct radeon_ttm_tt *gtt = (void *)ttm; > > - gtt = container_of(ttm, struct radeon_ttm_tt, ttm); > radeon_gart_unbind(gtt->rdev, gtt->offset, ttm->num_pages); > return 0; > } > > static void radeon_ttm_backend_destroy(struct ttm_tt *ttm) > { > - struct radeon_ttm_tt *gtt; > + struct radeon_ttm_tt *gtt = (void *)ttm; > > - gtt = container_of(ttm, struct radeon_ttm_tt, ttm); > + ttm_dma_tt_fini(>t->ttm); > kfree(gtt); > } > > @@ -570,17 +568,19 @@ struct ttm_tt *radeon_ttm_tt_create(struct ttm_bo_device *bdev, > if (gtt == NULL) { > return NULL; > } > - gtt->ttm.func = &radeon_backend_func; > + gtt->ttm.ttm.func = &radeon_backend_func; > gtt->rdev = rdev; > - if (ttm_tt_init(>t->ttm, bdev, size, page_flags, dummy_read_page)) { > + if (ttm_dma_tt_init(>t->ttm, bdev, size, page_flags, dummy_read_page)) { > + kfree(gtt); > return NULL; > } > - return >t->ttm; > + return >t->ttm.ttm; > } > > static int radeon_ttm_tt_populate(struct ttm_tt *ttm) > { > struct radeon_device *rdev; > + struct radeon_ttm_tt *gtt = (void *)ttm; > unsigned i; > int r; > > @@ -591,7 +591,7 @@ static int radeon_ttm_tt_populate(struct ttm_tt *ttm) > > #ifdef CONFIG_SWIOTLB > if (swiotlb_nr_tbl()) { > - return ttm_dma_populate(ttm, rdev->dev); > + return ttm_dma_populate(>t->ttm, rdev->dev); > } > #endif > > @@ -601,14 +601,14 @@ static int radeon_ttm_tt_populate(struct ttm_tt *ttm) > } > > for (i = 0; i < ttm->num_pages; i++) { > - ttm->dma_address[i] = pci_map_page(rdev->pdev, ttm->pages[i], > - 0, PAGE_SIZE, > - PCI_DMA_BIDIRECTIONAL); > - if (pci_dma_mapping_error(rdev->pdev, ttm->dma_address[i])) { > + gtt->ttm.dma_address[i] = pci_map_page(rdev->pdev, ttm->pages[i], > + 0, PAGE_SIZE, > + PCI_DMA_BIDIRECTIONAL); > + if (pci_dma_mapping_error(rdev->pdev, gtt->ttm.dma_address[i])) { > while (--i) { > - pci_unmap_page(rdev->pdev, ttm->dma_address[i], > + pci_unmap_page(rdev->pdev, gtt->ttm.dma_address[i], > PAGE_SIZE, PCI_DMA_BIDIRECTIONAL); > - ttm->dma_address[i] = 0; > + gtt->ttm.dma_address[i] = 0; > } > ttm_pool_unpopulate(ttm); > return -EFAULT; > @@ -620,20 +620,21 @@ static int radeon_ttm_tt_populate(struct ttm_tt *ttm) > static void radeon_ttm_tt_unpopulate(struct ttm_tt *ttm) > { > struct radeon_device *rdev; > + struct radeon_ttm_tt *gtt = (void *)ttm; > unsigned i; > > rdev = radeon_get_rdev(ttm->bdev); > > #ifdef CONFIG_SWIOTLB > if (swiotlb_nr_tbl()) { > - ttm_dma_unpopulate(ttm, rdev->dev); > + ttm_dma_unpopulate(>t->ttm, rdev->dev); > return; > } > #endif > > for (i = 0; i < ttm->num_pages; i++) { > - if (ttm->dma_address[i]) { > - pci_unmap_page(rdev->pdev, ttm->dma_address[i], > + if (gtt->ttm.dma_address[i]) { > + pci_unmap_page(rdev->pdev, gtt->ttm.dma_address[i], > PAGE_SIZE, PCI_DMA_BIDIRECTIONAL); > } > } > diff --git a/drivers/gpu/drm/ttm/ttm_page_alloc.c b/drivers/gpu/drm/ttm/ttm_page_alloc.c > index 8d6267e..499debd 100644 > --- a/drivers/gpu/drm/ttm/ttm_page_alloc.c > +++ b/drivers/gpu/drm/ttm/ttm_page_alloc.c > @@ -662,13 +662,61 @@ out: > return count; > } > > +/* Put all pages in pages list to correct pool to wait for reuse */ > +static void ttm_put_pages(struct page **pages, unsigned npages, int flags, > + enum ttm_caching_state cstate) > +{ > + unsigned long irq_flags; > + struct ttm_page_pool *pool = ttm_get_pool(flags, cstate); > + unsigned i; > + > + if (pool == NULL) { > + /* No pool for this memory type so free the pages */ > + for (i = 0; i < npages; i++) { > + if (pages[i]) { > + if (page_count(pages[i]) != 1) > + printk(KERN_ERR TTM_PFX > + "Erroneous page count. " > + "Leaking pages.\n"); > + __free_page(pages[i]); > + pages[i] = NULL; > + } > + } > + return; > + } > + > + spin_lock_irqsave(&pool->lock, irq_flags); > + for (i = 0; i < npages; i++) { > + if (pages[i]) { > + if (page_count(pages[i]) != 1) > + printk(KERN_ERR TTM_PFX > + "Erroneous page count. " > + "Leaking pages.\n"); > + list_add_tail(&pages[i]->lru, &pool->list); > + pages[i] = NULL; > + pool->npages++; > + } > + } > + /* Check that we don't go over the pool limit */ > + npages = 0; > + if (pool->npages > _manager->options.max_size) { > + npages = pool->npages - _manager->options.max_size; > + /* free at least NUM_PAGES_TO_ALLOC number of pages > + * to reduce calls to set_memory_wb */ > + if (npages < NUM_PAGES_TO_ALLOC) > + npages = NUM_PAGES_TO_ALLOC; > + } > + spin_unlock_irqrestore(&pool->lock, irq_flags); > + if (npages) > + ttm_page_pool_free(pool, npages); > +} > + > /* > * On success pages list will hold count number of correctly > * cached pages. > */ > -int ttm_get_pages(struct page **pages, int flags, > - enum ttm_caching_state cstate, unsigned npages, > - dma_addr_t *dma_address) > +static int ttm_get_pages(struct page **pages, unsigned npages, int flags, > + enum ttm_caching_state cstate) > { > struct ttm_page_pool *pool = ttm_get_pool(flags, cstate); > struct list_head plist; > @@ -736,7 +784,7 @@ int ttm_get_pages(struct page **pages, int flags, > printk(KERN_ERR TTM_PFX > "Failed to allocate extra pages " > "for large request."); > - ttm_put_pages(pages, count, flags, cstate, NULL); > + ttm_put_pages(pages, count, flags, cstate); > return r; > } > } > @@ -744,55 +792,6 @@ int ttm_get_pages(struct page **pages, int flags, > return 0; > } > > -/* Put all pages in pages list to correct pool to wait for reuse */ > -void ttm_put_pages(struct page **pages, unsigned npages, int flags, > - enum ttm_caching_state cstate, dma_addr_t *dma_address) > -{ > - unsigned long irq_flags; > - struct ttm_page_pool *pool = ttm_get_pool(flags, cstate); > - unsigned i; > - > - if (pool == NULL) { > - /* No pool for this memory type so free the pages */ > - for (i = 0; i < npages; i++) { > - if (pages[i]) { > - if (page_count(pages[i]) != 1) > - printk(KERN_ERR TTM_PFX > - "Erroneous page count. " > - "Leaking pages.\n"); > - __free_page(pages[i]); > - pages[i] = NULL; > - } > - } > - return; > - } > - > - spin_lock_irqsave(&pool->lock, irq_flags); > - for (i = 0; i < npages; i++) { > - if (pages[i]) { > - if (page_count(pages[i]) != 1) > - printk(KERN_ERR TTM_PFX > - "Erroneous page count. " > - "Leaking pages.\n"); > - list_add_tail(&pages[i]->lru, &pool->list); > - pages[i] = NULL; > - pool->npages++; > - } > - } > - /* Check that we don't go over the pool limit */ > - npages = 0; > - if (pool->npages > _manager->options.max_size) { > - npages = pool->npages - _manager->options.max_size; > - /* free at least NUM_PAGES_TO_ALLOC number of pages > - * to reduce calls to set_memory_wb */ > - if (npages < NUM_PAGES_TO_ALLOC) > - npages = NUM_PAGES_TO_ALLOC; > - } > - spin_unlock_irqrestore(&pool->lock, irq_flags); > - if (npages) > - ttm_page_pool_free(pool, npages); > -} > - > static void ttm_page_pool_init_locked(struct ttm_page_pool *pool, int flags, > char *name) > { > @@ -865,9 +864,9 @@ int ttm_pool_populate(struct ttm_tt *ttm) > return 0; > > for (i = 0; i < ttm->num_pages; ++i) { > - ret = ttm_get_pages(&ttm->pages[i], ttm->page_flags, > - ttm->caching_state, 1, > - &ttm->dma_address[i]); > + ret = ttm_get_pages(&ttm->pages[i], 1, > + ttm->page_flags, > + ttm->caching_state); > if (ret != 0) { > ttm_pool_unpopulate(ttm); > return -ENOMEM; > @@ -904,8 +903,7 @@ void ttm_pool_unpopulate(struct ttm_tt *ttm) > ttm->pages[i]); > ttm_put_pages(&ttm->pages[i], 1, > ttm->page_flags, > - ttm->caching_state, > - ttm->dma_address); > + ttm->caching_state); > } > } > ttm->state = tt_unpopulated; > diff --git a/drivers/gpu/drm/ttm/ttm_page_alloc_dma.c b/drivers/gpu/drm/ttm/ttm_page_alloc_dma.c > index 7a47793..6678abc 100644 > --- a/drivers/gpu/drm/ttm/ttm_page_alloc_dma.c > +++ b/drivers/gpu/drm/ttm/ttm_page_alloc_dma.c > @@ -789,7 +789,7 @@ out: > > /* > * @return count of pages still required to fulfill the request. > -*/ > + */ > static int ttm_dma_page_pool_fill_locked(struct dma_pool *pool, > unsigned long *irq_flags) > { > @@ -838,10 +838,11 @@ static int ttm_dma_page_pool_fill_locked(struct dma_pool *pool, > * allocates one page at a time. > */ > static int ttm_dma_pool_get_pages(struct dma_pool *pool, > - struct ttm_tt *ttm, > + struct ttm_dma_tt *ttm_dma, > unsigned index) > { > struct dma_page *d_page; > + struct ttm_tt *ttm = &ttm_dma->ttm; > unsigned long irq_flags; > int count, r = -ENOMEM; > > @@ -850,8 +851,8 @@ static int ttm_dma_pool_get_pages(struct dma_pool *pool, > if (count) { > d_page = list_first_entry(&pool->free_list, struct dma_page, page_list); > ttm->pages[index] = d_page->p; > - ttm->dma_address[index] = d_page->dma; > - list_move_tail(&d_page->page_list, &ttm->alloc_list); > + ttm_dma->dma_address[index] = d_page->dma; > + list_move_tail(&d_page->page_list, &ttm_dma->pages_list); > r = 0; > pool->npages_in_use += 1; > pool->npages_free -= 1; > @@ -864,8 +865,9 @@ static int ttm_dma_pool_get_pages(struct dma_pool *pool, > * On success pages list will hold count number of correctly > * cached pages. On failure will hold the negative return value (-ENOMEM, etc). > */ > -int ttm_dma_populate(struct ttm_tt *ttm, struct device *dev) > +int ttm_dma_populate(struct ttm_dma_tt *ttm_dma, struct device *dev) > { > + struct ttm_tt *ttm = &ttm_dma->ttm; > struct ttm_mem_global *mem_glob = ttm->glob->mem_glob; > struct dma_pool *pool; > enum pool_type type; > @@ -892,18 +894,18 @@ int ttm_dma_populate(struct ttm_tt *ttm, struct device *dev) > } > } > > - INIT_LIST_HEAD(&ttm->alloc_list); > + INIT_LIST_HEAD(&ttm_dma->pages_list); > for (i = 0; i < ttm->num_pages; ++i) { > - ret = ttm_dma_pool_get_pages(pool, ttm, i); > + ret = ttm_dma_pool_get_pages(pool, ttm_dma, i); > if (ret != 0) { > - ttm_dma_unpopulate(ttm, dev); > + ttm_dma_unpopulate(ttm_dma, dev); > return -ENOMEM; > } > > ret = ttm_mem_global_alloc_page(mem_glob, ttm->pages[i], > false, false); > if (unlikely(ret != 0)) { > - ttm_dma_unpopulate(ttm, dev); > + ttm_dma_unpopulate(ttm_dma, dev); > return -ENOMEM; > } > } > @@ -911,7 +913,7 @@ int ttm_dma_populate(struct ttm_tt *ttm, struct device *dev) > if (unlikely(ttm->page_flags & TTM_PAGE_FLAG_SWAPPED)) { > ret = ttm_tt_swapin(ttm); > if (unlikely(ret != 0)) { > - ttm_dma_unpopulate(ttm, dev); > + ttm_dma_unpopulate(ttm_dma, dev); > return ret; > } > } > @@ -937,8 +939,9 @@ static int ttm_dma_pool_get_num_unused_pages(void) > } > > /* Put all pages in pages list to correct pool to wait for reuse */ > -void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev) > +void ttm_dma_unpopulate(struct ttm_dma_tt *ttm_dma, struct device *dev) > { > + struct ttm_tt *ttm = &ttm_dma->ttm; > struct dma_pool *pool; > struct dma_page *d_page, *next; > enum pool_type type; > @@ -956,7 +959,7 @@ void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev) > ttm_to_type(ttm->page_flags, tt_cached)) == pool); > > /* make sure pages array match list and count number of pages */ > - list_for_each_entry(d_page, &ttm->alloc_list, page_list) { > + list_for_each_entry(d_page, &ttm_dma->pages_list, page_list) { > ttm->pages[count] = d_page->p; > count++; > } > @@ -967,7 +970,7 @@ void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev) > pool->nfrees += count; > } else { > pool->npages_free += count; > - list_splice(&ttm->alloc_list, &pool->free_list); > + list_splice(&ttm_dma->pages_list, &pool->free_list); > if (pool->npages_free > _manager->options.max_size) { > count = pool->npages_free - _manager->options.max_size; > } > @@ -975,7 +978,7 @@ void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev) > spin_unlock_irqrestore(&pool->lock, irq_flags); > > if (is_cached) { > - list_for_each_entry_safe(d_page, next, &ttm->alloc_list, page_list) { > + list_for_each_entry_safe(d_page, next, &ttm_dma->pages_list, page_list) { > ttm_mem_global_free_page(ttm->glob->mem_glob, > d_page->p); > ttm_dma_page_put(pool, d_page); > @@ -987,10 +990,10 @@ void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev) > } > } > > - INIT_LIST_HEAD(&ttm->alloc_list); > + INIT_LIST_HEAD(&ttm_dma->pages_list); > for (i = 0; i < ttm->num_pages; i++) { > ttm->pages[i] = NULL; > - ttm->dma_address[i] = 0; > + ttm_dma->dma_address[i] = 0; > } > > /* shrink pool if necessary */ > diff --git a/drivers/gpu/drm/ttm/ttm_tt.c b/drivers/gpu/drm/ttm/ttm_tt.c > index 1625739..58e1fa1 100644 > --- a/drivers/gpu/drm/ttm/ttm_tt.c > +++ b/drivers/gpu/drm/ttm/ttm_tt.c > @@ -48,17 +48,14 @@ > */ > static void ttm_tt_alloc_page_directory(struct ttm_tt *ttm) > { > - ttm->pages = drm_calloc_large(ttm->num_pages, sizeof(*ttm->pages)); > - ttm->dma_address = drm_calloc_large(ttm->num_pages, > - sizeof(*ttm->dma_address)); > + ttm->pages = drm_calloc_large(ttm->num_pages, sizeof(void*)); > } > > -static void ttm_tt_free_page_directory(struct ttm_tt *ttm) > +static void ttm_dma_tt_alloc_page_directory(struct ttm_dma_tt *ttm) > { > - drm_free_large(ttm->pages); > - ttm->pages = NULL; > - drm_free_large(ttm->dma_address); > - ttm->dma_address = NULL; > + ttm->ttm.pages = drm_calloc_large(ttm->ttm.num_pages, sizeof(void*)); > + ttm->dma_address = drm_calloc_large(ttm->ttm.num_pages, > + sizeof(*ttm->dma_address)); > } > > #ifdef CONFIG_X86 > @@ -173,7 +170,6 @@ void ttm_tt_destroy(struct ttm_tt *ttm) > > if (likely(ttm->pages != NULL)) { > ttm->bdev->driver->ttm_tt_unpopulate(ttm); > - ttm_tt_free_page_directory(ttm); > } > > if (!(ttm->page_flags & TTM_PAGE_FLAG_PERSISTENT_SWAP) && > @@ -196,9 +192,8 @@ int ttm_tt_init(struct ttm_tt *ttm, struct ttm_bo_device *bdev, > ttm->dummy_read_page = dummy_read_page; > ttm->state = tt_unpopulated; > > - INIT_LIST_HEAD(&ttm->alloc_list); > ttm_tt_alloc_page_directory(ttm); > - if (!ttm->pages || !ttm->dma_address) { > + if (!ttm->pages) { > ttm_tt_destroy(ttm); > printk(KERN_ERR TTM_PFX "Failed allocating page table\n"); > return -ENOMEM; > @@ -207,6 +202,49 @@ int ttm_tt_init(struct ttm_tt *ttm, struct ttm_bo_device *bdev, > } > EXPORT_SYMBOL(ttm_tt_init); > > +void ttm_tt_fini(struct ttm_tt *ttm) > +{ > + drm_free_large(ttm->pages); > + ttm->pages = NULL; > +} > +EXPORT_SYMBOL(ttm_tt_fini); > + > +int ttm_dma_tt_init(struct ttm_dma_tt *ttm_dma, struct ttm_bo_device *bdev, > + unsigned long size, uint32_t page_flags, > + struct page *dummy_read_page) > +{ > + struct ttm_tt *ttm = &ttm_dma->ttm; > + > + ttm->bdev = bdev; > + ttm->glob = bdev->glob; > + ttm->num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; > + ttm->caching_state = tt_cached; > + ttm->page_flags = page_flags; > + ttm->dummy_read_page = dummy_read_page; > + ttm->state = tt_unpopulated; > + > + INIT_LIST_HEAD(&ttm_dma->pages_list); > + ttm_dma_tt_alloc_page_directory(ttm_dma); > + if (!ttm->pages || !ttm_dma->dma_address) { > + ttm_tt_destroy(ttm); > + printk(KERN_ERR TTM_PFX "Failed allocating page table\n"); > + return -ENOMEM; > + } > + return 0; > +} > +EXPORT_SYMBOL(ttm_dma_tt_init); > + > +void ttm_dma_tt_fini(struct ttm_dma_tt *ttm_dma) > +{ > + struct ttm_tt *ttm = &ttm_dma->ttm; > + > + drm_free_large(ttm->pages); > + ttm->pages = NULL; > + drm_free_large(ttm_dma->dma_address); > + ttm_dma->dma_address = NULL; > +} > +EXPORT_SYMBOL(ttm_dma_tt_fini); > + > void ttm_tt_unbind(struct ttm_tt *ttm) > { > int ret; > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c b/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c > index 3986d74..1e2c0fb 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c > @@ -168,6 +168,7 @@ static void vmw_ttm_destroy(struct ttm_tt *ttm) > { > struct vmw_ttm_tt *vmw_be = container_of(ttm, struct vmw_ttm_tt, ttm); > > + ttm_tt_fini(ttm); > kfree(vmw_be); > } > > @@ -191,6 +192,7 @@ struct ttm_tt *vmw_ttm_tt_create(struct ttm_bo_device *bdev, > vmw_be->dev_priv = container_of(bdev, struct vmw_private, bdev); > > if (ttm_tt_init(&vmw_be->ttm, bdev, size, page_flags, dummy_read_page)) { > + kfree(vmw_be); > return NULL; > } > > diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h > index beef9ab..b2a0848 100644 > --- a/include/drm/ttm/ttm_bo_driver.h > +++ b/include/drm/ttm/ttm_bo_driver.h > @@ -103,8 +103,6 @@ enum ttm_caching_state { > * @swap_storage: Pointer to shmem struct file for swap storage. > * @caching_state: The current caching state of the pages. > * @state: The current binding state of the pages. > - * @dma_address: The DMA (bus) addresses of the pages (if TTM_PAGE_FLAG_DMA32) > - * @alloc_list: used by some page allocation backend > * > * This is a structure holding the pages, caching- and aperture binding > * status for a buffer object that isn't backed by fixed (VRAM / AGP) > @@ -127,8 +125,23 @@ struct ttm_tt { > tt_unbound, > tt_unpopulated, > } state; > +}; > + > +/** > + * struct ttm_dma_tt > + * > + * @ttm: Base ttm_tt struct. > + * @dma_address: The DMA (bus) addresses of the pages > + * @pages_list: used by some page allocation backend > + * > + * This is a structure holding the pages, caching- and aperture binding > + * status for a buffer object that isn't backed by fixed (VRAM / AGP) > + * memory. > + */ > +struct ttm_dma_tt { > + struct ttm_tt ttm; > dma_addr_t *dma_address; > - struct list_head alloc_list; > + struct list_head pages_list; > }; > > #define TTM_MEMTYPE_FLAG_FIXED (1 << 0) /* Fixed (on-card) PCI memory */ > @@ -595,6 +608,19 @@ ttm_flag_masked(uint32_t *old, uint32_t new, uint32_t mask) > extern int ttm_tt_init(struct ttm_tt *ttm, struct ttm_bo_device *bdev, > unsigned long size, uint32_t page_flags, > struct page *dummy_read_page); > +extern int ttm_dma_tt_init(struct ttm_dma_tt *ttm_dma, struct ttm_bo_device *bdev, > + unsigned long size, uint32_t page_flags, > + struct page *dummy_read_page); > + > +/** > + * ttm_tt_fini > + * > + * @ttm: the ttm_tt structure. > + * > + * Free memory of ttm_tt structure > + */ > +extern void ttm_tt_fini(struct ttm_tt *ttm); > +extern void ttm_dma_tt_fini(struct ttm_dma_tt *ttm_dma); > > /** > * ttm_ttm_bind: > diff --git a/include/drm/ttm/ttm_page_alloc.h b/include/drm/ttm/ttm_page_alloc.h > index 1e1337e..5fe2740 100644 > --- a/include/drm/ttm/ttm_page_alloc.h > +++ b/include/drm/ttm/ttm_page_alloc.h > @@ -30,35 +30,6 @@ > #include "ttm_memory.h" > > /** > - * Get count number of pages from pool to pages list. > - * > - * @pages: head of empty linked list where pages are filled. > - * @flags: ttm flags for page allocation. > - * @cstate: ttm caching state for the page. > - * @count: number of pages to allocate. > - * @dma_address: The DMA (bus) address of pages (if TTM_PAGE_FLAG_DMA32 set). > - */ > -int ttm_get_pages(struct page **pages, > - int flags, > - enum ttm_caching_state cstate, > - unsigned npages, > - dma_addr_t *dma_address); > -/** > - * Put linked list of pages to pool. > - * > - * @pages: list of pages to free. > - * @page_count: number of pages in the list. Zero can be passed for unknown > - * count. > - * @flags: ttm flags for page allocation. > - * @cstate: ttm caching state. > - * @dma_address: The DMA (bus) address of pages (if TTM_PAGE_FLAG_DMA32 set). > - */ > -void ttm_put_pages(struct page **pages, > - unsigned npages, > - int flags, > - enum ttm_caching_state cstate, > - dma_addr_t *dma_address); > -/** > * Initialize pool allocator. > */ > int ttm_page_alloc_init(struct ttm_mem_global *glob, unsigned max_pages); > @@ -107,8 +78,8 @@ void ttm_dma_page_alloc_fini(void); > */ > extern int ttm_dma_page_alloc_debugfs(struct seq_file *m, void *data); > > -int ttm_dma_populate(struct ttm_tt *ttm, struct device *dev); > -extern void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev); > +extern int ttm_dma_populate(struct ttm_dma_tt *ttm_dma, struct device *dev); > +extern void ttm_dma_unpopulate(struct ttm_dma_tt *ttm_dma, struct device *dev); > > #else > static inline int ttm_dma_page_alloc_init(struct ttm_mem_global *glob, > -- > 1.7.7.1 _______________________________________________ dri-devel mailing list dri-devel@xxxxxxxxxxxxxxxxxxxxx http://lists.freedesktop.org/mailman/listinfo/dri-devel