Am 19.01.2018 um 09:23 schrieb Chunming Zhou: > for(( i=1; i < 100; i++)) > do > echo "Hello, Welcome $i times " > sudo ./amdgpu_test -s 1 -t 5 > done > > with above stricpt, run in two terminals, will reproduce Felix's swap leeking issue. > > Change-Id: If637b5da1eabe69806f41f0c6f9056b1e9447efc > Signed-off-by: Chunming Zhou <david1.zhou at amd.com> Acked-by: Christian König <christian.koenig at amd.com> > --- > tests/amdgpu/amdgpu_test.h | 24 +++++++ > tests/amdgpu/basic_tests.c | 157 +++++++++++++++++++++++++++++++++++++++++++++ > 2 files changed, 181 insertions(+) > > diff --git a/tests/amdgpu/amdgpu_test.h b/tests/amdgpu/amdgpu_test.h > index ece93f4c..de724b33 100644 > --- a/tests/amdgpu/amdgpu_test.h > +++ b/tests/amdgpu/amdgpu_test.h > @@ -206,6 +206,30 @@ static inline int gpu_mem_free(amdgpu_bo_handle bo, > return 0; > } > > +static inline int > +amdgpu_bo_alloc_wrap(amdgpu_device_handle dev, unsigned size, > + unsigned alignment, unsigned heap, uint64_t flags, > + amdgpu_bo_handle *bo) > +{ > + struct amdgpu_bo_alloc_request request = {}; > + amdgpu_bo_handle buf_handle; > + int r; > + > + request.alloc_size = size; > + request.phys_alignment = alignment; > + request.preferred_heap = heap; > + request.flags = flags; > + > + r = amdgpu_bo_alloc(dev, &request, &buf_handle); > + if (r) > + return r; > + > + *bo = buf_handle; > + > + return 0; > +} > + > + > static inline int > amdgpu_bo_alloc_and_map(amdgpu_device_handle dev, unsigned size, > unsigned alignment, unsigned heap, uint64_t flags, > diff --git a/tests/amdgpu/basic_tests.c b/tests/amdgpu/basic_tests.c > index 2e283205..05368a27 100644 > --- a/tests/amdgpu/basic_tests.c > +++ b/tests/amdgpu/basic_tests.c > @@ -55,16 +55,24 @@ static void amdgpu_semaphore_test(void); > static void amdgpu_svm_test(void); > static void amdgpu_multi_svm_test(void); > static void amdgpu_va_range_test(void); > +static void amdgpu_bo_eviction_test(void); > > static void amdgpu_command_submission_write_linear_helper(unsigned ip_type); > static void amdgpu_command_submission_const_fill_helper(unsigned ip_type); > static void amdgpu_command_submission_copy_linear_helper(unsigned ip_type); > +static void amdgpu_test_exec_cs_helper(amdgpu_context_handle context_handle, > + unsigned ip_type, > + int instance, int pm4_dw, uint32_t *pm4_src, > + int res_cnt, amdgpu_bo_handle *resources, > + struct amdgpu_cs_ib_info *ib_info, > + struct amdgpu_cs_request *ibs_request); > > CU_TestInfo basic_tests[] = { > { "Query Info Test", amdgpu_query_info_test }, > { "Memory alloc Test", amdgpu_memory_alloc }, > { "Userptr Test", amdgpu_userptr_test }, > { "Find bo Test", amdgpu_find_bo_test }, > + { "bo eviction Test", amdgpu_bo_eviction_test }, > { "Command submission Test (GFX)", amdgpu_command_submission_gfx }, > { "Command submission Test (Compute)", amdgpu_command_submission_compute }, > { "Command submission Test (Multi-Fence)", amdgpu_command_submission_multi_fence }, > @@ -518,6 +526,155 @@ static void amdgpu_command_submission_gfx_cp_copy_data(void) > amdgpu_command_submission_copy_linear_helper(AMDGPU_HW_IP_GFX); > } > > +static void amdgpu_bo_eviction_test(void) > +{ > + const int sdma_write_length = 1024; > + const int pm4_dw = 256; > + amdgpu_context_handle context_handle; > + amdgpu_bo_handle bo1, bo2, vram_max[2], gtt_max[2]; > + amdgpu_bo_handle *resources; > + uint32_t *pm4; > + struct amdgpu_cs_ib_info *ib_info; > + struct amdgpu_cs_request *ibs_request; > + uint64_t bo1_mc, bo2_mc; > + volatile unsigned char *bo1_cpu, *bo2_cpu; > + int i, j, r, loop1, loop2; > + uint64_t gtt_flags[2] = {0, AMDGPU_GEM_CREATE_CPU_GTT_USWC}; > + amdgpu_va_handle bo1_va_handle, bo2_va_handle; > + struct amdgpu_heap_info vram_info, gtt_info; > + > + pm4 = calloc(pm4_dw, sizeof(*pm4)); > + CU_ASSERT_NOT_EQUAL(pm4, NULL); > + > + ib_info = calloc(1, sizeof(*ib_info)); > + CU_ASSERT_NOT_EQUAL(ib_info, NULL); > + > + ibs_request = calloc(1, sizeof(*ibs_request)); > + CU_ASSERT_NOT_EQUAL(ibs_request, NULL); > + > + r = amdgpu_cs_ctx_create(device_handle, &context_handle); > + CU_ASSERT_EQUAL(r, 0); > + > + /* prepare resource */ > + resources = calloc(4, sizeof(amdgpu_bo_handle)); > + CU_ASSERT_NOT_EQUAL(resources, NULL); > + > + r = amdgpu_query_heap_info(device_handle, AMDGPU_GEM_DOMAIN_VRAM, > + 0, &vram_info); > + CU_ASSERT_EQUAL(r, 0); > + > + r = amdgpu_bo_alloc_wrap(device_handle, vram_info.max_allocation, 4096, > + AMDGPU_GEM_DOMAIN_VRAM, 0, &vram_max[0]); > + CU_ASSERT_EQUAL(r, 0); > + r = amdgpu_bo_alloc_wrap(device_handle, vram_info.max_allocation, 4096, > + AMDGPU_GEM_DOMAIN_VRAM, 0, &vram_max[1]); > + CU_ASSERT_EQUAL(r, 0); > + > + r = amdgpu_query_heap_info(device_handle, AMDGPU_GEM_DOMAIN_GTT, > + 0, >t_info); > + CU_ASSERT_EQUAL(r, 0); > + > + r = amdgpu_bo_alloc_wrap(device_handle, gtt_info.max_allocation, 4096, > + AMDGPU_GEM_DOMAIN_GTT, 0, >t_max[0]); > + CU_ASSERT_EQUAL(r, 0); > + r = amdgpu_bo_alloc_wrap(device_handle, gtt_info.max_allocation, 4096, > + AMDGPU_GEM_DOMAIN_GTT, 0, >t_max[1]); > + CU_ASSERT_EQUAL(r, 0); > + > + > + > + loop1 = loop2 = 0; > + /* run 9 circle to test all mapping combination */ > + while(loop1 < 2) { > + while(loop2 < 2) { > + /* allocate UC bo1for sDMA use */ > + r = amdgpu_bo_alloc_and_map(device_handle, > + sdma_write_length, 4096, > + AMDGPU_GEM_DOMAIN_GTT, > + gtt_flags[loop1], &bo1, > + (void**)&bo1_cpu, &bo1_mc, > + &bo1_va_handle); > + CU_ASSERT_EQUAL(r, 0); > + > + /* set bo1 */ > + memset((void*)bo1_cpu, 0xaa, sdma_write_length); > + > + /* allocate UC bo2 for sDMA use */ > + r = amdgpu_bo_alloc_and_map(device_handle, > + sdma_write_length, 4096, > + AMDGPU_GEM_DOMAIN_GTT, > + gtt_flags[loop2], &bo2, > + (void**)&bo2_cpu, &bo2_mc, > + &bo2_va_handle); > + CU_ASSERT_EQUAL(r, 0); > + > + /* clear bo2 */ > + memset((void*)bo2_cpu, 0, sdma_write_length); > + > + resources[0] = bo1; > + resources[1] = bo2; > + resources[2] = vram_max[loop2]; > + resources[3] = gtt_max[loop2]; > + > + /* fulfill PM4: test DMA copy linear */ > + i = j = 0; > + if (family_id == AMDGPU_FAMILY_SI) { > + pm4[i++] = SDMA_PACKET_SI(SDMA_OPCODE_COPY_SI, 0, 0, 0, > + sdma_write_length); > + pm4[i++] = 0xffffffff & bo2_mc; > + pm4[i++] = 0xffffffff & bo1_mc; > + pm4[i++] = (0xffffffff00000000 & bo2_mc) >> 32; > + pm4[i++] = (0xffffffff00000000 & bo1_mc) >> 32; > + } else { > + pm4[i++] = SDMA_PACKET(SDMA_OPCODE_COPY, SDMA_COPY_SUB_OPCODE_LINEAR, 0); > + if (family_id >= AMDGPU_FAMILY_AI) > + pm4[i++] = sdma_write_length - 1; > + else > + pm4[i++] = sdma_write_length; > + pm4[i++] = 0; > + pm4[i++] = 0xffffffff & bo1_mc; > + pm4[i++] = (0xffffffff00000000 & bo1_mc) >> 32; > + pm4[i++] = 0xffffffff & bo2_mc; > + pm4[i++] = (0xffffffff00000000 & bo2_mc) >> 32; > + } > + > + amdgpu_test_exec_cs_helper(context_handle, > + AMDGPU_HW_IP_DMA, 0, > + i, pm4, > + 4, resources, > + ib_info, ibs_request); > + > + /* verify if SDMA test result meets with expected */ > + i = 0; > + while(i < sdma_write_length) { > + CU_ASSERT_EQUAL(bo2_cpu[i++], 0xaa); > + } > + r = amdgpu_bo_unmap_and_free(bo1, bo1_va_handle, bo1_mc, > + sdma_write_length); > + CU_ASSERT_EQUAL(r, 0); > + r = amdgpu_bo_unmap_and_free(bo2, bo2_va_handle, bo2_mc, > + sdma_write_length); > + CU_ASSERT_EQUAL(r, 0); > + loop2++; > + } > + loop2 = 0; > + loop1++; > + } > + amdgpu_bo_free(vram_max[0]); > + amdgpu_bo_free(vram_max[1]); > + amdgpu_bo_free(gtt_max[0]); > + amdgpu_bo_free(gtt_max[1]); > + /* clean resources */ > + free(resources); > + free(ibs_request); > + free(ib_info); > + free(pm4); > + > + /* end of test */ > + r = amdgpu_cs_ctx_free(context_handle); > + CU_ASSERT_EQUAL(r, 0); > +} > + > static void amdgpu_command_submission_gfx(void) > { > /* write data using the CP */