lloc_kernel (count=%zd, size=0x%llx) failed: %s ***va*call to uvm_gpu_dma_alloc*call to uvm_virt_to_page*call to mem_free_sysmem_dma_chunks*call to mem_free_vidmem_chunks*call to mem_free_sysmem_chunks*gpu_index*gpu_page_tree*attrs_out*protection*call to mem_can_be_mapped_on_gpu*call to vidmem_can_be_mapped*call to mem_can_be_mapped_on_cpu*call to uvm_range_allocator_deinit*call to uvm_range_allocator_init*call to tests*call to test_basic*test_basic(va_space)*/home/runner/work/bulk-builder/bulk-builder/kernel-open/nvidia-uvm/uvm_mem_test.c**/home/runner/work/bulk-builder/bulk-builder/kernel-open/nvidia-uvm/uvm_mem_test.c**test_basic(va_space)*call to test_all*test_all(va_space)**test_all(va_space)*call to test_basic_sysmem*test_basic_sysmem()**test_basic_sysmem()*call to test_basic_vidmem*test_basic_vidmem(gpu)**test_basic_vidmem(gpu)*call to test_basic_sysmem_dma*test_basic_sysmem_dma(gpu)**test_basic_sysmem_dma(gpu)*call to test_basic_dma_pool*test_basic_dma_pool(gpu)**test_basic_dma_pool(gpu)*call to test_huge_pages*test_huge_pages(va_space, gpu)**test_huge_pages(va_space, gpu)*call to test_huge_page_size*test_huge_page_size(va_space, gpu, page_size)**test_huge_page_size(va_space, gpu, page_size)*call to first_page_size*call to mem_alloc_sysmem_and_map_cpu_kernel*mem_alloc_sysmem_and_map_cpu_kernel(size, gpu, &mem)**mem_alloc_sysmem_and_map_cpu_kernel(size, gpu, &mem)*uvm_mem_map_gpu_phys(mem, gpu)**uvm_mem_map_gpu_phys(mem, gpu)*huge_gpu_va*IS_ALIGNED(huge_gpu_va, page_size)**IS_ALIGNED(huge_gpu_va, page_size)*(huge_gpu_va + page_size) < (1ull << gpu->address_space_tree.hal->num_va_bits())**(huge_gpu_va + page_size) < (1ull << gpu->address_space_tree.hal->num_va_bits())*uvm_page_table_range_vec_create(&gpu->address_space_tree, huge_gpu_va, page_size, page_size, UVM_PMM_ALLOC_FLAGS_NONE, &range_vec)**uvm_page_table_range_vec_create(&gpu->address_space_tree, huge_gpu_va, page_size, page_size, UVM_PMM_ALLOC_FLAGS_NONE, &range_vec)*gpu_phys_addr**gpu_phys_addr*uvm_page_table_range_vec_write_ptes(range_vec, UVM_MEMBAR_NONE, test_pte_maker, (void *)gpu_phys_addr)**uvm_page_table_range_vec_write_ptes(range_vec, UVM_MEMBAR_NONE, test_pte_maker, (void *)gpu_phys_addr)*call to check_huge_page_from_gpu*check_huge_page_from_gpu(gpu, mem, huge_gpu_va + (gpu_phys_addr % page_size))**check_huge_page_from_gpu(gpu, mem, huge_gpu_va + (gpu_phys_addr % page_size))**phys_addr*mem_alloc_sysmem_and_map_cpu_kernel(verif_size, gpu, &sys_mem)**mem_alloc_sysmem_and_map_cpu_kernel(verif_size, gpu, &sys_mem)*sys_mem*sys_verif**sys_verif*uvm_mem_map_gpu_kernel(sys_mem, gpu)**uvm_mem_map_gpu_kernel(sys_mem, gpu)*mem_gpu_address*sys_mem_gpu_address*Memcopy %llu bytes from virtual mem 0x%llx to virtual sys_mem 0x%llx**Memcopy %llu bytes from virtual mem 0x%llx to virtual sys_mem 0x%llx*uvm_push_begin(gpu->channel_manager, UVM_CHANNEL_TYPE_GPU_TO_CPU, &push, "Memcopy %llu bytes from virtual mem 0x%llx to virtual sys_mem 0x%llx", verif_size, mem_gpu_address.address, sys_mem_gpu_address.address)**uvm_push_begin(gpu->channel_manager, UVM_CHANNEL_TYPE_GPU_TO_CPU, &push, "Memcopy %llu bytes from virtual mem 0x%llx to virtual sys_mem 0x%llx", verif_size, mem_gpu_address.address, sys_mem_gpu_address.address)**expected_value*memcmp(sys_verif, expected_value, verif_size) == 0**memcmp(sys_verif, expected_value, verif_size) == 0*num_buffers*dma_buffers**dma_buffers***dma_buffers*gpu->conf_computing.dma_buffer_pool.num_dma_buffers >= num_buffers**gpu->conf_computing.dma_buffer_pool.num_dma_buffers >= num_buffers*i == num_buffers**i == num_buffers*uvm_mem_alloc_sysmem_dma(size, gpu, current->mm, &mem)**uvm_mem_alloc_sysmem_dma(size, gpu, current->mm, &mem)*mem->chunk_size == PAGE_SIZE**mem->chunk_size == PAGE_SIZE*uvm_mem_alloc_vidmem(page_size - 1, gpu, &mem) == NV_OK**uvm_mem_alloc_vidmem(page_size - 1, gpu, &mem) == NV_OK*mem->chunk_size >= PAGE_SIZE && mem->chunk_size <= max(page_size, (NvU64)PAGE_SIZE)**mem->chunk_size >= PAGE_SIZE && mem->chunk_size <= max(page_size, (NvU64)PAGE_SIZE)*mem->chunk_size < page_size || page_size == smallest_page_size**mem->chunk_size < page_size || page_size == smallest_page_size*uvm_mem_alloc_vidmem(page_size, gpu, &mem) == NV_OK**uvm_mem_alloc_vidmem(page_size, gpu, &mem) == NV_OK*mem->chunk_size == max(page_size, (NvU64)PAGE_SIZE)**mem->chunk_size == max(page_size, (NvU64)PAGE_SIZE)*mem->chunk_size == page_size**mem->chunk_size == page_size*uvm_mem_alloc_vidmem(5 * biggest_page_size - 1, gpu, &mem) == NV_OK**uvm_mem_alloc_vidmem(5 * biggest_page_size - 1, gpu, &mem) == NV_OK*mem->chunk_size == biggest_page_size**mem->chunk_size == biggest_page_size*allocation_count*all_mem**all_mem***all_mem*call to should_test_page_size*call to test_alloc_sysmem*