mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2026-04-15 15:23:23 -04:00
drm/xe: Decouple vram check from xe_bo_addr()
The output arg is_vram in xe_bo_addr() is unused by several callers. It's also not what the function is mainly doing. Remove the argument and let the interested callers to call xe_bo_is_vram(). Reviewed-by: Matt Roper <matthew.d.roper@intel.com> Link: https://lore.kernel.org/r/20230726160708.3967790-6-lucas.demarchi@intel.com Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com> Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
This commit is contained in:
committed by
Rodrigo Vivi
parent
621c1fbd9b
commit
937b4be72b
@@ -1526,12 +1526,11 @@ int xe_bo_pin(struct xe_bo *bo)
|
||||
if (IS_DGFX(xe) && !(IS_ENABLED(CONFIG_DRM_XE_DEBUG) &&
|
||||
bo->flags & XE_BO_INTERNAL_TEST)) {
|
||||
struct ttm_place *place = &(bo->placements[0]);
|
||||
bool vram;
|
||||
|
||||
if (mem_type_is_vram(place->mem_type)) {
|
||||
XE_BUG_ON(!(place->flags & TTM_PL_FLAG_CONTIGUOUS));
|
||||
|
||||
place->fpfn = (xe_bo_addr(bo, 0, PAGE_SIZE, &vram) -
|
||||
place->fpfn = (xe_bo_addr(bo, 0, PAGE_SIZE) -
|
||||
vram_region_gpu_offset(bo->ttm.resource)) >> PAGE_SHIFT;
|
||||
place->lpfn = place->fpfn + (bo->size >> PAGE_SHIFT);
|
||||
|
||||
@@ -1656,8 +1655,7 @@ bool xe_bo_is_xe_bo(struct ttm_buffer_object *bo)
|
||||
* address, such as printing debug information, but not in cases where memory is
|
||||
* written based on this result.
|
||||
*/
|
||||
dma_addr_t __xe_bo_addr(struct xe_bo *bo, u64 offset,
|
||||
size_t page_size, bool *is_vram)
|
||||
dma_addr_t __xe_bo_addr(struct xe_bo *bo, u64 offset, size_t page_size)
|
||||
{
|
||||
struct xe_res_cursor cur;
|
||||
u64 page;
|
||||
@@ -1666,9 +1664,7 @@ dma_addr_t __xe_bo_addr(struct xe_bo *bo, u64 offset,
|
||||
page = offset >> PAGE_SHIFT;
|
||||
offset &= (PAGE_SIZE - 1);
|
||||
|
||||
*is_vram = xe_bo_is_vram(bo);
|
||||
|
||||
if (!*is_vram && !xe_bo_is_stolen(bo)) {
|
||||
if (!xe_bo_is_vram(bo) && !xe_bo_is_stolen(bo)) {
|
||||
XE_BUG_ON(!bo->ttm.ttm);
|
||||
|
||||
xe_res_first_sg(xe_bo_get_sg(bo), page << PAGE_SHIFT,
|
||||
@@ -1683,12 +1679,11 @@ dma_addr_t __xe_bo_addr(struct xe_bo *bo, u64 offset,
|
||||
}
|
||||
}
|
||||
|
||||
dma_addr_t xe_bo_addr(struct xe_bo *bo, u64 offset,
|
||||
size_t page_size, bool *is_vram)
|
||||
dma_addr_t xe_bo_addr(struct xe_bo *bo, u64 offset, size_t page_size)
|
||||
{
|
||||
if (!READ_ONCE(bo->ttm.pin_count))
|
||||
xe_bo_assert_held(bo);
|
||||
return __xe_bo_addr(bo, offset, page_size, is_vram);
|
||||
return __xe_bo_addr(bo, offset, page_size);
|
||||
}
|
||||
|
||||
int xe_bo_vmap(struct xe_bo *bo)
|
||||
|
||||
@@ -216,17 +216,13 @@ static inline void xe_bo_unpin_map_no_vm(struct xe_bo *bo)
|
||||
}
|
||||
|
||||
bool xe_bo_is_xe_bo(struct ttm_buffer_object *bo);
|
||||
dma_addr_t __xe_bo_addr(struct xe_bo *bo, u64 offset,
|
||||
size_t page_size, bool *is_vram);
|
||||
dma_addr_t xe_bo_addr(struct xe_bo *bo, u64 offset,
|
||||
size_t page_size, bool *is_vram);
|
||||
dma_addr_t __xe_bo_addr(struct xe_bo *bo, u64 offset, size_t page_size);
|
||||
dma_addr_t xe_bo_addr(struct xe_bo *bo, u64 offset, size_t page_size);
|
||||
|
||||
static inline dma_addr_t
|
||||
xe_bo_main_addr(struct xe_bo *bo, size_t page_size)
|
||||
{
|
||||
bool is_vram;
|
||||
|
||||
return xe_bo_addr(bo, 0, page_size, &is_vram);
|
||||
return xe_bo_addr(bo, 0, page_size);
|
||||
}
|
||||
|
||||
static inline u32
|
||||
|
||||
@@ -31,12 +31,11 @@ u64 xe_ggtt_pte_encode(struct xe_bo *bo, u64 bo_offset)
|
||||
{
|
||||
struct xe_device *xe = xe_bo_device(bo);
|
||||
u64 pte;
|
||||
bool is_vram;
|
||||
|
||||
pte = xe_bo_addr(bo, bo_offset, XE_PAGE_SIZE, &is_vram);
|
||||
pte = xe_bo_addr(bo, bo_offset, XE_PAGE_SIZE);
|
||||
pte |= XE_PAGE_PRESENT;
|
||||
|
||||
if (is_vram)
|
||||
if (xe_bo_is_vram(bo))
|
||||
pte |= XE_GGTT_PTE_LM;
|
||||
|
||||
/* FIXME: vfunc + pass in caching rules */
|
||||
|
||||
@@ -132,7 +132,6 @@ static int xe_migrate_create_cleared_bo(struct xe_migrate *m, struct xe_vm *vm)
|
||||
struct xe_device *xe = vm->xe;
|
||||
size_t cleared_size;
|
||||
u64 vram_addr;
|
||||
bool is_vram;
|
||||
|
||||
if (!xe_device_has_flat_ccs(xe))
|
||||
return 0;
|
||||
@@ -147,8 +146,7 @@ static int xe_migrate_create_cleared_bo(struct xe_migrate *m, struct xe_vm *vm)
|
||||
return PTR_ERR(m->cleared_bo);
|
||||
|
||||
xe_map_memset(xe, &m->cleared_bo->vmap, 0, 0x00, cleared_size);
|
||||
vram_addr = xe_bo_addr(m->cleared_bo, 0, XE_PAGE_SIZE, &is_vram);
|
||||
XE_BUG_ON(!is_vram);
|
||||
vram_addr = xe_bo_addr(m->cleared_bo, 0, XE_PAGE_SIZE);
|
||||
m->cleared_vram_ofs = xe_migrate_vram_ofs(vram_addr);
|
||||
|
||||
return 0;
|
||||
@@ -221,15 +219,13 @@ static int xe_migrate_prepare_vm(struct xe_tile *tile, struct xe_migrate *m,
|
||||
level++;
|
||||
}
|
||||
} else {
|
||||
bool is_vram;
|
||||
u64 batch_addr = xe_bo_addr(batch, 0, XE_PAGE_SIZE, &is_vram);
|
||||
u64 batch_addr = xe_bo_addr(batch, 0, XE_PAGE_SIZE);
|
||||
|
||||
m->batch_base_ofs = xe_migrate_vram_ofs(batch_addr);
|
||||
|
||||
if (xe->info.supports_usm) {
|
||||
batch = tile->primary_gt->usm.bb_pool->bo;
|
||||
batch_addr = xe_bo_addr(batch, 0, XE_PAGE_SIZE,
|
||||
&is_vram);
|
||||
batch_addr = xe_bo_addr(batch, 0, XE_PAGE_SIZE);
|
||||
m->usm_batch_base_ofs = xe_migrate_vram_ofs(batch_addr);
|
||||
}
|
||||
}
|
||||
@@ -1000,12 +996,8 @@ static void write_pgtable(struct xe_tile *tile, struct xe_bb *bb, u64 ppgtt_ofs,
|
||||
*/
|
||||
XE_BUG_ON(update->qwords > 0x1ff);
|
||||
if (!ppgtt_ofs) {
|
||||
bool is_vram;
|
||||
|
||||
ppgtt_ofs = xe_migrate_vram_ofs(xe_bo_addr(update->pt_bo, 0,
|
||||
XE_PAGE_SIZE,
|
||||
&is_vram));
|
||||
XE_BUG_ON(!is_vram);
|
||||
XE_PAGE_SIZE));
|
||||
}
|
||||
|
||||
do {
|
||||
|
||||
@@ -61,13 +61,10 @@ u64 xe_pde_encode(struct xe_bo *bo, u64 bo_offset,
|
||||
const enum xe_cache_level level)
|
||||
{
|
||||
u64 pde;
|
||||
bool is_vram;
|
||||
|
||||
pde = xe_bo_addr(bo, bo_offset, XE_PAGE_SIZE, &is_vram);
|
||||
pde = xe_bo_addr(bo, bo_offset, XE_PAGE_SIZE);
|
||||
pde |= XE_PAGE_PRESENT | XE_PAGE_RW;
|
||||
|
||||
XE_WARN_ON(IS_DGFX(xe_bo_device(bo)) && !is_vram);
|
||||
|
||||
/* FIXME: I don't think the PPAT handling is correct for MTL */
|
||||
|
||||
if (level != XE_CACHE_NONE)
|
||||
@@ -128,10 +125,9 @@ u64 xe_pte_encode(struct xe_bo *bo, u64 offset, enum xe_cache_level cache,
|
||||
u32 pt_level)
|
||||
{
|
||||
u64 pte;
|
||||
bool is_vram;
|
||||
|
||||
pte = xe_bo_addr(bo, offset, XE_PAGE_SIZE, &is_vram);
|
||||
if (is_vram)
|
||||
pte = xe_bo_addr(bo, offset, XE_PAGE_SIZE);
|
||||
if (xe_bo_is_vram(bo))
|
||||
pte |= XE_PPGTT_PTE_LM;
|
||||
|
||||
return __pte_encode(pte, cache, NULL, pt_level);
|
||||
|
||||
@@ -3503,9 +3503,10 @@ int xe_analyze_vm(struct drm_printer *p, struct xe_vm *vm, int gt_id)
|
||||
return 0;
|
||||
}
|
||||
if (vm->pt_root[gt_id]) {
|
||||
addr = xe_bo_addr(vm->pt_root[gt_id]->bo, 0, XE_PAGE_SIZE,
|
||||
&is_vram);
|
||||
drm_printf(p, " VM root: A:0x%llx %s\n", addr, is_vram ? "VRAM" : "SYS");
|
||||
addr = xe_bo_addr(vm->pt_root[gt_id]->bo, 0, XE_PAGE_SIZE);
|
||||
is_vram = xe_bo_is_vram(vm->pt_root[gt_id]->bo);
|
||||
drm_printf(p, " VM root: A:0x%llx %s\n", addr,
|
||||
is_vram ? "VRAM" : "SYS");
|
||||
}
|
||||
|
||||
drm_gpuvm_for_each_va(gpuva, &vm->gpuvm) {
|
||||
@@ -3526,7 +3527,8 @@ int xe_analyze_vm(struct drm_printer *p, struct xe_vm *vm, int gt_id)
|
||||
addr = 0;
|
||||
}
|
||||
} else {
|
||||
addr = __xe_bo_addr(xe_vma_bo(vma), 0, XE_PAGE_SIZE, &is_vram);
|
||||
addr = __xe_bo_addr(xe_vma_bo(vma), 0, XE_PAGE_SIZE);
|
||||
is_vram = xe_bo_is_vram(xe_vma_bo(vma));
|
||||
}
|
||||
drm_printf(p, " [%016llx-%016llx] S:0x%016llx A:%016llx %s\n",
|
||||
xe_vma_start(vma), xe_vma_end(vma) - 1,
|
||||
|
||||
Reference in New Issue
Block a user