mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/netdev/net.git/
synced 2026-04-18 06:33:43 -04:00
Merge drm/drm-next into drm-misc-next
Backmerging to get a late RC of v6.10 before moving into v6.11. Signed-off-by: Thomas Zimmermann <tzimmermann@suse.de>
This commit is contained in:
@@ -333,7 +333,7 @@ void amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base *base,
|
||||
base->next = bo->vm_bo;
|
||||
bo->vm_bo = base;
|
||||
|
||||
if (bo->tbo.base.resv != vm->root.bo->tbo.base.resv)
|
||||
if (!amdgpu_vm_is_bo_always_valid(vm, bo))
|
||||
return;
|
||||
|
||||
dma_resv_assert_held(vm->root.bo->tbo.base.resv);
|
||||
@@ -434,7 +434,7 @@ uint64_t amdgpu_vm_generation(struct amdgpu_device *adev, struct amdgpu_vm *vm)
|
||||
if (!vm)
|
||||
return result;
|
||||
|
||||
result += vm->generation;
|
||||
result += lower_32_bits(vm->generation);
|
||||
/* Add one if the page tables will be re-generated on next CS */
|
||||
if (drm_sched_entity_error(&vm->delayed))
|
||||
++result;
|
||||
@@ -463,13 +463,14 @@ int amdgpu_vm_validate(struct amdgpu_device *adev, struct amdgpu_vm *vm,
|
||||
int (*validate)(void *p, struct amdgpu_bo *bo),
|
||||
void *param)
|
||||
{
|
||||
uint64_t new_vm_generation = amdgpu_vm_generation(adev, vm);
|
||||
struct amdgpu_vm_bo_base *bo_base;
|
||||
struct amdgpu_bo *shadow;
|
||||
struct amdgpu_bo *bo;
|
||||
int r;
|
||||
|
||||
if (drm_sched_entity_error(&vm->delayed)) {
|
||||
++vm->generation;
|
||||
if (vm->generation != new_vm_generation) {
|
||||
vm->generation = new_vm_generation;
|
||||
amdgpu_vm_bo_reset_state_machine(vm);
|
||||
amdgpu_vm_fini_entities(vm);
|
||||
r = amdgpu_vm_init_entities(adev, vm);
|
||||
@@ -1055,7 +1056,7 @@ int amdgpu_vm_update_range(struct amdgpu_device *adev, struct amdgpu_vm *vm,
|
||||
params.pages_addr = NULL;
|
||||
}
|
||||
|
||||
} else if (flags & (AMDGPU_PTE_VALID | AMDGPU_PTE_PRT)) {
|
||||
} else if (flags & (AMDGPU_PTE_VALID | AMDGPU_PTE_PRT_FLAG(adev))) {
|
||||
addr = vram_base + cursor.start;
|
||||
} else {
|
||||
addr = 0;
|
||||
@@ -1101,13 +1102,13 @@ static void amdgpu_vm_bo_get_memory(struct amdgpu_bo_va *bo_va,
|
||||
* For now ignore BOs which are currently locked and potentially
|
||||
* changing their location.
|
||||
*/
|
||||
if (bo->tbo.base.resv != vm->root.bo->tbo.base.resv &&
|
||||
if (!amdgpu_vm_is_bo_always_valid(vm, bo) &&
|
||||
!dma_resv_trylock(bo->tbo.base.resv))
|
||||
return;
|
||||
|
||||
amdgpu_bo_get_memory(bo, stats);
|
||||
if (bo->tbo.base.resv != vm->root.bo->tbo.base.resv)
|
||||
dma_resv_unlock(bo->tbo.base.resv);
|
||||
if (!amdgpu_vm_is_bo_always_valid(vm, bo))
|
||||
dma_resv_unlock(bo->tbo.base.resv);
|
||||
}
|
||||
|
||||
void amdgpu_vm_get_memory(struct amdgpu_vm *vm,
|
||||
@@ -1203,8 +1204,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
|
||||
uncached = false;
|
||||
}
|
||||
|
||||
if (clear || (bo && bo->tbo.base.resv ==
|
||||
vm->root.bo->tbo.base.resv))
|
||||
if (clear || amdgpu_vm_is_bo_always_valid(vm, bo))
|
||||
last_update = &vm->last_update;
|
||||
else
|
||||
last_update = &bo_va->last_pt_update;
|
||||
@@ -1246,7 +1246,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
|
||||
* the evicted list so that it gets validated again on the
|
||||
* next command submission.
|
||||
*/
|
||||
if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv) {
|
||||
if (amdgpu_vm_is_bo_always_valid(vm, bo)) {
|
||||
uint32_t mem_type = bo->tbo.resource->mem_type;
|
||||
|
||||
if (!(bo->preferred_domains &
|
||||
@@ -1369,7 +1369,7 @@ static void amdgpu_vm_free_mapping(struct amdgpu_device *adev,
|
||||
struct amdgpu_bo_va_mapping *mapping,
|
||||
struct dma_fence *fence)
|
||||
{
|
||||
if (mapping->flags & AMDGPU_PTE_PRT)
|
||||
if (mapping->flags & AMDGPU_PTE_PRT_FLAG(adev))
|
||||
amdgpu_vm_add_prt_cb(adev, fence);
|
||||
kfree(mapping);
|
||||
}
|
||||
@@ -1637,13 +1637,12 @@ static void amdgpu_vm_bo_insert_map(struct amdgpu_device *adev,
|
||||
list_add(&mapping->list, &bo_va->invalids);
|
||||
amdgpu_vm_it_insert(mapping, &vm->va);
|
||||
|
||||
if (mapping->flags & AMDGPU_PTE_PRT)
|
||||
if (mapping->flags & AMDGPU_PTE_PRT_FLAG(adev))
|
||||
amdgpu_vm_prt_get(adev);
|
||||
|
||||
if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv &&
|
||||
!bo_va->base.moved) {
|
||||
if (amdgpu_vm_is_bo_always_valid(vm, bo) && !bo_va->base.moved)
|
||||
amdgpu_vm_bo_moved(&bo_va->base);
|
||||
}
|
||||
|
||||
trace_amdgpu_vm_bo_map(bo_va, mapping);
|
||||
}
|
||||
|
||||
@@ -1939,10 +1938,10 @@ int amdgpu_vm_bo_clear_mappings(struct amdgpu_device *adev,
|
||||
struct amdgpu_bo *bo = before->bo_va->base.bo;
|
||||
|
||||
amdgpu_vm_it_insert(before, &vm->va);
|
||||
if (before->flags & AMDGPU_PTE_PRT)
|
||||
if (before->flags & AMDGPU_PTE_PRT_FLAG(adev))
|
||||
amdgpu_vm_prt_get(adev);
|
||||
|
||||
if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv &&
|
||||
if (amdgpu_vm_is_bo_always_valid(vm, bo) &&
|
||||
!before->bo_va->base.moved)
|
||||
amdgpu_vm_bo_moved(&before->bo_va->base);
|
||||
} else {
|
||||
@@ -1954,10 +1953,10 @@ int amdgpu_vm_bo_clear_mappings(struct amdgpu_device *adev,
|
||||
struct amdgpu_bo *bo = after->bo_va->base.bo;
|
||||
|
||||
amdgpu_vm_it_insert(after, &vm->va);
|
||||
if (after->flags & AMDGPU_PTE_PRT)
|
||||
if (after->flags & AMDGPU_PTE_PRT_FLAG(adev))
|
||||
amdgpu_vm_prt_get(adev);
|
||||
|
||||
if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv &&
|
||||
if (amdgpu_vm_is_bo_always_valid(vm, bo) &&
|
||||
!after->bo_va->base.moved)
|
||||
amdgpu_vm_bo_moved(&after->bo_va->base);
|
||||
} else {
|
||||
@@ -2037,7 +2036,7 @@ void amdgpu_vm_bo_del(struct amdgpu_device *adev,
|
||||
|
||||
if (bo) {
|
||||
dma_resv_assert_held(bo->tbo.base.resv);
|
||||
if (bo->tbo.base.resv == vm->root.bo->tbo.base.resv)
|
||||
if (amdgpu_vm_is_bo_always_valid(vm, bo))
|
||||
ttm_bo_set_bulk_move(&bo->tbo, NULL);
|
||||
|
||||
for (base = &bo_va->base.bo->vm_bo; *base;
|
||||
@@ -2131,7 +2130,7 @@ void amdgpu_vm_bo_invalidate(struct amdgpu_device *adev,
|
||||
for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next) {
|
||||
struct amdgpu_vm *vm = bo_base->vm;
|
||||
|
||||
if (evicted && bo->tbo.base.resv == vm->root.bo->tbo.base.resv) {
|
||||
if (evicted && amdgpu_vm_is_bo_always_valid(vm, bo)) {
|
||||
amdgpu_vm_bo_evicted(bo_base);
|
||||
continue;
|
||||
}
|
||||
@@ -2142,7 +2141,7 @@ void amdgpu_vm_bo_invalidate(struct amdgpu_device *adev,
|
||||
|
||||
if (bo->tbo.type == ttm_bo_type_kernel)
|
||||
amdgpu_vm_bo_relocated(bo_base);
|
||||
else if (bo->tbo.base.resv == vm->root.bo->tbo.base.resv)
|
||||
else if (amdgpu_vm_is_bo_always_valid(vm, bo))
|
||||
amdgpu_vm_bo_moved(bo_base);
|
||||
else
|
||||
amdgpu_vm_bo_invalidated(bo_base);
|
||||
@@ -2443,7 +2442,7 @@ int amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm,
|
||||
vm->last_update = dma_fence_get_stub();
|
||||
vm->last_unlocked = dma_fence_get_stub();
|
||||
vm->last_tlb_flush = dma_fence_get_stub();
|
||||
vm->generation = 0;
|
||||
vm->generation = amdgpu_vm_generation(adev, NULL);
|
||||
|
||||
mutex_init(&vm->eviction_lock);
|
||||
vm->evicting = false;
|
||||
@@ -2608,7 +2607,7 @@ void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)
|
||||
dma_fence_put(vm->last_tlb_flush);
|
||||
|
||||
list_for_each_entry_safe(mapping, tmp, &vm->freed, list) {
|
||||
if (mapping->flags & AMDGPU_PTE_PRT && prt_fini_needed) {
|
||||
if (mapping->flags & AMDGPU_PTE_PRT_FLAG(adev) && prt_fini_needed) {
|
||||
amdgpu_vm_prt_fini(adev, vm);
|
||||
prt_fini_needed = false;
|
||||
}
|
||||
@@ -3010,3 +3009,16 @@ void amdgpu_vm_update_fault_cache(struct amdgpu_device *adev,
|
||||
xa_unlock_irqrestore(&adev->vm_manager.pasids, flags);
|
||||
}
|
||||
|
||||
/**
|
||||
* amdgpu_vm_is_bo_always_valid - check if the BO is VM always valid
|
||||
*
|
||||
* @vm: VM to test against.
|
||||
* @bo: BO to be tested.
|
||||
*
|
||||
* Returns true if the BO shares the dma_resv object with the root PD and is
|
||||
* always guaranteed to be valid inside the VM.
|
||||
*/
|
||||
bool amdgpu_vm_is_bo_always_valid(struct amdgpu_vm *vm, struct amdgpu_bo *bo)
|
||||
{
|
||||
return bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv;
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user