mirror of https://gitee.com/openkylin/linux.git
drm/amdgpu: rework VM state machine lock handling v2
Only the moved state needs a separate spin lock protection. All other states are protected by reserving the VM anyway. v2: fix some more incorrect cases Signed-off-by: Christian König <christian.koenig@amd.com> Reviewed-by: Chunming Zhou <david1.zhou@amd.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
This commit is contained in:
parent
22cc6c5e19
commit
af4c0f650b
|
@ -119,9 +119,7 @@ static void amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base *base,
|
||||||
* is currently evicted. add the bo to the evicted list to make sure it
|
* is currently evicted. add the bo to the evicted list to make sure it
|
||||||
* is validated on next vm use to avoid fault.
|
* is validated on next vm use to avoid fault.
|
||||||
* */
|
* */
|
||||||
spin_lock(&vm->status_lock);
|
|
||||||
list_move_tail(&base->vm_status, &vm->evicted);
|
list_move_tail(&base->vm_status, &vm->evicted);
|
||||||
spin_unlock(&vm->status_lock);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -228,7 +226,6 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
|
||||||
struct ttm_bo_global *glob = adev->mman.bdev.glob;
|
struct ttm_bo_global *glob = adev->mman.bdev.glob;
|
||||||
int r;
|
int r;
|
||||||
|
|
||||||
spin_lock(&vm->status_lock);
|
|
||||||
while (!list_empty(&vm->evicted)) {
|
while (!list_empty(&vm->evicted)) {
|
||||||
struct amdgpu_vm_bo_base *bo_base;
|
struct amdgpu_vm_bo_base *bo_base;
|
||||||
struct amdgpu_bo *bo;
|
struct amdgpu_bo *bo;
|
||||||
|
@ -236,10 +233,8 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
|
||||||
bo_base = list_first_entry(&vm->evicted,
|
bo_base = list_first_entry(&vm->evicted,
|
||||||
struct amdgpu_vm_bo_base,
|
struct amdgpu_vm_bo_base,
|
||||||
vm_status);
|
vm_status);
|
||||||
spin_unlock(&vm->status_lock);
|
|
||||||
|
|
||||||
bo = bo_base->bo;
|
bo = bo_base->bo;
|
||||||
BUG_ON(!bo);
|
|
||||||
if (bo->parent) {
|
if (bo->parent) {
|
||||||
r = validate(param, bo);
|
r = validate(param, bo);
|
||||||
if (r)
|
if (r)
|
||||||
|
@ -259,13 +254,14 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_lock(&vm->status_lock);
|
if (bo->tbo.type != ttm_bo_type_kernel) {
|
||||||
if (bo->tbo.type != ttm_bo_type_kernel)
|
spin_lock(&vm->moved_lock);
|
||||||
list_move(&bo_base->vm_status, &vm->moved);
|
list_move(&bo_base->vm_status, &vm->moved);
|
||||||
else
|
spin_unlock(&vm->moved_lock);
|
||||||
|
} else {
|
||||||
list_move(&bo_base->vm_status, &vm->relocated);
|
list_move(&bo_base->vm_status, &vm->relocated);
|
||||||
}
|
}
|
||||||
spin_unlock(&vm->status_lock);
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
@ -279,13 +275,7 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
|
||||||
*/
|
*/
|
||||||
bool amdgpu_vm_ready(struct amdgpu_vm *vm)
|
bool amdgpu_vm_ready(struct amdgpu_vm *vm)
|
||||||
{
|
{
|
||||||
bool ready;
|
return list_empty(&vm->evicted);
|
||||||
|
|
||||||
spin_lock(&vm->status_lock);
|
|
||||||
ready = list_empty(&vm->evicted);
|
|
||||||
spin_unlock(&vm->status_lock);
|
|
||||||
|
|
||||||
return ready;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -477,9 +467,7 @@ static int amdgpu_vm_alloc_levels(struct amdgpu_device *adev,
|
||||||
pt->parent = amdgpu_bo_ref(parent->base.bo);
|
pt->parent = amdgpu_bo_ref(parent->base.bo);
|
||||||
|
|
||||||
amdgpu_vm_bo_base_init(&entry->base, vm, pt);
|
amdgpu_vm_bo_base_init(&entry->base, vm, pt);
|
||||||
spin_lock(&vm->status_lock);
|
|
||||||
list_move(&entry->base.vm_status, &vm->relocated);
|
list_move(&entry->base.vm_status, &vm->relocated);
|
||||||
spin_unlock(&vm->status_lock);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (level < AMDGPU_VM_PTB) {
|
if (level < AMDGPU_VM_PTB) {
|
||||||
|
@ -926,10 +914,8 @@ static void amdgpu_vm_invalidate_level(struct amdgpu_device *adev,
|
||||||
if (!entry->base.bo)
|
if (!entry->base.bo)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
spin_lock(&vm->status_lock);
|
|
||||||
if (list_empty(&entry->base.vm_status))
|
if (list_empty(&entry->base.vm_status))
|
||||||
list_add(&entry->base.vm_status, &vm->relocated);
|
list_add(&entry->base.vm_status, &vm->relocated);
|
||||||
spin_unlock(&vm->status_lock);
|
|
||||||
amdgpu_vm_invalidate_level(adev, vm, entry, level + 1);
|
amdgpu_vm_invalidate_level(adev, vm, entry, level + 1);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -974,7 +960,6 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev,
|
||||||
params.func = amdgpu_vm_do_set_ptes;
|
params.func = amdgpu_vm_do_set_ptes;
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_lock(&vm->status_lock);
|
|
||||||
while (!list_empty(&vm->relocated)) {
|
while (!list_empty(&vm->relocated)) {
|
||||||
struct amdgpu_vm_bo_base *bo_base, *parent;
|
struct amdgpu_vm_bo_base *bo_base, *parent;
|
||||||
struct amdgpu_vm_pt *pt, *entry;
|
struct amdgpu_vm_pt *pt, *entry;
|
||||||
|
@ -984,13 +969,10 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev,
|
||||||
struct amdgpu_vm_bo_base,
|
struct amdgpu_vm_bo_base,
|
||||||
vm_status);
|
vm_status);
|
||||||
list_del_init(&bo_base->vm_status);
|
list_del_init(&bo_base->vm_status);
|
||||||
spin_unlock(&vm->status_lock);
|
|
||||||
|
|
||||||
bo = bo_base->bo->parent;
|
bo = bo_base->bo->parent;
|
||||||
if (!bo) {
|
if (!bo)
|
||||||
spin_lock(&vm->status_lock);
|
|
||||||
continue;
|
continue;
|
||||||
}
|
|
||||||
|
|
||||||
parent = list_first_entry(&bo->va, struct amdgpu_vm_bo_base,
|
parent = list_first_entry(&bo->va, struct amdgpu_vm_bo_base,
|
||||||
bo_list);
|
bo_list);
|
||||||
|
@ -999,12 +981,10 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev,
|
||||||
|
|
||||||
amdgpu_vm_update_pde(¶ms, vm, pt, entry);
|
amdgpu_vm_update_pde(¶ms, vm, pt, entry);
|
||||||
|
|
||||||
spin_lock(&vm->status_lock);
|
|
||||||
if (!vm->use_cpu_for_update &&
|
if (!vm->use_cpu_for_update &&
|
||||||
(ndw - params.ib->length_dw) < 32)
|
(ndw - params.ib->length_dw) < 32)
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
spin_unlock(&vm->status_lock);
|
|
||||||
|
|
||||||
if (vm->use_cpu_for_update) {
|
if (vm->use_cpu_for_update) {
|
||||||
/* Flush HDP */
|
/* Flush HDP */
|
||||||
|
@ -1107,9 +1087,7 @@ static void amdgpu_vm_handle_huge_pages(struct amdgpu_pte_update_params *p,
|
||||||
if (entry->huge) {
|
if (entry->huge) {
|
||||||
/* Add the entry to the relocated list to update it. */
|
/* Add the entry to the relocated list to update it. */
|
||||||
entry->huge = false;
|
entry->huge = false;
|
||||||
spin_lock(&p->vm->status_lock);
|
|
||||||
list_move(&entry->base.vm_status, &p->vm->relocated);
|
list_move(&entry->base.vm_status, &p->vm->relocated);
|
||||||
spin_unlock(&p->vm->status_lock);
|
|
||||||
}
|
}
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
@ -1588,8 +1566,9 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev,
|
||||||
amdgpu_asic_flush_hdp(adev, NULL);
|
amdgpu_asic_flush_hdp(adev, NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_lock(&vm->status_lock);
|
spin_lock(&vm->moved_lock);
|
||||||
list_del_init(&bo_va->base.vm_status);
|
list_del_init(&bo_va->base.vm_status);
|
||||||
|
spin_unlock(&vm->moved_lock);
|
||||||
|
|
||||||
/* If the BO is not in its preferred location add it back to
|
/* If the BO is not in its preferred location add it back to
|
||||||
* the evicted list so that it gets validated again on the
|
* the evicted list so that it gets validated again on the
|
||||||
|
@ -1599,7 +1578,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev,
|
||||||
!(bo->preferred_domains &
|
!(bo->preferred_domains &
|
||||||
amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type)))
|
amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type)))
|
||||||
list_add_tail(&bo_va->base.vm_status, &vm->evicted);
|
list_add_tail(&bo_va->base.vm_status, &vm->evicted);
|
||||||
spin_unlock(&vm->status_lock);
|
|
||||||
|
|
||||||
list_splice_init(&bo_va->invalids, &bo_va->valids);
|
list_splice_init(&bo_va->invalids, &bo_va->valids);
|
||||||
bo_va->cleared = clear;
|
bo_va->cleared = clear;
|
||||||
|
@ -1811,14 +1789,14 @@ int amdgpu_vm_handle_moved(struct amdgpu_device *adev,
|
||||||
bool clear;
|
bool clear;
|
||||||
int r = 0;
|
int r = 0;
|
||||||
|
|
||||||
spin_lock(&vm->status_lock);
|
spin_lock(&vm->moved_lock);
|
||||||
while (!list_empty(&vm->moved)) {
|
while (!list_empty(&vm->moved)) {
|
||||||
struct amdgpu_bo_va *bo_va;
|
struct amdgpu_bo_va *bo_va;
|
||||||
struct reservation_object *resv;
|
struct reservation_object *resv;
|
||||||
|
|
||||||
bo_va = list_first_entry(&vm->moved,
|
bo_va = list_first_entry(&vm->moved,
|
||||||
struct amdgpu_bo_va, base.vm_status);
|
struct amdgpu_bo_va, base.vm_status);
|
||||||
spin_unlock(&vm->status_lock);
|
spin_unlock(&vm->moved_lock);
|
||||||
|
|
||||||
resv = bo_va->base.bo->tbo.resv;
|
resv = bo_va->base.bo->tbo.resv;
|
||||||
|
|
||||||
|
@ -1839,9 +1817,9 @@ int amdgpu_vm_handle_moved(struct amdgpu_device *adev,
|
||||||
if (!clear && resv != vm->root.base.bo->tbo.resv)
|
if (!clear && resv != vm->root.base.bo->tbo.resv)
|
||||||
reservation_object_unlock(resv);
|
reservation_object_unlock(resv);
|
||||||
|
|
||||||
spin_lock(&vm->status_lock);
|
spin_lock(&vm->moved_lock);
|
||||||
}
|
}
|
||||||
spin_unlock(&vm->status_lock);
|
spin_unlock(&vm->moved_lock);
|
||||||
|
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
@ -1903,10 +1881,10 @@ static void amdgpu_vm_bo_insert_map(struct amdgpu_device *adev,
|
||||||
amdgpu_vm_prt_get(adev);
|
amdgpu_vm_prt_get(adev);
|
||||||
|
|
||||||
if (bo && bo->tbo.resv == vm->root.base.bo->tbo.resv) {
|
if (bo && bo->tbo.resv == vm->root.base.bo->tbo.resv) {
|
||||||
spin_lock(&vm->status_lock);
|
spin_lock(&vm->moved_lock);
|
||||||
if (list_empty(&bo_va->base.vm_status))
|
if (list_empty(&bo_va->base.vm_status))
|
||||||
list_add(&bo_va->base.vm_status, &vm->moved);
|
list_add(&bo_va->base.vm_status, &vm->moved);
|
||||||
spin_unlock(&vm->status_lock);
|
spin_unlock(&vm->moved_lock);
|
||||||
}
|
}
|
||||||
trace_amdgpu_vm_bo_map(bo_va, mapping);
|
trace_amdgpu_vm_bo_map(bo_va, mapping);
|
||||||
}
|
}
|
||||||
|
@ -2216,9 +2194,9 @@ void amdgpu_vm_bo_rmv(struct amdgpu_device *adev,
|
||||||
|
|
||||||
list_del(&bo_va->base.bo_list);
|
list_del(&bo_va->base.bo_list);
|
||||||
|
|
||||||
spin_lock(&vm->status_lock);
|
spin_lock(&vm->moved_lock);
|
||||||
list_del(&bo_va->base.vm_status);
|
list_del(&bo_va->base.vm_status);
|
||||||
spin_unlock(&vm->status_lock);
|
spin_unlock(&vm->moved_lock);
|
||||||
|
|
||||||
list_for_each_entry_safe(mapping, next, &bo_va->valids, list) {
|
list_for_each_entry_safe(mapping, next, &bo_va->valids, list) {
|
||||||
list_del(&mapping->list);
|
list_del(&mapping->list);
|
||||||
|
@ -2261,28 +2239,24 @@ void amdgpu_vm_bo_invalidate(struct amdgpu_device *adev,
|
||||||
|
|
||||||
bo_base->moved = true;
|
bo_base->moved = true;
|
||||||
if (evicted && bo->tbo.resv == vm->root.base.bo->tbo.resv) {
|
if (evicted && bo->tbo.resv == vm->root.base.bo->tbo.resv) {
|
||||||
spin_lock(&bo_base->vm->status_lock);
|
|
||||||
if (bo->tbo.type == ttm_bo_type_kernel)
|
if (bo->tbo.type == ttm_bo_type_kernel)
|
||||||
list_move(&bo_base->vm_status, &vm->evicted);
|
list_move(&bo_base->vm_status, &vm->evicted);
|
||||||
else
|
else
|
||||||
list_move_tail(&bo_base->vm_status,
|
list_move_tail(&bo_base->vm_status,
|
||||||
&vm->evicted);
|
&vm->evicted);
|
||||||
spin_unlock(&bo_base->vm->status_lock);
|
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (bo->tbo.type == ttm_bo_type_kernel) {
|
if (bo->tbo.type == ttm_bo_type_kernel) {
|
||||||
spin_lock(&bo_base->vm->status_lock);
|
|
||||||
if (list_empty(&bo_base->vm_status))
|
if (list_empty(&bo_base->vm_status))
|
||||||
list_add(&bo_base->vm_status, &vm->relocated);
|
list_add(&bo_base->vm_status, &vm->relocated);
|
||||||
spin_unlock(&bo_base->vm->status_lock);
|
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_lock(&bo_base->vm->status_lock);
|
spin_lock(&bo_base->vm->moved_lock);
|
||||||
if (list_empty(&bo_base->vm_status))
|
if (list_empty(&bo_base->vm_status))
|
||||||
list_add(&bo_base->vm_status, &vm->moved);
|
list_add(&bo_base->vm_status, &vm->moved);
|
||||||
spin_unlock(&bo_base->vm->status_lock);
|
spin_unlock(&bo_base->vm->moved_lock);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2391,9 +2365,9 @@ int amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm,
|
||||||
vm->va = RB_ROOT_CACHED;
|
vm->va = RB_ROOT_CACHED;
|
||||||
for (i = 0; i < AMDGPU_MAX_VMHUBS; i++)
|
for (i = 0; i < AMDGPU_MAX_VMHUBS; i++)
|
||||||
vm->reserved_vmid[i] = NULL;
|
vm->reserved_vmid[i] = NULL;
|
||||||
spin_lock_init(&vm->status_lock);
|
|
||||||
INIT_LIST_HEAD(&vm->evicted);
|
INIT_LIST_HEAD(&vm->evicted);
|
||||||
INIT_LIST_HEAD(&vm->relocated);
|
INIT_LIST_HEAD(&vm->relocated);
|
||||||
|
spin_lock_init(&vm->moved_lock);
|
||||||
INIT_LIST_HEAD(&vm->moved);
|
INIT_LIST_HEAD(&vm->moved);
|
||||||
INIT_LIST_HEAD(&vm->freed);
|
INIT_LIST_HEAD(&vm->freed);
|
||||||
|
|
||||||
|
|
|
@ -168,9 +168,6 @@ struct amdgpu_vm {
|
||||||
/* tree of virtual addresses mapped */
|
/* tree of virtual addresses mapped */
|
||||||
struct rb_root_cached va;
|
struct rb_root_cached va;
|
||||||
|
|
||||||
/* protecting invalidated */
|
|
||||||
spinlock_t status_lock;
|
|
||||||
|
|
||||||
/* BOs who needs a validation */
|
/* BOs who needs a validation */
|
||||||
struct list_head evicted;
|
struct list_head evicted;
|
||||||
|
|
||||||
|
@ -179,6 +176,7 @@ struct amdgpu_vm {
|
||||||
|
|
||||||
/* BOs moved, but not yet updated in the PT */
|
/* BOs moved, but not yet updated in the PT */
|
||||||
struct list_head moved;
|
struct list_head moved;
|
||||||
|
spinlock_t moved_lock;
|
||||||
|
|
||||||
/* BO mappings freed, but not yet updated in the PT */
|
/* BO mappings freed, but not yet updated in the PT */
|
||||||
struct list_head freed;
|
struct list_head freed;
|
||||||
|
|
Loading…
Reference in New Issue