Unreference the shadow BOs in the error path as well and drop the NULL checks.
Signed-off-by: Christian König <christian.koenig@amd.com>
Acked-by: Edward O'Callaghan <funfunctor@folklore1984.net>
Reviewed-by: Alex Deucher <alexander.deucher@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
r = amdgpu_vm_clear_bo(adev, vm, pt);
if (r) {
r = amdgpu_vm_clear_bo(adev, vm, pt);
if (r) {
+ amdgpu_bo_unref(&pt->shadow);
amdgpu_bo_unref(&pt);
goto error_free;
}
amdgpu_bo_unref(&pt);
goto error_free;
}
return 0;
error_free_page_directory:
return 0;
error_free_page_directory:
+ amdgpu_bo_unref(&vm->page_directory->shadow);
amdgpu_bo_unref(&vm->page_directory);
vm->page_directory = NULL;
amdgpu_bo_unref(&vm->page_directory);
vm->page_directory = NULL;
}
for (i = 0; i < amdgpu_vm_num_pdes(adev); i++) {
}
for (i = 0; i < amdgpu_vm_num_pdes(adev); i++) {
- if (vm->page_tables[i].entry.robj &&
- vm->page_tables[i].entry.robj->shadow)
- amdgpu_bo_unref(&vm->page_tables[i].entry.robj->shadow);
- amdgpu_bo_unref(&vm->page_tables[i].entry.robj);
+ struct amdgpu_bo *pt = vm->page_tables[i].entry.robj;
+
+ if (!pt)
+ continue;
+
+ amdgpu_bo_unref(&pt->shadow);
+ amdgpu_bo_unref(&pt);
}
drm_free_large(vm->page_tables);
}
drm_free_large(vm->page_tables);
- if (vm->page_directory->shadow)
- amdgpu_bo_unref(&vm->page_directory->shadow);
+ amdgpu_bo_unref(&vm->page_directory->shadow);
amdgpu_bo_unref(&vm->page_directory);
fence_put(vm->page_directory_fence);
}
amdgpu_bo_unref(&vm->page_directory);
fence_put(vm->page_directory_fence);
}