mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-10-02 07:04:24 +00:00
drm/ttm: remove lazy parameter from ttm_bo_wait
Not used any more. Reviewed-by: Sinclair Yeh <syeh@vmware.com> Signed-off-by: Christian König <christian.koenig@amd.com> Reviewed-by: Alex Deucher <alexander.deucher@amd.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
This commit is contained in:
parent
dfd5e50ea4
commit
8aa6d4fc5f
13 changed files with 24 additions and 24 deletions
|
@ -1322,7 +1322,7 @@ nouveau_bo_move(struct ttm_buffer_object *bo, bool evict, bool intr,
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Fallback to software copy. */
|
/* Fallback to software copy. */
|
||||||
ret = ttm_bo_wait(bo, true, intr, no_wait_gpu);
|
ret = ttm_bo_wait(bo, intr, no_wait_gpu);
|
||||||
if (ret == 0)
|
if (ret == 0)
|
||||||
ret = ttm_bo_move_memcpy(bo, evict, no_wait_gpu, new_mem);
|
ret = ttm_bo_move_memcpy(bo, evict, no_wait_gpu, new_mem);
|
||||||
|
|
||||||
|
|
|
@ -126,7 +126,7 @@ nouveau_gem_object_unmap(struct nouveau_bo *nvbo, struct nvkm_vma *vma)
|
||||||
list_del(&vma->head);
|
list_del(&vma->head);
|
||||||
|
|
||||||
if (fobj && fobj->shared_count > 1)
|
if (fobj && fobj->shared_count > 1)
|
||||||
ttm_bo_wait(&nvbo->bo, true, false, false);
|
ttm_bo_wait(&nvbo->bo, false, false);
|
||||||
else if (fobj && fobj->shared_count == 1)
|
else if (fobj && fobj->shared_count == 1)
|
||||||
fence = rcu_dereference_protected(fobj->shared[0],
|
fence = rcu_dereference_protected(fobj->shared[0],
|
||||||
reservation_object_held(resv));
|
reservation_object_held(resv));
|
||||||
|
@ -651,7 +651,7 @@ nouveau_gem_pushbuf_reloc_apply(struct nouveau_cli *cli,
|
||||||
data |= r->vor;
|
data |= r->vor;
|
||||||
}
|
}
|
||||||
|
|
||||||
ret = ttm_bo_wait(&nvbo->bo, true, false, false);
|
ret = ttm_bo_wait(&nvbo->bo, false, false);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
NV_PRINTK(err, cli, "reloc wait_idle failed: %d\n", ret);
|
NV_PRINTK(err, cli, "reloc wait_idle failed: %d\n", ret);
|
||||||
break;
|
break;
|
||||||
|
|
|
@ -624,7 +624,7 @@ static int qxl_reap_surf(struct qxl_device *qdev, struct qxl_bo *surf, bool stal
|
||||||
if (stall)
|
if (stall)
|
||||||
mutex_unlock(&qdev->surf_evict_mutex);
|
mutex_unlock(&qdev->surf_evict_mutex);
|
||||||
|
|
||||||
ret = ttm_bo_wait(&surf->tbo, true, true, !stall);
|
ret = ttm_bo_wait(&surf->tbo, true, !stall);
|
||||||
|
|
||||||
if (stall)
|
if (stall)
|
||||||
mutex_lock(&qdev->surf_evict_mutex);
|
mutex_lock(&qdev->surf_evict_mutex);
|
||||||
|
|
|
@ -79,7 +79,7 @@ static inline int qxl_bo_wait(struct qxl_bo *bo, u32 *mem_type,
|
||||||
if (mem_type)
|
if (mem_type)
|
||||||
*mem_type = bo->tbo.mem.mem_type;
|
*mem_type = bo->tbo.mem.mem_type;
|
||||||
|
|
||||||
r = ttm_bo_wait(&bo->tbo, true, true, no_wait);
|
r = ttm_bo_wait(&bo->tbo, true, no_wait);
|
||||||
ttm_bo_unreserve(&bo->tbo);
|
ttm_bo_unreserve(&bo->tbo);
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
|
|
@ -838,7 +838,7 @@ int radeon_bo_wait(struct radeon_bo *bo, u32 *mem_type, bool no_wait)
|
||||||
if (mem_type)
|
if (mem_type)
|
||||||
*mem_type = bo->tbo.mem.mem_type;
|
*mem_type = bo->tbo.mem.mem_type;
|
||||||
|
|
||||||
r = ttm_bo_wait(&bo->tbo, true, true, no_wait);
|
r = ttm_bo_wait(&bo->tbo, true, no_wait);
|
||||||
ttm_bo_unreserve(&bo->tbo);
|
ttm_bo_unreserve(&bo->tbo);
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
|
|
@ -455,7 +455,7 @@ static void ttm_bo_cleanup_refs_or_queue(struct ttm_buffer_object *bo)
|
||||||
ret = __ttm_bo_reserve(bo, false, true, NULL);
|
ret = __ttm_bo_reserve(bo, false, true, NULL);
|
||||||
|
|
||||||
if (!ret) {
|
if (!ret) {
|
||||||
if (!ttm_bo_wait(bo, false, false, true)) {
|
if (!ttm_bo_wait(bo, false, true)) {
|
||||||
put_count = ttm_bo_del_from_lru(bo);
|
put_count = ttm_bo_del_from_lru(bo);
|
||||||
|
|
||||||
spin_unlock(&glob->lru_lock);
|
spin_unlock(&glob->lru_lock);
|
||||||
|
@ -508,7 +508,7 @@ static int ttm_bo_cleanup_refs_and_unlock(struct ttm_buffer_object *bo,
|
||||||
int put_count;
|
int put_count;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
ret = ttm_bo_wait(bo, false, false, true);
|
ret = ttm_bo_wait(bo, false, true);
|
||||||
|
|
||||||
if (ret && !no_wait_gpu) {
|
if (ret && !no_wait_gpu) {
|
||||||
long lret;
|
long lret;
|
||||||
|
@ -545,7 +545,7 @@ static int ttm_bo_cleanup_refs_and_unlock(struct ttm_buffer_object *bo,
|
||||||
* remove sync_obj with ttm_bo_wait, the wait should be
|
* remove sync_obj with ttm_bo_wait, the wait should be
|
||||||
* finished, and no new wait object should have been added.
|
* finished, and no new wait object should have been added.
|
||||||
*/
|
*/
|
||||||
ret = ttm_bo_wait(bo, false, false, true);
|
ret = ttm_bo_wait(bo, false, true);
|
||||||
WARN_ON(ret);
|
WARN_ON(ret);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -684,7 +684,7 @@ static int ttm_bo_evict(struct ttm_buffer_object *bo, bool interruptible,
|
||||||
struct ttm_placement placement;
|
struct ttm_placement placement;
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
|
||||||
ret = ttm_bo_wait(bo, false, interruptible, no_wait_gpu);
|
ret = ttm_bo_wait(bo, interruptible, no_wait_gpu);
|
||||||
|
|
||||||
if (unlikely(ret != 0)) {
|
if (unlikely(ret != 0)) {
|
||||||
if (ret != -ERESTARTSYS) {
|
if (ret != -ERESTARTSYS) {
|
||||||
|
@ -1006,7 +1006,7 @@ static int ttm_bo_move_buffer(struct ttm_buffer_object *bo,
|
||||||
* Have the driver move function wait for idle when necessary,
|
* Have the driver move function wait for idle when necessary,
|
||||||
* instead of doing it here.
|
* instead of doing it here.
|
||||||
*/
|
*/
|
||||||
ret = ttm_bo_wait(bo, false, interruptible, no_wait_gpu);
|
ret = ttm_bo_wait(bo, interruptible, no_wait_gpu);
|
||||||
if (ret)
|
if (ret)
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
@ -1567,7 +1567,7 @@ void ttm_bo_unmap_virtual(struct ttm_buffer_object *bo)
|
||||||
EXPORT_SYMBOL(ttm_bo_unmap_virtual);
|
EXPORT_SYMBOL(ttm_bo_unmap_virtual);
|
||||||
|
|
||||||
int ttm_bo_wait(struct ttm_buffer_object *bo,
|
int ttm_bo_wait(struct ttm_buffer_object *bo,
|
||||||
bool lazy, bool interruptible, bool no_wait)
|
bool interruptible, bool no_wait)
|
||||||
{
|
{
|
||||||
struct reservation_object_list *fobj;
|
struct reservation_object_list *fobj;
|
||||||
struct reservation_object *resv;
|
struct reservation_object *resv;
|
||||||
|
@ -1625,7 +1625,7 @@ int ttm_bo_synccpu_write_grab(struct ttm_buffer_object *bo, bool no_wait)
|
||||||
ret = ttm_bo_reserve(bo, true, no_wait, NULL);
|
ret = ttm_bo_reserve(bo, true, no_wait, NULL);
|
||||||
if (unlikely(ret != 0))
|
if (unlikely(ret != 0))
|
||||||
return ret;
|
return ret;
|
||||||
ret = ttm_bo_wait(bo, false, true, no_wait);
|
ret = ttm_bo_wait(bo, true, no_wait);
|
||||||
if (likely(ret == 0))
|
if (likely(ret == 0))
|
||||||
atomic_inc(&bo->cpu_writers);
|
atomic_inc(&bo->cpu_writers);
|
||||||
ttm_bo_unreserve(bo);
|
ttm_bo_unreserve(bo);
|
||||||
|
@ -1682,7 +1682,7 @@ static int ttm_bo_swapout(struct ttm_mem_shrink *shrink)
|
||||||
* Wait for GPU, then move to system cached.
|
* Wait for GPU, then move to system cached.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
ret = ttm_bo_wait(bo, false, false, false);
|
ret = ttm_bo_wait(bo, false, false);
|
||||||
|
|
||||||
if (unlikely(ret != 0))
|
if (unlikely(ret != 0))
|
||||||
goto out;
|
goto out;
|
||||||
|
|
|
@ -645,7 +645,7 @@ int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
|
||||||
|
|
||||||
reservation_object_add_excl_fence(bo->resv, fence);
|
reservation_object_add_excl_fence(bo->resv, fence);
|
||||||
if (evict) {
|
if (evict) {
|
||||||
ret = ttm_bo_wait(bo, false, false, false);
|
ret = ttm_bo_wait(bo, false, false);
|
||||||
if (ret)
|
if (ret)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
||||||
|
|
|
@ -54,7 +54,7 @@ static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
|
||||||
/*
|
/*
|
||||||
* Quick non-stalling check for idle.
|
* Quick non-stalling check for idle.
|
||||||
*/
|
*/
|
||||||
ret = ttm_bo_wait(bo, false, false, true);
|
ret = ttm_bo_wait(bo, false, true);
|
||||||
if (likely(ret == 0))
|
if (likely(ret == 0))
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
|
|
||||||
|
@ -68,14 +68,14 @@ static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
|
|
||||||
up_read(&vma->vm_mm->mmap_sem);
|
up_read(&vma->vm_mm->mmap_sem);
|
||||||
(void) ttm_bo_wait(bo, false, true, false);
|
(void) ttm_bo_wait(bo, true, false);
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Ordinary wait.
|
* Ordinary wait.
|
||||||
*/
|
*/
|
||||||
ret = ttm_bo_wait(bo, false, true, false);
|
ret = ttm_bo_wait(bo, true, false);
|
||||||
if (unlikely(ret != 0))
|
if (unlikely(ret != 0))
|
||||||
ret = (ret != -ERESTARTSYS) ? VM_FAULT_SIGBUS :
|
ret = (ret != -ERESTARTSYS) ? VM_FAULT_SIGBUS :
|
||||||
VM_FAULT_NOPAGE;
|
VM_FAULT_NOPAGE;
|
||||||
|
|
|
@ -158,7 +158,7 @@ int virtio_gpu_object_wait(struct virtio_gpu_object *bo, bool no_wait)
|
||||||
r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL);
|
r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL);
|
||||||
if (unlikely(r != 0))
|
if (unlikely(r != 0))
|
||||||
return r;
|
return r;
|
||||||
r = ttm_bo_wait(&bo->tbo, true, true, no_wait);
|
r = ttm_bo_wait(&bo->tbo, true, no_wait);
|
||||||
ttm_bo_unreserve(&bo->tbo);
|
ttm_bo_unreserve(&bo->tbo);
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
|
|
@ -839,7 +839,7 @@ static void vmw_move_notify(struct ttm_buffer_object *bo,
|
||||||
*/
|
*/
|
||||||
static void vmw_swap_notify(struct ttm_buffer_object *bo)
|
static void vmw_swap_notify(struct ttm_buffer_object *bo)
|
||||||
{
|
{
|
||||||
ttm_bo_wait(bo, false, false, false);
|
ttm_bo_wait(bo, false, false);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -423,7 +423,7 @@ static int vmw_cotable_resize(struct vmw_resource *res, size_t new_size)
|
||||||
bo = &buf->base;
|
bo = &buf->base;
|
||||||
WARN_ON_ONCE(ttm_bo_reserve(bo, false, true, NULL));
|
WARN_ON_ONCE(ttm_bo_reserve(bo, false, true, NULL));
|
||||||
|
|
||||||
ret = ttm_bo_wait(old_bo, false, false, false);
|
ret = ttm_bo_wait(old_bo, false, false);
|
||||||
if (unlikely(ret != 0)) {
|
if (unlikely(ret != 0)) {
|
||||||
DRM_ERROR("Failed waiting for cotable unbind.\n");
|
DRM_ERROR("Failed waiting for cotable unbind.\n");
|
||||||
goto out_wait;
|
goto out_wait;
|
||||||
|
|
|
@ -1512,7 +1512,7 @@ void vmw_resource_move_notify(struct ttm_buffer_object *bo,
|
||||||
list_del_init(&res->mob_head);
|
list_del_init(&res->mob_head);
|
||||||
}
|
}
|
||||||
|
|
||||||
(void) ttm_bo_wait(bo, false, false, false);
|
(void) ttm_bo_wait(bo, false, false);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1605,7 +1605,7 @@ void vmw_query_move_notify(struct ttm_buffer_object *bo,
|
||||||
if (fence != NULL)
|
if (fence != NULL)
|
||||||
vmw_fence_obj_unreference(&fence);
|
vmw_fence_obj_unreference(&fence);
|
||||||
|
|
||||||
(void) ttm_bo_wait(bo, false, false, false);
|
(void) ttm_bo_wait(bo, false, false);
|
||||||
} else
|
} else
|
||||||
mutex_unlock(&dev_priv->binding_mutex);
|
mutex_unlock(&dev_priv->binding_mutex);
|
||||||
|
|
||||||
|
|
|
@ -314,7 +314,7 @@ ttm_bo_reference(struct ttm_buffer_object *bo)
|
||||||
* Returns -EBUSY if no_wait is true and the buffer is busy.
|
* Returns -EBUSY if no_wait is true and the buffer is busy.
|
||||||
* Returns -ERESTARTSYS if interrupted by a signal.
|
* Returns -ERESTARTSYS if interrupted by a signal.
|
||||||
*/
|
*/
|
||||||
extern int ttm_bo_wait(struct ttm_buffer_object *bo, bool lazy,
|
extern int ttm_bo_wait(struct ttm_buffer_object *bo,
|
||||||
bool interruptible, bool no_wait);
|
bool interruptible, bool no_wait);
|
||||||
/**
|
/**
|
||||||
* ttm_bo_validate
|
* ttm_bo_validate
|
||||||
|
|
Loading…
Reference in a new issue