drm/xe: Take a reference in xe_exec_queue_last_fence_get()
Take a reference in xe_exec_queue_last_fence_get(). Also fix a reference counting underflow bug VM bind and unbind. Fixes:dd08ebf6c3
("drm/xe: Introduce a new DRM driver for Intel GPUs") Signed-off-by: Matthew Brost <matthew.brost@intel.com> Reviewed-by: Maarten Lankhorst <maarten.lankhorst@linux.intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20240201004849.2219558-2-matthew.brost@intel.com (cherry picked from commita856b67a84
) Signed-off-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
This commit is contained in:
parent
ddc7d4c584
commit
fc29b6d5ab
|
@ -926,20 +926,24 @@ void xe_exec_queue_last_fence_put_unlocked(struct xe_exec_queue *q)
|
|||
* @q: The exec queue
|
||||
* @vm: The VM the engine does a bind or exec for
|
||||
*
|
||||
* Get last fence, does not take a ref
|
||||
* Get last fence, takes a ref
|
||||
*
|
||||
* Returns: last fence if not signaled, dma fence stub if signaled
|
||||
*/
|
||||
struct dma_fence *xe_exec_queue_last_fence_get(struct xe_exec_queue *q,
|
||||
struct xe_vm *vm)
|
||||
{
|
||||
struct dma_fence *fence;
|
||||
|
||||
xe_exec_queue_last_fence_lockdep_assert(q, vm);
|
||||
|
||||
if (q->last_fence &&
|
||||
test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &q->last_fence->flags))
|
||||
xe_exec_queue_last_fence_put(q, vm);
|
||||
|
||||
return q->last_fence ? q->last_fence : dma_fence_get_stub();
|
||||
fence = q->last_fence ? q->last_fence : dma_fence_get_stub();
|
||||
dma_fence_get(fence);
|
||||
return fence;
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -1204,8 +1204,11 @@ static bool no_in_syncs(struct xe_vm *vm, struct xe_exec_queue *q,
|
|||
}
|
||||
if (q) {
|
||||
fence = xe_exec_queue_last_fence_get(q, vm);
|
||||
if (!test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
|
||||
if (!test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags)) {
|
||||
dma_fence_put(fence);
|
||||
return false;
|
||||
}
|
||||
dma_fence_put(fence);
|
||||
}
|
||||
|
||||
return true;
|
||||
|
|
|
@ -274,7 +274,6 @@ int xe_sched_job_last_fence_add_dep(struct xe_sched_job *job, struct xe_vm *vm)
|
|||
struct dma_fence *fence;
|
||||
|
||||
fence = xe_exec_queue_last_fence_get(job->q, vm);
|
||||
dma_fence_get(fence);
|
||||
|
||||
return drm_sched_job_add_dependency(&job->drm, fence);
|
||||
}
|
||||
|
|
|
@ -307,7 +307,6 @@ xe_sync_in_fence_get(struct xe_sync_entry *sync, int num_sync,
|
|||
/* Easy case... */
|
||||
if (!num_in_fence) {
|
||||
fence = xe_exec_queue_last_fence_get(q, vm);
|
||||
dma_fence_get(fence);
|
||||
return fence;
|
||||
}
|
||||
|
||||
|
@ -322,7 +321,6 @@ xe_sync_in_fence_get(struct xe_sync_entry *sync, int num_sync,
|
|||
}
|
||||
}
|
||||
fences[current_fence++] = xe_exec_queue_last_fence_get(q, vm);
|
||||
dma_fence_get(fences[current_fence - 1]);
|
||||
cf = dma_fence_array_create(num_in_fence, fences,
|
||||
vm->composite_fence_ctx,
|
||||
vm->composite_fence_seqno++,
|
||||
|
|
|
@ -1984,6 +1984,7 @@ static int xe_vm_prefetch(struct xe_vm *vm, struct xe_vma *vma,
|
|||
xe_exec_queue_last_fence_get(wait_exec_queue, vm);
|
||||
|
||||
xe_sync_entry_signal(&syncs[i], NULL, fence);
|
||||
dma_fence_put(fence);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in New Issue