drm: Get rid of drm_sched_job.id

Its only purpose was for trace events, but jobs can already be
uniquely identified using their fence.

The downside of using the fence is that it's only available
after 'drm_sched_job_arm' was called which is true for all trace
events that used job.id so they can safely switch to using it.

Suggested-by: Tvrtko Ursulin <tursulin@igalia.com>
Signed-off-by: Pierre-Eric Pelloux-Prayer <pierre-eric.pelloux-prayer@amd.com>
Reviewed-by: Christian König <christian.koenig@amd.com>
Reviewed-by: Arvind Yadav <arvind.yadav@amd.com>
Signed-off-by: Philipp Stanner <phasta@kernel.org>
Link: https://lore.kernel.org/r/20250526125505.2360-9-pierre-eric.pelloux-prayer@amd.com
This commit is contained in:
Pierre-Eric Pelloux-Prayer
2025-05-26 14:54:50 +02:00
committed by Philipp Stanner
parent f6743e6a00
commit 4f7fa5fa41
5 changed files with 14 additions and 32 deletions

View File

@@ -167,7 +167,6 @@ TRACE_EVENT(amdgpu_cs_ioctl,
TP_PROTO(struct amdgpu_job *job),
TP_ARGS(job),
TP_STRUCT__entry(
__field(uint64_t, sched_job_id)
__string(timeline, AMDGPU_JOB_GET_TIMELINE_NAME(job))
__field(unsigned int, context)
__field(unsigned int, seqno)
@@ -177,15 +176,14 @@ TRACE_EVENT(amdgpu_cs_ioctl,
),
TP_fast_assign(
__entry->sched_job_id = job->base.id;
__assign_str(timeline);
__entry->context = job->base.s_fence->finished.context;
__entry->seqno = job->base.s_fence->finished.seqno;
__assign_str(ring);
__entry->num_ibs = job->num_ibs;
),
TP_printk("sched_job=%llu, timeline=%s, context=%u, seqno=%u, ring_name=%s, num_ibs=%u",
__entry->sched_job_id, __get_str(timeline), __entry->context,
TP_printk("timeline=%s, context=%u, seqno=%u, ring_name=%s, num_ibs=%u",
__get_str(timeline), __entry->context,
__entry->seqno, __get_str(ring), __entry->num_ibs)
);
@@ -193,7 +191,6 @@ TRACE_EVENT(amdgpu_sched_run_job,
TP_PROTO(struct amdgpu_job *job),
TP_ARGS(job),
TP_STRUCT__entry(
__field(uint64_t, sched_job_id)
__string(timeline, AMDGPU_JOB_GET_TIMELINE_NAME(job))
__field(unsigned int, context)
__field(unsigned int, seqno)
@@ -202,15 +199,14 @@ TRACE_EVENT(amdgpu_sched_run_job,
),
TP_fast_assign(
__entry->sched_job_id = job->base.id;
__assign_str(timeline);
__entry->context = job->base.s_fence->finished.context;
__entry->seqno = job->base.s_fence->finished.seqno;
__assign_str(ring);
__entry->num_ibs = job->num_ibs;
),
TP_printk("sched_job=%llu, timeline=%s, context=%u, seqno=%u, ring_name=%s, num_ibs=%u",
__entry->sched_job_id, __get_str(timeline), __entry->context,
TP_printk("timeline=%s, context=%u, seqno=%u, ring_name=%s, num_ibs=%u",
__get_str(timeline), __entry->context,
__entry->seqno, __get_str(ring), __entry->num_ibs)
);
@@ -551,7 +547,6 @@ TRACE_EVENT(amdgpu_ib_pipe_sync,
TP_ARGS(sched_job, fence),
TP_STRUCT__entry(
__string(ring, sched_job->base.sched->name)
__field(uint64_t, id)
__field(struct dma_fence *, fence)
__field(uint64_t, ctx)
__field(unsigned, seqno)
@@ -559,13 +554,12 @@ TRACE_EVENT(amdgpu_ib_pipe_sync,
TP_fast_assign(
__assign_str(ring);
__entry->id = sched_job->base.id;
__entry->fence = fence;
__entry->ctx = fence->context;
__entry->seqno = fence->seqno;
),
TP_printk("job ring=%s, id=%llu, need pipe sync to fence=%p, context=%llu, seq=%u",
__get_str(ring), __entry->id,
TP_printk("job ring=%s need pipe sync to fence=%p, context=%llu, seq=%u",
__get_str(ring),
__entry->fence, __entry->ctx,
__entry->seqno)
);

View File

@@ -14,21 +14,19 @@ DECLARE_EVENT_CLASS(lima_task,
TP_PROTO(struct lima_sched_task *task),
TP_ARGS(task),
TP_STRUCT__entry(
__field(uint64_t, task_id)
__field(unsigned int, context)
__field(unsigned int, seqno)
__string(pipe, task->base.sched->name)
),
TP_fast_assign(
__entry->task_id = task->base.id;
__entry->context = task->base.s_fence->finished.context;
__entry->seqno = task->base.s_fence->finished.seqno;
__assign_str(pipe);
),
TP_printk("task=%llu, context=%u seqno=%u pipe=%s",
__entry->task_id, __entry->context, __entry->seqno,
TP_printk("context=%u seqno=%u pipe=%s",
__entry->context, __entry->seqno,
__get_str(pipe))
);

View File

@@ -36,7 +36,6 @@ DECLARE_EVENT_CLASS(drm_sched_job,
TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
TP_ARGS(sched_job, entity),
TP_STRUCT__entry(
__field(uint64_t, id)
__string(name, sched_job->sched->name)
__field(u32, job_count)
__field(int, hw_job_count)
@@ -47,7 +46,6 @@ DECLARE_EVENT_CLASS(drm_sched_job,
),
TP_fast_assign(
__entry->id = sched_job->id;
__assign_str(name);
__entry->job_count = spsc_queue_count(&entity->job_queue);
__entry->hw_job_count = atomic_read(
@@ -57,8 +55,8 @@ DECLARE_EVENT_CLASS(drm_sched_job,
__entry->fence_seqno = sched_job->s_fence->finished.seqno;
__entry->client_id = sched_job->s_fence->drm_client_id;
),
TP_printk("dev=%s, id=%llu, fence=%llu:%llu, ring=%s, job count:%u, hw job count:%d, client_id:%llu",
__get_str(dev), __entry->id,
TP_printk("dev=%s, fence=%llu:%llu, ring=%s, job count:%u, hw job count:%d, client_id:%llu",
__get_str(dev),
__entry->fence_context, __entry->fence_seqno, __get_str(name),
__entry->job_count, __entry->hw_job_count, __entry->client_id)
);
@@ -95,7 +93,6 @@ TRACE_EVENT(drm_sched_job_add_dep,
TP_STRUCT__entry(
__field(u64, fence_context)
__field(u64, fence_seqno)
__field(u64, id)
__field(u64, ctx)
__field(u64, seqno)
),
@@ -103,12 +100,11 @@ TRACE_EVENT(drm_sched_job_add_dep,
TP_fast_assign(
__entry->fence_context = sched_job->s_fence->finished.context;
__entry->fence_seqno = sched_job->s_fence->finished.seqno;
__entry->id = sched_job->id;
__entry->ctx = fence->context;
__entry->seqno = fence->seqno;
),
TP_printk("fence=%llu:%llu, id=%llu depends on fence=%llu:%llu",
__entry->fence_context, __entry->fence_seqno, __entry->id,
TP_printk("fence=%llu:%llu depends on fence=%llu:%llu",
__entry->fence_context, __entry->fence_seqno,
__entry->ctx, __entry->seqno)
);
@@ -118,7 +114,6 @@ TRACE_EVENT(drm_sched_job_unschedulable,
TP_STRUCT__entry(
__field(u64, fence_context)
__field(u64, fence_seqno)
__field(uint64_t, id)
__field(u64, ctx)
__field(u64, seqno)
),
@@ -126,12 +121,11 @@ TRACE_EVENT(drm_sched_job_unschedulable,
TP_fast_assign(
__entry->fence_context = sched_job->s_fence->finished.context;
__entry->fence_seqno = sched_job->s_fence->finished.seqno;
__entry->id = sched_job->id;
__entry->ctx = fence->context;
__entry->seqno = fence->seqno;
),
TP_printk("fence=%llu:%llu, id=%llu depends on unsignalled fence=%llu:%llu",
__entry->fence_context, __entry->fence_seqno, __entry->id,
TP_printk("fence=%llu:%llu depends on unsignalled fence=%llu:%llu",
__entry->fence_context, __entry->fence_seqno,
__entry->ctx, __entry->seqno)
);

View File

@@ -855,7 +855,6 @@ void drm_sched_job_arm(struct drm_sched_job *job)
job->sched = sched;
job->s_priority = entity->priority;
job->id = atomic64_inc_return(&sched->job_id_count);
drm_sched_fence_init(job->s_fence, job->entity);
}

View File

@@ -326,7 +326,6 @@ struct drm_sched_fence *to_drm_sched_fence(struct dma_fence *f);
* @finish_cb: the callback for the finished fence.
* @credits: the number of credits this job contributes to the scheduler
* @work: Helper to reschedule job kill to different context.
* @id: a unique id assigned to each job scheduled on the scheduler.
* @karma: increment on every hang caused by this job. If this exceeds the hang
* limit of the scheduler then the job is marked guilty and will not
* be scheduled further.
@@ -339,8 +338,6 @@ struct drm_sched_fence *to_drm_sched_fence(struct dma_fence *f);
* to schedule the job.
*/
struct drm_sched_job {
u64 id;
/**
* @submit_ts:
*