613 lines
22 KiB
Diff
613 lines
22 KiB
Diff
From a78422e9dff366b3a46ae44caf6ec8ded9c9fc2f Mon Sep 17 00:00:00 2001
|
|
From: Danilo Krummrich <dakr@redhat.com>
|
|
Date: Fri, 10 Nov 2023 01:16:33 +0100
|
|
Subject: [PATCH] drm/sched: implement dynamic job-flow control
|
|
|
|
Currently, job flow control is implemented simply by limiting the number
|
|
of jobs in flight. Therefore, a scheduler is initialized with a credit
|
|
limit that corresponds to the number of jobs which can be sent to the
|
|
hardware.
|
|
|
|
This implies that for each job, drivers need to account for the maximum
|
|
job size possible in order to not overflow the ring buffer.
|
|
|
|
However, there are drivers, such as Nouveau, where the job size has a
|
|
rather large range. For such drivers it can easily happen that job
|
|
submissions not even filling the ring by 1% can block subsequent
|
|
submissions, which, in the worst case, can lead to the ring run dry.
|
|
|
|
In order to overcome this issue, allow for tracking the actual job size
|
|
instead of the number of jobs. Therefore, add a field to track a job's
|
|
credit count, which represents the number of credits a job contributes
|
|
to the scheduler's credit limit.
|
|
|
|
Signed-off-by: Danilo Krummrich <dakr@redhat.com>
|
|
Reviewed-by: Luben Tuikov <ltuikov89@gmail.com>
|
|
Link: https://patchwork.freedesktop.org/patch/msgid/20231110001638.71750-1-dakr@redhat.com
|
|
---
|
|
Documentation/gpu/drm-mm.rst | 6 +
|
|
drivers/gpu/drm/amd/amdgpu/amdgpu_job.c | 2 +-
|
|
drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c | 2 +-
|
|
drivers/gpu/drm/etnaviv/etnaviv_gpu.c | 2 +-
|
|
drivers/gpu/drm/lima/lima_device.c | 2 +-
|
|
drivers/gpu/drm/lima/lima_sched.c | 2 +-
|
|
drivers/gpu/drm/msm/msm_gem_submit.c | 2 +-
|
|
drivers/gpu/drm/nouveau/nouveau_sched.c | 2 +-
|
|
drivers/gpu/drm/panfrost/panfrost_drv.c | 2 +-
|
|
drivers/gpu/drm/panfrost/panfrost_job.c | 2 +-
|
|
.../gpu/drm/scheduler/gpu_scheduler_trace.h | 2 +-
|
|
drivers/gpu/drm/scheduler/sched_main.c | 170 ++++++++++++++----
|
|
drivers/gpu/drm/v3d/v3d_gem.c | 2 +-
|
|
include/drm/gpu_scheduler.h | 28 ++-
|
|
14 files changed, 175 insertions(+), 51 deletions(-)
|
|
|
|
--- a/Documentation/gpu/drm-mm.rst
|
|
+++ b/Documentation/gpu/drm-mm.rst
|
|
@@ -552,6 +552,12 @@ Overview
|
|
.. kernel-doc:: drivers/gpu/drm/scheduler/sched_main.c
|
|
:doc: Overview
|
|
|
|
+Flow Control
|
|
+------------
|
|
+
|
|
+.. kernel-doc:: drivers/gpu/drm/scheduler/sched_main.c
|
|
+ :doc: Flow Control
|
|
+
|
|
Scheduler Function References
|
|
-----------------------------
|
|
|
|
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
|
|
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
|
|
@@ -115,7 +115,7 @@ int amdgpu_job_alloc(struct amdgpu_devic
|
|
if (!entity)
|
|
return 0;
|
|
|
|
- return drm_sched_job_init(&(*job)->base, entity, owner);
|
|
+ return drm_sched_job_init(&(*job)->base, entity, 1, owner);
|
|
}
|
|
|
|
int amdgpu_job_alloc_with_ib(struct amdgpu_device *adev,
|
|
--- a/drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
|
|
+++ b/drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
|
|
@@ -535,7 +535,7 @@ int etnaviv_ioctl_gem_submit(struct drm_
|
|
|
|
ret = drm_sched_job_init(&submit->sched_job,
|
|
&ctx->sched_entity[args->pipe],
|
|
- submit->ctx);
|
|
+ 1, submit->ctx);
|
|
if (ret)
|
|
goto err_submit_put;
|
|
|
|
--- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c
|
|
+++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.c
|
|
@@ -1917,7 +1917,7 @@ static int etnaviv_gpu_rpm_suspend(struc
|
|
u32 idle, mask;
|
|
|
|
/* If there are any jobs in the HW queue, we're not idle */
|
|
- if (atomic_read(&gpu->sched.hw_rq_count))
|
|
+ if (atomic_read(&gpu->sched.credit_count))
|
|
return -EBUSY;
|
|
|
|
/* Check whether the hardware (except FE and MC) is idle */
|
|
--- a/drivers/gpu/drm/lima/lima_device.c
|
|
+++ b/drivers/gpu/drm/lima/lima_device.c
|
|
@@ -514,7 +514,7 @@ int lima_device_suspend(struct device *d
|
|
|
|
/* check any task running */
|
|
for (i = 0; i < lima_pipe_num; i++) {
|
|
- if (atomic_read(&ldev->pipe[i].base.hw_rq_count))
|
|
+ if (atomic_read(&ldev->pipe[i].base.credit_count))
|
|
return -EBUSY;
|
|
}
|
|
|
|
--- a/drivers/gpu/drm/lima/lima_sched.c
|
|
+++ b/drivers/gpu/drm/lima/lima_sched.c
|
|
@@ -123,7 +123,7 @@ int lima_sched_task_init(struct lima_sch
|
|
for (i = 0; i < num_bos; i++)
|
|
drm_gem_object_get(&bos[i]->base.base);
|
|
|
|
- err = drm_sched_job_init(&task->base, &context->base, vm);
|
|
+ err = drm_sched_job_init(&task->base, &context->base, 1, vm);
|
|
if (err) {
|
|
kfree(task->bos);
|
|
return err;
|
|
--- a/drivers/gpu/drm/msm/msm_gem_submit.c
|
|
+++ b/drivers/gpu/drm/msm/msm_gem_submit.c
|
|
@@ -48,7 +48,7 @@ static struct msm_gem_submit *submit_cre
|
|
return ERR_PTR(ret);
|
|
}
|
|
|
|
- ret = drm_sched_job_init(&submit->base, queue->entity, queue);
|
|
+ ret = drm_sched_job_init(&submit->base, queue->entity, 1, queue);
|
|
if (ret) {
|
|
kfree(submit->hw_fence);
|
|
kfree(submit);
|
|
--- a/drivers/gpu/drm/nouveau/nouveau_sched.c
|
|
+++ b/drivers/gpu/drm/nouveau/nouveau_sched.c
|
|
@@ -89,7 +89,7 @@ nouveau_job_init(struct nouveau_job *job
|
|
|
|
}
|
|
|
|
- ret = drm_sched_job_init(&job->base, &entity->base, NULL);
|
|
+ ret = drm_sched_job_init(&job->base, &entity->base, 1, NULL);
|
|
if (ret)
|
|
goto err_free_chains;
|
|
|
|
--- a/drivers/gpu/drm/panfrost/panfrost_drv.c
|
|
+++ b/drivers/gpu/drm/panfrost/panfrost_drv.c
|
|
@@ -272,7 +272,7 @@ static int panfrost_ioctl_submit(struct
|
|
|
|
ret = drm_sched_job_init(&job->base,
|
|
&file_priv->sched_entity[slot],
|
|
- NULL);
|
|
+ 1, NULL);
|
|
if (ret)
|
|
goto out_put_job;
|
|
|
|
--- a/drivers/gpu/drm/panfrost/panfrost_job.c
|
|
+++ b/drivers/gpu/drm/panfrost/panfrost_job.c
|
|
@@ -939,7 +939,7 @@ int panfrost_job_is_idle(struct panfrost
|
|
|
|
for (i = 0; i < NUM_JOB_SLOTS; i++) {
|
|
/* If there are any jobs in the HW queue, we're not idle */
|
|
- if (atomic_read(&js->queue[i].sched.hw_rq_count))
|
|
+ if (atomic_read(&js->queue[i].sched.credit_count))
|
|
return false;
|
|
}
|
|
|
|
--- a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
|
|
+++ b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
|
|
@@ -51,7 +51,7 @@ DECLARE_EVENT_CLASS(drm_sched_job,
|
|
__assign_str(name, sched_job->sched->name);
|
|
__entry->job_count = spsc_queue_count(&entity->job_queue);
|
|
__entry->hw_job_count = atomic_read(
|
|
- &sched_job->sched->hw_rq_count);
|
|
+ &sched_job->sched->credit_count);
|
|
),
|
|
TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw job count:%d",
|
|
__entry->entity, __entry->id,
|
|
--- a/drivers/gpu/drm/scheduler/sched_main.c
|
|
+++ b/drivers/gpu/drm/scheduler/sched_main.c
|
|
@@ -48,6 +48,30 @@
|
|
* through the jobs entity pointer.
|
|
*/
|
|
|
|
+/**
|
|
+ * DOC: Flow Control
|
|
+ *
|
|
+ * The DRM GPU scheduler provides a flow control mechanism to regulate the rate
|
|
+ * in which the jobs fetched from scheduler entities are executed.
|
|
+ *
|
|
+ * In this context the &drm_gpu_scheduler keeps track of a driver specified
|
|
+ * credit limit representing the capacity of this scheduler and a credit count;
|
|
+ * every &drm_sched_job carries a driver specified number of credits.
|
|
+ *
|
|
+ * Once a job is executed (but not yet finished), the job's credits contribute
|
|
+ * to the scheduler's credit count until the job is finished. If by executing
|
|
+ * one more job the scheduler's credit count would exceed the scheduler's
|
|
+ * credit limit, the job won't be executed. Instead, the scheduler will wait
|
|
+ * until the credit count has decreased enough to not overflow its credit limit.
|
|
+ * This implies waiting for previously executed jobs.
|
|
+ *
|
|
+ * Optionally, drivers may register a callback (update_job_credits) provided by
|
|
+ * struct drm_sched_backend_ops to update the job's credits dynamically. The
|
|
+ * scheduler executes this callback every time the scheduler considers a job for
|
|
+ * execution and subsequently checks whether the job fits the scheduler's credit
|
|
+ * limit.
|
|
+ */
|
|
+
|
|
#include <linux/wait.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/completion.h>
|
|
@@ -75,6 +99,51 @@ int drm_sched_policy = DRM_SCHED_POLICY_
|
|
MODULE_PARM_DESC(sched_policy, "Specify the scheduling policy for entities on a run-queue, " __stringify(DRM_SCHED_POLICY_RR) " = Round Robin, " __stringify(DRM_SCHED_POLICY_FIFO) " = FIFO (default).");
|
|
module_param_named(sched_policy, drm_sched_policy, int, 0444);
|
|
|
|
+static u32 drm_sched_available_credits(struct drm_gpu_scheduler *sched)
|
|
+{
|
|
+ u32 credits;
|
|
+
|
|
+ drm_WARN_ON(sched, check_sub_overflow(sched->credit_limit,
|
|
+ atomic_read(&sched->credit_count),
|
|
+ &credits));
|
|
+
|
|
+ return credits;
|
|
+}
|
|
+
|
|
+/**
|
|
+ * drm_sched_can_queue -- Can we queue more to the hardware?
|
|
+ * @sched: scheduler instance
|
|
+ * @entity: the scheduler entity
|
|
+ *
|
|
+ * Return true if we can push at least one more job from @entity, false
|
|
+ * otherwise.
|
|
+ */
|
|
+static bool drm_sched_can_queue(struct drm_gpu_scheduler *sched,
|
|
+ struct drm_sched_entity *entity)
|
|
+{
|
|
+ struct drm_sched_job *s_job;
|
|
+
|
|
+ s_job = to_drm_sched_job(spsc_queue_peek(&entity->job_queue));
|
|
+ if (!s_job)
|
|
+ return false;
|
|
+
|
|
+ if (sched->ops->update_job_credits) {
|
|
+ s_job->credits = sched->ops->update_job_credits(s_job);
|
|
+
|
|
+ drm_WARN(sched, !s_job->credits,
|
|
+ "Jobs with zero credits bypass job-flow control.\n");
|
|
+ }
|
|
+
|
|
+ /* If a job exceeds the credit limit, truncate it to the credit limit
|
|
+ * itself to guarantee forward progress.
|
|
+ */
|
|
+ if (drm_WARN(sched, s_job->credits > sched->credit_limit,
|
|
+ "Jobs may not exceed the credit limit, truncate.\n"))
|
|
+ s_job->credits = sched->credit_limit;
|
|
+
|
|
+ return drm_sched_available_credits(sched) >= s_job->credits;
|
|
+}
|
|
+
|
|
static __always_inline bool drm_sched_entity_compare_before(struct rb_node *a,
|
|
const struct rb_node *b)
|
|
{
|
|
@@ -186,12 +255,18 @@ void drm_sched_rq_remove_entity(struct d
|
|
/**
|
|
* drm_sched_rq_select_entity_rr - Select an entity which could provide a job to run
|
|
*
|
|
+ * @sched: the gpu scheduler
|
|
* @rq: scheduler run queue to check.
|
|
*
|
|
- * Try to find a ready entity, returns NULL if none found.
|
|
+ * Try to find the next ready entity.
|
|
+ *
|
|
+ * Return an entity if one is found; return an error-pointer (!NULL) if an
|
|
+ * entity was ready, but the scheduler had insufficient credits to accommodate
|
|
+ * its job; return NULL, if no ready entity was found.
|
|
*/
|
|
static struct drm_sched_entity *
|
|
-drm_sched_rq_select_entity_rr(struct drm_sched_rq *rq)
|
|
+drm_sched_rq_select_entity_rr(struct drm_gpu_scheduler *sched,
|
|
+ struct drm_sched_rq *rq)
|
|
{
|
|
struct drm_sched_entity *entity;
|
|
|
|
@@ -201,6 +276,14 @@ drm_sched_rq_select_entity_rr(struct drm
|
|
if (entity) {
|
|
list_for_each_entry_continue(entity, &rq->entities, list) {
|
|
if (drm_sched_entity_is_ready(entity)) {
|
|
+ /* If we can't queue yet, preserve the current
|
|
+ * entity in terms of fairness.
|
|
+ */
|
|
+ if (!drm_sched_can_queue(sched, entity)) {
|
|
+ spin_unlock(&rq->lock);
|
|
+ return ERR_PTR(-ENOSPC);
|
|
+ }
|
|
+
|
|
rq->current_entity = entity;
|
|
reinit_completion(&entity->entity_idle);
|
|
spin_unlock(&rq->lock);
|
|
@@ -210,8 +293,15 @@ drm_sched_rq_select_entity_rr(struct drm
|
|
}
|
|
|
|
list_for_each_entry(entity, &rq->entities, list) {
|
|
-
|
|
if (drm_sched_entity_is_ready(entity)) {
|
|
+ /* If we can't queue yet, preserve the current entity in
|
|
+ * terms of fairness.
|
|
+ */
|
|
+ if (!drm_sched_can_queue(sched, entity)) {
|
|
+ spin_unlock(&rq->lock);
|
|
+ return ERR_PTR(-ENOSPC);
|
|
+ }
|
|
+
|
|
rq->current_entity = entity;
|
|
reinit_completion(&entity->entity_idle);
|
|
spin_unlock(&rq->lock);
|
|
@@ -230,12 +320,18 @@ drm_sched_rq_select_entity_rr(struct drm
|
|
/**
|
|
* drm_sched_rq_select_entity_fifo - Select an entity which provides a job to run
|
|
*
|
|
+ * @sched: the gpu scheduler
|
|
* @rq: scheduler run queue to check.
|
|
*
|
|
- * Find oldest waiting ready entity, returns NULL if none found.
|
|
+ * Find oldest waiting ready entity.
|
|
+ *
|
|
+ * Return an entity if one is found; return an error-pointer (!NULL) if an
|
|
+ * entity was ready, but the scheduler had insufficient credits to accommodate
|
|
+ * its job; return NULL, if no ready entity was found.
|
|
*/
|
|
static struct drm_sched_entity *
|
|
-drm_sched_rq_select_entity_fifo(struct drm_sched_rq *rq)
|
|
+drm_sched_rq_select_entity_fifo(struct drm_gpu_scheduler *sched,
|
|
+ struct drm_sched_rq *rq)
|
|
{
|
|
struct rb_node *rb;
|
|
|
|
@@ -245,6 +341,14 @@ drm_sched_rq_select_entity_fifo(struct d
|
|
|
|
entity = rb_entry(rb, struct drm_sched_entity, rb_tree_node);
|
|
if (drm_sched_entity_is_ready(entity)) {
|
|
+ /* If we can't queue yet, preserve the current entity in
|
|
+ * terms of fairness.
|
|
+ */
|
|
+ if (!drm_sched_can_queue(sched, entity)) {
|
|
+ spin_unlock(&rq->lock);
|
|
+ return ERR_PTR(-ENOSPC);
|
|
+ }
|
|
+
|
|
rq->current_entity = entity;
|
|
reinit_completion(&entity->entity_idle);
|
|
break;
|
|
@@ -302,7 +406,7 @@ static void drm_sched_job_done(struct dr
|
|
struct drm_sched_fence *s_fence = s_job->s_fence;
|
|
struct drm_gpu_scheduler *sched = s_fence->sched;
|
|
|
|
- atomic_dec(&sched->hw_rq_count);
|
|
+ atomic_sub(s_job->credits, &sched->credit_count);
|
|
atomic_dec(sched->score);
|
|
|
|
trace_drm_sched_process_job(s_fence);
|
|
@@ -519,7 +623,7 @@ void drm_sched_stop(struct drm_gpu_sched
|
|
&s_job->cb)) {
|
|
dma_fence_put(s_job->s_fence->parent);
|
|
s_job->s_fence->parent = NULL;
|
|
- atomic_dec(&sched->hw_rq_count);
|
|
+ atomic_sub(s_job->credits, &sched->credit_count);
|
|
} else {
|
|
/*
|
|
* remove job from pending_list.
|
|
@@ -580,7 +684,7 @@ void drm_sched_start(struct drm_gpu_sche
|
|
list_for_each_entry_safe(s_job, tmp, &sched->pending_list, list) {
|
|
struct dma_fence *fence = s_job->s_fence->parent;
|
|
|
|
- atomic_inc(&sched->hw_rq_count);
|
|
+ atomic_add(s_job->credits, &sched->credit_count);
|
|
|
|
if (!full_recovery)
|
|
continue;
|
|
@@ -664,6 +768,8 @@ EXPORT_SYMBOL(drm_sched_resubmit_jobs);
|
|
* drm_sched_job_init - init a scheduler job
|
|
* @job: scheduler job to init
|
|
* @entity: scheduler entity to use
|
|
+ * @credits: the number of credits this job contributes to the schedulers
|
|
+ * credit limit
|
|
* @owner: job owner for debugging
|
|
*
|
|
* Refer to drm_sched_entity_push_job() documentation
|
|
@@ -681,7 +787,7 @@ EXPORT_SYMBOL(drm_sched_resubmit_jobs);
|
|
*/
|
|
int drm_sched_job_init(struct drm_sched_job *job,
|
|
struct drm_sched_entity *entity,
|
|
- void *owner)
|
|
+ u32 credits, void *owner)
|
|
{
|
|
if (!entity->rq) {
|
|
/* This will most likely be followed by missing frames
|
|
@@ -700,7 +806,13 @@ int drm_sched_job_init(struct drm_sched_
|
|
*/
|
|
memset(job, 0, sizeof(*job));
|
|
|
|
+ if (unlikely(!credits)) {
|
|
+ pr_err("*ERROR* %s: credits cannot be 0!\n", __func__);
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
job->entity = entity;
|
|
+ job->credits = credits;
|
|
job->s_fence = drm_sched_fence_alloc(entity, owner);
|
|
if (!job->s_fence)
|
|
return -ENOMEM;
|
|
@@ -913,20 +1025,9 @@ void drm_sched_job_cleanup(struct drm_sc
|
|
EXPORT_SYMBOL(drm_sched_job_cleanup);
|
|
|
|
/**
|
|
- * drm_sched_can_queue -- Can we queue more to the hardware?
|
|
- * @sched: scheduler instance
|
|
- *
|
|
- * Return true if we can push more jobs to the hw, otherwise false.
|
|
- */
|
|
-static bool drm_sched_can_queue(struct drm_gpu_scheduler *sched)
|
|
-{
|
|
- return atomic_read(&sched->hw_rq_count) <
|
|
- sched->hw_submission_limit;
|
|
-}
|
|
-
|
|
-/**
|
|
* drm_sched_wakeup - Wake up the scheduler if it is ready to queue
|
|
* @sched: scheduler instance
|
|
+ * @entity: the scheduler entity
|
|
*
|
|
* Wake up the scheduler if we can queue jobs.
|
|
*/
|
|
@@ -934,7 +1035,7 @@ void drm_sched_wakeup(struct drm_gpu_sch
|
|
struct drm_sched_entity *entity)
|
|
{
|
|
if (drm_sched_entity_is_ready(entity))
|
|
- if (drm_sched_can_queue(sched))
|
|
+ if (drm_sched_can_queue(sched, entity))
|
|
drm_sched_run_job_queue(sched);
|
|
}
|
|
|
|
@@ -943,7 +1044,11 @@ void drm_sched_wakeup(struct drm_gpu_sch
|
|
*
|
|
* @sched: scheduler instance
|
|
*
|
|
- * Returns the entity to process or NULL if none are found.
|
|
+ * Return an entity to process or NULL if none are found.
|
|
+ *
|
|
+ * Note, that we break out of the for-loop when "entity" is non-null, which can
|
|
+ * also be an error-pointer--this assures we don't process lower priority
|
|
+ * run-queues. See comments in the respectively called functions.
|
|
*/
|
|
static struct drm_sched_entity *
|
|
drm_sched_select_entity(struct drm_gpu_scheduler *sched)
|
|
@@ -951,19 +1056,16 @@ drm_sched_select_entity(struct drm_gpu_s
|
|
struct drm_sched_entity *entity;
|
|
int i;
|
|
|
|
- if (!drm_sched_can_queue(sched))
|
|
- return NULL;
|
|
-
|
|
/* Kernel run queue has higher priority than normal run queue*/
|
|
for (i = sched->num_rqs - 1; i >= DRM_SCHED_PRIORITY_MIN; i--) {
|
|
entity = drm_sched_policy == DRM_SCHED_POLICY_FIFO ?
|
|
- drm_sched_rq_select_entity_fifo(sched->sched_rq[i]) :
|
|
- drm_sched_rq_select_entity_rr(sched->sched_rq[i]);
|
|
+ drm_sched_rq_select_entity_fifo(sched, sched->sched_rq[i]) :
|
|
+ drm_sched_rq_select_entity_rr(sched, sched->sched_rq[i]);
|
|
if (entity)
|
|
break;
|
|
}
|
|
|
|
- return entity;
|
|
+ return IS_ERR(entity) ? NULL : entity;
|
|
}
|
|
|
|
/**
|
|
@@ -1110,7 +1212,7 @@ static void drm_sched_run_job_work(struc
|
|
|
|
s_fence = sched_job->s_fence;
|
|
|
|
- atomic_inc(&sched->hw_rq_count);
|
|
+ atomic_add(sched_job->credits, &sched->credit_count);
|
|
drm_sched_job_begin(sched_job);
|
|
|
|
trace_drm_run_job(sched_job, entity);
|
|
@@ -1145,7 +1247,7 @@ static void drm_sched_run_job_work(struc
|
|
* @submit_wq: workqueue to use for submission. If NULL, an ordered wq is
|
|
* allocated and used
|
|
* @num_rqs: number of runqueues, one for each priority, up to DRM_SCHED_PRIORITY_COUNT
|
|
- * @hw_submission: number of hw submissions that can be in flight
|
|
+ * @credit_limit: the number of credits this scheduler can hold from all jobs
|
|
* @hang_limit: number of times to allow a job to hang before dropping it
|
|
* @timeout: timeout value in jiffies for the scheduler
|
|
* @timeout_wq: workqueue to use for timeout work. If NULL, the system_wq is
|
|
@@ -1159,14 +1261,14 @@ static void drm_sched_run_job_work(struc
|
|
int drm_sched_init(struct drm_gpu_scheduler *sched,
|
|
const struct drm_sched_backend_ops *ops,
|
|
struct workqueue_struct *submit_wq,
|
|
- u32 num_rqs, uint32_t hw_submission, unsigned int hang_limit,
|
|
+ u32 num_rqs, u32 credit_limit, unsigned int hang_limit,
|
|
long timeout, struct workqueue_struct *timeout_wq,
|
|
atomic_t *score, const char *name, struct device *dev)
|
|
{
|
|
int i, ret;
|
|
|
|
sched->ops = ops;
|
|
- sched->hw_submission_limit = hw_submission;
|
|
+ sched->credit_limit = credit_limit;
|
|
sched->name = name;
|
|
sched->timeout = timeout;
|
|
sched->timeout_wq = timeout_wq ? : system_wq;
|
|
@@ -1215,7 +1317,7 @@ int drm_sched_init(struct drm_gpu_schedu
|
|
init_waitqueue_head(&sched->job_scheduled);
|
|
INIT_LIST_HEAD(&sched->pending_list);
|
|
spin_lock_init(&sched->job_list_lock);
|
|
- atomic_set(&sched->hw_rq_count, 0);
|
|
+ atomic_set(&sched->credit_count, 0);
|
|
INIT_DELAYED_WORK(&sched->work_tdr, drm_sched_job_timedout);
|
|
INIT_WORK(&sched->work_run_job, drm_sched_run_job_work);
|
|
INIT_WORK(&sched->work_free_job, drm_sched_free_job_work);
|
|
--- a/drivers/gpu/drm/v3d/v3d_gem.c
|
|
+++ b/drivers/gpu/drm/v3d/v3d_gem.c
|
|
@@ -417,7 +417,7 @@ v3d_job_init(struct v3d_dev *v3d, struct
|
|
job->free = free;
|
|
|
|
ret = drm_sched_job_init(&job->base, &v3d_priv->sched_entity[queue],
|
|
- v3d_priv);
|
|
+ 1, v3d_priv);
|
|
if (ret)
|
|
goto fail;
|
|
|
|
--- a/include/drm/gpu_scheduler.h
|
|
+++ b/include/drm/gpu_scheduler.h
|
|
@@ -320,6 +320,7 @@ struct drm_sched_fence *to_drm_sched_fen
|
|
* @sched: the scheduler instance on which this job is scheduled.
|
|
* @s_fence: contains the fences for the scheduling of job.
|
|
* @finish_cb: the callback for the finished fence.
|
|
+ * @credits: the number of credits this job contributes to the scheduler
|
|
* @work: Helper to reschdeule job kill to different context.
|
|
* @id: a unique id assigned to each job scheduled on the scheduler.
|
|
* @karma: increment on every hang caused by this job. If this exceeds the hang
|
|
@@ -339,6 +340,8 @@ struct drm_sched_job {
|
|
struct drm_gpu_scheduler *sched;
|
|
struct drm_sched_fence *s_fence;
|
|
|
|
+ u32 credits;
|
|
+
|
|
/*
|
|
* work is used only after finish_cb has been used and will not be
|
|
* accessed anymore.
|
|
@@ -462,13 +465,27 @@ struct drm_sched_backend_ops {
|
|
* and it's time to clean it up.
|
|
*/
|
|
void (*free_job)(struct drm_sched_job *sched_job);
|
|
+
|
|
+ /**
|
|
+ * @update_job_credits: Called when the scheduler is considering this
|
|
+ * job for execution.
|
|
+ *
|
|
+ * This callback returns the number of credits the job would take if
|
|
+ * pushed to the hardware. Drivers may use this to dynamically update
|
|
+ * the job's credit count. For instance, deduct the number of credits
|
|
+ * for already signalled native fences.
|
|
+ *
|
|
+ * This callback is optional.
|
|
+ */
|
|
+ u32 (*update_job_credits)(struct drm_sched_job *sched_job);
|
|
};
|
|
|
|
/**
|
|
* struct drm_gpu_scheduler - scheduler instance-specific data
|
|
*
|
|
* @ops: backend operations provided by the driver.
|
|
- * @hw_submission_limit: the max size of the hardware queue.
|
|
+ * @credit_limit: the credit limit of this scheduler
|
|
+ * @credit_count: the current credit count of this scheduler
|
|
* @timeout: the time after which a job is removed from the scheduler.
|
|
* @name: name of the ring for which this scheduler is being used.
|
|
* @num_rqs: Number of run-queues. This is at most DRM_SCHED_PRIORITY_COUNT,
|
|
@@ -477,7 +494,6 @@ struct drm_sched_backend_ops {
|
|
* @job_scheduled: once @drm_sched_entity_do_release is called the scheduler
|
|
* waits on this wait queue until all the scheduled jobs are
|
|
* finished.
|
|
- * @hw_rq_count: the number of jobs currently in the hardware queue.
|
|
* @job_id_count: used to assign unique id to the each job.
|
|
* @submit_wq: workqueue used to queue @work_run_job and @work_free_job
|
|
* @timeout_wq: workqueue used to queue @work_tdr
|
|
@@ -501,13 +517,13 @@ struct drm_sched_backend_ops {
|
|
*/
|
|
struct drm_gpu_scheduler {
|
|
const struct drm_sched_backend_ops *ops;
|
|
- uint32_t hw_submission_limit;
|
|
+ u32 credit_limit;
|
|
+ atomic_t credit_count;
|
|
long timeout;
|
|
const char *name;
|
|
u32 num_rqs;
|
|
struct drm_sched_rq **sched_rq;
|
|
wait_queue_head_t job_scheduled;
|
|
- atomic_t hw_rq_count;
|
|
atomic64_t job_id_count;
|
|
struct workqueue_struct *submit_wq;
|
|
struct workqueue_struct *timeout_wq;
|
|
@@ -529,14 +545,14 @@ struct drm_gpu_scheduler {
|
|
int drm_sched_init(struct drm_gpu_scheduler *sched,
|
|
const struct drm_sched_backend_ops *ops,
|
|
struct workqueue_struct *submit_wq,
|
|
- u32 num_rqs, uint32_t hw_submission, unsigned int hang_limit,
|
|
+ u32 num_rqs, u32 credit_limit, unsigned int hang_limit,
|
|
long timeout, struct workqueue_struct *timeout_wq,
|
|
atomic_t *score, const char *name, struct device *dev);
|
|
|
|
void drm_sched_fini(struct drm_gpu_scheduler *sched);
|
|
int drm_sched_job_init(struct drm_sched_job *job,
|
|
struct drm_sched_entity *entity,
|
|
- void *owner);
|
|
+ u32 credits, void *owner);
|
|
void drm_sched_job_arm(struct drm_sched_job *job);
|
|
int drm_sched_job_add_dependency(struct drm_sched_job *job,
|
|
struct dma_fence *fence);
|