mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-22 10:58:18 +07:00
e9eaf82d97
Latency is in the eye of the beholder. In the case where a client stops and waits for the gpu, give that request chain a small priority boost (not so that it overtakes higher priority clients, to preserve the external ordering) so that ideally the wait completes earlier. v2: Tvrtko recommends to keep the boost-from-user-stall as small as possible and to allow new client flows to be preferred for interactivity over stalls. Testcase: igt/gem_sync/switch-default Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com> Cc: Joonas Lahtinen <joonas.lahtinen@linux.intel.com> Cc: Dmitry Rogozhkin <dmitry.v.rogozhkin@intel.com> Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20181001144755.7978-3-chris@chris-wilson.co.uk
109 lines
3.4 KiB
C
109 lines
3.4 KiB
C
/*
|
|
* SPDX-License-Identifier: MIT
|
|
*
|
|
* Copyright © 2018 Intel Corporation
|
|
*/
|
|
|
|
#ifndef _I915_SCHEDULER_H_
|
|
#define _I915_SCHEDULER_H_
|
|
|
|
#include <linux/bitops.h>
|
|
#include <linux/kernel.h>
|
|
|
|
#include <uapi/drm/i915_drm.h>
|
|
|
|
struct drm_i915_private;
|
|
struct i915_request;
|
|
struct intel_engine_cs;
|
|
|
|
enum {
|
|
I915_PRIORITY_MIN = I915_CONTEXT_MIN_USER_PRIORITY - 1,
|
|
I915_PRIORITY_NORMAL = I915_CONTEXT_DEFAULT_PRIORITY,
|
|
I915_PRIORITY_MAX = I915_CONTEXT_MAX_USER_PRIORITY + 1,
|
|
|
|
I915_PRIORITY_INVALID = INT_MIN
|
|
};
|
|
|
|
#define I915_USER_PRIORITY_SHIFT 2
|
|
#define I915_USER_PRIORITY(x) ((x) << I915_USER_PRIORITY_SHIFT)
|
|
|
|
#define I915_PRIORITY_COUNT BIT(I915_USER_PRIORITY_SHIFT)
|
|
#define I915_PRIORITY_MASK (I915_PRIORITY_COUNT - 1)
|
|
|
|
#define I915_PRIORITY_WAIT ((u8)BIT(0))
|
|
#define I915_PRIORITY_NEWCLIENT ((u8)BIT(1))
|
|
|
|
struct i915_sched_attr {
|
|
/**
|
|
* @priority: execution and service priority
|
|
*
|
|
* All clients are equal, but some are more equal than others!
|
|
*
|
|
* Requests from a context with a greater (more positive) value of
|
|
* @priority will be executed before those with a lower @priority
|
|
* value, forming a simple QoS.
|
|
*
|
|
* The &drm_i915_private.kernel_context is assigned the lowest priority.
|
|
*/
|
|
int priority;
|
|
};
|
|
|
|
/*
|
|
* "People assume that time is a strict progression of cause to effect, but
|
|
* actually, from a nonlinear, non-subjective viewpoint, it's more like a big
|
|
* ball of wibbly-wobbly, timey-wimey ... stuff." -The Doctor, 2015
|
|
*
|
|
* Requests exist in a complex web of interdependencies. Each request
|
|
* has to wait for some other request to complete before it is ready to be run
|
|
* (e.g. we have to wait until the pixels have been rendering into a texture
|
|
* before we can copy from it). We track the readiness of a request in terms
|
|
* of fences, but we also need to keep the dependency tree for the lifetime
|
|
* of the request (beyond the life of an individual fence). We use the tree
|
|
* at various points to reorder the requests whilst keeping the requests
|
|
* in order with respect to their various dependencies.
|
|
*
|
|
* There is no active component to the "scheduler". As we know the dependency
|
|
* DAG of each request, we are able to insert it into a sorted queue when it
|
|
* is ready, and are able to reorder its portion of the graph to accommodate
|
|
* dynamic priority changes.
|
|
*/
|
|
struct i915_sched_node {
|
|
struct list_head signalers_list; /* those before us, we depend upon */
|
|
struct list_head waiters_list; /* those after us, they depend upon us */
|
|
struct list_head link;
|
|
struct i915_sched_attr attr;
|
|
};
|
|
|
|
struct i915_dependency {
|
|
struct i915_sched_node *signaler;
|
|
struct list_head signal_link;
|
|
struct list_head wait_link;
|
|
struct list_head dfs_link;
|
|
unsigned long flags;
|
|
#define I915_DEPENDENCY_ALLOC BIT(0)
|
|
};
|
|
|
|
void i915_sched_node_init(struct i915_sched_node *node);
|
|
|
|
bool __i915_sched_node_add_dependency(struct i915_sched_node *node,
|
|
struct i915_sched_node *signal,
|
|
struct i915_dependency *dep,
|
|
unsigned long flags);
|
|
|
|
int i915_sched_node_add_dependency(struct drm_i915_private *i915,
|
|
struct i915_sched_node *node,
|
|
struct i915_sched_node *signal);
|
|
|
|
void i915_sched_node_fini(struct drm_i915_private *i915,
|
|
struct i915_sched_node *node);
|
|
|
|
void i915_schedule(struct i915_request *request,
|
|
const struct i915_sched_attr *attr);
|
|
|
|
void i915_schedule_bump_priority(struct i915_request *rq, unsigned int bump);
|
|
|
|
struct list_head *
|
|
i915_sched_lookup_priolist(struct intel_engine_cs *engine, int prio);
|
|
|
|
#endif /* _I915_SCHEDULER_H_ */
|