mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-22 15:45:09 +07:00
4c5896dc4c
For virtual engines, we need to keep the HW context alive while it remains in use. For regular HW contexts, they are created and kept alive until the end of the GEM context. For simplicity, generalise the requirements and keep an active reference to each HW context. Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20190318212347.30146-2-chris@chris-wilson.co.uk
88 lines
2.2 KiB
C
88 lines
2.2 KiB
C
/*
|
|
* SPDX-License-Identifier: MIT
|
|
*
|
|
* Copyright © 2019 Intel Corporation
|
|
*/
|
|
|
|
#ifndef __INTEL_CONTEXT_H__
|
|
#define __INTEL_CONTEXT_H__
|
|
|
|
#include <linux/lockdep.h>
|
|
|
|
#include "intel_context_types.h"
|
|
#include "intel_engine_types.h"
|
|
|
|
struct intel_context *intel_context_alloc(void);
|
|
void intel_context_free(struct intel_context *ce);
|
|
|
|
void intel_context_init(struct intel_context *ce,
|
|
struct i915_gem_context *ctx,
|
|
struct intel_engine_cs *engine);
|
|
|
|
/**
|
|
* intel_context_lookup - Find the matching HW context for this (ctx, engine)
|
|
* @ctx - the parent GEM context
|
|
* @engine - the target HW engine
|
|
*
|
|
* May return NULL if the HW context hasn't been instantiated (i.e. unused).
|
|
*/
|
|
struct intel_context *
|
|
intel_context_lookup(struct i915_gem_context *ctx,
|
|
struct intel_engine_cs *engine);
|
|
|
|
/**
|
|
* intel_context_pin_lock - Stablises the 'pinned' status of the HW context
|
|
* @ctx - the parent GEM context
|
|
* @engine - the target HW engine
|
|
*
|
|
* Acquire a lock on the pinned status of the HW context, such that the context
|
|
* can neither be bound to the GPU or unbound whilst the lock is held, i.e.
|
|
* intel_context_is_pinned() remains stable.
|
|
*/
|
|
struct intel_context *
|
|
intel_context_pin_lock(struct i915_gem_context *ctx,
|
|
struct intel_engine_cs *engine);
|
|
|
|
static inline bool
|
|
intel_context_is_pinned(struct intel_context *ce)
|
|
{
|
|
return atomic_read(&ce->pin_count);
|
|
}
|
|
|
|
static inline void intel_context_pin_unlock(struct intel_context *ce)
|
|
__releases(ce->pin_mutex)
|
|
{
|
|
mutex_unlock(&ce->pin_mutex);
|
|
}
|
|
|
|
struct intel_context *
|
|
__intel_context_insert(struct i915_gem_context *ctx,
|
|
struct intel_engine_cs *engine,
|
|
struct intel_context *ce);
|
|
void
|
|
__intel_context_remove(struct intel_context *ce);
|
|
|
|
struct intel_context *
|
|
intel_context_pin(struct i915_gem_context *ctx, struct intel_engine_cs *engine);
|
|
|
|
static inline void __intel_context_pin(struct intel_context *ce)
|
|
{
|
|
GEM_BUG_ON(!intel_context_is_pinned(ce));
|
|
atomic_inc(&ce->pin_count);
|
|
}
|
|
|
|
void intel_context_unpin(struct intel_context *ce);
|
|
|
|
static inline struct intel_context *intel_context_get(struct intel_context *ce)
|
|
{
|
|
kref_get(&ce->ref);
|
|
return ce;
|
|
}
|
|
|
|
static inline void intel_context_put(struct intel_context *ce)
|
|
{
|
|
kref_put(&ce->ref, ce->ops->destroy);
|
|
}
|
|
|
|
#endif /* __INTEL_CONTEXT_H__ */
|