mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-27 11:55:07 +07:00
0cb5670baa
We assume that the GPU is idle once receiving the seqno via the last
request's user interrupt. In execlist mode the corresponding context
completed interrupt can be delayed though and until this latter
interrupt arrives we consider the request to be pending on the ELSP
submit port. This can cause a problem during system suspend where this
last request will be seen by the resume code as still pending. Such
pending requests are normally replayed after a GPU reset, but during
resume we reset both SW and HW tracking of the ring head/tail pointers,
so replaying the pending request with its stale tail pointer will leave
the ring in an inconsistent state. A subsequent request submission can
lead then to the GPU executing from uninitialized area in the ring
behind the above stale tail pointer.
Fix this by making sure any pending request on the ELSP port is
completed before suspending. I used a polling wait since the completion
time I measured was <1ms and since normally we only need to wait during
system suspend. GPU idling during runtime suspend is scheduled with a
delay (currently 50-100ms) after the retirement of the last request at
which point the context completed interrupt must have arrived already.
The chance of this bug was increased by
commit 1c777c5d1d
Author: Imre Deak <imre.deak@intel.com>
Date: Wed Oct 12 17:46:37 2016 +0300
drm/i915/hsw: Fix GPU hang during resume from S3-devices state
but it could happen even without the explicit GPU reset, since we
disable interrupts afterwards during the suspend sequence.
v2:
- Do an unlocked poll-wait first. (Chris)
v3-4:
- s/intel_lr_engines_idle/intel_execlists_idle/ and move
i915.enable_execlists check to the new helper. (Chris)
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=98470
Signed-off-by: Imre Deak <imre.deak@intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1478510405-11799-3-git-send-email-imre.deak@intel.com
101 lines
4.1 KiB
C
101 lines
4.1 KiB
C
/*
|
|
* Copyright © 2014 Intel Corporation
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#ifndef _INTEL_LRC_H_
|
|
#define _INTEL_LRC_H_
|
|
|
|
#include "intel_ringbuffer.h"
|
|
|
|
#define GEN8_LR_CONTEXT_ALIGN 4096
|
|
|
|
/* Execlists regs */
|
|
#define RING_ELSP(engine) _MMIO((engine)->mmio_base + 0x230)
|
|
#define RING_EXECLIST_STATUS_LO(engine) _MMIO((engine)->mmio_base + 0x234)
|
|
#define RING_EXECLIST_STATUS_HI(engine) _MMIO((engine)->mmio_base + 0x234 + 4)
|
|
#define RING_CONTEXT_CONTROL(engine) _MMIO((engine)->mmio_base + 0x244)
|
|
#define CTX_CTRL_INHIBIT_SYN_CTX_SWITCH (1 << 3)
|
|
#define CTX_CTRL_ENGINE_CTX_RESTORE_INHIBIT (1 << 0)
|
|
#define CTX_CTRL_RS_CTX_ENABLE (1 << 1)
|
|
#define RING_CONTEXT_STATUS_BUF_BASE(engine) _MMIO((engine)->mmio_base + 0x370)
|
|
#define RING_CONTEXT_STATUS_BUF_LO(engine, i) _MMIO((engine)->mmio_base + 0x370 + (i) * 8)
|
|
#define RING_CONTEXT_STATUS_BUF_HI(engine, i) _MMIO((engine)->mmio_base + 0x370 + (i) * 8 + 4)
|
|
#define RING_CONTEXT_STATUS_PTR(engine) _MMIO((engine)->mmio_base + 0x3a0)
|
|
|
|
/* The docs specify that the write pointer wraps around after 5h, "After status
|
|
* is written out to the last available status QW at offset 5h, this pointer
|
|
* wraps to 0."
|
|
*
|
|
* Therefore, one must infer than even though there are 3 bits available, 6 and
|
|
* 7 appear to be * reserved.
|
|
*/
|
|
#define GEN8_CSB_ENTRIES 6
|
|
#define GEN8_CSB_PTR_MASK 0x7
|
|
#define GEN8_CSB_READ_PTR_MASK (GEN8_CSB_PTR_MASK << 8)
|
|
#define GEN8_CSB_WRITE_PTR_MASK (GEN8_CSB_PTR_MASK << 0)
|
|
#define GEN8_CSB_WRITE_PTR(csb_status) \
|
|
(((csb_status) & GEN8_CSB_WRITE_PTR_MASK) >> 0)
|
|
#define GEN8_CSB_READ_PTR(csb_status) \
|
|
(((csb_status) & GEN8_CSB_READ_PTR_MASK) >> 8)
|
|
|
|
enum {
|
|
INTEL_CONTEXT_SCHEDULE_IN = 0,
|
|
INTEL_CONTEXT_SCHEDULE_OUT,
|
|
};
|
|
|
|
/* Logical Rings */
|
|
int intel_logical_ring_alloc_request_extras(struct drm_i915_gem_request *request);
|
|
int intel_logical_ring_reserve_space(struct drm_i915_gem_request *request);
|
|
void intel_logical_ring_stop(struct intel_engine_cs *engine);
|
|
void intel_logical_ring_cleanup(struct intel_engine_cs *engine);
|
|
int logical_render_ring_init(struct intel_engine_cs *engine);
|
|
int logical_xcs_ring_init(struct intel_engine_cs *engine);
|
|
|
|
int intel_engines_init(struct drm_device *dev);
|
|
|
|
/* Logical Ring Contexts */
|
|
|
|
/* One extra page is added before LRC for GuC as shared data */
|
|
#define LRC_GUCSHR_PN (0)
|
|
#define LRC_PPHWSP_PN (LRC_GUCSHR_PN + 1)
|
|
#define LRC_STATE_PN (LRC_PPHWSP_PN + 1)
|
|
|
|
struct i915_gem_context;
|
|
|
|
uint32_t intel_lr_context_size(struct intel_engine_cs *engine);
|
|
void intel_lr_context_unpin(struct i915_gem_context *ctx,
|
|
struct intel_engine_cs *engine);
|
|
|
|
struct drm_i915_private;
|
|
|
|
void intel_lr_context_resume(struct drm_i915_private *dev_priv);
|
|
uint64_t intel_lr_context_descriptor(struct i915_gem_context *ctx,
|
|
struct intel_engine_cs *engine);
|
|
|
|
/* Execlists */
|
|
int intel_sanitize_enable_execlists(struct drm_i915_private *dev_priv,
|
|
int enable_execlists);
|
|
void intel_execlists_enable_submission(struct drm_i915_private *dev_priv);
|
|
bool intel_execlists_idle(struct drm_i915_private *dev_priv);
|
|
|
|
#endif /* _INTEL_LRC_H_ */
|