mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-27 13:35:09 +07:00
b1c4f7fead
UAPI Changes: - uAPI "Fixes:" patch for the upcoming kernel 5.1, included here too We have an Ack from the media folks (only current user) for this late tweak Cross-subsystem Changes: - ALSA: hda: Fix racy display power access (Takashi, Chris) Driver Changes: - DDI and MIPI-DSI clocks fixes for Icelake (Vandita) - Fix Icelake frequency change/locking (RPS) (Mika) - Temporarily disable ppGTT read-only bit on Icelake (Mika) - Add missing Icelake W/As (Mika) - Enable 12 deep CSB status FIFO on Icelake (Mika) - Inherit more Icelake code for Elkhartlake (Bob, Jani) - Handle catastrophic error on engine reset (Mika) - Shortcut readiness to reset check (Mika) - Regression fix for GEM_BUSY causing us to report a mixed uabi-class request as not busy (Chris) - Revert back to max link rate and lane count on eDP (Jani) - Fix pipe BPP readout for BXT/GLK DSI (Ville) - Set DP min_bpp to 8*3 for non-RGB output formats (Ville) - Enable coarse preemption boundaries for Gen8 (Chris) - Do not enable FEC without DSC (Ville) - Restore correct BXT DDI latency optim setting calculation (Ville) - Always reset context's RING registers to avoid running workload twice during reset (Chris) - Set GPU wedged on driver unload (Janusz) - Consolidate two similar barries from timeline into one (Chris) - Only reset the pinned kernel contexts on resume (Chris) - Wakeref tracking improvements (Chris, Imre) - Lockdep fixes for shrinker interactions (Chris) - Bump ready tasks ahead of busywaits in prep of semaphore use (Chris) - Huge step in splitting display code into fine grained files (Jani) - Refactor the IRQ init/reset macros for code saving (Paulo) - Convert IRQ initialization code to uncore MMIO access (Paulo) - Convert workarounds code to use uncore MMIO access (Chris) - Nuke drm_crtc_state and use intel_atomic_state instead (Manasi) - Update SKL clock-gating WA (Radhakrishna, Ville) - Isolate GuC reset code flow (Chris) - Expose force_dsc_enable through debugfs (Manasi) - Header standalone compile testing framework (Jani) - Code cleanups to reduce driver footprint (Chris) - PSR code fixes and cleanups (Jose) - Sparse and kerneldoc updates (Chris) - Suppress spurious combo PHY B warning (Vile) Signed-off-by: Dave Airlie <airlied@redhat.com> From: Joonas Lahtinen <joonas.lahtinen@linux.intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20190418080426.GA6409@jlahtine-desk.ger.corp.intel.com
281 lines
7.6 KiB
C
281 lines
7.6 KiB
C
/*
|
|
* Copyright(c) 2011-2016 Intel Corporation. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*
|
|
* Authors:
|
|
* Zhi Wang <zhi.a.wang@intel.com>
|
|
* Zhenyu Wang <zhenyuw@linux.intel.com>
|
|
* Xiao Zheng <xiao.zheng@intel.com>
|
|
*
|
|
* Contributors:
|
|
* Min He <min.he@intel.com>
|
|
* Bing Niu <bing.niu@intel.com>
|
|
*
|
|
*/
|
|
|
|
#ifndef _GVT_GTT_H_
|
|
#define _GVT_GTT_H_
|
|
|
|
#define I915_GTT_PAGE_SHIFT 12
|
|
|
|
struct intel_vgpu_mm;
|
|
|
|
#define INTEL_GVT_INVALID_ADDR (~0UL)
|
|
|
|
struct intel_gvt_gtt_entry {
|
|
u64 val64;
|
|
int type;
|
|
};
|
|
|
|
struct intel_gvt_gtt_pte_ops {
|
|
int (*get_entry)(void *pt,
|
|
struct intel_gvt_gtt_entry *e,
|
|
unsigned long index,
|
|
bool hypervisor_access,
|
|
unsigned long gpa,
|
|
struct intel_vgpu *vgpu);
|
|
int (*set_entry)(void *pt,
|
|
struct intel_gvt_gtt_entry *e,
|
|
unsigned long index,
|
|
bool hypervisor_access,
|
|
unsigned long gpa,
|
|
struct intel_vgpu *vgpu);
|
|
bool (*test_present)(struct intel_gvt_gtt_entry *e);
|
|
void (*clear_present)(struct intel_gvt_gtt_entry *e);
|
|
void (*set_present)(struct intel_gvt_gtt_entry *e);
|
|
bool (*test_pse)(struct intel_gvt_gtt_entry *e);
|
|
void (*clear_pse)(struct intel_gvt_gtt_entry *e);
|
|
bool (*test_ips)(struct intel_gvt_gtt_entry *e);
|
|
void (*clear_ips)(struct intel_gvt_gtt_entry *e);
|
|
bool (*test_64k_splited)(struct intel_gvt_gtt_entry *e);
|
|
void (*clear_64k_splited)(struct intel_gvt_gtt_entry *e);
|
|
void (*set_64k_splited)(struct intel_gvt_gtt_entry *e);
|
|
void (*set_pfn)(struct intel_gvt_gtt_entry *e, unsigned long pfn);
|
|
unsigned long (*get_pfn)(struct intel_gvt_gtt_entry *e);
|
|
};
|
|
|
|
struct intel_gvt_gtt_gma_ops {
|
|
unsigned long (*gma_to_ggtt_pte_index)(unsigned long gma);
|
|
unsigned long (*gma_to_pte_index)(unsigned long gma);
|
|
unsigned long (*gma_to_pde_index)(unsigned long gma);
|
|
unsigned long (*gma_to_l3_pdp_index)(unsigned long gma);
|
|
unsigned long (*gma_to_l4_pdp_index)(unsigned long gma);
|
|
unsigned long (*gma_to_pml4_index)(unsigned long gma);
|
|
};
|
|
|
|
struct intel_gvt_gtt {
|
|
struct intel_gvt_gtt_pte_ops *pte_ops;
|
|
struct intel_gvt_gtt_gma_ops *gma_ops;
|
|
int (*mm_alloc_page_table)(struct intel_vgpu_mm *mm);
|
|
void (*mm_free_page_table)(struct intel_vgpu_mm *mm);
|
|
struct list_head oos_page_use_list_head;
|
|
struct list_head oos_page_free_list_head;
|
|
struct mutex ppgtt_mm_lock;
|
|
struct list_head ppgtt_mm_lru_list_head;
|
|
|
|
struct page *scratch_page;
|
|
unsigned long scratch_mfn;
|
|
};
|
|
|
|
typedef enum {
|
|
GTT_TYPE_INVALID = -1,
|
|
|
|
GTT_TYPE_GGTT_PTE,
|
|
|
|
GTT_TYPE_PPGTT_PTE_4K_ENTRY,
|
|
GTT_TYPE_PPGTT_PTE_64K_ENTRY,
|
|
GTT_TYPE_PPGTT_PTE_2M_ENTRY,
|
|
GTT_TYPE_PPGTT_PTE_1G_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_PTE_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_PDE_ENTRY,
|
|
GTT_TYPE_PPGTT_PDP_ENTRY,
|
|
GTT_TYPE_PPGTT_PML4_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_ROOT_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_ROOT_L3_ENTRY,
|
|
GTT_TYPE_PPGTT_ROOT_L4_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_PTE_PT,
|
|
GTT_TYPE_PPGTT_PDE_PT,
|
|
GTT_TYPE_PPGTT_PDP_PT,
|
|
GTT_TYPE_PPGTT_PML4_PT,
|
|
|
|
GTT_TYPE_MAX,
|
|
} intel_gvt_gtt_type_t;
|
|
|
|
enum intel_gvt_mm_type {
|
|
INTEL_GVT_MM_GGTT,
|
|
INTEL_GVT_MM_PPGTT,
|
|
};
|
|
|
|
#define GVT_RING_CTX_NR_PDPS GEN8_3LVL_PDPES
|
|
|
|
struct intel_gvt_partial_pte {
|
|
unsigned long offset;
|
|
u64 data;
|
|
struct list_head list;
|
|
};
|
|
|
|
struct intel_vgpu_mm {
|
|
enum intel_gvt_mm_type type;
|
|
struct intel_vgpu *vgpu;
|
|
|
|
struct kref ref;
|
|
atomic_t pincount;
|
|
|
|
union {
|
|
struct {
|
|
intel_gvt_gtt_type_t root_entry_type;
|
|
/*
|
|
* The 4 PDPs in ring context. For 48bit addressing,
|
|
* only PDP0 is valid and point to PML4. For 32it
|
|
* addressing, all 4 are used as true PDPs.
|
|
*/
|
|
u64 guest_pdps[GVT_RING_CTX_NR_PDPS];
|
|
u64 shadow_pdps[GVT_RING_CTX_NR_PDPS];
|
|
bool shadowed;
|
|
|
|
struct list_head list;
|
|
struct list_head lru_list;
|
|
} ppgtt_mm;
|
|
struct {
|
|
void *virtual_ggtt;
|
|
struct list_head partial_pte_list;
|
|
} ggtt_mm;
|
|
};
|
|
};
|
|
|
|
struct intel_vgpu_mm *intel_vgpu_create_ppgtt_mm(struct intel_vgpu *vgpu,
|
|
intel_gvt_gtt_type_t root_entry_type, u64 pdps[]);
|
|
|
|
static inline void intel_vgpu_mm_get(struct intel_vgpu_mm *mm)
|
|
{
|
|
kref_get(&mm->ref);
|
|
}
|
|
|
|
void _intel_vgpu_mm_release(struct kref *mm_ref);
|
|
|
|
static inline void intel_vgpu_mm_put(struct intel_vgpu_mm *mm)
|
|
{
|
|
kref_put(&mm->ref, _intel_vgpu_mm_release);
|
|
}
|
|
|
|
static inline void intel_vgpu_destroy_mm(struct intel_vgpu_mm *mm)
|
|
{
|
|
intel_vgpu_mm_put(mm);
|
|
}
|
|
|
|
struct intel_vgpu_guest_page;
|
|
|
|
struct intel_vgpu_scratch_pt {
|
|
struct page *page;
|
|
unsigned long page_mfn;
|
|
};
|
|
|
|
struct intel_vgpu_gtt {
|
|
struct intel_vgpu_mm *ggtt_mm;
|
|
unsigned long active_ppgtt_mm_bitmap;
|
|
struct list_head ppgtt_mm_list_head;
|
|
struct radix_tree_root spt_tree;
|
|
struct list_head oos_page_list_head;
|
|
struct list_head post_shadow_list_head;
|
|
struct intel_vgpu_scratch_pt scratch_pt[GTT_TYPE_MAX];
|
|
};
|
|
|
|
extern int intel_vgpu_init_gtt(struct intel_vgpu *vgpu);
|
|
extern void intel_vgpu_clean_gtt(struct intel_vgpu *vgpu);
|
|
void intel_vgpu_reset_ggtt(struct intel_vgpu *vgpu, bool invalidate_old);
|
|
void intel_vgpu_invalidate_ppgtt(struct intel_vgpu *vgpu);
|
|
|
|
extern int intel_gvt_init_gtt(struct intel_gvt *gvt);
|
|
void intel_vgpu_reset_gtt(struct intel_vgpu *vgpu);
|
|
extern void intel_gvt_clean_gtt(struct intel_gvt *gvt);
|
|
|
|
extern struct intel_vgpu_mm *intel_gvt_find_ppgtt_mm(struct intel_vgpu *vgpu,
|
|
int page_table_level, void *root_entry);
|
|
|
|
struct intel_vgpu_oos_page {
|
|
struct intel_vgpu_ppgtt_spt *spt;
|
|
struct list_head list;
|
|
struct list_head vm_list;
|
|
int id;
|
|
void *mem;
|
|
};
|
|
|
|
#define GTT_ENTRY_NUM_IN_ONE_PAGE 512
|
|
|
|
/* Represent a vgpu shadow page table. */
|
|
struct intel_vgpu_ppgtt_spt {
|
|
atomic_t refcount;
|
|
struct intel_vgpu *vgpu;
|
|
|
|
struct {
|
|
intel_gvt_gtt_type_t type;
|
|
bool pde_ips; /* for 64KB PTEs */
|
|
void *vaddr;
|
|
struct page *page;
|
|
unsigned long mfn;
|
|
} shadow_page;
|
|
|
|
struct {
|
|
intel_gvt_gtt_type_t type;
|
|
bool pde_ips; /* for 64KB PTEs */
|
|
unsigned long gfn;
|
|
unsigned long write_cnt;
|
|
struct intel_vgpu_oos_page *oos_page;
|
|
} guest_page;
|
|
|
|
DECLARE_BITMAP(post_shadow_bitmap, GTT_ENTRY_NUM_IN_ONE_PAGE);
|
|
struct list_head post_shadow_list;
|
|
};
|
|
|
|
int intel_vgpu_sync_oos_pages(struct intel_vgpu *vgpu);
|
|
|
|
int intel_vgpu_flush_post_shadow(struct intel_vgpu *vgpu);
|
|
|
|
int intel_vgpu_pin_mm(struct intel_vgpu_mm *mm);
|
|
|
|
void intel_vgpu_unpin_mm(struct intel_vgpu_mm *mm);
|
|
|
|
unsigned long intel_vgpu_gma_to_gpa(struct intel_vgpu_mm *mm,
|
|
unsigned long gma);
|
|
|
|
struct intel_vgpu_mm *intel_vgpu_find_ppgtt_mm(struct intel_vgpu *vgpu,
|
|
u64 pdps[]);
|
|
|
|
struct intel_vgpu_mm *intel_vgpu_get_ppgtt_mm(struct intel_vgpu *vgpu,
|
|
intel_gvt_gtt_type_t root_entry_type, u64 pdps[]);
|
|
|
|
int intel_vgpu_put_ppgtt_mm(struct intel_vgpu *vgpu, u64 pdps[]);
|
|
|
|
int intel_vgpu_emulate_ggtt_mmio_read(struct intel_vgpu *vgpu,
|
|
unsigned int off, void *p_data, unsigned int bytes);
|
|
|
|
int intel_vgpu_emulate_ggtt_mmio_write(struct intel_vgpu *vgpu,
|
|
unsigned int off, void *p_data, unsigned int bytes);
|
|
|
|
#endif /* _GVT_GTT_H_ */
|