mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
bec3df930f
The PPGTT in context image can be overridden by LRI cmd with another PPGTT's pdps. In such case, the load mm is used instead of the one in the context image. So we need to load its shadow mm in GVT and replace ppgtt pointers in command. This feature is used by guest IGD driver to share gfx VM between different contexts. Verified by IGT "gem_ctx_clone" test. v4: - consolidate shadow mm handlers (Yan) - fix cmd shadow mm pin error path v3: (Zhenyu Wang) - Cleanup PDP register offset check - Add debug check for guest context ppgtt update - Skip 3-level ppgtt guest handling code. The reason is that all guests now use 4-level ppgtt table and the only left case for 3-level table is ancient aliasing ppgtt case. But those guest kernel has no use of PPGTT LRI command. So 3-level ppgtt guest for this feature becomes simply un-testable. v2: (Zhenyu Wang) - Change to list for handling possible multiple ppgtt table loads in one submission. Make sure shadow mm is to replace for each one. Reviewed-by: Yan Zhao <yan.y.zhao@intel.com> Cc: Yan Zhao <yan.y.zhao@intel.com> Signed-off-by: Tina Zhang <tina.zhang@intel.com> Signed-off-by: Zhenyu Wang <zhenyuw@linux.intel.com> Link: http://patchwork.freedesktop.org/patch/msgid/20200508031409.2562-1-zhenyuw@linux.intel.com
283 lines
7.7 KiB
C
283 lines
7.7 KiB
C
/*
|
|
* Copyright(c) 2011-2016 Intel Corporation. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*
|
|
* Authors:
|
|
* Zhi Wang <zhi.a.wang@intel.com>
|
|
* Zhenyu Wang <zhenyuw@linux.intel.com>
|
|
* Xiao Zheng <xiao.zheng@intel.com>
|
|
*
|
|
* Contributors:
|
|
* Min He <min.he@intel.com>
|
|
* Bing Niu <bing.niu@intel.com>
|
|
*
|
|
*/
|
|
|
|
#ifndef _GVT_GTT_H_
|
|
#define _GVT_GTT_H_
|
|
|
|
#define I915_GTT_PAGE_SHIFT 12
|
|
|
|
struct intel_vgpu_mm;
|
|
|
|
#define INTEL_GVT_INVALID_ADDR (~0UL)
|
|
|
|
struct intel_gvt_gtt_entry {
|
|
u64 val64;
|
|
int type;
|
|
};
|
|
|
|
struct intel_gvt_gtt_pte_ops {
|
|
int (*get_entry)(void *pt,
|
|
struct intel_gvt_gtt_entry *e,
|
|
unsigned long index,
|
|
bool hypervisor_access,
|
|
unsigned long gpa,
|
|
struct intel_vgpu *vgpu);
|
|
int (*set_entry)(void *pt,
|
|
struct intel_gvt_gtt_entry *e,
|
|
unsigned long index,
|
|
bool hypervisor_access,
|
|
unsigned long gpa,
|
|
struct intel_vgpu *vgpu);
|
|
bool (*test_present)(struct intel_gvt_gtt_entry *e);
|
|
void (*clear_present)(struct intel_gvt_gtt_entry *e);
|
|
void (*set_present)(struct intel_gvt_gtt_entry *e);
|
|
bool (*test_pse)(struct intel_gvt_gtt_entry *e);
|
|
void (*clear_pse)(struct intel_gvt_gtt_entry *e);
|
|
bool (*test_ips)(struct intel_gvt_gtt_entry *e);
|
|
void (*clear_ips)(struct intel_gvt_gtt_entry *e);
|
|
bool (*test_64k_splited)(struct intel_gvt_gtt_entry *e);
|
|
void (*clear_64k_splited)(struct intel_gvt_gtt_entry *e);
|
|
void (*set_64k_splited)(struct intel_gvt_gtt_entry *e);
|
|
void (*set_pfn)(struct intel_gvt_gtt_entry *e, unsigned long pfn);
|
|
unsigned long (*get_pfn)(struct intel_gvt_gtt_entry *e);
|
|
};
|
|
|
|
struct intel_gvt_gtt_gma_ops {
|
|
unsigned long (*gma_to_ggtt_pte_index)(unsigned long gma);
|
|
unsigned long (*gma_to_pte_index)(unsigned long gma);
|
|
unsigned long (*gma_to_pde_index)(unsigned long gma);
|
|
unsigned long (*gma_to_l3_pdp_index)(unsigned long gma);
|
|
unsigned long (*gma_to_l4_pdp_index)(unsigned long gma);
|
|
unsigned long (*gma_to_pml4_index)(unsigned long gma);
|
|
};
|
|
|
|
struct intel_gvt_gtt {
|
|
struct intel_gvt_gtt_pte_ops *pte_ops;
|
|
struct intel_gvt_gtt_gma_ops *gma_ops;
|
|
int (*mm_alloc_page_table)(struct intel_vgpu_mm *mm);
|
|
void (*mm_free_page_table)(struct intel_vgpu_mm *mm);
|
|
struct list_head oos_page_use_list_head;
|
|
struct list_head oos_page_free_list_head;
|
|
struct mutex ppgtt_mm_lock;
|
|
struct list_head ppgtt_mm_lru_list_head;
|
|
|
|
struct page *scratch_page;
|
|
unsigned long scratch_mfn;
|
|
};
|
|
|
|
enum intel_gvt_gtt_type {
|
|
GTT_TYPE_INVALID = 0,
|
|
|
|
GTT_TYPE_GGTT_PTE,
|
|
|
|
GTT_TYPE_PPGTT_PTE_4K_ENTRY,
|
|
GTT_TYPE_PPGTT_PTE_64K_ENTRY,
|
|
GTT_TYPE_PPGTT_PTE_2M_ENTRY,
|
|
GTT_TYPE_PPGTT_PTE_1G_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_PTE_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_PDE_ENTRY,
|
|
GTT_TYPE_PPGTT_PDP_ENTRY,
|
|
GTT_TYPE_PPGTT_PML4_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_ROOT_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_ROOT_L3_ENTRY,
|
|
GTT_TYPE_PPGTT_ROOT_L4_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_ENTRY,
|
|
|
|
GTT_TYPE_PPGTT_PTE_PT,
|
|
GTT_TYPE_PPGTT_PDE_PT,
|
|
GTT_TYPE_PPGTT_PDP_PT,
|
|
GTT_TYPE_PPGTT_PML4_PT,
|
|
|
|
GTT_TYPE_MAX,
|
|
};
|
|
|
|
enum intel_gvt_mm_type {
|
|
INTEL_GVT_MM_GGTT,
|
|
INTEL_GVT_MM_PPGTT,
|
|
};
|
|
|
|
#define GVT_RING_CTX_NR_PDPS GEN8_3LVL_PDPES
|
|
|
|
struct intel_gvt_partial_pte {
|
|
unsigned long offset;
|
|
u64 data;
|
|
struct list_head list;
|
|
};
|
|
|
|
struct intel_vgpu_mm {
|
|
enum intel_gvt_mm_type type;
|
|
struct intel_vgpu *vgpu;
|
|
|
|
struct kref ref;
|
|
atomic_t pincount;
|
|
|
|
union {
|
|
struct {
|
|
enum intel_gvt_gtt_type root_entry_type;
|
|
/*
|
|
* The 4 PDPs in ring context. For 48bit addressing,
|
|
* only PDP0 is valid and point to PML4. For 32it
|
|
* addressing, all 4 are used as true PDPs.
|
|
*/
|
|
u64 guest_pdps[GVT_RING_CTX_NR_PDPS];
|
|
u64 shadow_pdps[GVT_RING_CTX_NR_PDPS];
|
|
bool shadowed;
|
|
|
|
struct list_head list;
|
|
struct list_head lru_list;
|
|
struct list_head link; /* possible LRI shadow mm list */
|
|
} ppgtt_mm;
|
|
struct {
|
|
void *virtual_ggtt;
|
|
struct list_head partial_pte_list;
|
|
} ggtt_mm;
|
|
};
|
|
};
|
|
|
|
struct intel_vgpu_mm *intel_vgpu_create_ppgtt_mm(struct intel_vgpu *vgpu,
|
|
enum intel_gvt_gtt_type root_entry_type, u64 pdps[]);
|
|
|
|
static inline void intel_vgpu_mm_get(struct intel_vgpu_mm *mm)
|
|
{
|
|
kref_get(&mm->ref);
|
|
}
|
|
|
|
void _intel_vgpu_mm_release(struct kref *mm_ref);
|
|
|
|
static inline void intel_vgpu_mm_put(struct intel_vgpu_mm *mm)
|
|
{
|
|
kref_put(&mm->ref, _intel_vgpu_mm_release);
|
|
}
|
|
|
|
static inline void intel_vgpu_destroy_mm(struct intel_vgpu_mm *mm)
|
|
{
|
|
intel_vgpu_mm_put(mm);
|
|
}
|
|
|
|
struct intel_vgpu_guest_page;
|
|
|
|
struct intel_vgpu_scratch_pt {
|
|
struct page *page;
|
|
unsigned long page_mfn;
|
|
};
|
|
|
|
struct intel_vgpu_gtt {
|
|
struct intel_vgpu_mm *ggtt_mm;
|
|
unsigned long active_ppgtt_mm_bitmap;
|
|
struct list_head ppgtt_mm_list_head;
|
|
struct radix_tree_root spt_tree;
|
|
struct list_head oos_page_list_head;
|
|
struct list_head post_shadow_list_head;
|
|
struct intel_vgpu_scratch_pt scratch_pt[GTT_TYPE_MAX];
|
|
};
|
|
|
|
int intel_vgpu_init_gtt(struct intel_vgpu *vgpu);
|
|
void intel_vgpu_clean_gtt(struct intel_vgpu *vgpu);
|
|
void intel_vgpu_reset_ggtt(struct intel_vgpu *vgpu, bool invalidate_old);
|
|
void intel_vgpu_invalidate_ppgtt(struct intel_vgpu *vgpu);
|
|
|
|
int intel_gvt_init_gtt(struct intel_gvt *gvt);
|
|
void intel_vgpu_reset_gtt(struct intel_vgpu *vgpu);
|
|
void intel_gvt_clean_gtt(struct intel_gvt *gvt);
|
|
|
|
struct intel_vgpu_mm *intel_gvt_find_ppgtt_mm(struct intel_vgpu *vgpu,
|
|
int page_table_level,
|
|
void *root_entry);
|
|
|
|
struct intel_vgpu_oos_page {
|
|
struct intel_vgpu_ppgtt_spt *spt;
|
|
struct list_head list;
|
|
struct list_head vm_list;
|
|
int id;
|
|
void *mem;
|
|
};
|
|
|
|
#define GTT_ENTRY_NUM_IN_ONE_PAGE 512
|
|
|
|
/* Represent a vgpu shadow page table. */
|
|
struct intel_vgpu_ppgtt_spt {
|
|
atomic_t refcount;
|
|
struct intel_vgpu *vgpu;
|
|
|
|
struct {
|
|
enum intel_gvt_gtt_type type;
|
|
bool pde_ips; /* for 64KB PTEs */
|
|
void *vaddr;
|
|
struct page *page;
|
|
unsigned long mfn;
|
|
} shadow_page;
|
|
|
|
struct {
|
|
enum intel_gvt_gtt_type type;
|
|
bool pde_ips; /* for 64KB PTEs */
|
|
unsigned long gfn;
|
|
unsigned long write_cnt;
|
|
struct intel_vgpu_oos_page *oos_page;
|
|
} guest_page;
|
|
|
|
DECLARE_BITMAP(post_shadow_bitmap, GTT_ENTRY_NUM_IN_ONE_PAGE);
|
|
struct list_head post_shadow_list;
|
|
};
|
|
|
|
int intel_vgpu_sync_oos_pages(struct intel_vgpu *vgpu);
|
|
|
|
int intel_vgpu_flush_post_shadow(struct intel_vgpu *vgpu);
|
|
|
|
int intel_vgpu_pin_mm(struct intel_vgpu_mm *mm);
|
|
|
|
void intel_vgpu_unpin_mm(struct intel_vgpu_mm *mm);
|
|
|
|
unsigned long intel_vgpu_gma_to_gpa(struct intel_vgpu_mm *mm,
|
|
unsigned long gma);
|
|
|
|
struct intel_vgpu_mm *intel_vgpu_find_ppgtt_mm(struct intel_vgpu *vgpu,
|
|
u64 pdps[]);
|
|
|
|
struct intel_vgpu_mm *intel_vgpu_get_ppgtt_mm(struct intel_vgpu *vgpu,
|
|
enum intel_gvt_gtt_type root_entry_type, u64 pdps[]);
|
|
|
|
int intel_vgpu_put_ppgtt_mm(struct intel_vgpu *vgpu, u64 pdps[]);
|
|
|
|
int intel_vgpu_emulate_ggtt_mmio_read(struct intel_vgpu *vgpu,
|
|
unsigned int off, void *p_data, unsigned int bytes);
|
|
|
|
int intel_vgpu_emulate_ggtt_mmio_write(struct intel_vgpu *vgpu,
|
|
unsigned int off, void *p_data, unsigned int bytes);
|
|
|
|
#endif /* _GVT_GTT_H_ */
|