mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-25 23:15:26 +07:00
c7d9f77d33
Add support for using multiple hugepage sizes simultaneously on mainline. Currently, support for 256M has been added which can be used along with 8M pages. Page tables are set like this (e.g. for 256M page): VA + (8M * x) -> PA + (8M * x) (sz bit = 256M) where x in [0, 31] and TSB is set similarly: VA + (4M * x) -> PA + (4M * x) (sz bit = 256M) where x in [0, 63] - Testing Tested on Sonoma (which supports 256M pages) by running stream benchmark instances in parallel: one instance uses 8M pages and another uses 256M pages, consuming 48G each. Boot params used: default_hugepagesz=256M hugepagesz=256M hugepages=300 hugepagesz=8M hugepages=10000 Signed-off-by: Nitin Gupta <nitin.m.gupta@oracle.com> Signed-off-by: David S. Miller <davem@davemloft.net>
71 lines
1.7 KiB
C
71 lines
1.7 KiB
C
#ifndef _SPARC64_TLBFLUSH_H
|
|
#define _SPARC64_TLBFLUSH_H
|
|
|
|
#include <asm/mmu_context.h>
|
|
|
|
/* TSB flush operations. */
|
|
|
|
#define TLB_BATCH_NR 192
|
|
|
|
struct tlb_batch {
|
|
unsigned int hugepage_shift;
|
|
struct mm_struct *mm;
|
|
unsigned long tlb_nr;
|
|
unsigned long active;
|
|
unsigned long vaddrs[TLB_BATCH_NR];
|
|
};
|
|
|
|
void flush_tsb_kernel_range(unsigned long start, unsigned long end);
|
|
void flush_tsb_user(struct tlb_batch *tb);
|
|
void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr,
|
|
unsigned int hugepage_shift);
|
|
|
|
/* TLB flush operations. */
|
|
|
|
static inline void flush_tlb_mm(struct mm_struct *mm)
|
|
{
|
|
}
|
|
|
|
static inline void flush_tlb_page(struct vm_area_struct *vma,
|
|
unsigned long vmaddr)
|
|
{
|
|
}
|
|
|
|
static inline void flush_tlb_range(struct vm_area_struct *vma,
|
|
unsigned long start, unsigned long end)
|
|
{
|
|
}
|
|
|
|
void flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
|
|
|
#define __HAVE_ARCH_ENTER_LAZY_MMU_MODE
|
|
|
|
void flush_tlb_pending(void);
|
|
void arch_enter_lazy_mmu_mode(void);
|
|
void arch_leave_lazy_mmu_mode(void);
|
|
#define arch_flush_lazy_mmu_mode() do {} while (0)
|
|
|
|
/* Local cpu only. */
|
|
void __flush_tlb_all(void);
|
|
void __flush_tlb_page(unsigned long context, unsigned long vaddr);
|
|
void __flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
|
|
|
#ifndef CONFIG_SMP
|
|
|
|
static inline void global_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr)
|
|
{
|
|
__flush_tlb_page(CTX_HWBITS(mm->context), vaddr);
|
|
}
|
|
|
|
#else /* CONFIG_SMP */
|
|
|
|
void smp_flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
|
void smp_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr);
|
|
|
|
#define global_flush_tlb_page(mm, vaddr) \
|
|
smp_flush_tlb_page(mm, vaddr)
|
|
|
|
#endif /* ! CONFIG_SMP */
|
|
|
|
#endif /* _SPARC64_TLBFLUSH_H */
|