mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
e31cf2f4ca
Patch series "mm: consolidate definitions of page table accessors", v2. The low level page table accessors (pXY_index(), pXY_offset()) are duplicated across all architectures and sometimes more than once. For instance, we have 31 definition of pgd_offset() for 25 supported architectures. Most of these definitions are actually identical and typically it boils down to, e.g. static inline unsigned long pmd_index(unsigned long address) { return (address >> PMD_SHIFT) & (PTRS_PER_PMD - 1); } static inline pmd_t *pmd_offset(pud_t *pud, unsigned long address) { return (pmd_t *)pud_page_vaddr(*pud) + pmd_index(address); } These definitions can be shared among 90% of the arches provided XYZ_SHIFT, PTRS_PER_XYZ and xyz_page_vaddr() are defined. For architectures that really need a custom version there is always possibility to override the generic version with the usual ifdefs magic. These patches introduce include/linux/pgtable.h that replaces include/asm-generic/pgtable.h and add the definitions of the page table accessors to the new header. This patch (of 12): The linux/mm.h header includes <asm/pgtable.h> to allow inlining of the functions involving page table manipulations, e.g. pte_alloc() and pmd_alloc(). So, there is no point to explicitly include <asm/pgtable.h> in the files that include <linux/mm.h>. The include statements in such cases are remove with a simple loop: for f in $(git grep -l "include <linux/mm.h>") ; do sed -i -e '/include <asm\/pgtable.h>/ d' $f done Signed-off-by: Mike Rapoport <rppt@linux.ibm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Cc: Arnd Bergmann <arnd@arndb.de> Cc: Borislav Petkov <bp@alien8.de> Cc: Brian Cain <bcain@codeaurora.org> Cc: Catalin Marinas <catalin.marinas@arm.com> Cc: Chris Zankel <chris@zankel.net> Cc: "David S. Miller" <davem@davemloft.net> Cc: Geert Uytterhoeven <geert@linux-m68k.org> Cc: Greentime Hu <green.hu@gmail.com> Cc: Greg Ungerer <gerg@linux-m68k.org> Cc: Guan Xuetao <gxt@pku.edu.cn> Cc: Guo Ren <guoren@kernel.org> Cc: Heiko Carstens <heiko.carstens@de.ibm.com> Cc: Helge Deller <deller@gmx.de> Cc: Ingo Molnar <mingo@redhat.com> Cc: Ley Foon Tan <ley.foon.tan@intel.com> Cc: Mark Salter <msalter@redhat.com> Cc: Matthew Wilcox <willy@infradead.org> Cc: Matt Turner <mattst88@gmail.com> Cc: Max Filippov <jcmvbkbc@gmail.com> Cc: Michael Ellerman <mpe@ellerman.id.au> Cc: Michal Simek <monstr@monstr.eu> Cc: Mike Rapoport <rppt@kernel.org> Cc: Nick Hu <nickhu@andestech.com> Cc: Paul Walmsley <paul.walmsley@sifive.com> Cc: Richard Weinberger <richard@nod.at> Cc: Rich Felker <dalias@libc.org> Cc: Russell King <linux@armlinux.org.uk> Cc: Stafford Horne <shorne@gmail.com> Cc: Thomas Bogendoerfer <tsbogend@alpha.franken.de> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Tony Luck <tony.luck@intel.com> Cc: Vincent Chen <deanbo422@gmail.com> Cc: Vineet Gupta <vgupta@synopsys.com> Cc: Will Deacon <will@kernel.org> Cc: Yoshinori Sato <ysato@users.sourceforge.jp> Link: http://lkml.kernel.org/r/20200514170327.31389-1-rppt@kernel.org Link: http://lkml.kernel.org/r/20200514170327.31389-2-rppt@kernel.org Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
200 lines
5.4 KiB
C
200 lines
5.4 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Dump R4x00 TLB for debugging purposes.
|
|
*
|
|
* Copyright (C) 1994, 1995 by Waldorf Electronics, written by Ralf Baechle.
|
|
* Copyright (C) 1999 by Silicon Graphics, Inc.
|
|
*/
|
|
#include <linux/kernel.h>
|
|
#include <linux/mm.h>
|
|
|
|
#include <asm/hazards.h>
|
|
#include <asm/mipsregs.h>
|
|
#include <asm/mmu_context.h>
|
|
#include <asm/page.h>
|
|
#include <asm/tlbdebug.h>
|
|
|
|
void dump_tlb_regs(void)
|
|
{
|
|
const int field = 2 * sizeof(unsigned long);
|
|
|
|
pr_info("Index : %0x\n", read_c0_index());
|
|
pr_info("PageMask : %0x\n", read_c0_pagemask());
|
|
if (cpu_has_guestid)
|
|
pr_info("GuestCtl1: %0x\n", read_c0_guestctl1());
|
|
pr_info("EntryHi : %0*lx\n", field, read_c0_entryhi());
|
|
pr_info("EntryLo0 : %0*lx\n", field, read_c0_entrylo0());
|
|
pr_info("EntryLo1 : %0*lx\n", field, read_c0_entrylo1());
|
|
pr_info("Wired : %0x\n", read_c0_wired());
|
|
switch (current_cpu_type()) {
|
|
case CPU_R10000:
|
|
case CPU_R12000:
|
|
case CPU_R14000:
|
|
case CPU_R16000:
|
|
pr_info("FrameMask: %0x\n", read_c0_framemask());
|
|
break;
|
|
}
|
|
if (cpu_has_small_pages || cpu_has_rixi || cpu_has_xpa)
|
|
pr_info("PageGrain: %0x\n", read_c0_pagegrain());
|
|
if (cpu_has_htw) {
|
|
pr_info("PWField : %0*lx\n", field, read_c0_pwfield());
|
|
pr_info("PWSize : %0*lx\n", field, read_c0_pwsize());
|
|
pr_info("PWCtl : %0x\n", read_c0_pwctl());
|
|
}
|
|
}
|
|
|
|
static inline const char *msk2str(unsigned int mask)
|
|
{
|
|
switch (mask) {
|
|
case PM_4K: return "4kb";
|
|
case PM_16K: return "16kb";
|
|
case PM_64K: return "64kb";
|
|
case PM_256K: return "256kb";
|
|
#ifdef CONFIG_CPU_CAVIUM_OCTEON
|
|
case PM_8K: return "8kb";
|
|
case PM_32K: return "32kb";
|
|
case PM_128K: return "128kb";
|
|
case PM_512K: return "512kb";
|
|
case PM_2M: return "2Mb";
|
|
case PM_8M: return "8Mb";
|
|
case PM_32M: return "32Mb";
|
|
#endif
|
|
#ifndef CONFIG_CPU_VR41XX
|
|
case PM_1M: return "1Mb";
|
|
case PM_4M: return "4Mb";
|
|
case PM_16M: return "16Mb";
|
|
case PM_64M: return "64Mb";
|
|
case PM_256M: return "256Mb";
|
|
case PM_1G: return "1Gb";
|
|
#endif
|
|
}
|
|
return "";
|
|
}
|
|
|
|
static void dump_tlb(int first, int last)
|
|
{
|
|
unsigned long s_entryhi, entryhi, asid, mmid;
|
|
unsigned long long entrylo0, entrylo1, pa;
|
|
unsigned int s_index, s_pagemask, s_guestctl1 = 0;
|
|
unsigned int pagemask, guestctl1 = 0, c0, c1, i;
|
|
unsigned long asidmask = cpu_asid_mask(¤t_cpu_data);
|
|
int asidwidth = DIV_ROUND_UP(ilog2(asidmask) + 1, 4);
|
|
unsigned long uninitialized_var(s_mmid);
|
|
#ifdef CONFIG_32BIT
|
|
bool xpa = cpu_has_xpa && (read_c0_pagegrain() & PG_ELPA);
|
|
int pwidth = xpa ? 11 : 8;
|
|
int vwidth = 8;
|
|
#else
|
|
bool xpa = false;
|
|
int pwidth = 11;
|
|
int vwidth = 11;
|
|
#endif
|
|
|
|
s_pagemask = read_c0_pagemask();
|
|
s_entryhi = read_c0_entryhi();
|
|
s_index = read_c0_index();
|
|
|
|
if (cpu_has_mmid)
|
|
asid = s_mmid = read_c0_memorymapid();
|
|
else
|
|
asid = s_entryhi & asidmask;
|
|
|
|
if (cpu_has_guestid)
|
|
s_guestctl1 = read_c0_guestctl1();
|
|
|
|
for (i = first; i <= last; i++) {
|
|
write_c0_index(i);
|
|
mtc0_tlbr_hazard();
|
|
tlb_read();
|
|
tlb_read_hazard();
|
|
pagemask = read_c0_pagemask();
|
|
entryhi = read_c0_entryhi();
|
|
entrylo0 = read_c0_entrylo0();
|
|
entrylo1 = read_c0_entrylo1();
|
|
|
|
if (cpu_has_mmid)
|
|
mmid = read_c0_memorymapid();
|
|
else
|
|
mmid = entryhi & asidmask;
|
|
|
|
if (cpu_has_guestid)
|
|
guestctl1 = read_c0_guestctl1();
|
|
|
|
/* EHINV bit marks entire entry as invalid */
|
|
if (cpu_has_tlbinv && entryhi & MIPS_ENTRYHI_EHINV)
|
|
continue;
|
|
/*
|
|
* Prior to tlbinv, unused entries have a virtual address of
|
|
* CKSEG0.
|
|
*/
|
|
if ((entryhi & ~0x1ffffUL) == CKSEG0)
|
|
continue;
|
|
/*
|
|
* ASID takes effect in absence of G (global) bit.
|
|
* We check both G bits, even though architecturally they should
|
|
* match one another, because some revisions of the SB1 core may
|
|
* leave only a single G bit set after a machine check exception
|
|
* due to duplicate TLB entry.
|
|
*/
|
|
if (!((entrylo0 | entrylo1) & ENTRYLO_G) && (mmid != asid))
|
|
continue;
|
|
|
|
/*
|
|
* Only print entries in use
|
|
*/
|
|
printk("Index: %2d pgmask=%s ", i, msk2str(pagemask));
|
|
|
|
c0 = (entrylo0 & ENTRYLO_C) >> ENTRYLO_C_SHIFT;
|
|
c1 = (entrylo1 & ENTRYLO_C) >> ENTRYLO_C_SHIFT;
|
|
|
|
pr_cont("va=%0*lx asid=%0*lx",
|
|
vwidth, (entryhi & ~0x1fffUL),
|
|
asidwidth, mmid);
|
|
if (cpu_has_guestid)
|
|
pr_cont(" gid=%02lx",
|
|
(guestctl1 & MIPS_GCTL1_RID)
|
|
>> MIPS_GCTL1_RID_SHIFT);
|
|
/* RI/XI are in awkward places, so mask them off separately */
|
|
pa = entrylo0 & ~(MIPS_ENTRYLO_RI | MIPS_ENTRYLO_XI);
|
|
if (xpa)
|
|
pa |= (unsigned long long)readx_c0_entrylo0() << 30;
|
|
pa = (pa << 6) & PAGE_MASK;
|
|
pr_cont("\n\t[");
|
|
if (cpu_has_rixi)
|
|
pr_cont("ri=%d xi=%d ",
|
|
(entrylo0 & MIPS_ENTRYLO_RI) ? 1 : 0,
|
|
(entrylo0 & MIPS_ENTRYLO_XI) ? 1 : 0);
|
|
pr_cont("pa=%0*llx c=%d d=%d v=%d g=%d] [",
|
|
pwidth, pa, c0,
|
|
(entrylo0 & ENTRYLO_D) ? 1 : 0,
|
|
(entrylo0 & ENTRYLO_V) ? 1 : 0,
|
|
(entrylo0 & ENTRYLO_G) ? 1 : 0);
|
|
/* RI/XI are in awkward places, so mask them off separately */
|
|
pa = entrylo1 & ~(MIPS_ENTRYLO_RI | MIPS_ENTRYLO_XI);
|
|
if (xpa)
|
|
pa |= (unsigned long long)readx_c0_entrylo1() << 30;
|
|
pa = (pa << 6) & PAGE_MASK;
|
|
if (cpu_has_rixi)
|
|
pr_cont("ri=%d xi=%d ",
|
|
(entrylo1 & MIPS_ENTRYLO_RI) ? 1 : 0,
|
|
(entrylo1 & MIPS_ENTRYLO_XI) ? 1 : 0);
|
|
pr_cont("pa=%0*llx c=%d d=%d v=%d g=%d]\n",
|
|
pwidth, pa, c1,
|
|
(entrylo1 & ENTRYLO_D) ? 1 : 0,
|
|
(entrylo1 & ENTRYLO_V) ? 1 : 0,
|
|
(entrylo1 & ENTRYLO_G) ? 1 : 0);
|
|
}
|
|
printk("\n");
|
|
|
|
write_c0_entryhi(s_entryhi);
|
|
write_c0_index(s_index);
|
|
write_c0_pagemask(s_pagemask);
|
|
if (cpu_has_guestid)
|
|
write_c0_guestctl1(s_guestctl1);
|
|
}
|
|
|
|
void dump_tlb_all(void)
|
|
{
|
|
dump_tlb(0, current_cpu_data.tlbsize - 1);
|
|
}
|