mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-05 22:36:40 +07:00
f32af63ed1
The test to check whether we have _PAGE_SPECIAL defined is broken, since we always define it, just not always to a meaninful value :-) That broke 8xx and 40x under some circumstances. This fixes it by adding _PAGE_SPECIAL for both of these since they had a free PTE bit, and removing the condition around advertising it. Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
66 lines
2.9 KiB
C
66 lines
2.9 KiB
C
#ifndef _ASM_POWERPC_PTE_40x_H
|
|
#define _ASM_POWERPC_PTE_40x_H
|
|
#ifdef __KERNEL__
|
|
|
|
/*
|
|
* At present, all PowerPC 400-class processors share a similar TLB
|
|
* architecture. The instruction and data sides share a unified,
|
|
* 64-entry, fully-associative TLB which is maintained totally under
|
|
* software control. In addition, the instruction side has a
|
|
* hardware-managed, 4-entry, fully-associative TLB which serves as a
|
|
* first level to the shared TLB. These two TLBs are known as the UTLB
|
|
* and ITLB, respectively (see "mmu.h" for definitions).
|
|
*
|
|
* There are several potential gotchas here. The 40x hardware TLBLO
|
|
* field looks like this:
|
|
*
|
|
* 0 1 2 3 4 ... 18 19 20 21 22 23 24 25 26 27 28 29 30 31
|
|
* RPN..................... 0 0 EX WR ZSEL....... W I M G
|
|
*
|
|
* Where possible we make the Linux PTE bits match up with this
|
|
*
|
|
* - bits 20 and 21 must be cleared, because we use 4k pages (40x can
|
|
* support down to 1k pages), this is done in the TLBMiss exception
|
|
* handler.
|
|
* - We use only zones 0 (for kernel pages) and 1 (for user pages)
|
|
* of the 16 available. Bit 24-26 of the TLB are cleared in the TLB
|
|
* miss handler. Bit 27 is PAGE_USER, thus selecting the correct
|
|
* zone.
|
|
* - PRESENT *must* be in the bottom two bits because swap cache
|
|
* entries use the top 30 bits. Because 40x doesn't support SMP
|
|
* anyway, M is irrelevant so we borrow it for PAGE_PRESENT. Bit 30
|
|
* is cleared in the TLB miss handler before the TLB entry is loaded.
|
|
* - All other bits of the PTE are loaded into TLBLO without
|
|
* modification, leaving us only the bits 20, 21, 24, 25, 26, 30 for
|
|
* software PTE bits. We actually use use bits 21, 24, 25, and
|
|
* 30 respectively for the software bits: ACCESSED, DIRTY, RW, and
|
|
* PRESENT.
|
|
*/
|
|
|
|
#define _PAGE_GUARDED 0x001 /* G: page is guarded from prefetch */
|
|
#define _PAGE_FILE 0x001 /* when !present: nonlinear file mapping */
|
|
#define _PAGE_PRESENT 0x002 /* software: PTE contains a translation */
|
|
#define _PAGE_NO_CACHE 0x004 /* I: caching is inhibited */
|
|
#define _PAGE_WRITETHRU 0x008 /* W: caching is write-through */
|
|
#define _PAGE_USER 0x010 /* matches one of the zone permission bits */
|
|
#define _PAGE_SPECIAL 0x020 /* software: Special page */
|
|
#define _PAGE_RW 0x040 /* software: Writes permitted */
|
|
#define _PAGE_DIRTY 0x080 /* software: dirty page */
|
|
#define _PAGE_HWWRITE 0x100 /* hardware: Dirty & RW, set in exception */
|
|
#define _PAGE_EXEC 0x200 /* hardware: EX permission */
|
|
#define _PAGE_ACCESSED 0x400 /* software: R: page referenced */
|
|
|
|
#define _PMD_PRESENT 0x400 /* PMD points to page of PTEs */
|
|
#define _PMD_BAD 0x802
|
|
#define _PMD_SIZE 0x0e0 /* size field, != 0 for large-page PMD entry */
|
|
#define _PMD_SIZE_4M 0x0c0
|
|
#define _PMD_SIZE_16M 0x0e0
|
|
|
|
#define PMD_PAGE_SIZE(pmdval) (1024 << (((pmdval) & _PMD_SIZE) >> 4))
|
|
|
|
/* Until my rework is finished, 40x still needs atomic PTE updates */
|
|
#define PTE_ATOMIC_UPDATES 1
|
|
|
|
#endif /* __KERNEL__ */
|
|
#endif /* _ASM_POWERPC_PTE_40x_H */
|