mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-25 17:05:24 +07:00
8e6d08e0a1
This patch introduces the SMP support for the OpenRISC architecture. The SMP architecture requires cores which have multi-core features which have been introduced a few years back including: - New SPRS SPR_COREID SPR_NUMCORES - Shadow SPRs - Atomic Instructions - Cache Coherency - A wired in IPI controller This patch adds all of the SMP specific changes to core infrastructure, it looks big but it needs to go all together as its hard to split this one up. Boot loader spinning of second cpu is not supported yet, it's assumed that Linux is booted straight after cpu reset. The bulk of these changes are trivial changes to refactor to use per cpu data structures throughout. The addition of the smp.c and changes in time.c are the changes. Some specific notes: MM changes ---------- The reason why this is created as an array, and not with DEFINE_PER_CPU is that doing it this way, we'll save a load in the tlb-miss handler (the load from __per_cpu_offset). TLB Flush --------- The SMP implementation of flush_tlb_* works by sending out a function-call IPI to all the non-local cpus by using the generic on_each_cpu() function. Currently, all flush_tlb_* functions will result in a flush_tlb_all(), which has always been the behaviour in the UP case. CPU INFO -------- This creates a per cpu cpuinfo struct and fills it out accordingly for each activated cpu. show_cpuinfo is also updated to reflect new version information in later versions of the spec. SMP API ------- This imitates the arm64 implementation by having a smp_cross_call callback that can be set by set_smp_cross_call to initiate an IPI and a handle_IPI function that is expected to be called from an IPI irqchip driver. Signed-off-by: Stefan Kristiansson <stefan.kristiansson@saunalahti.fi> [shorne@gmail.com: added cpu stop, checkpatch fixes, wrote commit message] Signed-off-by: Stafford Horne <shorne@gmail.com>
193 lines
5.0 KiB
C
193 lines
5.0 KiB
C
/*
|
|
* OpenRISC tlb.c
|
|
*
|
|
* Linux architectural port borrowing liberally from similar works of
|
|
* others. All original copyrights apply as per the original source
|
|
* declaration.
|
|
*
|
|
* Modifications for the OpenRISC architecture:
|
|
* Copyright (C) 2003 Matjaz Breskvar <phoenix@bsemi.com>
|
|
* Copyright (C) 2010-2011 Julius Baxter <julius.baxter@orsoc.se>
|
|
* Copyright (C) 2010-2011 Jonas Bonn <jonas@southpole.se>
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version
|
|
* 2 of the License, or (at your option) any later version.
|
|
*/
|
|
|
|
#include <linux/sched.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/string.h>
|
|
#include <linux/types.h>
|
|
#include <linux/ptrace.h>
|
|
#include <linux/mman.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/init.h>
|
|
|
|
#include <asm/segment.h>
|
|
#include <asm/tlbflush.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/mmu_context.h>
|
|
#include <asm/spr_defs.h>
|
|
|
|
#define NO_CONTEXT -1
|
|
|
|
#define NUM_DTLB_SETS (1 << ((mfspr(SPR_IMMUCFGR) & SPR_IMMUCFGR_NTS) >> \
|
|
SPR_DMMUCFGR_NTS_OFF))
|
|
#define NUM_ITLB_SETS (1 << ((mfspr(SPR_IMMUCFGR) & SPR_IMMUCFGR_NTS) >> \
|
|
SPR_IMMUCFGR_NTS_OFF))
|
|
#define DTLB_OFFSET(addr) (((addr) >> PAGE_SHIFT) & (NUM_DTLB_SETS-1))
|
|
#define ITLB_OFFSET(addr) (((addr) >> PAGE_SHIFT) & (NUM_ITLB_SETS-1))
|
|
/*
|
|
* Invalidate all TLB entries.
|
|
*
|
|
* This comes down to setting the 'valid' bit for all xTLBMR registers to 0.
|
|
* Easiest way to accomplish this is to just zero out the xTLBMR register
|
|
* completely.
|
|
*
|
|
*/
|
|
|
|
void local_flush_tlb_all(void)
|
|
{
|
|
int i;
|
|
unsigned long num_tlb_sets;
|
|
|
|
/* Determine number of sets for IMMU. */
|
|
/* FIXME: Assumption is I & D nsets equal. */
|
|
num_tlb_sets = NUM_ITLB_SETS;
|
|
|
|
for (i = 0; i < num_tlb_sets; i++) {
|
|
mtspr_off(SPR_DTLBMR_BASE(0), i, 0);
|
|
mtspr_off(SPR_ITLBMR_BASE(0), i, 0);
|
|
}
|
|
}
|
|
|
|
#define have_dtlbeir (mfspr(SPR_DMMUCFGR) & SPR_DMMUCFGR_TEIRI)
|
|
#define have_itlbeir (mfspr(SPR_IMMUCFGR) & SPR_IMMUCFGR_TEIRI)
|
|
|
|
/*
|
|
* Invalidate a single page. This is what the xTLBEIR register is for.
|
|
*
|
|
* There's no point in checking the vma for PAGE_EXEC to determine whether it's
|
|
* the data or instruction TLB that should be flushed... that would take more
|
|
* than the few instructions that the following compiles down to!
|
|
*
|
|
* The case where we don't have the xTLBEIR register really only works for
|
|
* MMU's with a single way and is hard-coded that way.
|
|
*/
|
|
|
|
#define flush_dtlb_page_eir(addr) mtspr(SPR_DTLBEIR, addr)
|
|
#define flush_dtlb_page_no_eir(addr) \
|
|
mtspr_off(SPR_DTLBMR_BASE(0), DTLB_OFFSET(addr), 0);
|
|
|
|
#define flush_itlb_page_eir(addr) mtspr(SPR_ITLBEIR, addr)
|
|
#define flush_itlb_page_no_eir(addr) \
|
|
mtspr_off(SPR_ITLBMR_BASE(0), ITLB_OFFSET(addr), 0);
|
|
|
|
void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long addr)
|
|
{
|
|
if (have_dtlbeir)
|
|
flush_dtlb_page_eir(addr);
|
|
else
|
|
flush_dtlb_page_no_eir(addr);
|
|
|
|
if (have_itlbeir)
|
|
flush_itlb_page_eir(addr);
|
|
else
|
|
flush_itlb_page_no_eir(addr);
|
|
}
|
|
|
|
void local_flush_tlb_range(struct vm_area_struct *vma,
|
|
unsigned long start, unsigned long end)
|
|
{
|
|
int addr;
|
|
bool dtlbeir;
|
|
bool itlbeir;
|
|
|
|
dtlbeir = have_dtlbeir;
|
|
itlbeir = have_itlbeir;
|
|
|
|
for (addr = start; addr < end; addr += PAGE_SIZE) {
|
|
if (dtlbeir)
|
|
flush_dtlb_page_eir(addr);
|
|
else
|
|
flush_dtlb_page_no_eir(addr);
|
|
|
|
if (itlbeir)
|
|
flush_itlb_page_eir(addr);
|
|
else
|
|
flush_itlb_page_no_eir(addr);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Invalidate the selected mm context only.
|
|
*
|
|
* FIXME: Due to some bug here, we're flushing everything for now.
|
|
* This should be changed to loop over over mm and call flush_tlb_range.
|
|
*/
|
|
|
|
void local_flush_tlb_mm(struct mm_struct *mm)
|
|
{
|
|
|
|
/* Was seeing bugs with the mm struct passed to us. Scrapped most of
|
|
this function. */
|
|
/* Several architctures do this */
|
|
local_flush_tlb_all();
|
|
}
|
|
|
|
/* called in schedule() just before actually doing the switch_to */
|
|
|
|
void switch_mm(struct mm_struct *prev, struct mm_struct *next,
|
|
struct task_struct *next_tsk)
|
|
{
|
|
/* remember the pgd for the fault handlers
|
|
* this is similar to the pgd register in some other CPU's.
|
|
* we need our own copy of it because current and active_mm
|
|
* might be invalid at points where we still need to derefer
|
|
* the pgd.
|
|
*/
|
|
current_pgd[smp_processor_id()] = next->pgd;
|
|
|
|
/* We don't have context support implemented, so flush all
|
|
* entries belonging to previous map
|
|
*/
|
|
|
|
if (prev != next)
|
|
local_flush_tlb_mm(prev);
|
|
|
|
}
|
|
|
|
/*
|
|
* Initialize the context related info for a new mm_struct
|
|
* instance.
|
|
*/
|
|
|
|
int init_new_context(struct task_struct *tsk, struct mm_struct *mm)
|
|
{
|
|
mm->context = NO_CONTEXT;
|
|
return 0;
|
|
}
|
|
|
|
/* called by __exit_mm to destroy the used MMU context if any before
|
|
* destroying the mm itself. this is only called when the last user of the mm
|
|
* drops it.
|
|
*/
|
|
|
|
void destroy_context(struct mm_struct *mm)
|
|
{
|
|
flush_tlb_mm(mm);
|
|
|
|
}
|
|
|
|
/* called once during VM initialization, from init.c */
|
|
|
|
void __init tlb_init(void)
|
|
{
|
|
/* Do nothing... */
|
|
/* invalidate the entire TLB */
|
|
/* flush_tlb_all(); */
|
|
}
|