2007-11-09 15:08:54 +07:00
|
|
|
/*
|
|
|
|
* include/asm-sh/processor.h
|
|
|
|
*
|
|
|
|
* Copyright (C) 1999, 2000 Niibe Yutaka
|
|
|
|
* Copyright (C) 2002, 2003 Paul Mundt
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef __ASM_SH_PROCESSOR_32_H
|
|
|
|
#define __ASM_SH_PROCESSOR_32_H
|
|
|
|
#ifdef __KERNEL__
|
|
|
|
|
|
|
|
#include <linux/compiler.h>
|
2008-09-04 16:53:58 +07:00
|
|
|
#include <linux/linkage.h>
|
2007-11-09 15:08:54 +07:00
|
|
|
#include <asm/page.h>
|
|
|
|
#include <asm/types.h>
|
2010-01-05 17:06:45 +07:00
|
|
|
#include <asm/hw_breakpoint.h>
|
2007-11-09 15:08:54 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Default implementation of macro that returns current
|
|
|
|
* instruction pointer ("program counter").
|
|
|
|
*/
|
2008-07-28 20:32:03 +07:00
|
|
|
#define current_text_addr() ({ void *pc; __asm__("mova 1f, %0\n.align 2\n1:":"=z" (pc)); pc; })
|
2007-11-09 15:08:54 +07:00
|
|
|
|
|
|
|
/* Core Processor Version Register */
|
|
|
|
#define CCN_PVR 0xff000030
|
|
|
|
#define CCN_CVR 0xff000040
|
|
|
|
#define CCN_PRR 0xff000044
|
|
|
|
|
|
|
|
/*
|
|
|
|
* User space process size: 2GB.
|
|
|
|
*
|
|
|
|
* Since SH7709 and SH7750 have "area 7", we can't use 0x7c000000--0x7fffffff
|
|
|
|
*/
|
|
|
|
#define TASK_SIZE 0x7c000000UL
|
|
|
|
|
2008-02-08 19:19:26 +07:00
|
|
|
#define STACK_TOP TASK_SIZE
|
|
|
|
#define STACK_TOP_MAX STACK_TOP
|
|
|
|
|
2007-11-09 15:08:54 +07:00
|
|
|
/* This decides where the kernel will search for a free chunk of vm
|
|
|
|
* space during mmap's.
|
|
|
|
*/
|
|
|
|
#define TASK_UNMAPPED_BASE (TASK_SIZE / 3)
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Bit of SR register
|
|
|
|
*
|
|
|
|
* FD-bit:
|
|
|
|
* When it's set, it means the processor doesn't have right to use FPU,
|
|
|
|
* and it results exception when the floating operation is executed.
|
|
|
|
*
|
|
|
|
* IMASK-bit:
|
|
|
|
* Interrupt level mask
|
|
|
|
*/
|
|
|
|
#define SR_DSP 0x00001000
|
|
|
|
#define SR_IMASK 0x000000f0
|
2008-03-26 17:02:47 +07:00
|
|
|
#define SR_FD 0x00008000
|
sh: Minor optimisations to FPU handling
A number of small optimisations to FPU handling, in particular:
- move the task USEDFPU flag from the thread_info flags field (which
is accessed asynchronously to the thread) to a new status field,
which is only accessed by the thread itself. This allows locking to
be removed in most cases, or can be reduced to a preempt_lock().
This mimics the i386 behaviour.
- move the modification of regs->sr and thread_info->status flags out
of save_fpu() to __unlazy_fpu(). This gives the compiler a better
chance to optimise things, as well as making save_fpu() symmetrical
with restore_fpu() and init_fpu().
- implement prepare_to_copy(), so that when creating a thread, we can
unlazy the FPU prior to copying the thread data structures.
Also make sure that the FPU is disabled while in the kernel, in
particular while booting, and for newly created kernel threads,
In a very artificial benchmark, the execution time for 2500000
context switches was reduced from 50 to 45 seconds.
Signed-off-by: Stuart Menefy <stuart.menefy@st.com>
Signed-off-by: Paul Mundt <lethal@linux-sh.org>
2009-09-26 00:25:10 +07:00
|
|
|
#define SR_MD 0x40000000
|
2007-11-09 15:08:54 +07:00
|
|
|
|
2009-04-04 00:32:33 +07:00
|
|
|
/*
|
|
|
|
* DSP structure and data
|
|
|
|
*/
|
|
|
|
struct sh_dsp_struct {
|
|
|
|
unsigned long dsp_regs[14];
|
|
|
|
long status;
|
|
|
|
};
|
|
|
|
|
2007-11-09 15:08:54 +07:00
|
|
|
/*
|
|
|
|
* FPU structure and data
|
|
|
|
*/
|
|
|
|
|
|
|
|
struct sh_fpu_hard_struct {
|
|
|
|
unsigned long fp_regs[16];
|
|
|
|
unsigned long xfp_regs[16];
|
|
|
|
unsigned long fpscr;
|
|
|
|
unsigned long fpul;
|
|
|
|
|
|
|
|
long status; /* software status information */
|
|
|
|
};
|
|
|
|
|
|
|
|
/* Dummy fpu emulator */
|
|
|
|
struct sh_fpu_soft_struct {
|
|
|
|
unsigned long fp_regs[16];
|
|
|
|
unsigned long xfp_regs[16];
|
|
|
|
unsigned long fpscr;
|
|
|
|
unsigned long fpul;
|
|
|
|
|
|
|
|
unsigned char lookahead;
|
|
|
|
unsigned long entry_pc;
|
|
|
|
};
|
|
|
|
|
2010-01-13 10:51:40 +07:00
|
|
|
union thread_xstate {
|
|
|
|
struct sh_fpu_hard_struct hardfpu;
|
|
|
|
struct sh_fpu_soft_struct softfpu;
|
2007-11-09 15:08:54 +07:00
|
|
|
};
|
|
|
|
|
|
|
|
struct thread_struct {
|
|
|
|
/* Saved registers when thread is descheduled */
|
|
|
|
unsigned long sp;
|
|
|
|
unsigned long pc;
|
|
|
|
|
2010-02-23 10:56:30 +07:00
|
|
|
/* Various thread flags, see SH_THREAD_xxx */
|
|
|
|
unsigned long flags;
|
|
|
|
|
2009-11-09 14:27:40 +07:00
|
|
|
/* Save middle states of ptrace breakpoints */
|
2010-02-23 10:56:30 +07:00
|
|
|
struct perf_event *ptrace_bps[HBP_NUM];
|
2009-04-04 00:32:33 +07:00
|
|
|
|
|
|
|
#ifdef CONFIG_SH_DSP
|
|
|
|
/* Dsp status information */
|
|
|
|
struct sh_dsp_struct dsp_status;
|
|
|
|
#endif
|
2007-11-09 15:08:54 +07:00
|
|
|
|
2010-01-13 10:51:40 +07:00
|
|
|
/* Extended processor state */
|
|
|
|
union thread_xstate *xstate;
|
2007-11-09 15:08:54 +07:00
|
|
|
};
|
|
|
|
|
|
|
|
#define INIT_THREAD { \
|
|
|
|
.sp = sizeof(init_stack) + (long) &init_stack, \
|
2010-02-23 10:56:30 +07:00
|
|
|
.flags = 0, \
|
2007-11-09 15:08:54 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Forward declaration, a strange C thing */
|
|
|
|
struct task_struct;
|
2010-01-12 16:52:00 +07:00
|
|
|
|
|
|
|
extern void start_thread(struct pt_regs *regs, unsigned long new_pc, unsigned long new_sp);
|
2007-11-09 15:08:54 +07:00
|
|
|
|
|
|
|
/* Free all resources held by a thread. */
|
|
|
|
extern void release_thread(struct task_struct *);
|
|
|
|
|
|
|
|
/* Prepare to copy thread state - unlazy all lazy status */
|
sh: Minor optimisations to FPU handling
A number of small optimisations to FPU handling, in particular:
- move the task USEDFPU flag from the thread_info flags field (which
is accessed asynchronously to the thread) to a new status field,
which is only accessed by the thread itself. This allows locking to
be removed in most cases, or can be reduced to a preempt_lock().
This mimics the i386 behaviour.
- move the modification of regs->sr and thread_info->status flags out
of save_fpu() to __unlazy_fpu(). This gives the compiler a better
chance to optimise things, as well as making save_fpu() symmetrical
with restore_fpu() and init_fpu().
- implement prepare_to_copy(), so that when creating a thread, we can
unlazy the FPU prior to copying the thread data structures.
Also make sure that the FPU is disabled while in the kernel, in
particular while booting, and for newly created kernel threads,
In a very artificial benchmark, the execution time for 2500000
context switches was reduced from 50 to 45 seconds.
Signed-off-by: Stuart Menefy <stuart.menefy@st.com>
Signed-off-by: Paul Mundt <lethal@linux-sh.org>
2009-09-26 00:25:10 +07:00
|
|
|
void prepare_to_copy(struct task_struct *tsk);
|
2007-11-09 15:08:54 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* create a kernel thread without removing it from tasklists
|
|
|
|
*/
|
|
|
|
extern int kernel_thread(int (*fn)(void *), void * arg, unsigned long flags);
|
|
|
|
|
|
|
|
/* Copy and release all segment info associated with a VM */
|
|
|
|
#define copy_segments(p, mm) do { } while(0)
|
|
|
|
#define release_segments(mm) do { } while(0)
|
|
|
|
|
|
|
|
/*
|
|
|
|
* FPU lazy state save handling.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static __inline__ void disable_fpu(void)
|
|
|
|
{
|
|
|
|
unsigned long __dummy;
|
|
|
|
|
|
|
|
/* Set FD flag in SR */
|
|
|
|
__asm__ __volatile__("stc sr, %0\n\t"
|
|
|
|
"or %1, %0\n\t"
|
|
|
|
"ldc %0, sr"
|
|
|
|
: "=&r" (__dummy)
|
|
|
|
: "r" (SR_FD));
|
|
|
|
}
|
|
|
|
|
|
|
|
static __inline__ void enable_fpu(void)
|
|
|
|
{
|
|
|
|
unsigned long __dummy;
|
|
|
|
|
|
|
|
/* Clear out FD flag in SR */
|
|
|
|
__asm__ __volatile__("stc sr, %0\n\t"
|
|
|
|
"and %1, %0\n\t"
|
|
|
|
"ldc %0, sr"
|
|
|
|
: "=&r" (__dummy)
|
|
|
|
: "r" (~SR_FD));
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Double presision, NANS as NANS, rounding to nearest, no exceptions */
|
|
|
|
#define FPSCR_INIT 0x00080000
|
|
|
|
|
|
|
|
#define FPSCR_CAUSE_MASK 0x0001f000 /* Cause bits */
|
|
|
|
#define FPSCR_FLAG_MASK 0x0000007c /* Flag bits */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Return saved PC of a blocked thread.
|
|
|
|
*/
|
|
|
|
#define thread_saved_pc(tsk) (tsk->thread.pc)
|
|
|
|
|
|
|
|
void show_trace(struct task_struct *tsk, unsigned long *sp,
|
|
|
|
struct pt_regs *regs);
|
2008-12-17 13:56:06 +07:00
|
|
|
|
|
|
|
#ifdef CONFIG_DUMP_CODE
|
2008-11-26 12:31:03 +07:00
|
|
|
void show_code(struct pt_regs *regs);
|
2008-12-17 13:56:06 +07:00
|
|
|
#else
|
|
|
|
static inline void show_code(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2007-11-09 15:08:54 +07:00
|
|
|
extern unsigned long get_wchan(struct task_struct *p);
|
|
|
|
|
|
|
|
#define KSTK_EIP(tsk) (task_pt_regs(tsk)->pc)
|
|
|
|
#define KSTK_ESP(tsk) (task_pt_regs(tsk)->regs[15])
|
|
|
|
|
2009-02-27 14:42:05 +07:00
|
|
|
#if defined(CONFIG_CPU_SH2A) || defined(CONFIG_CPU_SH4)
|
2011-01-11 12:39:35 +07:00
|
|
|
|
2007-11-09 15:08:54 +07:00
|
|
|
#define PREFETCH_STRIDE L1_CACHE_BYTES
|
|
|
|
#define ARCH_HAS_PREFETCH
|
|
|
|
#define ARCH_HAS_PREFETCHW
|
2011-01-11 12:39:35 +07:00
|
|
|
|
|
|
|
static inline void prefetch(const void *x)
|
2007-11-09 15:08:54 +07:00
|
|
|
{
|
2010-11-17 13:50:17 +07:00
|
|
|
__builtin_prefetch(x, 0, 3);
|
2007-11-09 15:08:54 +07:00
|
|
|
}
|
|
|
|
|
2011-01-11 12:39:35 +07:00
|
|
|
static inline void prefetchw(const void *x)
|
2010-11-17 13:50:17 +07:00
|
|
|
{
|
|
|
|
__builtin_prefetch(x, 1, 3);
|
|
|
|
}
|
2007-11-09 15:08:54 +07:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#endif /* __KERNEL__ */
|
|
|
|
#endif /* __ASM_SH_PROCESSOR_32_H */
|