mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-19 08:26:53 +07:00
fff7fb0b2d
The binary GCD algorithm is based on the following facts: 1. If a and b are all evens, then gcd(a,b) = 2 * gcd(a/2, b/2) 2. If a is even and b is odd, then gcd(a,b) = gcd(a/2, b) 3. If a and b are all odds, then gcd(a,b) = gcd((a-b)/2, b) = gcd((a+b)/2, b) Even on x86 machines with reasonable division hardware, the binary algorithm runs about 25% faster (80% the execution time) than the division-based Euclidian algorithm. On platforms like Alpha and ARMv6 where division is a function call to emulation code, it's even more significant. There are two variants of the code here, depending on whether a fast __ffs (find least significant set bit) instruction is available. This allows the unpredictable branches in the bit-at-a-time shifting loop to be eliminated. If fast __ffs is not available, the "even/odd" GCD variant is used. I use the following code to benchmark: #include <stdio.h> #include <stdlib.h> #include <stdint.h> #include <string.h> #include <time.h> #include <unistd.h> #define swap(a, b) \ do { \ a ^= b; \ b ^= a; \ a ^= b; \ } while (0) unsigned long gcd0(unsigned long a, unsigned long b) { unsigned long r; if (a < b) { swap(a, b); } if (b == 0) return a; while ((r = a % b) != 0) { a = b; b = r; } return b; } unsigned long gcd1(unsigned long a, unsigned long b) { unsigned long r = a | b; if (!a || !b) return r; b >>= __builtin_ctzl(b); for (;;) { a >>= __builtin_ctzl(a); if (a == b) return a << __builtin_ctzl(r); if (a < b) swap(a, b); a -= b; } } unsigned long gcd2(unsigned long a, unsigned long b) { unsigned long r = a | b; if (!a || !b) return r; r &= -r; while (!(b & r)) b >>= 1; for (;;) { while (!(a & r)) a >>= 1; if (a == b) return a; if (a < b) swap(a, b); a -= b; a >>= 1; if (a & r) a += b; a >>= 1; } } unsigned long gcd3(unsigned long a, unsigned long b) { unsigned long r = a | b; if (!a || !b) return r; b >>= __builtin_ctzl(b); if (b == 1) return r & -r; for (;;) { a >>= __builtin_ctzl(a); if (a == 1) return r & -r; if (a == b) return a << __builtin_ctzl(r); if (a < b) swap(a, b); a -= b; } } unsigned long gcd4(unsigned long a, unsigned long b) { unsigned long r = a | b; if (!a || !b) return r; r &= -r; while (!(b & r)) b >>= 1; if (b == r) return r; for (;;) { while (!(a & r)) a >>= 1; if (a == r) return r; if (a == b) return a; if (a < b) swap(a, b); a -= b; a >>= 1; if (a & r) a += b; a >>= 1; } } static unsigned long (*gcd_func[])(unsigned long a, unsigned long b) = { gcd0, gcd1, gcd2, gcd3, gcd4, }; #define TEST_ENTRIES (sizeof(gcd_func) / sizeof(gcd_func[0])) #if defined(__x86_64__) #define rdtscll(val) do { \ unsigned long __a,__d; \ __asm__ __volatile__("rdtsc" : "=a" (__a), "=d" (__d)); \ (val) = ((unsigned long long)__a) | (((unsigned long long)__d)<<32); \ } while(0) static unsigned long long benchmark_gcd_func(unsigned long (*gcd)(unsigned long, unsigned long), unsigned long a, unsigned long b, unsigned long *res) { unsigned long long start, end; unsigned long long ret; unsigned long gcd_res; rdtscll(start); gcd_res = gcd(a, b); rdtscll(end); if (end >= start) ret = end - start; else ret = ~0ULL - start + 1 + end; *res = gcd_res; return ret; } #else static inline struct timespec read_time(void) { struct timespec time; clock_gettime(CLOCK_PROCESS_CPUTIME_ID, &time); return time; } static inline unsigned long long diff_time(struct timespec start, struct timespec end) { struct timespec temp; if ((end.tv_nsec - start.tv_nsec) < 0) { temp.tv_sec = end.tv_sec - start.tv_sec - 1; temp.tv_nsec = 1000000000ULL + end.tv_nsec - start.tv_nsec; } else { temp.tv_sec = end.tv_sec - start.tv_sec; temp.tv_nsec = end.tv_nsec - start.tv_nsec; } return temp.tv_sec * 1000000000ULL + temp.tv_nsec; } static unsigned long long benchmark_gcd_func(unsigned long (*gcd)(unsigned long, unsigned long), unsigned long a, unsigned long b, unsigned long *res) { struct timespec start, end; unsigned long gcd_res; start = read_time(); gcd_res = gcd(a, b); end = read_time(); *res = gcd_res; return diff_time(start, end); } #endif static inline unsigned long get_rand() { if (sizeof(long) == 8) return (unsigned long)rand() << 32 | rand(); else return rand(); } int main(int argc, char **argv) { unsigned int seed = time(0); int loops = 100; int repeats = 1000; unsigned long (*res)[TEST_ENTRIES]; unsigned long long elapsed[TEST_ENTRIES]; int i, j, k; for (;;) { int opt = getopt(argc, argv, "n:r:s:"); /* End condition always first */ if (opt == -1) break; switch (opt) { case 'n': loops = atoi(optarg); break; case 'r': repeats = atoi(optarg); break; case 's': seed = strtoul(optarg, NULL, 10); break; default: /* You won't actually get here. */ break; } } res = malloc(sizeof(unsigned long) * TEST_ENTRIES * loops); memset(elapsed, 0, sizeof(elapsed)); srand(seed); for (j = 0; j < loops; j++) { unsigned long a = get_rand(); /* Do we have args? */ unsigned long b = argc > optind ? strtoul(argv[optind], NULL, 10) : get_rand(); unsigned long long min_elapsed[TEST_ENTRIES]; for (k = 0; k < repeats; k++) { for (i = 0; i < TEST_ENTRIES; i++) { unsigned long long tmp = benchmark_gcd_func(gcd_func[i], a, b, &res[j][i]); if (k == 0 || min_elapsed[i] > tmp) min_elapsed[i] = tmp; } } for (i = 0; i < TEST_ENTRIES; i++) elapsed[i] += min_elapsed[i]; } for (i = 0; i < TEST_ENTRIES; i++) printf("gcd%d: elapsed %llu\n", i, elapsed[i]); k = 0; srand(seed); for (j = 0; j < loops; j++) { unsigned long a = get_rand(); unsigned long b = argc > optind ? strtoul(argv[optind], NULL, 10) : get_rand(); for (i = 1; i < TEST_ENTRIES; i++) { if (res[j][i] != res[j][0]) break; } if (i < TEST_ENTRIES) { if (k == 0) { k = 1; fprintf(stderr, "Error:\n"); } fprintf(stderr, "gcd(%lu, %lu): ", a, b); for (i = 0; i < TEST_ENTRIES; i++) fprintf(stderr, "%ld%s", res[j][i], i < TEST_ENTRIES - 1 ? ", " : "\n"); } } if (k == 0) fprintf(stderr, "PASS\n"); free(res); return 0; } Compiled with "-O2", on "VirtualBox 4.4.0-22-generic #38-Ubuntu x86_64" got: zhaoxiuzeng@zhaoxiuzeng-VirtualBox:~/develop$ ./gcd -r 500000 -n 10 gcd0: elapsed 10174 gcd1: elapsed 2120 gcd2: elapsed 2902 gcd3: elapsed 2039 gcd4: elapsed 2812 PASS zhaoxiuzeng@zhaoxiuzeng-VirtualBox:~/develop$ ./gcd -r 500000 -n 10 gcd0: elapsed 9309 gcd1: elapsed 2280 gcd2: elapsed 2822 gcd3: elapsed 2217 gcd4: elapsed 2710 PASS zhaoxiuzeng@zhaoxiuzeng-VirtualBox:~/develop$ ./gcd -r 500000 -n 10 gcd0: elapsed 9589 gcd1: elapsed 2098 gcd2: elapsed 2815 gcd3: elapsed 2030 gcd4: elapsed 2718 PASS zhaoxiuzeng@zhaoxiuzeng-VirtualBox:~/develop$ ./gcd -r 500000 -n 10 gcd0: elapsed 9914 gcd1: elapsed 2309 gcd2: elapsed 2779 gcd3: elapsed 2228 gcd4: elapsed 2709 PASS [akpm@linux-foundation.org: avoid #defining a CONFIG_ variable] Signed-off-by: Zhaoxiu Zeng <zhaoxiu.zeng@gmail.com> Signed-off-by: George Spelvin <linux@horizon.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
570 lines
17 KiB
C
570 lines
17 KiB
C
/*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*
|
|
* Copyright (C) 2003, 2004 Ralf Baechle
|
|
* Copyright (C) 2004 Maciej W. Rozycki
|
|
*/
|
|
#ifndef __ASM_CPU_FEATURES_H
|
|
#define __ASM_CPU_FEATURES_H
|
|
|
|
#include <asm/cpu.h>
|
|
#include <asm/cpu-info.h>
|
|
#include <cpu-feature-overrides.h>
|
|
|
|
/*
|
|
* SMP assumption: Options of CPU 0 are a superset of all processors.
|
|
* This is true for all known MIPS systems.
|
|
*/
|
|
#ifndef cpu_has_tlb
|
|
#define cpu_has_tlb (cpu_data[0].options & MIPS_CPU_TLB)
|
|
#endif
|
|
#ifndef cpu_has_ftlb
|
|
#define cpu_has_ftlb (cpu_data[0].options & MIPS_CPU_FTLB)
|
|
#endif
|
|
#ifndef cpu_has_tlbinv
|
|
#define cpu_has_tlbinv (cpu_data[0].options & MIPS_CPU_TLBINV)
|
|
#endif
|
|
#ifndef cpu_has_segments
|
|
#define cpu_has_segments (cpu_data[0].options & MIPS_CPU_SEGMENTS)
|
|
#endif
|
|
#ifndef cpu_has_eva
|
|
#define cpu_has_eva (cpu_data[0].options & MIPS_CPU_EVA)
|
|
#endif
|
|
#ifndef cpu_has_htw
|
|
#define cpu_has_htw (cpu_data[0].options & MIPS_CPU_HTW)
|
|
#endif
|
|
#ifndef cpu_has_ldpte
|
|
#define cpu_has_ldpte (cpu_data[0].options & MIPS_CPU_LDPTE)
|
|
#endif
|
|
#ifndef cpu_has_rixiex
|
|
#define cpu_has_rixiex (cpu_data[0].options & MIPS_CPU_RIXIEX)
|
|
#endif
|
|
#ifndef cpu_has_maar
|
|
#define cpu_has_maar (cpu_data[0].options & MIPS_CPU_MAAR)
|
|
#endif
|
|
#ifndef cpu_has_rw_llb
|
|
#define cpu_has_rw_llb (cpu_data[0].options & MIPS_CPU_RW_LLB)
|
|
#endif
|
|
|
|
/*
|
|
* For the moment we don't consider R6000 and R8000 so we can assume that
|
|
* anything that doesn't support R4000-style exceptions and interrupts is
|
|
* R3000-like. Users should still treat these two macro definitions as
|
|
* opaque.
|
|
*/
|
|
#ifndef cpu_has_3kex
|
|
#define cpu_has_3kex (!cpu_has_4kex)
|
|
#endif
|
|
#ifndef cpu_has_4kex
|
|
#define cpu_has_4kex (cpu_data[0].options & MIPS_CPU_4KEX)
|
|
#endif
|
|
#ifndef cpu_has_3k_cache
|
|
#define cpu_has_3k_cache (cpu_data[0].options & MIPS_CPU_3K_CACHE)
|
|
#endif
|
|
#define cpu_has_6k_cache 0
|
|
#define cpu_has_8k_cache 0
|
|
#ifndef cpu_has_4k_cache
|
|
#define cpu_has_4k_cache (cpu_data[0].options & MIPS_CPU_4K_CACHE)
|
|
#endif
|
|
#ifndef cpu_has_tx39_cache
|
|
#define cpu_has_tx39_cache (cpu_data[0].options & MIPS_CPU_TX39_CACHE)
|
|
#endif
|
|
#ifndef cpu_has_octeon_cache
|
|
#define cpu_has_octeon_cache 0
|
|
#endif
|
|
/* Don't override `cpu_has_fpu' to 1 or the "nofpu" option won't work. */
|
|
#ifndef cpu_has_fpu
|
|
#define cpu_has_fpu (current_cpu_data.options & MIPS_CPU_FPU)
|
|
#define raw_cpu_has_fpu (raw_current_cpu_data.options & MIPS_CPU_FPU)
|
|
#else
|
|
#define raw_cpu_has_fpu cpu_has_fpu
|
|
#endif
|
|
#ifndef cpu_has_32fpr
|
|
#define cpu_has_32fpr (cpu_data[0].options & MIPS_CPU_32FPR)
|
|
#endif
|
|
#ifndef cpu_has_counter
|
|
#define cpu_has_counter (cpu_data[0].options & MIPS_CPU_COUNTER)
|
|
#endif
|
|
#ifndef cpu_has_watch
|
|
#define cpu_has_watch (cpu_data[0].options & MIPS_CPU_WATCH)
|
|
#endif
|
|
#ifndef cpu_has_divec
|
|
#define cpu_has_divec (cpu_data[0].options & MIPS_CPU_DIVEC)
|
|
#endif
|
|
#ifndef cpu_has_vce
|
|
#define cpu_has_vce (cpu_data[0].options & MIPS_CPU_VCE)
|
|
#endif
|
|
#ifndef cpu_has_cache_cdex_p
|
|
#define cpu_has_cache_cdex_p (cpu_data[0].options & MIPS_CPU_CACHE_CDEX_P)
|
|
#endif
|
|
#ifndef cpu_has_cache_cdex_s
|
|
#define cpu_has_cache_cdex_s (cpu_data[0].options & MIPS_CPU_CACHE_CDEX_S)
|
|
#endif
|
|
#ifndef cpu_has_prefetch
|
|
#define cpu_has_prefetch (cpu_data[0].options & MIPS_CPU_PREFETCH)
|
|
#endif
|
|
#ifndef cpu_has_mcheck
|
|
#define cpu_has_mcheck (cpu_data[0].options & MIPS_CPU_MCHECK)
|
|
#endif
|
|
#ifndef cpu_has_ejtag
|
|
#define cpu_has_ejtag (cpu_data[0].options & MIPS_CPU_EJTAG)
|
|
#endif
|
|
#ifndef cpu_has_llsc
|
|
#define cpu_has_llsc (cpu_data[0].options & MIPS_CPU_LLSC)
|
|
#endif
|
|
#ifndef cpu_has_bp_ghist
|
|
#define cpu_has_bp_ghist (cpu_data[0].options & MIPS_CPU_BP_GHIST)
|
|
#endif
|
|
#ifndef kernel_uses_llsc
|
|
#define kernel_uses_llsc cpu_has_llsc
|
|
#endif
|
|
#ifndef cpu_has_guestctl0ext
|
|
#define cpu_has_guestctl0ext (cpu_data[0].options & MIPS_CPU_GUESTCTL0EXT)
|
|
#endif
|
|
#ifndef cpu_has_guestctl1
|
|
#define cpu_has_guestctl1 (cpu_data[0].options & MIPS_CPU_GUESTCTL1)
|
|
#endif
|
|
#ifndef cpu_has_guestctl2
|
|
#define cpu_has_guestctl2 (cpu_data[0].options & MIPS_CPU_GUESTCTL2)
|
|
#endif
|
|
#ifndef cpu_has_guestid
|
|
#define cpu_has_guestid (cpu_data[0].options & MIPS_CPU_GUESTID)
|
|
#endif
|
|
#ifndef cpu_has_drg
|
|
#define cpu_has_drg (cpu_data[0].options & MIPS_CPU_DRG)
|
|
#endif
|
|
#ifndef cpu_has_mips16
|
|
#define cpu_has_mips16 (cpu_data[0].ases & MIPS_ASE_MIPS16)
|
|
#endif
|
|
#ifndef cpu_has_mdmx
|
|
#define cpu_has_mdmx (cpu_data[0].ases & MIPS_ASE_MDMX)
|
|
#endif
|
|
#ifndef cpu_has_mips3d
|
|
#define cpu_has_mips3d (cpu_data[0].ases & MIPS_ASE_MIPS3D)
|
|
#endif
|
|
#ifndef cpu_has_smartmips
|
|
#define cpu_has_smartmips (cpu_data[0].ases & MIPS_ASE_SMARTMIPS)
|
|
#endif
|
|
|
|
#ifndef cpu_has_rixi
|
|
#define cpu_has_rixi (cpu_data[0].options & MIPS_CPU_RIXI)
|
|
#endif
|
|
|
|
#ifndef cpu_has_mmips
|
|
# ifdef CONFIG_SYS_SUPPORTS_MICROMIPS
|
|
# define cpu_has_mmips (cpu_data[0].options & MIPS_CPU_MICROMIPS)
|
|
# else
|
|
# define cpu_has_mmips 0
|
|
# endif
|
|
#endif
|
|
|
|
#ifndef cpu_has_lpa
|
|
#define cpu_has_lpa (cpu_data[0].options & MIPS_CPU_LPA)
|
|
#endif
|
|
#ifndef cpu_has_mvh
|
|
#define cpu_has_mvh (cpu_data[0].options & MIPS_CPU_MVH)
|
|
#endif
|
|
#ifndef cpu_has_xpa
|
|
#define cpu_has_xpa (cpu_has_lpa && cpu_has_mvh)
|
|
#endif
|
|
#ifndef cpu_has_vtag_icache
|
|
#define cpu_has_vtag_icache (cpu_data[0].icache.flags & MIPS_CACHE_VTAG)
|
|
#endif
|
|
#ifndef cpu_has_dc_aliases
|
|
#define cpu_has_dc_aliases (cpu_data[0].dcache.flags & MIPS_CACHE_ALIASES)
|
|
#endif
|
|
#ifndef cpu_has_ic_fills_f_dc
|
|
#define cpu_has_ic_fills_f_dc (cpu_data[0].icache.flags & MIPS_CACHE_IC_F_DC)
|
|
#endif
|
|
#ifndef cpu_has_pindexed_dcache
|
|
#define cpu_has_pindexed_dcache (cpu_data[0].dcache.flags & MIPS_CACHE_PINDEX)
|
|
#endif
|
|
#ifndef cpu_has_local_ebase
|
|
#define cpu_has_local_ebase 1
|
|
#endif
|
|
|
|
/*
|
|
* I-Cache snoops remote store. This only matters on SMP. Some multiprocessors
|
|
* such as the R10000 have I-Caches that snoop local stores; the embedded ones
|
|
* don't. For maintaining I-cache coherency this means we need to flush the
|
|
* D-cache all the way back to whever the I-cache does refills from, so the
|
|
* I-cache has a chance to see the new data at all. Then we have to flush the
|
|
* I-cache also.
|
|
* Note we may have been rescheduled and may no longer be running on the CPU
|
|
* that did the store so we can't optimize this into only doing the flush on
|
|
* the local CPU.
|
|
*/
|
|
#ifndef cpu_icache_snoops_remote_store
|
|
#ifdef CONFIG_SMP
|
|
#define cpu_icache_snoops_remote_store (cpu_data[0].icache.flags & MIPS_IC_SNOOPS_REMOTE)
|
|
#else
|
|
#define cpu_icache_snoops_remote_store 1
|
|
#endif
|
|
#endif
|
|
|
|
/* __builtin_constant_p(cpu_has_mips_r) && cpu_has_mips_r */
|
|
#if !((defined(cpu_has_mips32r1) && cpu_has_mips32r1) || \
|
|
(defined(cpu_has_mips32r2) && cpu_has_mips32r2) || \
|
|
(defined(cpu_has_mips32r6) && cpu_has_mips32r6) || \
|
|
(defined(cpu_has_mips64r1) && cpu_has_mips64r1) || \
|
|
(defined(cpu_has_mips64r2) && cpu_has_mips64r2) || \
|
|
(defined(cpu_has_mips64r6) && cpu_has_mips64r6))
|
|
#define CPU_NO_EFFICIENT_FFS 1
|
|
#endif
|
|
|
|
#ifndef cpu_has_mips_1
|
|
# define cpu_has_mips_1 (!cpu_has_mips_r6)
|
|
#endif
|
|
#ifndef cpu_has_mips_2
|
|
# define cpu_has_mips_2 (cpu_data[0].isa_level & MIPS_CPU_ISA_II)
|
|
#endif
|
|
#ifndef cpu_has_mips_3
|
|
# define cpu_has_mips_3 (cpu_data[0].isa_level & MIPS_CPU_ISA_III)
|
|
#endif
|
|
#ifndef cpu_has_mips_4
|
|
# define cpu_has_mips_4 (cpu_data[0].isa_level & MIPS_CPU_ISA_IV)
|
|
#endif
|
|
#ifndef cpu_has_mips_5
|
|
# define cpu_has_mips_5 (cpu_data[0].isa_level & MIPS_CPU_ISA_V)
|
|
#endif
|
|
#ifndef cpu_has_mips32r1
|
|
# define cpu_has_mips32r1 (cpu_data[0].isa_level & MIPS_CPU_ISA_M32R1)
|
|
#endif
|
|
#ifndef cpu_has_mips32r2
|
|
# define cpu_has_mips32r2 (cpu_data[0].isa_level & MIPS_CPU_ISA_M32R2)
|
|
#endif
|
|
#ifndef cpu_has_mips32r6
|
|
# define cpu_has_mips32r6 (cpu_data[0].isa_level & MIPS_CPU_ISA_M32R6)
|
|
#endif
|
|
#ifndef cpu_has_mips64r1
|
|
# define cpu_has_mips64r1 (cpu_data[0].isa_level & MIPS_CPU_ISA_M64R1)
|
|
#endif
|
|
#ifndef cpu_has_mips64r2
|
|
# define cpu_has_mips64r2 (cpu_data[0].isa_level & MIPS_CPU_ISA_M64R2)
|
|
#endif
|
|
#ifndef cpu_has_mips64r6
|
|
# define cpu_has_mips64r6 (cpu_data[0].isa_level & MIPS_CPU_ISA_M64R6)
|
|
#endif
|
|
|
|
/*
|
|
* Shortcuts ...
|
|
*/
|
|
#define cpu_has_mips_2_3_4_5 (cpu_has_mips_2 | cpu_has_mips_3_4_5)
|
|
#define cpu_has_mips_3_4_5 (cpu_has_mips_3 | cpu_has_mips_4_5)
|
|
#define cpu_has_mips_4_5 (cpu_has_mips_4 | cpu_has_mips_5)
|
|
|
|
#define cpu_has_mips_2_3_4_5_r (cpu_has_mips_2 | cpu_has_mips_3_4_5_r)
|
|
#define cpu_has_mips_3_4_5_r (cpu_has_mips_3 | cpu_has_mips_4_5_r)
|
|
#define cpu_has_mips_4_5_r (cpu_has_mips_4 | cpu_has_mips_5_r)
|
|
#define cpu_has_mips_5_r (cpu_has_mips_5 | cpu_has_mips_r)
|
|
|
|
#define cpu_has_mips_3_4_5_64_r2_r6 \
|
|
(cpu_has_mips_3 | cpu_has_mips_4_5_64_r2_r6)
|
|
#define cpu_has_mips_4_5_64_r2_r6 \
|
|
(cpu_has_mips_4_5 | cpu_has_mips64r1 | \
|
|
cpu_has_mips_r2 | cpu_has_mips_r6)
|
|
|
|
#define cpu_has_mips32 (cpu_has_mips32r1 | cpu_has_mips32r2 | cpu_has_mips32r6)
|
|
#define cpu_has_mips64 (cpu_has_mips64r1 | cpu_has_mips64r2 | cpu_has_mips64r6)
|
|
#define cpu_has_mips_r1 (cpu_has_mips32r1 | cpu_has_mips64r1)
|
|
#define cpu_has_mips_r2 (cpu_has_mips32r2 | cpu_has_mips64r2)
|
|
#define cpu_has_mips_r6 (cpu_has_mips32r6 | cpu_has_mips64r6)
|
|
#define cpu_has_mips_r (cpu_has_mips32r1 | cpu_has_mips32r2 | \
|
|
cpu_has_mips32r6 | cpu_has_mips64r1 | \
|
|
cpu_has_mips64r2 | cpu_has_mips64r6)
|
|
|
|
/* MIPSR2 and MIPSR6 have a lot of similarities */
|
|
#define cpu_has_mips_r2_r6 (cpu_has_mips_r2 | cpu_has_mips_r6)
|
|
|
|
/*
|
|
* cpu_has_mips_r2_exec_hazard - return if IHB is required on current processor
|
|
*
|
|
* Returns non-zero value if the current processor implementation requires
|
|
* an IHB instruction to deal with an instruction hazard as per MIPS R2
|
|
* architecture specification, zero otherwise.
|
|
*/
|
|
#ifndef cpu_has_mips_r2_exec_hazard
|
|
#define cpu_has_mips_r2_exec_hazard \
|
|
({ \
|
|
int __res; \
|
|
\
|
|
switch (current_cpu_type()) { \
|
|
case CPU_M14KC: \
|
|
case CPU_74K: \
|
|
case CPU_1074K: \
|
|
case CPU_PROAPTIV: \
|
|
case CPU_P5600: \
|
|
case CPU_M5150: \
|
|
case CPU_QEMU_GENERIC: \
|
|
case CPU_CAVIUM_OCTEON: \
|
|
case CPU_CAVIUM_OCTEON_PLUS: \
|
|
case CPU_CAVIUM_OCTEON2: \
|
|
case CPU_CAVIUM_OCTEON3: \
|
|
__res = 0; \
|
|
break; \
|
|
\
|
|
default: \
|
|
__res = 1; \
|
|
} \
|
|
\
|
|
__res; \
|
|
})
|
|
#endif
|
|
|
|
/*
|
|
* MIPS32, MIPS64, VR5500, IDT32332, IDT32334 and maybe a few other
|
|
* pre-MIPS32/MIPS64 processors have CLO, CLZ. The IDT RC64574 is 64-bit and
|
|
* has CLO and CLZ but not DCLO nor DCLZ. For 64-bit kernels
|
|
* cpu_has_clo_clz also indicates the availability of DCLO and DCLZ.
|
|
*/
|
|
#ifndef cpu_has_clo_clz
|
|
#define cpu_has_clo_clz cpu_has_mips_r
|
|
#endif
|
|
|
|
/*
|
|
* MIPS32 R2, MIPS64 R2, Loongson 3A and Octeon have WSBH.
|
|
* MIPS64 R2, Loongson 3A and Octeon have WSBH, DSBH and DSHD.
|
|
* This indicates the availability of WSBH and in case of 64 bit CPUs also
|
|
* DSBH and DSHD.
|
|
*/
|
|
#ifndef cpu_has_wsbh
|
|
#define cpu_has_wsbh cpu_has_mips_r2
|
|
#endif
|
|
|
|
#ifndef cpu_has_dsp
|
|
#define cpu_has_dsp (cpu_data[0].ases & MIPS_ASE_DSP)
|
|
#endif
|
|
|
|
#ifndef cpu_has_dsp2
|
|
#define cpu_has_dsp2 (cpu_data[0].ases & MIPS_ASE_DSP2P)
|
|
#endif
|
|
|
|
#ifndef cpu_has_dsp3
|
|
#define cpu_has_dsp3 (cpu_data[0].ases & MIPS_ASE_DSP3)
|
|
#endif
|
|
|
|
#ifndef cpu_has_mipsmt
|
|
#define cpu_has_mipsmt (cpu_data[0].ases & MIPS_ASE_MIPSMT)
|
|
#endif
|
|
|
|
#ifndef cpu_has_vp
|
|
#define cpu_has_vp (cpu_data[0].options & MIPS_CPU_VP)
|
|
#endif
|
|
|
|
#ifndef cpu_has_userlocal
|
|
#define cpu_has_userlocal (cpu_data[0].options & MIPS_CPU_ULRI)
|
|
#endif
|
|
|
|
#ifdef CONFIG_32BIT
|
|
# ifndef cpu_has_nofpuex
|
|
# define cpu_has_nofpuex (cpu_data[0].options & MIPS_CPU_NOFPUEX)
|
|
# endif
|
|
# ifndef cpu_has_64bits
|
|
# define cpu_has_64bits (cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT)
|
|
# endif
|
|
# ifndef cpu_has_64bit_zero_reg
|
|
# define cpu_has_64bit_zero_reg (cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT)
|
|
# endif
|
|
# ifndef cpu_has_64bit_gp_regs
|
|
# define cpu_has_64bit_gp_regs 0
|
|
# endif
|
|
# ifndef cpu_has_64bit_addresses
|
|
# define cpu_has_64bit_addresses 0
|
|
# endif
|
|
# ifndef cpu_vmbits
|
|
# define cpu_vmbits 31
|
|
# endif
|
|
#endif
|
|
|
|
#ifdef CONFIG_64BIT
|
|
# ifndef cpu_has_nofpuex
|
|
# define cpu_has_nofpuex 0
|
|
# endif
|
|
# ifndef cpu_has_64bits
|
|
# define cpu_has_64bits 1
|
|
# endif
|
|
# ifndef cpu_has_64bit_zero_reg
|
|
# define cpu_has_64bit_zero_reg 1
|
|
# endif
|
|
# ifndef cpu_has_64bit_gp_regs
|
|
# define cpu_has_64bit_gp_regs 1
|
|
# endif
|
|
# ifndef cpu_has_64bit_addresses
|
|
# define cpu_has_64bit_addresses 1
|
|
# endif
|
|
# ifndef cpu_vmbits
|
|
# define cpu_vmbits cpu_data[0].vmbits
|
|
# define __NEED_VMBITS_PROBE
|
|
# endif
|
|
#endif
|
|
|
|
#if defined(CONFIG_CPU_MIPSR2_IRQ_VI) && !defined(cpu_has_vint)
|
|
# define cpu_has_vint (cpu_data[0].options & MIPS_CPU_VINT)
|
|
#elif !defined(cpu_has_vint)
|
|
# define cpu_has_vint 0
|
|
#endif
|
|
|
|
#if defined(CONFIG_CPU_MIPSR2_IRQ_EI) && !defined(cpu_has_veic)
|
|
# define cpu_has_veic (cpu_data[0].options & MIPS_CPU_VEIC)
|
|
#elif !defined(cpu_has_veic)
|
|
# define cpu_has_veic 0
|
|
#endif
|
|
|
|
#ifndef cpu_has_inclusive_pcaches
|
|
#define cpu_has_inclusive_pcaches (cpu_data[0].options & MIPS_CPU_INCLUSIVE_CACHES)
|
|
#endif
|
|
|
|
#ifndef cpu_dcache_line_size
|
|
#define cpu_dcache_line_size() cpu_data[0].dcache.linesz
|
|
#endif
|
|
#ifndef cpu_icache_line_size
|
|
#define cpu_icache_line_size() cpu_data[0].icache.linesz
|
|
#endif
|
|
#ifndef cpu_scache_line_size
|
|
#define cpu_scache_line_size() cpu_data[0].scache.linesz
|
|
#endif
|
|
|
|
#ifndef cpu_hwrena_impl_bits
|
|
#define cpu_hwrena_impl_bits 0
|
|
#endif
|
|
|
|
#ifndef cpu_has_perf_cntr_intr_bit
|
|
#define cpu_has_perf_cntr_intr_bit (cpu_data[0].options & MIPS_CPU_PCI)
|
|
#endif
|
|
|
|
#ifndef cpu_has_vz
|
|
#define cpu_has_vz (cpu_data[0].ases & MIPS_ASE_VZ)
|
|
#endif
|
|
|
|
#if defined(CONFIG_CPU_HAS_MSA) && !defined(cpu_has_msa)
|
|
# define cpu_has_msa (cpu_data[0].ases & MIPS_ASE_MSA)
|
|
#elif !defined(cpu_has_msa)
|
|
# define cpu_has_msa 0
|
|
#endif
|
|
|
|
#ifndef cpu_has_fre
|
|
# define cpu_has_fre (cpu_data[0].options & MIPS_CPU_FRE)
|
|
#endif
|
|
|
|
#ifndef cpu_has_cdmm
|
|
# define cpu_has_cdmm (cpu_data[0].options & MIPS_CPU_CDMM)
|
|
#endif
|
|
|
|
#ifndef cpu_has_small_pages
|
|
# define cpu_has_small_pages (cpu_data[0].options & MIPS_CPU_SP)
|
|
#endif
|
|
|
|
#ifndef cpu_has_nan_legacy
|
|
#define cpu_has_nan_legacy (cpu_data[0].options & MIPS_CPU_NAN_LEGACY)
|
|
#endif
|
|
#ifndef cpu_has_nan_2008
|
|
#define cpu_has_nan_2008 (cpu_data[0].options & MIPS_CPU_NAN_2008)
|
|
#endif
|
|
|
|
#ifndef cpu_has_ebase_wg
|
|
# define cpu_has_ebase_wg (cpu_data[0].options & MIPS_CPU_EBASE_WG)
|
|
#endif
|
|
|
|
#ifndef cpu_has_badinstr
|
|
# define cpu_has_badinstr (cpu_data[0].options & MIPS_CPU_BADINSTR)
|
|
#endif
|
|
|
|
#ifndef cpu_has_badinstrp
|
|
# define cpu_has_badinstrp (cpu_data[0].options & MIPS_CPU_BADINSTRP)
|
|
#endif
|
|
|
|
#ifndef cpu_has_contextconfig
|
|
# define cpu_has_contextconfig (cpu_data[0].options & MIPS_CPU_CTXTC)
|
|
#endif
|
|
|
|
#ifndef cpu_has_perf
|
|
# define cpu_has_perf (cpu_data[0].options & MIPS_CPU_PERF)
|
|
#endif
|
|
|
|
/*
|
|
* Guest capabilities
|
|
*/
|
|
#ifndef cpu_guest_has_conf1
|
|
#define cpu_guest_has_conf1 (cpu_data[0].guest.conf & (1 << 1))
|
|
#endif
|
|
#ifndef cpu_guest_has_conf2
|
|
#define cpu_guest_has_conf2 (cpu_data[0].guest.conf & (1 << 2))
|
|
#endif
|
|
#ifndef cpu_guest_has_conf3
|
|
#define cpu_guest_has_conf3 (cpu_data[0].guest.conf & (1 << 3))
|
|
#endif
|
|
#ifndef cpu_guest_has_conf4
|
|
#define cpu_guest_has_conf4 (cpu_data[0].guest.conf & (1 << 4))
|
|
#endif
|
|
#ifndef cpu_guest_has_conf5
|
|
#define cpu_guest_has_conf5 (cpu_data[0].guest.conf & (1 << 5))
|
|
#endif
|
|
#ifndef cpu_guest_has_conf6
|
|
#define cpu_guest_has_conf6 (cpu_data[0].guest.conf & (1 << 6))
|
|
#endif
|
|
#ifndef cpu_guest_has_conf7
|
|
#define cpu_guest_has_conf7 (cpu_data[0].guest.conf & (1 << 7))
|
|
#endif
|
|
#ifndef cpu_guest_has_fpu
|
|
#define cpu_guest_has_fpu (cpu_data[0].guest.options & MIPS_CPU_FPU)
|
|
#endif
|
|
#ifndef cpu_guest_has_watch
|
|
#define cpu_guest_has_watch (cpu_data[0].guest.options & MIPS_CPU_WATCH)
|
|
#endif
|
|
#ifndef cpu_guest_has_contextconfig
|
|
#define cpu_guest_has_contextconfig (cpu_data[0].guest.options & MIPS_CPU_CTXTC)
|
|
#endif
|
|
#ifndef cpu_guest_has_segments
|
|
#define cpu_guest_has_segments (cpu_data[0].guest.options & MIPS_CPU_SEGMENTS)
|
|
#endif
|
|
#ifndef cpu_guest_has_badinstr
|
|
#define cpu_guest_has_badinstr (cpu_data[0].guest.options & MIPS_CPU_BADINSTR)
|
|
#endif
|
|
#ifndef cpu_guest_has_badinstrp
|
|
#define cpu_guest_has_badinstrp (cpu_data[0].guest.options & MIPS_CPU_BADINSTRP)
|
|
#endif
|
|
#ifndef cpu_guest_has_htw
|
|
#define cpu_guest_has_htw (cpu_data[0].guest.options & MIPS_CPU_HTW)
|
|
#endif
|
|
#ifndef cpu_guest_has_msa
|
|
#define cpu_guest_has_msa (cpu_data[0].guest.ases & MIPS_ASE_MSA)
|
|
#endif
|
|
#ifndef cpu_guest_has_kscr
|
|
#define cpu_guest_has_kscr(n) (cpu_data[0].guest.kscratch_mask & (1u << (n)))
|
|
#endif
|
|
#ifndef cpu_guest_has_rw_llb
|
|
#define cpu_guest_has_rw_llb (cpu_has_mips_r6 || (cpu_data[0].guest.options & MIPS_CPU_RW_LLB))
|
|
#endif
|
|
#ifndef cpu_guest_has_perf
|
|
#define cpu_guest_has_perf (cpu_data[0].guest.options & MIPS_CPU_PERF)
|
|
#endif
|
|
#ifndef cpu_guest_has_maar
|
|
#define cpu_guest_has_maar (cpu_data[0].guest.options & MIPS_CPU_MAAR)
|
|
#endif
|
|
|
|
/*
|
|
* Guest dynamic capabilities
|
|
*/
|
|
#ifndef cpu_guest_has_dyn_fpu
|
|
#define cpu_guest_has_dyn_fpu (cpu_data[0].guest.options_dyn & MIPS_CPU_FPU)
|
|
#endif
|
|
#ifndef cpu_guest_has_dyn_watch
|
|
#define cpu_guest_has_dyn_watch (cpu_data[0].guest.options_dyn & MIPS_CPU_WATCH)
|
|
#endif
|
|
#ifndef cpu_guest_has_dyn_contextconfig
|
|
#define cpu_guest_has_dyn_contextconfig (cpu_data[0].guest.options_dyn & MIPS_CPU_CTXTC)
|
|
#endif
|
|
#ifndef cpu_guest_has_dyn_perf
|
|
#define cpu_guest_has_dyn_perf (cpu_data[0].guest.options_dyn & MIPS_CPU_PERF)
|
|
#endif
|
|
#ifndef cpu_guest_has_dyn_msa
|
|
#define cpu_guest_has_dyn_msa (cpu_data[0].guest.ases_dyn & MIPS_ASE_MSA)
|
|
#endif
|
|
#ifndef cpu_guest_has_dyn_maar
|
|
#define cpu_guest_has_dyn_maar (cpu_data[0].guest.options_dyn & MIPS_CPU_MAAR)
|
|
#endif
|
|
|
|
#endif /* __ASM_CPU_FEATURES_H */
|