mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-26 00:05:35 +07:00
a308a71022
- clean up various obsolete ioremap and iounmap variants - add a new generic ioremap implementation and switch csky, nds32 and riscv over to it -----BEGIN PGP SIGNATURE----- iQI/BAABCgApFiEEgdbnc3r/njty3Iq9D55TZVIEUYMFAl3cKcsLHGhjaEBsc3Qu ZGUACgkQD55TZVIEUYO1CRAAwFQigsbi0CqqshPWnP0owKV+HA4Xfz/lQZsd7SM/ BVXhKyDJQum6gp73dW025HCfjidTknsbdCUIP/LNUgAnop3lOlnB31/munDnJJ1H 6hB1pc+zB9VgbOe0A6TxtxPRm5aE33k1hZIZS99lOh7mY3FvF7mbkkbVoCjdS3Cq a9bTX+X+esfUQ5GgaIc2zmz2GLkyFXIeVGs8/CoOX58ESCWQcVZrsQRompo4SgrI jqwf47NzdmK8hW4mZ+jdQUiWiAmNs5+2om7Bvi/deFAIFUo1/hLHvQzqEGramq/j 5SPHax2gWAN3uWYP91QISkUAJWFydwgmUDoTO1M04ov4xLuBrqIQmc43tLjHo2UT RwMozWJWN+gkB9zTIboqMPi2qcuDaWcCij7LwHl5zLxPTcOKsrALarL55BQ8MipQ x6fpvskrQQvlArNTsRWFRUq0mCtkzE3wMZ9RR3AIETQL2hlAzB1S4gzhD+Z6WTYY pXNgkunonVGxwyN/7iJTEl/mvF/+MynGcWqhrwHZLqncyhn/WJJ2USH3nAD1+yjp v8v6UUeMXIjUsGAyfTjXy/WXAfwRuSC038AAFcmWKDdh08h4XvPHRficT4U8wr34 7WzGizHP9f1CqrhYL/4exhPY9X2Yb7HhsFd0bZGG0rRvSillPUp0b8s++m12QuQU +VY= =ooiA -----END PGP SIGNATURE----- Merge tag 'ioremap-5.5' of git://git.infradead.org/users/hch/ioremap Pull generic ioremap support from Christoph Hellwig: "This adds the remaining bits for an entirely generic ioremap and iounmap to lib/ioremap.c. To facilitate that, it cleans up the giant mess of weird ioremap variants we had with no users outside the arch code. For now just the three newest ports use the code, but there is more than a handful others that can be converted without too much work. Summary: - clean up various obsolete ioremap and iounmap variants - add a new generic ioremap implementation and switch csky, nds32 and riscv over to it" * tag 'ioremap-5.5' of git://git.infradead.org/users/hch/ioremap: (21 commits) nds32: use generic ioremap csky: use generic ioremap csky: remove ioremap_cache riscv: use the generic ioremap code lib: provide a simple generic ioremap implementation sh: remove __iounmap nios2: remove __iounmap hexagon: remove __iounmap m68k: rename __iounmap and mark it static arch: rely on asm-generic/io.h for default ioremap_* definitions asm-generic: don't provide ioremap for CONFIG_MMU asm-generic: ioremap_uc should behave the same with and without MMU xtensa: clean up ioremap x86: Clean up ioremap() parisc: remove __ioremap nios2: remove __ioremap alpha: remove the unused __ioremap wrapper hexagon: clean up ioremap ia64: rename ioremap_nocache to ioremap_uc unicore32: remove ioremap_cached ...
156 lines
5.2 KiB
C
156 lines
5.2 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* {read,write}{b,w,l,q} based on arch/arm64/include/asm/io.h
|
|
* which was based on arch/arm/include/io.h
|
|
*
|
|
* Copyright (C) 1996-2000 Russell King
|
|
* Copyright (C) 2012 ARM Ltd.
|
|
* Copyright (C) 2014 Regents of the University of California
|
|
*/
|
|
|
|
#ifndef _ASM_RISCV_MMIO_H
|
|
#define _ASM_RISCV_MMIO_H
|
|
|
|
#include <linux/types.h>
|
|
#include <asm/mmiowb.h>
|
|
|
|
#ifndef CONFIG_MMU
|
|
#define pgprot_noncached(x) (x)
|
|
#endif /* CONFIG_MMU */
|
|
|
|
/* Generic IO read/write. These perform native-endian accesses. */
|
|
#define __raw_writeb __raw_writeb
|
|
static inline void __raw_writeb(u8 val, volatile void __iomem *addr)
|
|
{
|
|
asm volatile("sb %0, 0(%1)" : : "r" (val), "r" (addr));
|
|
}
|
|
|
|
#define __raw_writew __raw_writew
|
|
static inline void __raw_writew(u16 val, volatile void __iomem *addr)
|
|
{
|
|
asm volatile("sh %0, 0(%1)" : : "r" (val), "r" (addr));
|
|
}
|
|
|
|
#define __raw_writel __raw_writel
|
|
static inline void __raw_writel(u32 val, volatile void __iomem *addr)
|
|
{
|
|
asm volatile("sw %0, 0(%1)" : : "r" (val), "r" (addr));
|
|
}
|
|
|
|
#ifdef CONFIG_64BIT
|
|
#define __raw_writeq __raw_writeq
|
|
static inline void __raw_writeq(u64 val, volatile void __iomem *addr)
|
|
{
|
|
asm volatile("sd %0, 0(%1)" : : "r" (val), "r" (addr));
|
|
}
|
|
#endif
|
|
|
|
#define __raw_readb __raw_readb
|
|
static inline u8 __raw_readb(const volatile void __iomem *addr)
|
|
{
|
|
u8 val;
|
|
|
|
asm volatile("lb %0, 0(%1)" : "=r" (val) : "r" (addr));
|
|
return val;
|
|
}
|
|
|
|
#define __raw_readw __raw_readw
|
|
static inline u16 __raw_readw(const volatile void __iomem *addr)
|
|
{
|
|
u16 val;
|
|
|
|
asm volatile("lh %0, 0(%1)" : "=r" (val) : "r" (addr));
|
|
return val;
|
|
}
|
|
|
|
#define __raw_readl __raw_readl
|
|
static inline u32 __raw_readl(const volatile void __iomem *addr)
|
|
{
|
|
u32 val;
|
|
|
|
asm volatile("lw %0, 0(%1)" : "=r" (val) : "r" (addr));
|
|
return val;
|
|
}
|
|
|
|
#ifdef CONFIG_64BIT
|
|
#define __raw_readq __raw_readq
|
|
static inline u64 __raw_readq(const volatile void __iomem *addr)
|
|
{
|
|
u64 val;
|
|
|
|
asm volatile("ld %0, 0(%1)" : "=r" (val) : "r" (addr));
|
|
return val;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* Unordered I/O memory access primitives. These are even more relaxed than
|
|
* the relaxed versions, as they don't even order accesses between successive
|
|
* operations to the I/O regions.
|
|
*/
|
|
#define readb_cpu(c) ({ u8 __r = __raw_readb(c); __r; })
|
|
#define readw_cpu(c) ({ u16 __r = le16_to_cpu((__force __le16)__raw_readw(c)); __r; })
|
|
#define readl_cpu(c) ({ u32 __r = le32_to_cpu((__force __le32)__raw_readl(c)); __r; })
|
|
|
|
#define writeb_cpu(v, c) ((void)__raw_writeb((v), (c)))
|
|
#define writew_cpu(v, c) ((void)__raw_writew((__force u16)cpu_to_le16(v), (c)))
|
|
#define writel_cpu(v, c) ((void)__raw_writel((__force u32)cpu_to_le32(v), (c)))
|
|
|
|
#ifdef CONFIG_64BIT
|
|
#define readq_cpu(c) ({ u64 __r = le64_to_cpu((__force __le64)__raw_readq(c)); __r; })
|
|
#define writeq_cpu(v, c) ((void)__raw_writeq((__force u64)cpu_to_le64(v), (c)))
|
|
#endif
|
|
|
|
/*
|
|
* Relaxed I/O memory access primitives. These follow the Device memory
|
|
* ordering rules but do not guarantee any ordering relative to Normal memory
|
|
* accesses. These are defined to order the indicated access (either a read or
|
|
* write) with all other I/O memory accesses. Since the platform specification
|
|
* defines that all I/O regions are strongly ordered on channel 2, no explicit
|
|
* fences are required to enforce this ordering.
|
|
*/
|
|
/* FIXME: These are now the same as asm-generic */
|
|
#define __io_rbr() do {} while (0)
|
|
#define __io_rar() do {} while (0)
|
|
#define __io_rbw() do {} while (0)
|
|
#define __io_raw() do {} while (0)
|
|
|
|
#define readb_relaxed(c) ({ u8 __v; __io_rbr(); __v = readb_cpu(c); __io_rar(); __v; })
|
|
#define readw_relaxed(c) ({ u16 __v; __io_rbr(); __v = readw_cpu(c); __io_rar(); __v; })
|
|
#define readl_relaxed(c) ({ u32 __v; __io_rbr(); __v = readl_cpu(c); __io_rar(); __v; })
|
|
|
|
#define writeb_relaxed(v, c) ({ __io_rbw(); writeb_cpu((v), (c)); __io_raw(); })
|
|
#define writew_relaxed(v, c) ({ __io_rbw(); writew_cpu((v), (c)); __io_raw(); })
|
|
#define writel_relaxed(v, c) ({ __io_rbw(); writel_cpu((v), (c)); __io_raw(); })
|
|
|
|
#ifdef CONFIG_64BIT
|
|
#define readq_relaxed(c) ({ u64 __v; __io_rbr(); __v = readq_cpu(c); __io_rar(); __v; })
|
|
#define writeq_relaxed(v, c) ({ __io_rbw(); writeq_cpu((v), (c)); __io_raw(); })
|
|
#endif
|
|
|
|
/*
|
|
* I/O memory access primitives. Reads are ordered relative to any
|
|
* following Normal memory access. Writes are ordered relative to any prior
|
|
* Normal memory access. The memory barriers here are necessary as RISC-V
|
|
* doesn't define any ordering between the memory space and the I/O space.
|
|
*/
|
|
#define __io_br() do {} while (0)
|
|
#define __io_ar(v) __asm__ __volatile__ ("fence i,r" : : : "memory")
|
|
#define __io_bw() __asm__ __volatile__ ("fence w,o" : : : "memory")
|
|
#define __io_aw() mmiowb_set_pending()
|
|
|
|
#define readb(c) ({ u8 __v; __io_br(); __v = readb_cpu(c); __io_ar(__v); __v; })
|
|
#define readw(c) ({ u16 __v; __io_br(); __v = readw_cpu(c); __io_ar(__v); __v; })
|
|
#define readl(c) ({ u32 __v; __io_br(); __v = readl_cpu(c); __io_ar(__v); __v; })
|
|
|
|
#define writeb(v, c) ({ __io_bw(); writeb_cpu((v), (c)); __io_aw(); })
|
|
#define writew(v, c) ({ __io_bw(); writew_cpu((v), (c)); __io_aw(); })
|
|
#define writel(v, c) ({ __io_bw(); writel_cpu((v), (c)); __io_aw(); })
|
|
|
|
#ifdef CONFIG_64BIT
|
|
#define readq(c) ({ u64 __v; __io_br(); __v = readq_cpu(c); __io_ar(__v); __v; })
|
|
#define writeq(v, c) ({ __io_bw(); writeq_cpu((v), (c)); __io_aw(); })
|
|
#endif
|
|
|
|
#endif /* _ASM_RISCV_MMIO_H */
|