mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-04 14:26:41 +07:00
b01aec9b2c
So first of all, this atomic_scrub() function's naming is bad. It looks like an atomic_t helper. Change it to edac_atomic_scrub(). The bigger problem is that this function is arch-specific and every new arch which doesn't necessarily need that functionality still needs to define it, otherwise EDAC doesn't compile. So instead of doing that and including arch-specific headers, have each arch define an EDAC_ATOMIC_SCRUB symbol which can be used in edac_mc.c for ifdeffery. Much cleaner. And we already are doing this with another symbol - EDAC_SUPPORT. This is also much cleaner than having CONFIG_EDAC enumerate all the arches which need/have EDAC support and drivers. This way I can kill the useless edac.h header in tile too. Acked-by: Ralf Baechle <ralf@linux-mips.org> Acked-by: Michael Ellerman <mpe@ellerman.id.au> Acked-by: Chris Metcalf <cmetcalf@ezchip.com> Acked-by: Ingo Molnar <mingo@kernel.org> Acked-by: Russell King <rmk+kernel@arm.linux.org.uk> Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org> Cc: Doug Thompson <dougthompson@xmission.com> Cc: linux-arm-kernel@lists.infradead.org Cc: linux-edac@vger.kernel.org Cc: linux-kernel@vger.kernel.org Cc: linux-mips@linux-mips.org Cc: linuxppc-dev@lists.ozlabs.org Cc: "Maciej W. Rozycki" <macro@codesourcery.com> Cc: Markos Chandras <markos.chandras@imgtec.com> Cc: Mauro Carvalho Chehab <mchehab@osg.samsung.com> Cc: Paul Mackerras <paulus@samba.org> Cc: "Steven J. Hill" <Steven.Hill@imgtec.com> Cc: x86@kernel.org Signed-off-by: Borislav Petkov <bp@suse.de>
37 lines
780 B
C
37 lines
780 B
C
#ifndef ASM_EDAC_H
|
|
#define ASM_EDAC_H
|
|
|
|
#include <asm/compiler.h>
|
|
|
|
/* ECC atomic, DMA, SMP and interrupt safe scrub function */
|
|
|
|
static inline void edac_atomic_scrub(void *va, u32 size)
|
|
{
|
|
unsigned long *virt_addr = va;
|
|
unsigned long temp;
|
|
u32 i;
|
|
|
|
for (i = 0; i < size / sizeof(unsigned long); i++) {
|
|
/*
|
|
* Very carefully read and write to memory atomically
|
|
* so we are interrupt, DMA and SMP safe.
|
|
*
|
|
* Intel: asm("lock; addl $0, %0"::"m"(*virt_addr));
|
|
*/
|
|
|
|
__asm__ __volatile__ (
|
|
" .set mips2 \n"
|
|
"1: ll %0, %1 # edac_atomic_scrub \n"
|
|
" addu %0, $0 \n"
|
|
" sc %0, %1 \n"
|
|
" beqz %0, 1b \n"
|
|
" .set mips0 \n"
|
|
: "=&r" (temp), "=" GCC_OFF_SMALL_ASM() (*virt_addr)
|
|
: GCC_OFF_SMALL_ASM() (*virt_addr));
|
|
|
|
virt_addr++;
|
|
}
|
|
}
|
|
|
|
#endif
|