mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-23 22:13:50 +07:00
e335e3eb82
Get rid of INLINE_SPIN_UNLOCK entirely replacing it with UNINLINE_SPIN_UNLOCK instead of the reverse meaning. Whoever wants to change the default spinlock inlining behavior and uninline the spinlocks for some weird reason, such as spinlock debugging, paravirt etc. can now all just select UNINLINE_SPIN_UNLOCK Original discussion at: https://lkml.org/lkml/2012/3/21/357 Suggested-by: Linus Torvalds <torvalds@linux-foundation.org> Signed-off-by: Raghavendra K T <raghavendra.kt@linux.vnet.ibm.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Ralf Baechle <ralf@linux-mips.org> Cc: Chris Metcalf <cmetcalf@tilera.com> Cc: Chris Zankel <chris@zankel.net> Cc: linux-mips@linux-mips.org Link: http://lkml.kernel.org/r/20120322095502.30866.75756.sendpatchset@codeblue [ tidied up the changelog a bit ] Signed-off-by: Ingo Molnar <mingo@kernel.org>
197 lines
5.4 KiB
C
197 lines
5.4 KiB
C
#ifndef __LINUX_SPINLOCK_API_SMP_H
|
|
#define __LINUX_SPINLOCK_API_SMP_H
|
|
|
|
#ifndef __LINUX_SPINLOCK_H
|
|
# error "please don't include this file directly"
|
|
#endif
|
|
|
|
/*
|
|
* include/linux/spinlock_api_smp.h
|
|
*
|
|
* spinlock API declarations on SMP (and debug)
|
|
* (implemented in kernel/spinlock.c)
|
|
*
|
|
* portions Copyright 2005, Red Hat, Inc., Ingo Molnar
|
|
* Released under the General Public License (GPL).
|
|
*/
|
|
|
|
int in_lock_functions(unsigned long addr);
|
|
|
|
#define assert_raw_spin_locked(x) BUG_ON(!raw_spin_is_locked(x))
|
|
|
|
void __lockfunc _raw_spin_lock(raw_spinlock_t *lock) __acquires(lock);
|
|
void __lockfunc _raw_spin_lock_nested(raw_spinlock_t *lock, int subclass)
|
|
__acquires(lock);
|
|
void __lockfunc
|
|
_raw_spin_lock_nest_lock(raw_spinlock_t *lock, struct lockdep_map *map)
|
|
__acquires(lock);
|
|
void __lockfunc _raw_spin_lock_bh(raw_spinlock_t *lock) __acquires(lock);
|
|
void __lockfunc _raw_spin_lock_irq(raw_spinlock_t *lock)
|
|
__acquires(lock);
|
|
|
|
unsigned long __lockfunc _raw_spin_lock_irqsave(raw_spinlock_t *lock)
|
|
__acquires(lock);
|
|
unsigned long __lockfunc
|
|
_raw_spin_lock_irqsave_nested(raw_spinlock_t *lock, int subclass)
|
|
__acquires(lock);
|
|
int __lockfunc _raw_spin_trylock(raw_spinlock_t *lock);
|
|
int __lockfunc _raw_spin_trylock_bh(raw_spinlock_t *lock);
|
|
void __lockfunc _raw_spin_unlock(raw_spinlock_t *lock) __releases(lock);
|
|
void __lockfunc _raw_spin_unlock_bh(raw_spinlock_t *lock) __releases(lock);
|
|
void __lockfunc _raw_spin_unlock_irq(raw_spinlock_t *lock) __releases(lock);
|
|
void __lockfunc
|
|
_raw_spin_unlock_irqrestore(raw_spinlock_t *lock, unsigned long flags)
|
|
__releases(lock);
|
|
|
|
#ifdef CONFIG_INLINE_SPIN_LOCK
|
|
#define _raw_spin_lock(lock) __raw_spin_lock(lock)
|
|
#endif
|
|
|
|
#ifdef CONFIG_INLINE_SPIN_LOCK_BH
|
|
#define _raw_spin_lock_bh(lock) __raw_spin_lock_bh(lock)
|
|
#endif
|
|
|
|
#ifdef CONFIG_INLINE_SPIN_LOCK_IRQ
|
|
#define _raw_spin_lock_irq(lock) __raw_spin_lock_irq(lock)
|
|
#endif
|
|
|
|
#ifdef CONFIG_INLINE_SPIN_LOCK_IRQSAVE
|
|
#define _raw_spin_lock_irqsave(lock) __raw_spin_lock_irqsave(lock)
|
|
#endif
|
|
|
|
#ifdef CONFIG_INLINE_SPIN_TRYLOCK
|
|
#define _raw_spin_trylock(lock) __raw_spin_trylock(lock)
|
|
#endif
|
|
|
|
#ifdef CONFIG_INLINE_SPIN_TRYLOCK_BH
|
|
#define _raw_spin_trylock_bh(lock) __raw_spin_trylock_bh(lock)
|
|
#endif
|
|
|
|
#ifndef CONFIG_UNINLINE_SPIN_UNLOCK
|
|
#define _raw_spin_unlock(lock) __raw_spin_unlock(lock)
|
|
#endif
|
|
|
|
#ifdef CONFIG_INLINE_SPIN_UNLOCK_BH
|
|
#define _raw_spin_unlock_bh(lock) __raw_spin_unlock_bh(lock)
|
|
#endif
|
|
|
|
#ifdef CONFIG_INLINE_SPIN_UNLOCK_IRQ
|
|
#define _raw_spin_unlock_irq(lock) __raw_spin_unlock_irq(lock)
|
|
#endif
|
|
|
|
#ifdef CONFIG_INLINE_SPIN_UNLOCK_IRQRESTORE
|
|
#define _raw_spin_unlock_irqrestore(lock, flags) __raw_spin_unlock_irqrestore(lock, flags)
|
|
#endif
|
|
|
|
static inline int __raw_spin_trylock(raw_spinlock_t *lock)
|
|
{
|
|
preempt_disable();
|
|
if (do_raw_spin_trylock(lock)) {
|
|
spin_acquire(&lock->dep_map, 0, 1, _RET_IP_);
|
|
return 1;
|
|
}
|
|
preempt_enable();
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* If lockdep is enabled then we use the non-preemption spin-ops
|
|
* even on CONFIG_PREEMPT, because lockdep assumes that interrupts are
|
|
* not re-enabled during lock-acquire (which the preempt-spin-ops do):
|
|
*/
|
|
#if !defined(CONFIG_GENERIC_LOCKBREAK) || defined(CONFIG_DEBUG_LOCK_ALLOC)
|
|
|
|
static inline unsigned long __raw_spin_lock_irqsave(raw_spinlock_t *lock)
|
|
{
|
|
unsigned long flags;
|
|
|
|
local_irq_save(flags);
|
|
preempt_disable();
|
|
spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
|
|
/*
|
|
* On lockdep we dont want the hand-coded irq-enable of
|
|
* do_raw_spin_lock_flags() code, because lockdep assumes
|
|
* that interrupts are not re-enabled during lock-acquire:
|
|
*/
|
|
#ifdef CONFIG_LOCKDEP
|
|
LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock);
|
|
#else
|
|
do_raw_spin_lock_flags(lock, &flags);
|
|
#endif
|
|
return flags;
|
|
}
|
|
|
|
static inline void __raw_spin_lock_irq(raw_spinlock_t *lock)
|
|
{
|
|
local_irq_disable();
|
|
preempt_disable();
|
|
spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
|
|
LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock);
|
|
}
|
|
|
|
static inline void __raw_spin_lock_bh(raw_spinlock_t *lock)
|
|
{
|
|
local_bh_disable();
|
|
preempt_disable();
|
|
spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
|
|
LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock);
|
|
}
|
|
|
|
static inline void __raw_spin_lock(raw_spinlock_t *lock)
|
|
{
|
|
preempt_disable();
|
|
spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
|
|
LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock);
|
|
}
|
|
|
|
#endif /* CONFIG_PREEMPT */
|
|
|
|
static inline void __raw_spin_unlock(raw_spinlock_t *lock)
|
|
{
|
|
spin_release(&lock->dep_map, 1, _RET_IP_);
|
|
do_raw_spin_unlock(lock);
|
|
preempt_enable();
|
|
}
|
|
|
|
static inline void __raw_spin_unlock_irqrestore(raw_spinlock_t *lock,
|
|
unsigned long flags)
|
|
{
|
|
spin_release(&lock->dep_map, 1, _RET_IP_);
|
|
do_raw_spin_unlock(lock);
|
|
local_irq_restore(flags);
|
|
preempt_enable();
|
|
}
|
|
|
|
static inline void __raw_spin_unlock_irq(raw_spinlock_t *lock)
|
|
{
|
|
spin_release(&lock->dep_map, 1, _RET_IP_);
|
|
do_raw_spin_unlock(lock);
|
|
local_irq_enable();
|
|
preempt_enable();
|
|
}
|
|
|
|
static inline void __raw_spin_unlock_bh(raw_spinlock_t *lock)
|
|
{
|
|
spin_release(&lock->dep_map, 1, _RET_IP_);
|
|
do_raw_spin_unlock(lock);
|
|
preempt_enable_no_resched();
|
|
local_bh_enable_ip((unsigned long)__builtin_return_address(0));
|
|
}
|
|
|
|
static inline int __raw_spin_trylock_bh(raw_spinlock_t *lock)
|
|
{
|
|
local_bh_disable();
|
|
preempt_disable();
|
|
if (do_raw_spin_trylock(lock)) {
|
|
spin_acquire(&lock->dep_map, 0, 1, _RET_IP_);
|
|
return 1;
|
|
}
|
|
preempt_enable_no_resched();
|
|
local_bh_enable_ip((unsigned long)__builtin_return_address(0));
|
|
return 0;
|
|
}
|
|
|
|
#include <linux/rwlock_api_smp.h>
|
|
|
|
#endif /* __LINUX_SPINLOCK_API_SMP_H */
|