mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2024-12-28 16:52:18 +00:00
5c2e7736e2
When PREEMPT_RT=y, spin locks are mapped to rt_mutex types, so using
spinlock_check() + __raw_spin_lock_init() to initialize spin locks is
incorrect, and would cause build errors.
Introduce __spin_lock_init() to initialize a spin lock with lockdep
rquired information for PREEMPT_RT builds, and use it in the Rust
helper.
Fixes: d2d6422f8b
("x86: Allow to enable PREEMPT_RT.")
Closes: https://lore.kernel.org/oe-kbuild-all/202409251238.vetlgXE9-lkp@intel.com/
Reported-by: kernel test robot <lkp@intel.com>
Signed-off-by: Eder Zulian <ezulian@redhat.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Reviewed-by: Boqun Feng <boqun.feng@gmail.com>
Tested-by: Boqun Feng <boqun.feng@gmail.com>
Link: https://lore.kernel.org/r/20241107163223.2092690-2-ezulian@redhat.com
156 lines
4.1 KiB
C
156 lines
4.1 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
#ifndef __LINUX_SPINLOCK_RT_H
|
|
#define __LINUX_SPINLOCK_RT_H
|
|
|
|
#ifndef __LINUX_INSIDE_SPINLOCK_H
|
|
#error Do not include directly. Use spinlock.h
|
|
#endif
|
|
|
|
#ifdef CONFIG_DEBUG_LOCK_ALLOC
|
|
extern void __rt_spin_lock_init(spinlock_t *lock, const char *name,
|
|
struct lock_class_key *key, bool percpu);
|
|
#else
|
|
static inline void __rt_spin_lock_init(spinlock_t *lock, const char *name,
|
|
struct lock_class_key *key, bool percpu)
|
|
{
|
|
}
|
|
#endif
|
|
|
|
#define __spin_lock_init(slock, name, key, percpu) \
|
|
do { \
|
|
rt_mutex_base_init(&(slock)->lock); \
|
|
__rt_spin_lock_init(slock, name, key, percpu); \
|
|
} while (0)
|
|
|
|
#define _spin_lock_init(slock, percpu) \
|
|
do { \
|
|
static struct lock_class_key __key; \
|
|
__spin_lock_init(slock, #slock, &__key, percpu); \
|
|
} while (0)
|
|
|
|
#define spin_lock_init(slock) _spin_lock_init(slock, false)
|
|
#define local_spin_lock_init(slock) _spin_lock_init(slock, true)
|
|
|
|
extern void rt_spin_lock(spinlock_t *lock) __acquires(lock);
|
|
extern void rt_spin_lock_nested(spinlock_t *lock, int subclass) __acquires(lock);
|
|
extern void rt_spin_lock_nest_lock(spinlock_t *lock, struct lockdep_map *nest_lock) __acquires(lock);
|
|
extern void rt_spin_unlock(spinlock_t *lock) __releases(lock);
|
|
extern void rt_spin_lock_unlock(spinlock_t *lock);
|
|
extern int rt_spin_trylock_bh(spinlock_t *lock);
|
|
extern int rt_spin_trylock(spinlock_t *lock);
|
|
|
|
static __always_inline void spin_lock(spinlock_t *lock)
|
|
{
|
|
rt_spin_lock(lock);
|
|
}
|
|
|
|
#ifdef CONFIG_LOCKDEP
|
|
# define __spin_lock_nested(lock, subclass) \
|
|
rt_spin_lock_nested(lock, subclass)
|
|
|
|
# define __spin_lock_nest_lock(lock, nest_lock) \
|
|
do { \
|
|
typecheck(struct lockdep_map *, &(nest_lock)->dep_map); \
|
|
rt_spin_lock_nest_lock(lock, &(nest_lock)->dep_map); \
|
|
} while (0)
|
|
# define __spin_lock_irqsave_nested(lock, flags, subclass) \
|
|
do { \
|
|
typecheck(unsigned long, flags); \
|
|
flags = 0; \
|
|
__spin_lock_nested(lock, subclass); \
|
|
} while (0)
|
|
|
|
#else
|
|
/*
|
|
* Always evaluate the 'subclass' argument to avoid that the compiler
|
|
* warns about set-but-not-used variables when building with
|
|
* CONFIG_DEBUG_LOCK_ALLOC=n and with W=1.
|
|
*/
|
|
# define __spin_lock_nested(lock, subclass) spin_lock(((void)(subclass), (lock)))
|
|
# define __spin_lock_nest_lock(lock, subclass) spin_lock(((void)(subclass), (lock)))
|
|
# define __spin_lock_irqsave_nested(lock, flags, subclass) \
|
|
spin_lock_irqsave(((void)(subclass), (lock)), flags)
|
|
#endif
|
|
|
|
#define spin_lock_nested(lock, subclass) \
|
|
__spin_lock_nested(lock, subclass)
|
|
|
|
#define spin_lock_nest_lock(lock, nest_lock) \
|
|
__spin_lock_nest_lock(lock, nest_lock)
|
|
|
|
#define spin_lock_irqsave_nested(lock, flags, subclass) \
|
|
__spin_lock_irqsave_nested(lock, flags, subclass)
|
|
|
|
static __always_inline void spin_lock_bh(spinlock_t *lock)
|
|
{
|
|
/* Investigate: Drop bh when blocking ? */
|
|
local_bh_disable();
|
|
rt_spin_lock(lock);
|
|
}
|
|
|
|
static __always_inline void spin_lock_irq(spinlock_t *lock)
|
|
{
|
|
rt_spin_lock(lock);
|
|
}
|
|
|
|
#define spin_lock_irqsave(lock, flags) \
|
|
do { \
|
|
typecheck(unsigned long, flags); \
|
|
flags = 0; \
|
|
spin_lock(lock); \
|
|
} while (0)
|
|
|
|
static __always_inline void spin_unlock(spinlock_t *lock)
|
|
{
|
|
rt_spin_unlock(lock);
|
|
}
|
|
|
|
static __always_inline void spin_unlock_bh(spinlock_t *lock)
|
|
{
|
|
rt_spin_unlock(lock);
|
|
local_bh_enable();
|
|
}
|
|
|
|
static __always_inline void spin_unlock_irq(spinlock_t *lock)
|
|
{
|
|
rt_spin_unlock(lock);
|
|
}
|
|
|
|
static __always_inline void spin_unlock_irqrestore(spinlock_t *lock,
|
|
unsigned long flags)
|
|
{
|
|
rt_spin_unlock(lock);
|
|
}
|
|
|
|
#define spin_trylock(lock) \
|
|
__cond_lock(lock, rt_spin_trylock(lock))
|
|
|
|
#define spin_trylock_bh(lock) \
|
|
__cond_lock(lock, rt_spin_trylock_bh(lock))
|
|
|
|
#define spin_trylock_irq(lock) \
|
|
__cond_lock(lock, rt_spin_trylock(lock))
|
|
|
|
#define spin_trylock_irqsave(lock, flags) \
|
|
({ \
|
|
int __locked; \
|
|
\
|
|
typecheck(unsigned long, flags); \
|
|
flags = 0; \
|
|
__locked = spin_trylock(lock); \
|
|
__locked; \
|
|
})
|
|
|
|
#define spin_is_contended(lock) (((void)(lock), 0))
|
|
|
|
static inline int spin_is_locked(spinlock_t *lock)
|
|
{
|
|
return rt_mutex_base_is_locked(&lock->lock);
|
|
}
|
|
|
|
#define assert_spin_locked(lock) BUG_ON(!spin_is_locked(lock))
|
|
|
|
#include <linux/rwlock_rt.h>
|
|
|
|
#endif
|