Use __attribute__((__cleanup__(func))) to buid pointer and lock guards. Actual usage in the next patch Signed-off-by: Peter Zijlstra (Intel) <peterz@xxxxxxxxxxxxx> --- include/linux/compiler_attributes.h | 6 + include/linux/guards.h | 142 ++++++++++++++++++++++++++++++++++++ include/linux/irqflags.h | 7 + include/linux/mutex.h | 5 + include/linux/preempt.h | 4 + include/linux/rcupdate.h | 3 include/linux/sched/task.h | 2 include/linux/spinlock.h | 27 ++++++ scripts/checkpatch.pl | 2 9 files changed, 197 insertions(+), 1 deletion(-) --- a/include/linux/compiler_attributes.h +++ b/include/linux/compiler_attributes.h @@ -77,6 +77,12 @@ #define __attribute_const__ __attribute__((__const__)) /* + * gcc: https://gcc.gnu.org/onlinedocs/gcc/Common-Variable-Attributes.html#index-cleanup-variable-attribute + * clang: https://clang.llvm.org/docs/AttributeReference.html#cleanup + */ +#define __cleanup(func) __attribute__((__cleanup__(func))) + +/* * Optional: only supported since gcc >= 9 * Optional: not supported by clang * --- /dev/null +++ b/include/linux/guards.h @@ -0,0 +1,142 @@ +/* SPDX-License-Identifier: GPL-2.0 */ +#ifndef __LINUX_GUARDS_H +#define __LINUX_GUARDS_H + +#include <linux/compiler_attributes.h> + +/* + * Pointer Guards are special pointers (variables) with a scope bound cleanup + * function. + * + * Various guard types can be created using: + * + * DEFINE_PTR_GUARD(guard_type, typename, cleanup-exp) + * + * After which they can be used like so: + * + * ptr_guard(guard_type, name) = find_get_object(foo); + * + * Where the return type of find_get_object() should match the guard_type's + * 'typname *'. And when @name goes out of scope cleanup-exp is ran (inserted + * by the compiler) when !NULL at that time. Also see the __cleanup attribute. + */ + +#define DEFINE_PTR_GUARD(_type, _Type, _Put) \ +typedef _Type *ptr_guard_##_type##_t; \ + \ +static inline void ptr_guard_##_type##_cleanup(_Type **_ptr) \ +{ \ + _Type *_G = *_ptr; \ + if (_G) \ + _Put(_G); \ +} + +#define ptr_guard(_type, _name) \ + ptr_guard_##_type##_t _name __cleanup(ptr_guard_##_type##_cleanup) + + +/* + * Lock Guards are like the pointer guards above except they also have + * a fixed initializor to cover both the Lock and Unlock of the lock type. + * + * Lock guards types can be created using: + * + * DEFINE_LOCK_GUARD_0(guard_type, Lock, Unlock, [extra guard members]) + * DEFINE_LOCK_GUARD_1(guard_type, typename, Lock, Unlock, ...) + * DEFINE_LOCK_GUARD_2(guard_type, typename, Lock, Unlock, ...) + * + * Where the _n suffix indicates the number of arguments of 'typename *' the + * Lock function requires. + * + * Once defined, the lock guards can be used in one of two ways: + * + * guard(guard_type, name, var...); + * + * or: + * + * scoped (guard_type, var...) { + * ... + * } + * + * The first creates a named variable that is initialized with the Lock + * function and will call the Unlock function when it goes out of scope. + * + * The second creates an explicit scope, using a for-loop with an implicit + * named _scope variable. Again, Lock is called before the scope is entered and + * Unlock will be called when the scope is left. + * + * Both Lock and Unlock are expressions and can access the guard object through + * the _G pointer. The guard object will have _n implicit members called of + * type 'typename *' called 'lock' and 'lock2' as well as any additional + * members specified in the definition. + */ + +#define DEFINE_LOCK_GUARD_0(_type, _Lock, _Unlock, ...) \ +typedef struct { \ + __VA_ARGS__; \ +} lock_guard_##_type##_t; \ + \ +static inline void lock_guard_##_type##_cleanup(lock_guard_##_type##_t *_G) \ +{ \ + _Unlock; \ +} \ + \ +static inline lock_guard_##_type##_t lock_guard_##_type##_init(void) \ +{ \ + lock_guard_##_type##_t _g = { }, *_G __maybe_unused = &_g; \ + _Lock; \ + return _g; \ +} + +#define DEFINE_LOCK_GUARD_1(_type, _Type, _Lock, _Unlock, ...) \ +typedef struct { \ + _Type *lock; \ + __VA_ARGS__; \ +} lock_guard_##_type##_t; \ + \ +static inline void lock_guard_##_type##_cleanup(lock_guard_##_type##_t *_G) \ +{ \ + _Unlock; \ +} \ + \ +static inline lock_guard_##_type##_t lock_guard_##_type##_init(_Type *lock) \ +{ \ + lock_guard_##_type##_t _g = { .lock = lock }, *_G = &_g; \ + _Lock; \ + return _g; \ +} + +#define DEFINE_LOCK_GUARD_2(_type, _Type, _Lock, _Unlock, ...) \ +typedef struct { \ + _Type *lock; \ + _Type *lock2; \ + __VA_ARGS__; \ +} lock_guard_##_type##_t; \ + \ +static inline void lock_guard_##_type##_cleanup(lock_guard_##_type##_t *_G) \ +{ \ + _Unlock; \ +} \ + \ +static inline lock_guard_##_type##_t \ +lock_guard_##_type##_init(_Type *lock, _Type *lock2) \ +{ \ + lock_guard_##_type##_t _g = { .lock = lock, .lock2 = lock2 }, *_G = &_g;\ + _Lock; \ + return _g; \ +} + +#define variable_scope(_type, _enter, _exit) \ + for (_type *_done = NULL, _scope __cleanup(_exit) = _enter; \ + !_done; _done = (void *)8) + +#define scoped(_type, _var...) \ + variable_scope(lock_guard_##_type##_t, \ + lock_guard_##_type##_init(_var), \ + lock_guard_##_type##_cleanup) + +#define guard(_type, _name, _var...) \ + lock_guard_##_type##_t __cleanup(lock_guard_##_type##_cleanup) _name = \ + lock_guard_##_type##_init(_var) + +#endif /* __LINUX_GUARDS_H */ --- a/include/linux/irqflags.h +++ b/include/linux/irqflags.h @@ -13,6 +13,7 @@ #define _LINUX_TRACE_IRQFLAGS_H #include <linux/typecheck.h> +#include <linux/guards.h> #include <asm/irqflags.h> #include <asm/percpu.h> @@ -267,4 +268,10 @@ extern void warn_bogus_irq_restore(void) #define irqs_disabled_flags(flags) raw_irqs_disabled_flags(flags) +DEFINE_LOCK_GUARD_0(irq, local_irq_disable(), local_irq_enable()) +DEFINE_LOCK_GUARD_0(irqsave, + local_irq_save(_G->flags), + local_irq_restore(_G->flags), + unsigned long flags) + #endif --- a/include/linux/mutex.h +++ b/include/linux/mutex.h @@ -19,6 +19,7 @@ #include <asm/processor.h> #include <linux/osq_lock.h> #include <linux/debug_locks.h> +#include <linux/guards.h> #ifdef CONFIG_DEBUG_LOCK_ALLOC # define __DEP_MAP_MUTEX_INITIALIZER(lockname) \ @@ -219,4 +220,8 @@ extern void mutex_unlock(struct mutex *l extern int atomic_dec_and_mutex_lock(atomic_t *cnt, struct mutex *lock); +DEFINE_LOCK_GUARD_1(mutex, struct mutex, + mutex_lock(_G->lock), + mutex_unlock(_G->lock)) + #endif /* __LINUX_MUTEX_H */ --- a/include/linux/preempt.h +++ b/include/linux/preempt.h @@ -8,6 +8,7 @@ */ #include <linux/linkage.h> +#include <linux/guards.h> #include <linux/list.h> /* @@ -463,4 +464,7 @@ static __always_inline void preempt_enab preempt_enable(); } +DEFINE_LOCK_GUARD_0(preempt, preempt_disable(), preempt_enable()) +DEFINE_LOCK_GUARD_0(migrate, migrate_disable(), migrate_enable()) + #endif /* __LINUX_PREEMPT_H */ --- a/include/linux/rcupdate.h +++ b/include/linux/rcupdate.h @@ -27,6 +27,7 @@ #include <linux/preempt.h> #include <linux/bottom_half.h> #include <linux/lockdep.h> +#include <linux/guards.h> #include <asm/processor.h> #include <linux/cpumask.h> #include <linux/context_tracking_irq.h> @@ -1095,4 +1096,6 @@ rcu_head_after_call_rcu(struct rcu_head extern int rcu_expedited; extern int rcu_normal; +DEFINE_LOCK_GUARD_0(rcu, rcu_read_lock(), rcu_read_unlock()) + #endif /* __LINUX_RCUPDATE_H */ --- a/include/linux/sched/task.h +++ b/include/linux/sched/task.h @@ -126,6 +126,8 @@ static inline void put_task_struct(struc __put_task_struct(t); } +DEFINE_PTR_GUARD(put_task, struct task_struct, put_task_struct) + static inline void put_task_struct_many(struct task_struct *t, int nr) { if (refcount_sub_and_test(nr, &t->usage)) --- a/include/linux/spinlock.h +++ b/include/linux/spinlock.h @@ -61,6 +61,7 @@ #include <linux/stringify.h> #include <linux/bottom_half.h> #include <linux/lockdep.h> +#include <linux/guards.h> #include <asm/barrier.h> #include <asm/mmiowb.h> @@ -502,5 +503,31 @@ int __alloc_bucket_spinlocks(spinlock_t void free_bucket_spinlocks(spinlock_t *locks); +DEFINE_LOCK_GUARD_1(raw_spinlock, raw_spinlock_t, + raw_spin_lock(_G->lock), + raw_spin_unlock(_G->lock)) + +DEFINE_LOCK_GUARD_1(raw_spinlock_irq, raw_spinlock_t, + raw_spin_lock_irq(_G->lock), + raw_spin_unlock_irq(_G->lock)) + +DEFINE_LOCK_GUARD_1(raw_spinlock_irqsave, raw_spinlock_t, + raw_spin_lock_irqsave(_G->lock, _G->flags), + raw_spin_unlock_irqrestore(_G->lock, _G->flags), + unsigned long flags) + +DEFINE_LOCK_GUARD_1(spinlock, spinlock_t, + spin_lock(_G->lock), + spin_unlock(_G->lock)) + +DEFINE_LOCK_GUARD_1(spinlock_irq, spinlock_t, + spin_lock_irq(_G->lock), + spin_unlock_irq(_G->lock)) + +DEFINE_LOCK_GUARD_1(spinlock_irqsave, spinlock_t, + spin_lock_irqsave(_G->lock, _G->flags), + spin_unlock_irqrestore(_G->lock, _G->flags), + unsigned long flags) + #undef __LINUX_INSIDE_SPINLOCK_H #endif /* __LINUX_SPINLOCK_H */ --- a/scripts/checkpatch.pl +++ b/scripts/checkpatch.pl @@ -5046,7 +5046,7 @@ sub process { if|for|while|switch|return|case| volatile|__volatile__| __attribute__|format|__extension__| - asm|__asm__)$/x) + asm|__asm__|scoped)$/x) { # cpp #define statements have non-optional spaces, ie # if there is a space between the name and the open