atomic: Add irqsave variant of atomic_dec_and_lock()
There are in-tree users of atomic_dec_and_lock() which must acquire the spin lock with interrupts disabled. To workaround the lack of an irqsave variant of atomic_dec_and_lock() they use local_irq_save() at the call site. This causes extra code and creates in some places unneeded long interrupt disabled times. These places need also extra treatment for PREEMPT_RT due to the disconnect of the irq disabling and the lock function. Implement the missing irqsave variant of the function. Signed-off-by: Anna-Maria Gleixner <anna-maria@linutronix.de> Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de> Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org> Link: https://lkml.kernel.org/r20180612161621.22645-3-bigeasy@linutronix.de
This commit is contained in:
parent
f2ae679411
commit
ccfbb5bed4
|
@ -427,6 +427,11 @@ extern int _atomic_dec_and_lock(atomic_t *atomic, spinlock_t *lock);
|
||||||
#define atomic_dec_and_lock(atomic, lock) \
|
#define atomic_dec_and_lock(atomic, lock) \
|
||||||
__cond_lock(lock, _atomic_dec_and_lock(atomic, lock))
|
__cond_lock(lock, _atomic_dec_and_lock(atomic, lock))
|
||||||
|
|
||||||
|
extern int _atomic_dec_and_lock_irqsave(atomic_t *atomic, spinlock_t *lock,
|
||||||
|
unsigned long *flags);
|
||||||
|
#define atomic_dec_and_lock_irqsave(atomic, lock, flags) \
|
||||||
|
__cond_lock(lock, _atomic_dec_and_lock_irqsave(atomic, lock, &(flags)))
|
||||||
|
|
||||||
int alloc_bucket_spinlocks(spinlock_t **locks, unsigned int *lock_mask,
|
int alloc_bucket_spinlocks(spinlock_t **locks, unsigned int *lock_mask,
|
||||||
size_t max_size, unsigned int cpu_mult,
|
size_t max_size, unsigned int cpu_mult,
|
||||||
gfp_t gfp);
|
gfp_t gfp);
|
||||||
|
|
|
@ -33,3 +33,19 @@ int _atomic_dec_and_lock(atomic_t *atomic, spinlock_t *lock)
|
||||||
}
|
}
|
||||||
|
|
||||||
EXPORT_SYMBOL(_atomic_dec_and_lock);
|
EXPORT_SYMBOL(_atomic_dec_and_lock);
|
||||||
|
|
||||||
|
int _atomic_dec_and_lock_irqsave(atomic_t *atomic, spinlock_t *lock,
|
||||||
|
unsigned long *flags)
|
||||||
|
{
|
||||||
|
/* Subtract 1 from counter unless that drops it to 0 (ie. it was 1) */
|
||||||
|
if (atomic_add_unless(atomic, -1, 1))
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
/* Otherwise do it the slow way */
|
||||||
|
spin_lock_irqsave(lock, *flags);
|
||||||
|
if (atomic_dec_and_test(atomic))
|
||||||
|
return 1;
|
||||||
|
spin_unlock_irqrestore(lock, *flags);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(_atomic_dec_and_lock_irqsave);
|
||||||
|
|
Loading…
Reference in New Issue