#ifndef _ASM_GENERIC_ATOMIC_LONG_H #define _ASM_GENERIC_ATOMIC_LONG_H /* * Copyright (C) 2005 Silicon Graphics, Inc. * Christoph Lameter * * Allows to provide arch independent atomic definitions without the need to * edit all arch specific atomic.h files. */ #include /* * Suppport for atomic_long_t * * Casts for parameters are avoided for existing atomic functions in order to * avoid issues with cast-as-lval under gcc 4.x and other limitations that the * macros of a platform may have. */ #if BITS_PER_LONG == 64 typedef atomic64_t atomic_long_t; #define ATOMIC_LONG_INIT(i) ATOMIC64_INIT(i) #define ATOMIC_LONG_PFX(x) atomic64 ## x #else typedef atomic_t atomic_long_t; #define ATOMIC_LONG_INIT(i) ATOMIC_INIT(i) #define ATOMIC_LONG_PFX(x) atomic ## x #endif #define ATOMIC_LONG_READ_OP(mo) \ static inline long atomic_long_read##mo(const atomic_long_t *l) \ { \ ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; \ \ return (long)ATOMIC_LONG_PFX(_read##mo)(v); \ } ATOMIC_LONG_READ_OP() ATOMIC_LONG_READ_OP(_acquire) #undef ATOMIC_LONG_READ_OP #define ATOMIC_LONG_SET_OP(mo) \ static inline void atomic_long_set##mo(atomic_long_t *l, long i) \ { \ ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; \ \ ATOMIC_LONG_PFX(_set##mo)(v, i); \ } ATOMIC_LONG_SET_OP() ATOMIC_LONG_SET_OP(_release) #undef ATOMIC_LONG_SET_OP #define ATOMIC_LONG_ADD_SUB_OP(op, mo) \ static inline long \ atomic_long_##op##_return##mo(long i, atomic_long_t *l) \ { \ ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; \ \ return (long)ATOMIC_LONG_PFX(_##op##_return##mo)(i, v); \ } ATOMIC_LONG_ADD_SUB_OP(add,) ATOMIC_LONG_ADD_SUB_OP(add, _relaxed) ATOMIC_LONG_ADD_SUB_OP(add, _acquire) ATOMIC_LONG_ADD_SUB_OP(add, _release) ATOMIC_LONG_ADD_SUB_OP(sub,) ATOMIC_LONG_ADD_SUB_OP(sub, _relaxed) ATOMIC_LONG_ADD_SUB_OP(sub, _acquire) ATOMIC_LONG_ADD_SUB_OP(sub, _release) #undef ATOMIC_LONG_ADD_SUB_OP #define atomic_long_cmpxchg_relaxed(l, old, new) \ (ATOMIC_LONG_PFX(_cmpxchg_relaxed)((ATOMIC_LONG_PFX(_t) *)(l), \ (old), (new))) #define atomic_long_cmpxchg_acquire(l, old, new) \ (ATOMIC_LONG_PFX(_cmpxchg_acquire)((ATOMIC_LONG_PFX(_t) *)(l), \ (old), (new))) #define atomic_long_cmpxchg_release(l, old, new) \ (ATOMIC_LONG_PFX(_cmpxchg_release)((ATOMIC_LONG_PFX(_t) *)(l), \ (old), (new))) #define atomic_long_cmpxchg(l, old, new) \ (ATOMIC_LONG_PFX(_cmpxchg)((ATOMIC_LONG_PFX(_t) *)(l), (old), (new))) #define atomic_long_xchg_relaxed(v, new) \ (ATOMIC_LONG_PFX(_xchg_relaxed)((ATOMIC_LONG_PFX(_t) *)(v), (new))) #define atomic_long_xchg_acquire(v, new) \ (ATOMIC_LONG_PFX(_xchg_acquire)((ATOMIC_LONG_PFX(_t) *)(v), (new))) #define atomic_long_xchg_release(v, new) \ (ATOMIC_LONG_PFX(_xchg_release)((ATOMIC_LONG_PFX(_t) *)(v), (new))) #define atomic_long_xchg(v, new) \ (ATOMIC_LONG_PFX(_xchg)((ATOMIC_LONG_PFX(_t) *)(v), (new))) static __always_inline void atomic_long_inc(atomic_long_t *l) { ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; ATOMIC_LONG_PFX(_inc)(v); } static __always_inline void atomic_long_dec(atomic_long_t *l) { ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; ATOMIC_LONG_PFX(_dec)(v); } #define ATOMIC_LONG_FETCH_OP(op, mo) \ static inline long \ atomic_long_fetch_##op##mo(long i, atomic_long_t *l) \ { \ ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; \ \ return (long)ATOMIC_LONG_PFX(_fetch_##op##mo)(i, v); \ } ATOMIC_LONG_FETCH_OP(add, ) ATOMIC_LONG_FETCH_OP(add, _relaxed) ATOMIC_LONG_FETCH_OP(add, _acquire) ATOMIC_LONG_FETCH_OP(add, _release) ATOMIC_LONG_FETCH_OP(sub, ) ATOMIC_LONG_FETCH_OP(sub, _relaxed) ATOMIC_LONG_FETCH_OP(sub, _acquire) ATOMIC_LONG_FETCH_OP(sub, _release) ATOMIC_LONG_FETCH_OP(and, ) ATOMIC_LONG_FETCH_OP(and, _relaxed) ATOMIC_LONG_FETCH_OP(and, _acquire) ATOMIC_LONG_FETCH_OP(and, _release) ATOMIC_LONG_FETCH_OP(andnot, ) ATOMIC_LONG_FETCH_OP(andnot, _relaxed) ATOMIC_LONG_FETCH_OP(andnot, _acquire) ATOMIC_LONG_FETCH_OP(andnot, _release) ATOMIC_LONG_FETCH_OP(or, ) ATOMIC_LONG_FETCH_OP(or, _relaxed) ATOMIC_LONG_FETCH_OP(or, _acquire) ATOMIC_LONG_FETCH_OP(or, _release) ATOMIC_LONG_FETCH_OP(xor, ) ATOMIC_LONG_FETCH_OP(xor, _relaxed) ATOMIC_LONG_FETCH_OP(xor, _acquire) ATOMIC_LONG_FETCH_OP(xor, _release) #undef ATOMIC_LONG_FETCH_OP #define ATOMIC_LONG_FETCH_INC_DEC_OP(op, mo) \ static inline long \ atomic_long_fetch_##op##mo(atomic_long_t *l) \ { \ ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; \ \ return (long)ATOMIC_LONG_PFX(_fetch_##op##mo)(v); \ } ATOMIC_LONG_FETCH_INC_DEC_OP(inc,) ATOMIC_LONG_FETCH_INC_DEC_OP(inc, _relaxed) ATOMIC_LONG_FETCH_INC_DEC_OP(inc, _acquire) ATOMIC_LONG_FETCH_INC_DEC_OP(inc, _release) ATOMIC_LONG_FETCH_INC_DEC_OP(dec,) ATOMIC_LONG_FETCH_INC_DEC_OP(dec, _relaxed) ATOMIC_LONG_FETCH_INC_DEC_OP(dec, _acquire) ATOMIC_LONG_FETCH_INC_DEC_OP(dec, _release) #undef ATOMIC_LONG_FETCH_INC_DEC_OP #define ATOMIC_LONG_OP(op) \ static __always_inline void \ atomic_long_##op(long i, atomic_long_t *l) \ { \ ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; \ \ ATOMIC_LONG_PFX(_##op)(i, v); \ } ATOMIC_LONG_OP(add) ATOMIC_LONG_OP(sub) ATOMIC_LONG_OP(and) ATOMIC_LONG_OP(andnot) ATOMIC_LONG_OP(or) ATOMIC_LONG_OP(xor) #undef ATOMIC_LONG_OP static inline int atomic_long_sub_and_test(long i, atomic_long_t *l) { ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; return ATOMIC_LONG_PFX(_sub_and_test)(i, v); } static inline int atomic_long_dec_and_test(atomic_long_t *l) { ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; return ATOMIC_LONG_PFX(_dec_and_test)(v); } static inline int atomic_long_inc_and_test(atomic_long_t *l) { ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; return ATOMIC_LONG_PFX(_inc_and_test)(v); } static inline int atomic_long_add_negative(long i, atomic_long_t *l) { ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; return ATOMIC_LONG_PFX(_add_negative)(i, v); } #define ATOMIC_LONG_INC_DEC_OP(op, mo) \ static inline long \ atomic_long_##op##_return##mo(atomic_long_t *l) \ { \ ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; \ \ return (long)ATOMIC_LONG_PFX(_##op##_return##mo)(v); \ } ATOMIC_LONG_INC_DEC_OP(inc,) ATOMIC_LONG_INC_DEC_OP(inc, _relaxed) ATOMIC_LONG_INC_DEC_OP(inc, _acquire) ATOMIC_LONG_INC_DEC_OP(inc, _release) ATOMIC_LONG_INC_DEC_OP(dec,) ATOMIC_LONG_INC_DEC_OP(dec, _relaxed) ATOMIC_LONG_INC_DEC_OP(dec, _acquire) ATOMIC_LONG_INC_DEC_OP(dec, _release) #undef ATOMIC_LONG_INC_DEC_OP static inline long atomic_long_add_unless(atomic_long_t *l, long a, long u) { ATOMIC_LONG_PFX(_t) *v = (ATOMIC_LONG_PFX(_t) *)l; return (long)ATOMIC_LONG_PFX(_add_unless)(v, a, u); } #define atomic_long_inc_not_zero(l) \ ATOMIC_LONG_PFX(_inc_not_zero)((ATOMIC_LONG_PFX(_t) *)(l)) #endif /* _ASM_GENERIC_ATOMIC_LONG_H */ d, and the previous affinity did not match the current newly created one, making it appear that the user modified the thread's affinity when it did not, and the thread failed to migrate again. Cc: stable@vger.kernel.org Fixes: 0330f7aa8ee6 ("tracing: Have hwlat trace migrate across tracing_cpumask CPUs") Signed-off-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
Diffstat (limited to 'sound/pci/au88x0')