locking,arch,cris: Fold atomic_ops
authorPeter Zijlstra <peterz@infradead.org>
Sun, 23 Mar 2014 17:19:25 +0000 (18:19 +0100)
committerIngo Molnar <mingo@kernel.org>
Thu, 14 Aug 2014 10:48:06 +0000 (12:48 +0200)
Many of the atomic op implementations are the same except for one
instruction; fold the lot into a few CPP macros and reduce LoC.

This also prepares for easy addition of new ops.

Signed-off-by: Peter Zijlstra <peterz@infradead.org>
Acked-by: Jesper Nilsson <jesper.nilsson@axis.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mikael Starvik <starvik@axis.com>
Cc: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: linux-cris-kernel@axis.com
Link: http://lkml.kernel.org/r/20140508135852.104572724@infradead.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
arch/cris/include/asm/atomic.h

index aa429ba..0033f9d 100644 (file)
 
 /* These should be written in asm but we do it in C for now. */
 
-static inline void atomic_add(int i, volatile atomic_t *v)
-{
-       unsigned long flags;
-       cris_atomic_save(v, flags);
-       v->counter += i;
-       cris_atomic_restore(v, flags);
+#define ATOMIC_OP(op, c_op)                                            \
+static inline void atomic_##op(int i, volatile atomic_t *v)            \
+{                                                                      \
+       unsigned long flags;                                            \
+       cris_atomic_save(v, flags);                                     \
+       v->counter c_op i;                                              \
+       cris_atomic_restore(v, flags);                                  \
+}                                                                      \
+
+#define ATOMIC_OP_RETURN(op, c_op)                                     \
+static inline int atomic_##op##_return(int i, volatile atomic_t *v)    \
+{                                                                      \
+       unsigned long flags;                                            \
+       int retval;                                                     \
+       cris_atomic_save(v, flags);                                     \
+       retval = (v->counter c_op i);                                   \
+       cris_atomic_restore(v, flags);                                  \
+       return retval;                                                  \
 }
 
-static inline void atomic_sub(int i, volatile atomic_t *v)
-{
-       unsigned long flags;
-       cris_atomic_save(v, flags);
-       v->counter -= i;
-       cris_atomic_restore(v, flags);
-}
+#define ATOMIC_OPS(op, c_op) ATOMIC_OP(op, c_op) ATOMIC_OP_RETURN(op, c_op)
 
-static inline int atomic_add_return(int i, volatile atomic_t *v)
-{
-       unsigned long flags;
-       int retval;
-       cris_atomic_save(v, flags);
-       retval = (v->counter += i);
-       cris_atomic_restore(v, flags);
-       return retval;
-}
+ATOMIC_OPS(add, +=)
+ATOMIC_OPS(sub, -=)
 
-#define atomic_add_negative(a, v)      (atomic_add_return((a), (v)) < 0)
+#undef ATOMIC_OPS
+#undef ATOMIC_OP_RETURN
+#undef ATOMIC_OP
 
-static inline int atomic_sub_return(int i, volatile atomic_t *v)
-{
-       unsigned long flags;
-       int retval;
-       cris_atomic_save(v, flags);
-       retval = (v->counter -= i);
-       cris_atomic_restore(v, flags);
-       return retval;
-}
+#define atomic_add_negative(a, v)      (atomic_add_return((a), (v)) < 0)
 
 static inline int atomic_sub_and_test(int i, volatile atomic_t *v)
 {