locking/barriers, arch/arm64: Implement LDXR+WFE based smp_cond_load_acquire()
authorWill Deacon <will.deacon@arm.com>
Mon, 27 Jun 2016 17:43:54 +0000 (18:43 +0100)
committerIngo Molnar <mingo@kernel.org>
Thu, 7 Jul 2016 07:10:53 +0000 (09:10 +0200)
smp_cond_load_acquire() is used to spin on a variable until some
expression involving that variable becomes true.

On arm64, we can build this using the LDXR and WFE instructions, since
clearing of the exclusive monitor as a result of the variable being
changed by another CPU generates an event, which will wake us up out of WFE.

This patch implements smp_cond_load_acquire() using LDXR and WFE, which
themselves are contained in an internal __cmpwait() function.

Signed-off-by: Will Deacon <will.deacon@arm.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: catalin.marinas@arm.com
Cc: linux-arm-kernel@lists.infradead.org
Link: http://lkml.kernel.org/r/1467049434-30451-1-git-send-email-will.deacon@arm.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
arch/arm64/include/asm/barrier.h
arch/arm64/include/asm/cmpxchg.h

index dae5c49..4eea7f6 100644 (file)
@@ -91,6 +91,19 @@ do {                                                                 \
        __u.__val;                                                      \
 })
 
+#define smp_cond_load_acquire(ptr, cond_expr)                          \
+({                                                                     \
+       typeof(ptr) __PTR = (ptr);                                      \
+       typeof(*ptr) VAL;                                               \
+       for (;;) {                                                      \
+               VAL = smp_load_acquire(__PTR);                          \
+               if (cond_expr)                                          \
+                       break;                                          \
+               __cmpwait_relaxed(__PTR, VAL);                          \
+       }                                                               \
+       VAL;                                                            \
+})
+
 #include <asm-generic/barrier.h>
 
 #endif /* __ASSEMBLY__ */
index 510c7b4..bd86a79 100644 (file)
@@ -224,4 +224,55 @@ __CMPXCHG_GEN(_mb)
        __ret;                                                          \
 })
 
+#define __CMPWAIT_CASE(w, sz, name)                                    \
+static inline void __cmpwait_case_##name(volatile void *ptr,           \
+                                        unsigned long val)             \
+{                                                                      \
+       unsigned long tmp;                                              \
+                                                                       \
+       asm volatile(                                                   \
+       "       ldxr" #sz "\t%" #w "[tmp], %[v]\n"              \
+       "       eor     %" #w "[tmp], %" #w "[tmp], %" #w "[val]\n"     \
+       "       cbnz    %" #w "[tmp], 1f\n"                             \
+       "       wfe\n"                                                  \
+       "1:"                                                            \
+       : [tmp] "=&r" (tmp), [v] "+Q" (*(unsigned long *)ptr)           \
+       : [val] "r" (val));                                             \
+}
+
+__CMPWAIT_CASE(w, b, 1);
+__CMPWAIT_CASE(w, h, 2);
+__CMPWAIT_CASE(w,  , 4);
+__CMPWAIT_CASE( ,  , 8);
+
+#undef __CMPWAIT_CASE
+
+#define __CMPWAIT_GEN(sfx)                                             \
+static inline void __cmpwait##sfx(volatile void *ptr,                  \
+                                 unsigned long val,                    \
+                                 int size)                             \
+{                                                                      \
+       switch (size) {                                                 \
+       case 1:                                                         \
+               return __cmpwait_case##sfx##_1(ptr, (u8)val);           \
+       case 2:                                                         \
+               return __cmpwait_case##sfx##_2(ptr, (u16)val);          \
+       case 4:                                                         \
+               return __cmpwait_case##sfx##_4(ptr, val);               \
+       case 8:                                                         \
+               return __cmpwait_case##sfx##_8(ptr, val);               \
+       default:                                                        \
+               BUILD_BUG();                                            \
+       }                                                               \
+                                                                       \
+       unreachable();                                                  \
+}
+
+__CMPWAIT_GEN()
+
+#undef __CMPWAIT_GEN
+
+#define __cmpwait_relaxed(ptr, val) \
+       __cmpwait((ptr), (unsigned long)(val), sizeof(*(ptr)))
+
 #endif /* __ASM_CMPXCHG_H */