diff options
author | Jiri Olsa <jolsa@redhat.com> | 2009-07-08 12:10:31 +0000 |
---|---|---|
committer | David S. Miller <davem@davemloft.net> | 2009-07-09 17:06:58 -0700 |
commit | ad46276952f1af34cd91d46d49ba13d347d56367 (patch) | |
tree | 55cf35156794ab34d8a607c25fd044c37231f9e4 /arch/x86 | |
parent | a57de0b4336e48db2811a2030bb68dba8dd09d88 (diff) | |
download | linux-ad46276952f1af34cd91d46d49ba13d347d56367.tar.bz2 |
memory barrier: adding smp_mb__after_lock
Adding smp_mb__after_lock define to be used as a smp_mb call after
a lock.
Making it nop for x86, since {read|write|spin}_lock() on x86 are
full memory barriers.
Signed-off-by: Jiri Olsa <jolsa@redhat.com>
Signed-off-by: Eric Dumazet <eric.dumazet@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'arch/x86')
-rw-r--r-- | arch/x86/include/asm/spinlock.h | 4 |
1 files changed, 4 insertions, 0 deletions
diff --git a/arch/x86/include/asm/spinlock.h b/arch/x86/include/asm/spinlock.h index b7e5db876399..4e77853321db 100644 --- a/arch/x86/include/asm/spinlock.h +++ b/arch/x86/include/asm/spinlock.h @@ -302,4 +302,8 @@ static inline void __raw_write_unlock(raw_rwlock_t *rw) #define _raw_read_relax(lock) cpu_relax() #define _raw_write_relax(lock) cpu_relax() +/* The {read|write|spin}_lock() on x86 are full memory barriers. */ +static inline void smp_mb__after_lock(void) { } +#define ARCH_HAS_SMP_MB_AFTER_LOCK + #endif /* _ASM_X86_SPINLOCK_H */ |