blob: 0d0d9cdd330986a141dbcc386c3cc4b2e067c2d8 [file] [log] [blame]
Jeremy Fitzhardingee9826382011-08-18 11:48:06 -07001#ifndef ASM_X86_CMPXCHG_H
2#define ASM_X86_CMPXCHG_H
3
4#include <asm/alternative.h> /* Provides LOCK_PREFIX */
5
6/* Non-existant functions to indicate usage errors at link time. */
7extern void __xchg_wrong_size(void);
8extern void __cmpxchg_wrong_size(void);
Jeremy Fitzhardinge433b3522011-06-21 12:00:55 -07009extern void __xadd_wrong_size(void);
Jeremy Fitzhardingee9826382011-08-18 11:48:06 -070010
11/*
12 * Constants for operation sizes. On 32-bit, the 64-bit size it set to
13 * -1 because sizeof will never return -1, thereby making those switch
14 * case statements guaranteeed dead code which the compiler will
15 * eliminate, and allowing the "missing symbol in the default case" to
16 * indicate a usage error.
17 */
18#define __X86_CASE_B 1
19#define __X86_CASE_W 2
20#define __X86_CASE_L 4
21#ifdef CONFIG_64BIT
22#define __X86_CASE_Q 8
23#else
24#define __X86_CASE_Q -1 /* sizeof will never return -1 */
25#endif
26
27/*
28 * Note: no "lock" prefix even on SMP: xchg always implies lock anyway.
29 * Since this is generally used to protect other memory information, we
30 * use "asm volatile" and "memory" clobbers to prevent gcc from moving
31 * information around.
32 */
33#define __xchg(x, ptr, size) \
34({ \
35 __typeof(*(ptr)) __x = (x); \
36 switch (size) { \
37 case __X86_CASE_B: \
38 { \
39 volatile u8 *__ptr = (volatile u8 *)(ptr); \
40 asm volatile("xchgb %0,%1" \
41 : "=q" (__x), "+m" (*__ptr) \
42 : "0" (__x) \
43 : "memory"); \
44 break; \
45 } \
46 case __X86_CASE_W: \
47 { \
48 volatile u16 *__ptr = (volatile u16 *)(ptr); \
49 asm volatile("xchgw %0,%1" \
50 : "=r" (__x), "+m" (*__ptr) \
51 : "0" (__x) \
52 : "memory"); \
53 break; \
54 } \
55 case __X86_CASE_L: \
56 { \
57 volatile u32 *__ptr = (volatile u32 *)(ptr); \
58 asm volatile("xchgl %0,%1" \
59 : "=r" (__x), "+m" (*__ptr) \
60 : "0" (__x) \
61 : "memory"); \
62 break; \
63 } \
64 case __X86_CASE_Q: \
65 { \
66 volatile u64 *__ptr = (volatile u64 *)(ptr); \
67 asm volatile("xchgq %0,%1" \
68 : "=r" (__x), "+m" (*__ptr) \
69 : "0" (__x) \
70 : "memory"); \
71 break; \
72 } \
73 default: \
74 __xchg_wrong_size(); \
75 } \
76 __x; \
77})
78
79#define xchg(ptr, v) \
80 __xchg((v), (ptr), sizeof(*ptr))
81
82/*
83 * Atomic compare and exchange. Compare OLD with MEM, if identical,
84 * store NEW in MEM. Return the initial value in MEM. Success is
85 * indicated by comparing RETURN with OLD.
86 */
87#define __raw_cmpxchg(ptr, old, new, size, lock) \
88({ \
89 __typeof__(*(ptr)) __ret; \
90 __typeof__(*(ptr)) __old = (old); \
91 __typeof__(*(ptr)) __new = (new); \
92 switch (size) { \
93 case __X86_CASE_B: \
94 { \
95 volatile u8 *__ptr = (volatile u8 *)(ptr); \
96 asm volatile(lock "cmpxchgb %2,%1" \
97 : "=a" (__ret), "+m" (*__ptr) \
98 : "q" (__new), "0" (__old) \
99 : "memory"); \
100 break; \
101 } \
102 case __X86_CASE_W: \
103 { \
104 volatile u16 *__ptr = (volatile u16 *)(ptr); \
105 asm volatile(lock "cmpxchgw %2,%1" \
106 : "=a" (__ret), "+m" (*__ptr) \
107 : "r" (__new), "0" (__old) \
108 : "memory"); \
109 break; \
110 } \
111 case __X86_CASE_L: \
112 { \
113 volatile u32 *__ptr = (volatile u32 *)(ptr); \
114 asm volatile(lock "cmpxchgl %2,%1" \
115 : "=a" (__ret), "+m" (*__ptr) \
116 : "r" (__new), "0" (__old) \
117 : "memory"); \
118 break; \
119 } \
120 case __X86_CASE_Q: \
121 { \
122 volatile u64 *__ptr = (volatile u64 *)(ptr); \
123 asm volatile(lock "cmpxchgq %2,%1" \
124 : "=a" (__ret), "+m" (*__ptr) \
125 : "r" (__new), "0" (__old) \
126 : "memory"); \
127 break; \
128 } \
129 default: \
130 __cmpxchg_wrong_size(); \
131 } \
132 __ret; \
133})
134
135#define __cmpxchg(ptr, old, new, size) \
136 __raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX)
137
138#define __sync_cmpxchg(ptr, old, new, size) \
139 __raw_cmpxchg((ptr), (old), (new), (size), "lock; ")
140
141#define __cmpxchg_local(ptr, old, new, size) \
142 __raw_cmpxchg((ptr), (old), (new), (size), "")
143
Thomas Gleixner96a388d2007-10-11 11:20:03 +0200144#ifdef CONFIG_X86_32
145# include "cmpxchg_32.h"
146#else
147# include "cmpxchg_64.h"
148#endif
Jeremy Fitzhardingee9826382011-08-18 11:48:06 -0700149
150#ifdef __HAVE_ARCH_CMPXCHG
151#define cmpxchg(ptr, old, new) \
152 __cmpxchg((ptr), (old), (new), sizeof(*ptr))
153
154#define sync_cmpxchg(ptr, old, new) \
155 __sync_cmpxchg((ptr), (old), (new), sizeof(*ptr))
156
157#define cmpxchg_local(ptr, old, new) \
158 __cmpxchg_local((ptr), (old), (new), sizeof(*ptr))
159#endif
160
Jeremy Fitzhardinge433b3522011-06-21 12:00:55 -0700161#define __xadd(ptr, inc, lock) \
162 ({ \
163 __typeof__ (*(ptr)) __ret = (inc); \
164 switch (sizeof(*(ptr))) { \
165 case __X86_CASE_B: \
166 asm volatile (lock "xaddb %b0, %1\n" \
167 : "+r" (__ret), "+m" (*(ptr)) \
168 : : "memory", "cc"); \
169 break; \
170 case __X86_CASE_W: \
171 asm volatile (lock "xaddw %w0, %1\n" \
172 : "+r" (__ret), "+m" (*(ptr)) \
173 : : "memory", "cc"); \
174 break; \
175 case __X86_CASE_L: \
176 asm volatile (lock "xaddl %0, %1\n" \
177 : "+r" (__ret), "+m" (*(ptr)) \
178 : : "memory", "cc"); \
179 break; \
180 case __X86_CASE_Q: \
181 asm volatile (lock "xaddq %q0, %1\n" \
182 : "+r" (__ret), "+m" (*(ptr)) \
183 : : "memory", "cc"); \
184 break; \
185 default: \
186 __xadd_wrong_size(); \
187 } \
188 __ret; \
189 })
190
191/*
192 * xadd() adds "inc" to "*ptr" and atomically returns the previous
193 * value of "*ptr".
194 *
195 * xadd() is locked when multiple CPUs are online
196 * xadd_sync() is always locked
197 * xadd_local() is never locked
198 */
199#define xadd(ptr, inc) __xadd((ptr), (inc), LOCK_PREFIX)
200#define xadd_sync(ptr, inc) __xadd((ptr), (inc), "lock; ")
201#define xadd_local(ptr, inc) __xadd((ptr), (inc), "")
202
Jeremy Fitzhardingee9826382011-08-18 11:48:06 -0700203#endif /* ASM_X86_CMPXCHG_H */