blob: 49eade13161cd4e4aec0280285d101e542526890 [file] [log] [blame]
Jeremy Fitzhardingee9826382011-08-18 11:48:06 -07001#ifndef ASM_X86_CMPXCHG_H
2#define ASM_X86_CMPXCHG_H
3
Jeremy Fitzhardinge61e2cd02011-08-29 14:47:58 -07004#include <linux/compiler.h>
Jeremy Fitzhardingee9826382011-08-18 11:48:06 -07005#include <asm/alternative.h> /* Provides LOCK_PREFIX */
6
Jeremy Fitzhardinge61e2cd02011-08-29 14:47:58 -07007/*
8 * Non-existant functions to indicate usage errors at link time
9 * (or compile-time if the compiler implements __compiletime_error().
10 */
11extern void __xchg_wrong_size(void)
12 __compiletime_error("Bad argument size for xchg");
13extern void __cmpxchg_wrong_size(void)
14 __compiletime_error("Bad argument size for cmpxchg");
15extern void __xadd_wrong_size(void)
16 __compiletime_error("Bad argument size for xadd");
Jeremy Fitzhardinge3d94ae02011-09-28 11:49:28 -070017extern void __add_wrong_size(void)
18 __compiletime_error("Bad argument size for add");
Jeremy Fitzhardingee9826382011-08-18 11:48:06 -070019
20/*
21 * Constants for operation sizes. On 32-bit, the 64-bit size it set to
22 * -1 because sizeof will never return -1, thereby making those switch
23 * case statements guaranteeed dead code which the compiler will
24 * eliminate, and allowing the "missing symbol in the default case" to
25 * indicate a usage error.
26 */
27#define __X86_CASE_B 1
28#define __X86_CASE_W 2
29#define __X86_CASE_L 4
30#ifdef CONFIG_64BIT
31#define __X86_CASE_Q 8
32#else
33#define __X86_CASE_Q -1 /* sizeof will never return -1 */
34#endif
35
36/*
37 * Note: no "lock" prefix even on SMP: xchg always implies lock anyway.
38 * Since this is generally used to protect other memory information, we
39 * use "asm volatile" and "memory" clobbers to prevent gcc from moving
40 * information around.
41 */
42#define __xchg(x, ptr, size) \
43({ \
44 __typeof(*(ptr)) __x = (x); \
45 switch (size) { \
46 case __X86_CASE_B: \
47 { \
48 volatile u8 *__ptr = (volatile u8 *)(ptr); \
49 asm volatile("xchgb %0,%1" \
50 : "=q" (__x), "+m" (*__ptr) \
51 : "0" (__x) \
52 : "memory"); \
53 break; \
54 } \
55 case __X86_CASE_W: \
56 { \
57 volatile u16 *__ptr = (volatile u16 *)(ptr); \
58 asm volatile("xchgw %0,%1" \
59 : "=r" (__x), "+m" (*__ptr) \
60 : "0" (__x) \
61 : "memory"); \
62 break; \
63 } \
64 case __X86_CASE_L: \
65 { \
66 volatile u32 *__ptr = (volatile u32 *)(ptr); \
67 asm volatile("xchgl %0,%1" \
68 : "=r" (__x), "+m" (*__ptr) \
69 : "0" (__x) \
70 : "memory"); \
71 break; \
72 } \
73 case __X86_CASE_Q: \
74 { \
75 volatile u64 *__ptr = (volatile u64 *)(ptr); \
76 asm volatile("xchgq %0,%1" \
77 : "=r" (__x), "+m" (*__ptr) \
78 : "0" (__x) \
79 : "memory"); \
80 break; \
81 } \
82 default: \
83 __xchg_wrong_size(); \
84 } \
85 __x; \
86})
87
88#define xchg(ptr, v) \
89 __xchg((v), (ptr), sizeof(*ptr))
90
91/*
92 * Atomic compare and exchange. Compare OLD with MEM, if identical,
93 * store NEW in MEM. Return the initial value in MEM. Success is
94 * indicated by comparing RETURN with OLD.
95 */
96#define __raw_cmpxchg(ptr, old, new, size, lock) \
97({ \
98 __typeof__(*(ptr)) __ret; \
99 __typeof__(*(ptr)) __old = (old); \
100 __typeof__(*(ptr)) __new = (new); \
101 switch (size) { \
102 case __X86_CASE_B: \
103 { \
104 volatile u8 *__ptr = (volatile u8 *)(ptr); \
105 asm volatile(lock "cmpxchgb %2,%1" \
106 : "=a" (__ret), "+m" (*__ptr) \
107 : "q" (__new), "0" (__old) \
108 : "memory"); \
109 break; \
110 } \
111 case __X86_CASE_W: \
112 { \
113 volatile u16 *__ptr = (volatile u16 *)(ptr); \
114 asm volatile(lock "cmpxchgw %2,%1" \
115 : "=a" (__ret), "+m" (*__ptr) \
116 : "r" (__new), "0" (__old) \
117 : "memory"); \
118 break; \
119 } \
120 case __X86_CASE_L: \
121 { \
122 volatile u32 *__ptr = (volatile u32 *)(ptr); \
123 asm volatile(lock "cmpxchgl %2,%1" \
124 : "=a" (__ret), "+m" (*__ptr) \
125 : "r" (__new), "0" (__old) \
126 : "memory"); \
127 break; \
128 } \
129 case __X86_CASE_Q: \
130 { \
131 volatile u64 *__ptr = (volatile u64 *)(ptr); \
132 asm volatile(lock "cmpxchgq %2,%1" \
133 : "=a" (__ret), "+m" (*__ptr) \
134 : "r" (__new), "0" (__old) \
135 : "memory"); \
136 break; \
137 } \
138 default: \
139 __cmpxchg_wrong_size(); \
140 } \
141 __ret; \
142})
143
144#define __cmpxchg(ptr, old, new, size) \
145 __raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX)
146
147#define __sync_cmpxchg(ptr, old, new, size) \
148 __raw_cmpxchg((ptr), (old), (new), (size), "lock; ")
149
150#define __cmpxchg_local(ptr, old, new, size) \
151 __raw_cmpxchg((ptr), (old), (new), (size), "")
152
Thomas Gleixner96a388d2007-10-11 11:20:03 +0200153#ifdef CONFIG_X86_32
154# include "cmpxchg_32.h"
155#else
156# include "cmpxchg_64.h"
157#endif
Jeremy Fitzhardingee9826382011-08-18 11:48:06 -0700158
159#ifdef __HAVE_ARCH_CMPXCHG
160#define cmpxchg(ptr, old, new) \
161 __cmpxchg((ptr), (old), (new), sizeof(*ptr))
162
163#define sync_cmpxchg(ptr, old, new) \
164 __sync_cmpxchg((ptr), (old), (new), sizeof(*ptr))
165
166#define cmpxchg_local(ptr, old, new) \
167 __cmpxchg_local((ptr), (old), (new), sizeof(*ptr))
168#endif
169
Jeremy Fitzhardinge433b3522011-06-21 12:00:55 -0700170#define __xadd(ptr, inc, lock) \
171 ({ \
172 __typeof__ (*(ptr)) __ret = (inc); \
173 switch (sizeof(*(ptr))) { \
174 case __X86_CASE_B: \
175 asm volatile (lock "xaddb %b0, %1\n" \
176 : "+r" (__ret), "+m" (*(ptr)) \
177 : : "memory", "cc"); \
178 break; \
179 case __X86_CASE_W: \
180 asm volatile (lock "xaddw %w0, %1\n" \
181 : "+r" (__ret), "+m" (*(ptr)) \
182 : : "memory", "cc"); \
183 break; \
184 case __X86_CASE_L: \
185 asm volatile (lock "xaddl %0, %1\n" \
186 : "+r" (__ret), "+m" (*(ptr)) \
187 : : "memory", "cc"); \
188 break; \
189 case __X86_CASE_Q: \
190 asm volatile (lock "xaddq %q0, %1\n" \
191 : "+r" (__ret), "+m" (*(ptr)) \
192 : : "memory", "cc"); \
193 break; \
194 default: \
195 __xadd_wrong_size(); \
196 } \
197 __ret; \
198 })
199
200/*
201 * xadd() adds "inc" to "*ptr" and atomically returns the previous
202 * value of "*ptr".
203 *
204 * xadd() is locked when multiple CPUs are online
205 * xadd_sync() is always locked
206 * xadd_local() is never locked
207 */
208#define xadd(ptr, inc) __xadd((ptr), (inc), LOCK_PREFIX)
209#define xadd_sync(ptr, inc) __xadd((ptr), (inc), "lock; ")
210#define xadd_local(ptr, inc) __xadd((ptr), (inc), "")
211
Jeremy Fitzhardinge3d94ae02011-09-28 11:49:28 -0700212#define __add(ptr, inc, lock) \
213 ({ \
214 __typeof__ (*(ptr)) __ret = (inc); \
215 switch (sizeof(*(ptr))) { \
216 case __X86_CASE_B: \
217 asm volatile (lock "addb %b1, %0\n" \
218 : "+m" (*(ptr)) : "ri" (inc) \
219 : "memory", "cc"); \
220 break; \
221 case __X86_CASE_W: \
222 asm volatile (lock "addw %w1, %0\n" \
223 : "+m" (*(ptr)) : "ri" (inc) \
224 : "memory", "cc"); \
225 break; \
226 case __X86_CASE_L: \
227 asm volatile (lock "addl %1, %0\n" \
228 : "+m" (*(ptr)) : "ri" (inc) \
229 : "memory", "cc"); \
230 break; \
231 case __X86_CASE_Q: \
232 asm volatile (lock "addq %1, %0\n" \
233 : "+m" (*(ptr)) : "ri" (inc) \
234 : "memory", "cc"); \
235 break; \
236 default: \
237 __add_wrong_size(); \
238 } \
239 __ret; \
240 })
241
242/*
243 * add_*() adds "inc" to "*ptr"
244 *
245 * __add() takes a lock prefix
246 * add_smp() is locked when multiple CPUs are online
247 * add_sync() is always locked
248 */
249#define add_smp(ptr, inc) __add((ptr), (inc), LOCK_PREFIX)
250#define add_sync(ptr, inc) __add((ptr), (inc), "lock; ")
251
Jeremy Fitzhardingee9826382011-08-18 11:48:06 -0700252#endif /* ASM_X86_CMPXCHG_H */