Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1 | /* |
| 2 | * This file is subject to the terms and conditions of the GNU General Public |
| 3 | * License. See the file "COPYING" in the main directory of this archive |
| 4 | * for more details. |
| 5 | * |
| 6 | * Copyright (C) 1996, 1997, 1998, 1999, 2000, 03, 04 by Ralf Baechle |
| 7 | * Copyright (C) 1999, 2000 Silicon Graphics, Inc. |
Maciej W. Rozycki | 619b6e1 | 2007-10-23 12:43:25 +0100 | [diff] [blame] | 8 | * Copyright (C) 2007 Maciej W. Rozycki |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 9 | * Copyright (C) 2014, Imagination Technologies Ltd. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 10 | */ |
| 11 | #ifndef _ASM_UACCESS_H |
| 12 | #define _ASM_UACCESS_H |
| 13 | |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 14 | #include <linux/kernel.h> |
| 15 | #include <linux/errno.h> |
| 16 | #include <linux/thread_info.h> |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 17 | #include <asm/asm-eva.h> |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 18 | |
| 19 | /* |
| 20 | * The fs value determines whether argument validity checking should be |
| 21 | * performed or not. If get_fs() == USER_DS, checking is performed, with |
| 22 | * get_fs() == KERNEL_DS, checking is bypassed. |
| 23 | * |
| 24 | * For historical reasons, these macros are grossly misnamed. |
| 25 | */ |
Ralf Baechle | 875d43e | 2005-09-03 15:56:16 -0700 | [diff] [blame] | 26 | #ifdef CONFIG_32BIT |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 27 | |
Sanjay Lal | 9843b03 | 2012-11-21 18:34:03 -0800 | [diff] [blame] | 28 | #ifdef CONFIG_KVM_GUEST |
| 29 | #define __UA_LIMIT 0x40000000UL |
| 30 | #else |
| 31 | #define __UA_LIMIT 0x80000000UL |
| 32 | #endif |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 33 | |
| 34 | #define __UA_ADDR ".word" |
| 35 | #define __UA_LA "la" |
| 36 | #define __UA_ADDU "addu" |
| 37 | #define __UA_t0 "$8" |
| 38 | #define __UA_t1 "$9" |
| 39 | |
Ralf Baechle | 875d43e | 2005-09-03 15:56:16 -0700 | [diff] [blame] | 40 | #endif /* CONFIG_32BIT */ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 41 | |
Ralf Baechle | 875d43e | 2005-09-03 15:56:16 -0700 | [diff] [blame] | 42 | #ifdef CONFIG_64BIT |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 43 | |
David Daney | 949e51b | 2010-10-14 11:32:33 -0700 | [diff] [blame] | 44 | extern u64 __ua_limit; |
| 45 | |
| 46 | #define __UA_LIMIT __ua_limit |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 47 | |
| 48 | #define __UA_ADDR ".dword" |
| 49 | #define __UA_LA "dla" |
| 50 | #define __UA_ADDU "daddu" |
| 51 | #define __UA_t0 "$12" |
| 52 | #define __UA_t1 "$13" |
| 53 | |
Ralf Baechle | 875d43e | 2005-09-03 15:56:16 -0700 | [diff] [blame] | 54 | #endif /* CONFIG_64BIT */ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 55 | |
| 56 | /* |
| 57 | * USER_DS is a bitmask that has the bits set that may not be set in a valid |
| 58 | * userspace address. Note that we limit 32-bit userspace to 0x7fff8000 but |
| 59 | * the arithmetic we're doing only works if the limit is a power of two, so |
| 60 | * we use 0x80000000 here on 32-bit kernels. If a process passes an invalid |
| 61 | * address in this range it's the process's problem, not ours :-) |
| 62 | */ |
| 63 | |
Sanjay Lal | 9843b03 | 2012-11-21 18:34:03 -0800 | [diff] [blame] | 64 | #ifdef CONFIG_KVM_GUEST |
| 65 | #define KERNEL_DS ((mm_segment_t) { 0x80000000UL }) |
| 66 | #define USER_DS ((mm_segment_t) { 0xC0000000UL }) |
| 67 | #else |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 68 | #define KERNEL_DS ((mm_segment_t) { 0UL }) |
| 69 | #define USER_DS ((mm_segment_t) { __UA_LIMIT }) |
Sanjay Lal | 9843b03 | 2012-11-21 18:34:03 -0800 | [diff] [blame] | 70 | #endif |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 71 | |
| 72 | #define VERIFY_READ 0 |
| 73 | #define VERIFY_WRITE 1 |
| 74 | |
| 75 | #define get_ds() (KERNEL_DS) |
| 76 | #define get_fs() (current_thread_info()->addr_limit) |
| 77 | #define set_fs(x) (current_thread_info()->addr_limit = (x)) |
| 78 | |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 79 | #define segment_eq(a, b) ((a).seg == (b).seg) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 80 | |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 81 | /* |
| 82 | * eva_kernel_access() - determine whether kernel memory access on an EVA system |
| 83 | * |
| 84 | * Determines whether memory accesses should be performed to kernel memory |
| 85 | * on a system using Extended Virtual Addressing (EVA). |
| 86 | * |
| 87 | * Return: true if a kernel memory access on an EVA system, else false. |
| 88 | */ |
| 89 | static inline bool eva_kernel_access(void) |
| 90 | { |
Masahiro Yamada | 97f2645 | 2016-08-03 13:45:50 -0700 | [diff] [blame] | 91 | if (!IS_ENABLED(CONFIG_EVA)) |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 92 | return false; |
| 93 | |
| 94 | return segment_eq(get_fs(), get_ds()); |
| 95 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 96 | |
| 97 | /* |
Adam Buchbinder | 92a76f6 | 2016-02-25 00:44:58 -0800 | [diff] [blame] | 98 | * Is a address valid? This does a straightforward calculation rather |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 99 | * than tests. |
| 100 | * |
| 101 | * Address valid if: |
| 102 | * - "addr" doesn't have any high-bits set |
| 103 | * - AND "size" doesn't have any high-bits set |
| 104 | * - AND "addr+size" doesn't have any high-bits set |
| 105 | * - OR we are in kernel mode. |
| 106 | * |
| 107 | * __ua_size() is a trick to avoid runtime checking of positive constant |
| 108 | * sizes; for those we already know at compile time that the size is ok. |
| 109 | */ |
| 110 | #define __ua_size(size) \ |
| 111 | ((__builtin_constant_p(size) && (signed long) (size) > 0) ? 0 : (size)) |
| 112 | |
| 113 | /* |
| 114 | * access_ok: - Checks if a user space pointer is valid |
| 115 | * @type: Type of access: %VERIFY_READ or %VERIFY_WRITE. Note that |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 116 | * %VERIFY_WRITE is a superset of %VERIFY_READ - if it is safe |
| 117 | * to write to a block, it is always safe to read from it. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 118 | * @addr: User space pointer to start of block to check |
| 119 | * @size: Size of block to check |
| 120 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 121 | * Context: User context only. This function may sleep if pagefaults are |
| 122 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 123 | * |
| 124 | * Checks if a pointer to a block of memory in user space is valid. |
| 125 | * |
| 126 | * Returns true (nonzero) if the memory block may be valid, false (zero) |
| 127 | * if it is definitely invalid. |
| 128 | * |
| 129 | * Note that, depending on architecture, this function probably just |
| 130 | * checks that the pointer is in the user space range - after calling |
| 131 | * this function, memory access functions may still return -EFAULT. |
| 132 | */ |
| 133 | |
| 134 | #define __access_mask get_fs().seg |
| 135 | |
Ralf Baechle | ed01b3d | 2009-04-27 16:46:21 +0200 | [diff] [blame] | 136 | #define __access_ok(addr, size, mask) \ |
| 137 | ({ \ |
| 138 | unsigned long __addr = (unsigned long) (addr); \ |
| 139 | unsigned long __size = size; \ |
| 140 | unsigned long __mask = mask; \ |
| 141 | unsigned long __ok; \ |
| 142 | \ |
| 143 | __chk_user_ptr(addr); \ |
| 144 | __ok = (signed long)(__mask & (__addr | (__addr + __size) | \ |
| 145 | __ua_size(__size))); \ |
| 146 | __ok == 0; \ |
Ralf Baechle | d0aab92 | 2009-04-27 15:31:34 +0200 | [diff] [blame] | 147 | }) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 148 | |
| 149 | #define access_ok(type, addr, size) \ |
Ralf Baechle | d0aab92 | 2009-04-27 15:31:34 +0200 | [diff] [blame] | 150 | likely(__access_ok((addr), (size), __access_mask)) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 151 | |
| 152 | /* |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 153 | * put_user: - Write a simple value into user space. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 154 | * @x: Value to copy to user space. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 155 | * @ptr: Destination address, in user space. |
| 156 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 157 | * Context: User context only. This function may sleep if pagefaults are |
| 158 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 159 | * |
| 160 | * This macro copies a single simple value from kernel space to user |
| 161 | * space. It supports simple types like char and int, but not larger |
| 162 | * data types like structures or arrays. |
| 163 | * |
| 164 | * @ptr must have pointer-to-simple-variable type, and @x must be assignable |
| 165 | * to the result of dereferencing @ptr. |
| 166 | * |
| 167 | * Returns zero on success, or -EFAULT on error. |
| 168 | */ |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 169 | #define put_user(x,ptr) \ |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 170 | __put_user_check((x), (ptr), sizeof(*(ptr))) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 171 | |
| 172 | /* |
| 173 | * get_user: - Get a simple variable from user space. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 174 | * @x: Variable to store result. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 175 | * @ptr: Source address, in user space. |
| 176 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 177 | * Context: User context only. This function may sleep if pagefaults are |
| 178 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 179 | * |
| 180 | * This macro copies a single simple variable from user space to kernel |
| 181 | * space. It supports simple types like char and int, but not larger |
| 182 | * data types like structures or arrays. |
| 183 | * |
| 184 | * @ptr must have pointer-to-simple-variable type, and the result of |
| 185 | * dereferencing @ptr must be assignable to @x without a cast. |
| 186 | * |
| 187 | * Returns zero on success, or -EFAULT on error. |
| 188 | * On error, the variable @x is set to zero. |
| 189 | */ |
| 190 | #define get_user(x,ptr) \ |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 191 | __get_user_check((x), (ptr), sizeof(*(ptr))) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 192 | |
| 193 | /* |
| 194 | * __put_user: - Write a simple value into user space, with less checking. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 195 | * @x: Value to copy to user space. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 196 | * @ptr: Destination address, in user space. |
| 197 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 198 | * Context: User context only. This function may sleep if pagefaults are |
| 199 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 200 | * |
| 201 | * This macro copies a single simple value from kernel space to user |
| 202 | * space. It supports simple types like char and int, but not larger |
| 203 | * data types like structures or arrays. |
| 204 | * |
| 205 | * @ptr must have pointer-to-simple-variable type, and @x must be assignable |
| 206 | * to the result of dereferencing @ptr. |
| 207 | * |
| 208 | * Caller must check the pointer with access_ok() before calling this |
| 209 | * function. |
| 210 | * |
| 211 | * Returns zero on success, or -EFAULT on error. |
| 212 | */ |
| 213 | #define __put_user(x,ptr) \ |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 214 | __put_user_nocheck((x), (ptr), sizeof(*(ptr))) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 215 | |
| 216 | /* |
| 217 | * __get_user: - Get a simple variable from user space, with less checking. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 218 | * @x: Variable to store result. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 219 | * @ptr: Source address, in user space. |
| 220 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 221 | * Context: User context only. This function may sleep if pagefaults are |
| 222 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 223 | * |
| 224 | * This macro copies a single simple variable from user space to kernel |
| 225 | * space. It supports simple types like char and int, but not larger |
| 226 | * data types like structures or arrays. |
| 227 | * |
| 228 | * @ptr must have pointer-to-simple-variable type, and the result of |
| 229 | * dereferencing @ptr must be assignable to @x without a cast. |
| 230 | * |
| 231 | * Caller must check the pointer with access_ok() before calling this |
| 232 | * function. |
| 233 | * |
| 234 | * Returns zero on success, or -EFAULT on error. |
| 235 | * On error, the variable @x is set to zero. |
| 236 | */ |
| 237 | #define __get_user(x,ptr) \ |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 238 | __get_user_nocheck((x), (ptr), sizeof(*(ptr))) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 239 | |
| 240 | struct __large_struct { unsigned long buf[100]; }; |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 241 | #define __m(x) (*(struct __large_struct __user *)(x)) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 242 | |
| 243 | /* |
| 244 | * Yuck. We need two variants, one for 64bit operation and one |
| 245 | * for 32 bit mode and old iron. |
| 246 | */ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 247 | #ifndef CONFIG_EVA |
| 248 | #define __get_kernel_common(val, size, ptr) __get_user_common(val, size, ptr) |
| 249 | #else |
| 250 | /* |
| 251 | * Kernel specific functions for EVA. We need to use normal load instructions |
| 252 | * to read data from kernel when operating in EVA mode. We use these macros to |
| 253 | * avoid redefining __get_user_asm for EVA. |
| 254 | */ |
| 255 | #undef _loadd |
| 256 | #undef _loadw |
| 257 | #undef _loadh |
| 258 | #undef _loadb |
| 259 | #ifdef CONFIG_32BIT |
| 260 | #define _loadd _loadw |
| 261 | #else |
| 262 | #define _loadd(reg, addr) "ld " reg ", " addr |
| 263 | #endif |
| 264 | #define _loadw(reg, addr) "lw " reg ", " addr |
| 265 | #define _loadh(reg, addr) "lh " reg ", " addr |
| 266 | #define _loadb(reg, addr) "lb " reg ", " addr |
| 267 | |
| 268 | #define __get_kernel_common(val, size, ptr) \ |
| 269 | do { \ |
| 270 | switch (size) { \ |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 271 | case 1: __get_data_asm(val, _loadb, ptr); break; \ |
| 272 | case 2: __get_data_asm(val, _loadh, ptr); break; \ |
| 273 | case 4: __get_data_asm(val, _loadw, ptr); break; \ |
| 274 | case 8: __GET_DW(val, _loadd, ptr); break; \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 275 | default: __get_user_unknown(); break; \ |
| 276 | } \ |
| 277 | } while (0) |
| 278 | #endif |
| 279 | |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 280 | #ifdef CONFIG_32BIT |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 281 | #define __GET_DW(val, insn, ptr) __get_data_asm_ll32(val, insn, ptr) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 282 | #endif |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 283 | #ifdef CONFIG_64BIT |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 284 | #define __GET_DW(val, insn, ptr) __get_data_asm(val, insn, ptr) |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 285 | #endif |
| 286 | |
| 287 | extern void __get_user_unknown(void); |
| 288 | |
| 289 | #define __get_user_common(val, size, ptr) \ |
| 290 | do { \ |
| 291 | switch (size) { \ |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 292 | case 1: __get_data_asm(val, user_lb, ptr); break; \ |
| 293 | case 2: __get_data_asm(val, user_lh, ptr); break; \ |
| 294 | case 4: __get_data_asm(val, user_lw, ptr); break; \ |
| 295 | case 8: __GET_DW(val, user_ld, ptr); break; \ |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 296 | default: __get_user_unknown(); break; \ |
| 297 | } \ |
| 298 | } while (0) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 299 | |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 300 | #define __get_user_nocheck(x, ptr, size) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 301 | ({ \ |
Ralf Baechle | 8d2d91e | 2008-10-11 16:18:50 +0100 | [diff] [blame] | 302 | int __gu_err; \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 303 | \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 304 | if (eva_kernel_access()) { \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 305 | __get_kernel_common((x), size, ptr); \ |
| 306 | } else { \ |
| 307 | __chk_user_ptr(ptr); \ |
| 308 | __get_user_common((x), size, ptr); \ |
| 309 | } \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 310 | __gu_err; \ |
| 311 | }) |
| 312 | |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 313 | #define __get_user_check(x, ptr, size) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 314 | ({ \ |
Ralf Baechle | 8d2d91e | 2008-10-11 16:18:50 +0100 | [diff] [blame] | 315 | int __gu_err = -EFAULT; \ |
Atsushi Nemoto | 8ecbbca | 2006-02-14 15:57:50 +0900 | [diff] [blame] | 316 | const __typeof__(*(ptr)) __user * __gu_ptr = (ptr); \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 317 | \ |
Ralf Baechle | ef41f46 | 2009-04-28 14:17:54 +0200 | [diff] [blame] | 318 | might_fault(); \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 319 | if (likely(access_ok(VERIFY_READ, __gu_ptr, size))) { \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 320 | if (eva_kernel_access()) \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 321 | __get_kernel_common((x), size, __gu_ptr); \ |
| 322 | else \ |
| 323 | __get_user_common((x), size, __gu_ptr); \ |
Ralf Baechle | 640465b | 2014-11-18 18:47:13 +0100 | [diff] [blame] | 324 | } else \ |
| 325 | (x) = 0; \ |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 326 | \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 327 | __gu_err; \ |
| 328 | }) |
| 329 | |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 330 | #define __get_data_asm(val, insn, addr) \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 331 | { \ |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 332 | long __gu_tmp; \ |
| 333 | \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 334 | __asm__ __volatile__( \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 335 | "1: "insn("%1", "%3")" \n" \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 336 | "2: \n" \ |
Steven J. Hill | 1658f91 | 2013-03-25 13:22:59 -0500 | [diff] [blame] | 337 | " .insn \n" \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 338 | " .section .fixup,\"ax\" \n" \ |
| 339 | "3: li %0, %4 \n" \ |
Ralf Baechle | 640465b | 2014-11-18 18:47:13 +0100 | [diff] [blame] | 340 | " move %1, $0 \n" \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 341 | " j 2b \n" \ |
| 342 | " .previous \n" \ |
| 343 | " .section __ex_table,\"a\" \n" \ |
| 344 | " "__UA_ADDR "\t1b, 3b \n" \ |
| 345 | " .previous \n" \ |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 346 | : "=r" (__gu_err), "=r" (__gu_tmp) \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 347 | : "0" (0), "o" (__m(addr)), "i" (-EFAULT)); \ |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 348 | \ |
Atsushi Nemoto | 8ecbbca | 2006-02-14 15:57:50 +0900 | [diff] [blame] | 349 | (val) = (__typeof__(*(addr))) __gu_tmp; \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 350 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 351 | |
| 352 | /* |
| 353 | * Get a long long 64 using 32 bit registers. |
| 354 | */ |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 355 | #define __get_data_asm_ll32(val, insn, addr) \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 356 | { \ |
Ralf Baechle | cb66fb3 | 2007-02-13 11:45:24 +0000 | [diff] [blame] | 357 | union { \ |
| 358 | unsigned long long l; \ |
| 359 | __typeof__(*(addr)) t; \ |
| 360 | } __gu_tmp; \ |
Ralf Baechle | cd1fb9e | 2007-02-12 23:12:38 +0000 | [diff] [blame] | 361 | \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 362 | __asm__ __volatile__( \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 363 | "1: " insn("%1", "(%3)")" \n" \ |
| 364 | "2: " insn("%D1", "4(%3)")" \n" \ |
Steven J. Hill | 1658f91 | 2013-03-25 13:22:59 -0500 | [diff] [blame] | 365 | "3: \n" \ |
| 366 | " .insn \n" \ |
| 367 | " .section .fixup,\"ax\" \n" \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 368 | "4: li %0, %4 \n" \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 369 | " move %1, $0 \n" \ |
| 370 | " move %D1, $0 \n" \ |
| 371 | " j 3b \n" \ |
| 372 | " .previous \n" \ |
| 373 | " .section __ex_table,\"a\" \n" \ |
| 374 | " " __UA_ADDR " 1b, 4b \n" \ |
| 375 | " " __UA_ADDR " 2b, 4b \n" \ |
| 376 | " .previous \n" \ |
Ralf Baechle | cb66fb3 | 2007-02-13 11:45:24 +0000 | [diff] [blame] | 377 | : "=r" (__gu_err), "=&r" (__gu_tmp.l) \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 378 | : "0" (0), "r" (addr), "i" (-EFAULT)); \ |
Ralf Baechle | cb66fb3 | 2007-02-13 11:45:24 +0000 | [diff] [blame] | 379 | \ |
| 380 | (val) = __gu_tmp.t; \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 381 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 382 | |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 383 | #ifndef CONFIG_EVA |
| 384 | #define __put_kernel_common(ptr, size) __put_user_common(ptr, size) |
| 385 | #else |
| 386 | /* |
| 387 | * Kernel specific functions for EVA. We need to use normal load instructions |
| 388 | * to read data from kernel when operating in EVA mode. We use these macros to |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 389 | * avoid redefining __get_data_asm for EVA. |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 390 | */ |
| 391 | #undef _stored |
| 392 | #undef _storew |
| 393 | #undef _storeh |
| 394 | #undef _storeb |
| 395 | #ifdef CONFIG_32BIT |
| 396 | #define _stored _storew |
| 397 | #else |
| 398 | #define _stored(reg, addr) "ld " reg ", " addr |
| 399 | #endif |
| 400 | |
| 401 | #define _storew(reg, addr) "sw " reg ", " addr |
| 402 | #define _storeh(reg, addr) "sh " reg ", " addr |
| 403 | #define _storeb(reg, addr) "sb " reg ", " addr |
| 404 | |
| 405 | #define __put_kernel_common(ptr, size) \ |
| 406 | do { \ |
| 407 | switch (size) { \ |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 408 | case 1: __put_data_asm(_storeb, ptr); break; \ |
| 409 | case 2: __put_data_asm(_storeh, ptr); break; \ |
| 410 | case 4: __put_data_asm(_storew, ptr); break; \ |
| 411 | case 8: __PUT_DW(_stored, ptr); break; \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 412 | default: __put_user_unknown(); break; \ |
| 413 | } \ |
| 414 | } while(0) |
| 415 | #endif |
| 416 | |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 417 | /* |
| 418 | * Yuck. We need two variants, one for 64bit operation and one |
| 419 | * for 32 bit mode and old iron. |
| 420 | */ |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 421 | #ifdef CONFIG_32BIT |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 422 | #define __PUT_DW(insn, ptr) __put_data_asm_ll32(insn, ptr) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 423 | #endif |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 424 | #ifdef CONFIG_64BIT |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 425 | #define __PUT_DW(insn, ptr) __put_data_asm(insn, ptr) |
Ralf Baechle | 4feb8f8 | 2006-01-23 16:15:30 +0000 | [diff] [blame] | 426 | #endif |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 427 | |
Markos Chandras | ec56b1d | 2013-12-17 14:42:23 +0000 | [diff] [blame] | 428 | #define __put_user_common(ptr, size) \ |
| 429 | do { \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 430 | switch (size) { \ |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 431 | case 1: __put_data_asm(user_sb, ptr); break; \ |
| 432 | case 2: __put_data_asm(user_sh, ptr); break; \ |
| 433 | case 4: __put_data_asm(user_sw, ptr); break; \ |
| 434 | case 8: __PUT_DW(user_sd, ptr); break; \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 435 | default: __put_user_unknown(); break; \ |
| 436 | } \ |
Markos Chandras | ec56b1d | 2013-12-17 14:42:23 +0000 | [diff] [blame] | 437 | } while (0) |
| 438 | |
| 439 | #define __put_user_nocheck(x, ptr, size) \ |
| 440 | ({ \ |
| 441 | __typeof__(*(ptr)) __pu_val; \ |
| 442 | int __pu_err = 0; \ |
| 443 | \ |
Markos Chandras | ec56b1d | 2013-12-17 14:42:23 +0000 | [diff] [blame] | 444 | __pu_val = (x); \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 445 | if (eva_kernel_access()) { \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 446 | __put_kernel_common(ptr, size); \ |
| 447 | } else { \ |
| 448 | __chk_user_ptr(ptr); \ |
| 449 | __put_user_common(ptr, size); \ |
| 450 | } \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 451 | __pu_err; \ |
| 452 | }) |
| 453 | |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 454 | #define __put_user_check(x, ptr, size) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 455 | ({ \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 456 | __typeof__(*(ptr)) __user *__pu_addr = (ptr); \ |
| 457 | __typeof__(*(ptr)) __pu_val = (x); \ |
Ralf Baechle | 8d2d91e | 2008-10-11 16:18:50 +0100 | [diff] [blame] | 458 | int __pu_err = -EFAULT; \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 459 | \ |
Ralf Baechle | ef41f46 | 2009-04-28 14:17:54 +0200 | [diff] [blame] | 460 | might_fault(); \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 461 | if (likely(access_ok(VERIFY_WRITE, __pu_addr, size))) { \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 462 | if (eva_kernel_access()) \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 463 | __put_kernel_common(__pu_addr, size); \ |
| 464 | else \ |
| 465 | __put_user_common(__pu_addr, size); \ |
| 466 | } \ |
Markos Chandras | ec56b1d | 2013-12-17 14:42:23 +0000 | [diff] [blame] | 467 | \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 468 | __pu_err; \ |
| 469 | }) |
| 470 | |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 471 | #define __put_data_asm(insn, ptr) \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 472 | { \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 473 | __asm__ __volatile__( \ |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 474 | "1: "insn("%z2", "%3")" # __put_data_asm \n" \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 475 | "2: \n" \ |
Steven J. Hill | 1658f91 | 2013-03-25 13:22:59 -0500 | [diff] [blame] | 476 | " .insn \n" \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 477 | " .section .fixup,\"ax\" \n" \ |
| 478 | "3: li %0, %4 \n" \ |
| 479 | " j 2b \n" \ |
| 480 | " .previous \n" \ |
| 481 | " .section __ex_table,\"a\" \n" \ |
| 482 | " " __UA_ADDR " 1b, 3b \n" \ |
| 483 | " .previous \n" \ |
| 484 | : "=r" (__pu_err) \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 485 | : "0" (0), "Jr" (__pu_val), "o" (__m(ptr)), \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 486 | "i" (-EFAULT)); \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 487 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 488 | |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 489 | #define __put_data_asm_ll32(insn, ptr) \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 490 | { \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 491 | __asm__ __volatile__( \ |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 492 | "1: "insn("%2", "(%3)")" # __put_data_asm_ll32 \n" \ |
Markos Chandras | ac1d859 | 2013-12-11 11:25:33 +0000 | [diff] [blame] | 493 | "2: "insn("%D2", "4(%3)")" \n" \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 494 | "3: \n" \ |
Steven J. Hill | 1658f91 | 2013-03-25 13:22:59 -0500 | [diff] [blame] | 495 | " .insn \n" \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 496 | " .section .fixup,\"ax\" \n" \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 497 | "4: li %0, %4 \n" \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 498 | " j 3b \n" \ |
| 499 | " .previous \n" \ |
| 500 | " .section __ex_table,\"a\" \n" \ |
| 501 | " " __UA_ADDR " 1b, 4b \n" \ |
| 502 | " " __UA_ADDR " 2b, 4b \n" \ |
| 503 | " .previous" \ |
| 504 | : "=r" (__pu_err) \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 505 | : "0" (0), "r" (__pu_val), "r" (ptr), \ |
| 506 | "i" (-EFAULT)); \ |
| 507 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 508 | |
| 509 | extern void __put_user_unknown(void); |
| 510 | |
| 511 | /* |
Leonid Yegoshin | 18e9001 | 2013-12-17 15:20:24 +0000 | [diff] [blame] | 512 | * ul{b,h,w} are macros and there are no equivalent macros for EVA. |
| 513 | * EVA unaligned access is handled in the ADE exception handler. |
| 514 | */ |
| 515 | #ifndef CONFIG_EVA |
| 516 | /* |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 517 | * put_user_unaligned: - Write a simple value into user space. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 518 | * @x: Value to copy to user space. |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 519 | * @ptr: Destination address, in user space. |
| 520 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 521 | * Context: User context only. This function may sleep if pagefaults are |
| 522 | * enabled. |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 523 | * |
| 524 | * This macro copies a single simple value from kernel space to user |
| 525 | * space. It supports simple types like char and int, but not larger |
| 526 | * data types like structures or arrays. |
| 527 | * |
| 528 | * @ptr must have pointer-to-simple-variable type, and @x must be assignable |
| 529 | * to the result of dereferencing @ptr. |
| 530 | * |
| 531 | * Returns zero on success, or -EFAULT on error. |
| 532 | */ |
| 533 | #define put_user_unaligned(x,ptr) \ |
| 534 | __put_user_unaligned_check((x),(ptr),sizeof(*(ptr))) |
| 535 | |
| 536 | /* |
| 537 | * get_user_unaligned: - Get a simple variable from user space. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 538 | * @x: Variable to store result. |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 539 | * @ptr: Source address, in user space. |
| 540 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 541 | * Context: User context only. This function may sleep if pagefaults are |
| 542 | * enabled. |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 543 | * |
| 544 | * This macro copies a single simple variable from user space to kernel |
| 545 | * space. It supports simple types like char and int, but not larger |
| 546 | * data types like structures or arrays. |
| 547 | * |
| 548 | * @ptr must have pointer-to-simple-variable type, and the result of |
| 549 | * dereferencing @ptr must be assignable to @x without a cast. |
| 550 | * |
| 551 | * Returns zero on success, or -EFAULT on error. |
| 552 | * On error, the variable @x is set to zero. |
| 553 | */ |
| 554 | #define get_user_unaligned(x,ptr) \ |
| 555 | __get_user_unaligned_check((x),(ptr),sizeof(*(ptr))) |
| 556 | |
| 557 | /* |
| 558 | * __put_user_unaligned: - Write a simple value into user space, with less checking. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 559 | * @x: Value to copy to user space. |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 560 | * @ptr: Destination address, in user space. |
| 561 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 562 | * Context: User context only. This function may sleep if pagefaults are |
| 563 | * enabled. |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 564 | * |
| 565 | * This macro copies a single simple value from kernel space to user |
| 566 | * space. It supports simple types like char and int, but not larger |
| 567 | * data types like structures or arrays. |
| 568 | * |
| 569 | * @ptr must have pointer-to-simple-variable type, and @x must be assignable |
| 570 | * to the result of dereferencing @ptr. |
| 571 | * |
| 572 | * Caller must check the pointer with access_ok() before calling this |
| 573 | * function. |
| 574 | * |
| 575 | * Returns zero on success, or -EFAULT on error. |
| 576 | */ |
| 577 | #define __put_user_unaligned(x,ptr) \ |
| 578 | __put_user_unaligned_nocheck((x),(ptr),sizeof(*(ptr))) |
| 579 | |
| 580 | /* |
| 581 | * __get_user_unaligned: - Get a simple variable from user space, with less checking. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 582 | * @x: Variable to store result. |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 583 | * @ptr: Source address, in user space. |
| 584 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 585 | * Context: User context only. This function may sleep if pagefaults are |
| 586 | * enabled. |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 587 | * |
| 588 | * This macro copies a single simple variable from user space to kernel |
| 589 | * space. It supports simple types like char and int, but not larger |
| 590 | * data types like structures or arrays. |
| 591 | * |
| 592 | * @ptr must have pointer-to-simple-variable type, and the result of |
| 593 | * dereferencing @ptr must be assignable to @x without a cast. |
| 594 | * |
| 595 | * Caller must check the pointer with access_ok() before calling this |
| 596 | * function. |
| 597 | * |
| 598 | * Returns zero on success, or -EFAULT on error. |
| 599 | * On error, the variable @x is set to zero. |
| 600 | */ |
| 601 | #define __get_user_unaligned(x,ptr) \ |
Al Viro | 930c0f7 | 2015-12-25 12:09:30 -0500 | [diff] [blame] | 602 | __get_user_unaligned_nocheck((x),(ptr),sizeof(*(ptr))) |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 603 | |
| 604 | /* |
| 605 | * Yuck. We need two variants, one for 64bit operation and one |
| 606 | * for 32 bit mode and old iron. |
| 607 | */ |
| 608 | #ifdef CONFIG_32BIT |
| 609 | #define __GET_USER_UNALIGNED_DW(val, ptr) \ |
| 610 | __get_user_unaligned_asm_ll32(val, ptr) |
| 611 | #endif |
| 612 | #ifdef CONFIG_64BIT |
| 613 | #define __GET_USER_UNALIGNED_DW(val, ptr) \ |
| 614 | __get_user_unaligned_asm(val, "uld", ptr) |
| 615 | #endif |
| 616 | |
| 617 | extern void __get_user_unaligned_unknown(void); |
| 618 | |
| 619 | #define __get_user_unaligned_common(val, size, ptr) \ |
| 620 | do { \ |
| 621 | switch (size) { \ |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 622 | case 1: __get_data_asm(val, "lb", ptr); break; \ |
Al Viro | 930c0f7 | 2015-12-25 12:09:30 -0500 | [diff] [blame] | 623 | case 2: __get_data_unaligned_asm(val, "ulh", ptr); break; \ |
| 624 | case 4: __get_data_unaligned_asm(val, "ulw", ptr); break; \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 625 | case 8: __GET_USER_UNALIGNED_DW(val, ptr); break; \ |
| 626 | default: __get_user_unaligned_unknown(); break; \ |
| 627 | } \ |
| 628 | } while (0) |
| 629 | |
| 630 | #define __get_user_unaligned_nocheck(x,ptr,size) \ |
| 631 | ({ \ |
| 632 | int __gu_err; \ |
| 633 | \ |
| 634 | __get_user_unaligned_common((x), size, ptr); \ |
| 635 | __gu_err; \ |
| 636 | }) |
| 637 | |
| 638 | #define __get_user_unaligned_check(x,ptr,size) \ |
| 639 | ({ \ |
| 640 | int __gu_err = -EFAULT; \ |
| 641 | const __typeof__(*(ptr)) __user * __gu_ptr = (ptr); \ |
| 642 | \ |
| 643 | if (likely(access_ok(VERIFY_READ, __gu_ptr, size))) \ |
| 644 | __get_user_unaligned_common((x), size, __gu_ptr); \ |
| 645 | \ |
| 646 | __gu_err; \ |
| 647 | }) |
| 648 | |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 649 | #define __get_data_unaligned_asm(val, insn, addr) \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 650 | { \ |
| 651 | long __gu_tmp; \ |
| 652 | \ |
| 653 | __asm__ __volatile__( \ |
| 654 | "1: " insn " %1, %3 \n" \ |
| 655 | "2: \n" \ |
Steven J. Hill | 1658f91 | 2013-03-25 13:22:59 -0500 | [diff] [blame] | 656 | " .insn \n" \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 657 | " .section .fixup,\"ax\" \n" \ |
| 658 | "3: li %0, %4 \n" \ |
Ralf Baechle | 640465b | 2014-11-18 18:47:13 +0100 | [diff] [blame] | 659 | " move %1, $0 \n" \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 660 | " j 2b \n" \ |
| 661 | " .previous \n" \ |
| 662 | " .section __ex_table,\"a\" \n" \ |
| 663 | " "__UA_ADDR "\t1b, 3b \n" \ |
| 664 | " "__UA_ADDR "\t1b + 4, 3b \n" \ |
| 665 | " .previous \n" \ |
| 666 | : "=r" (__gu_err), "=r" (__gu_tmp) \ |
| 667 | : "0" (0), "o" (__m(addr)), "i" (-EFAULT)); \ |
| 668 | \ |
| 669 | (val) = (__typeof__(*(addr))) __gu_tmp; \ |
| 670 | } |
| 671 | |
| 672 | /* |
| 673 | * Get a long long 64 using 32 bit registers. |
| 674 | */ |
| 675 | #define __get_user_unaligned_asm_ll32(val, addr) \ |
| 676 | { \ |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 677 | unsigned long long __gu_tmp; \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 678 | \ |
| 679 | __asm__ __volatile__( \ |
| 680 | "1: ulw %1, (%3) \n" \ |
| 681 | "2: ulw %D1, 4(%3) \n" \ |
| 682 | " move %0, $0 \n" \ |
Steven J. Hill | 1658f91 | 2013-03-25 13:22:59 -0500 | [diff] [blame] | 683 | "3: \n" \ |
| 684 | " .insn \n" \ |
| 685 | " .section .fixup,\"ax\" \n" \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 686 | "4: li %0, %4 \n" \ |
| 687 | " move %1, $0 \n" \ |
| 688 | " move %D1, $0 \n" \ |
| 689 | " j 3b \n" \ |
| 690 | " .previous \n" \ |
| 691 | " .section __ex_table,\"a\" \n" \ |
| 692 | " " __UA_ADDR " 1b, 4b \n" \ |
| 693 | " " __UA_ADDR " 1b + 4, 4b \n" \ |
| 694 | " " __UA_ADDR " 2b, 4b \n" \ |
| 695 | " " __UA_ADDR " 2b + 4, 4b \n" \ |
| 696 | " .previous \n" \ |
| 697 | : "=r" (__gu_err), "=&r" (__gu_tmp) \ |
| 698 | : "0" (0), "r" (addr), "i" (-EFAULT)); \ |
| 699 | (val) = (__typeof__(*(addr))) __gu_tmp; \ |
| 700 | } |
| 701 | |
| 702 | /* |
| 703 | * Yuck. We need two variants, one for 64bit operation and one |
| 704 | * for 32 bit mode and old iron. |
| 705 | */ |
| 706 | #ifdef CONFIG_32BIT |
| 707 | #define __PUT_USER_UNALIGNED_DW(ptr) __put_user_unaligned_asm_ll32(ptr) |
| 708 | #endif |
| 709 | #ifdef CONFIG_64BIT |
| 710 | #define __PUT_USER_UNALIGNED_DW(ptr) __put_user_unaligned_asm("usd", ptr) |
| 711 | #endif |
| 712 | |
Markos Chandras | ec56b1d | 2013-12-17 14:42:23 +0000 | [diff] [blame] | 713 | #define __put_user_unaligned_common(ptr, size) \ |
| 714 | do { \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 715 | switch (size) { \ |
Markos Chandras | 0081ad2 | 2014-01-06 12:48:28 +0000 | [diff] [blame] | 716 | case 1: __put_data_asm("sb", ptr); break; \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 717 | case 2: __put_user_unaligned_asm("ush", ptr); break; \ |
| 718 | case 4: __put_user_unaligned_asm("usw", ptr); break; \ |
| 719 | case 8: __PUT_USER_UNALIGNED_DW(ptr); break; \ |
| 720 | default: __put_user_unaligned_unknown(); break; \ |
Markos Chandras | ec56b1d | 2013-12-17 14:42:23 +0000 | [diff] [blame] | 721 | } while (0) |
| 722 | |
| 723 | #define __put_user_unaligned_nocheck(x,ptr,size) \ |
| 724 | ({ \ |
| 725 | __typeof__(*(ptr)) __pu_val; \ |
| 726 | int __pu_err = 0; \ |
| 727 | \ |
| 728 | __pu_val = (x); \ |
| 729 | __put_user_unaligned_common(ptr, size); \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 730 | __pu_err; \ |
| 731 | }) |
| 732 | |
| 733 | #define __put_user_unaligned_check(x,ptr,size) \ |
| 734 | ({ \ |
| 735 | __typeof__(*(ptr)) __user *__pu_addr = (ptr); \ |
| 736 | __typeof__(*(ptr)) __pu_val = (x); \ |
| 737 | int __pu_err = -EFAULT; \ |
| 738 | \ |
Markos Chandras | ec56b1d | 2013-12-17 14:42:23 +0000 | [diff] [blame] | 739 | if (likely(access_ok(VERIFY_WRITE, __pu_addr, size))) \ |
| 740 | __put_user_unaligned_common(__pu_addr, size); \ |
| 741 | \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 742 | __pu_err; \ |
| 743 | }) |
| 744 | |
| 745 | #define __put_user_unaligned_asm(insn, ptr) \ |
| 746 | { \ |
| 747 | __asm__ __volatile__( \ |
| 748 | "1: " insn " %z2, %3 # __put_user_unaligned_asm\n" \ |
| 749 | "2: \n" \ |
Steven J. Hill | 1658f91 | 2013-03-25 13:22:59 -0500 | [diff] [blame] | 750 | " .insn \n" \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 751 | " .section .fixup,\"ax\" \n" \ |
| 752 | "3: li %0, %4 \n" \ |
| 753 | " j 2b \n" \ |
| 754 | " .previous \n" \ |
| 755 | " .section __ex_table,\"a\" \n" \ |
| 756 | " " __UA_ADDR " 1b, 3b \n" \ |
| 757 | " .previous \n" \ |
| 758 | : "=r" (__pu_err) \ |
| 759 | : "0" (0), "Jr" (__pu_val), "o" (__m(ptr)), \ |
| 760 | "i" (-EFAULT)); \ |
| 761 | } |
| 762 | |
| 763 | #define __put_user_unaligned_asm_ll32(ptr) \ |
| 764 | { \ |
| 765 | __asm__ __volatile__( \ |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 766 | "1: sw %2, (%3) # __put_user_unaligned_asm_ll32 \n" \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 767 | "2: sw %D2, 4(%3) \n" \ |
| 768 | "3: \n" \ |
Steven J. Hill | 1658f91 | 2013-03-25 13:22:59 -0500 | [diff] [blame] | 769 | " .insn \n" \ |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 770 | " .section .fixup,\"ax\" \n" \ |
| 771 | "4: li %0, %4 \n" \ |
| 772 | " j 3b \n" \ |
| 773 | " .previous \n" \ |
| 774 | " .section __ex_table,\"a\" \n" \ |
| 775 | " " __UA_ADDR " 1b, 4b \n" \ |
| 776 | " " __UA_ADDR " 1b + 4, 4b \n" \ |
| 777 | " " __UA_ADDR " 2b, 4b \n" \ |
| 778 | " " __UA_ADDR " 2b + 4, 4b \n" \ |
| 779 | " .previous" \ |
| 780 | : "=r" (__pu_err) \ |
| 781 | : "0" (0), "r" (__pu_val), "r" (ptr), \ |
| 782 | "i" (-EFAULT)); \ |
| 783 | } |
| 784 | |
| 785 | extern void __put_user_unaligned_unknown(void); |
Leonid Yegoshin | 18e9001 | 2013-12-17 15:20:24 +0000 | [diff] [blame] | 786 | #endif |
Ralf Baechle | 71ec6cc | 2006-10-31 02:52:56 +0000 | [diff] [blame] | 787 | |
| 788 | /* |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 789 | * We're generating jump to subroutines which will be outside the range of |
| 790 | * jump instructions |
| 791 | */ |
| 792 | #ifdef MODULE |
| 793 | #define __MODULE_JAL(destination) \ |
| 794 | ".set\tnoat\n\t" \ |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 795 | __UA_LA "\t$1, " #destination "\n\t" \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 796 | "jalr\t$1\n\t" \ |
| 797 | ".set\tat\n\t" |
| 798 | #else |
| 799 | #define __MODULE_JAL(destination) \ |
| 800 | "jal\t" #destination "\n\t" |
| 801 | #endif |
| 802 | |
Markos Chandras | 5856381 | 2014-11-17 09:30:23 +0000 | [diff] [blame] | 803 | #if defined(CONFIG_CPU_DADDI_WORKAROUNDS) || (defined(CONFIG_EVA) && \ |
| 804 | defined(CONFIG_CPU_HAS_PREFETCH)) |
Maciej W. Rozycki | 619b6e1 | 2007-10-23 12:43:25 +0100 | [diff] [blame] | 805 | #define DADDI_SCRATCH "$3" |
Markos Chandras | 5856381 | 2014-11-17 09:30:23 +0000 | [diff] [blame] | 806 | #else |
| 807 | #define DADDI_SCRATCH "$0" |
Maciej W. Rozycki | 619b6e1 | 2007-10-23 12:43:25 +0100 | [diff] [blame] | 808 | #endif |
| 809 | |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 810 | extern size_t __copy_user(void *__to, const void *__from, size_t __n); |
| 811 | |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 812 | #ifndef CONFIG_EVA |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 813 | #define __invoke_copy_to_user(to, from, n) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 814 | ({ \ |
Ralf Baechle | 49a89ef | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 815 | register void __user *__cu_to_r __asm__("$4"); \ |
| 816 | register const void *__cu_from_r __asm__("$5"); \ |
| 817 | register long __cu_len_r __asm__("$6"); \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 818 | \ |
| 819 | __cu_to_r = (to); \ |
| 820 | __cu_from_r = (from); \ |
| 821 | __cu_len_r = (n); \ |
| 822 | __asm__ __volatile__( \ |
| 823 | __MODULE_JAL(__copy_user) \ |
| 824 | : "+r" (__cu_to_r), "+r" (__cu_from_r), "+r" (__cu_len_r) \ |
| 825 | : \ |
David Daney | bb0757e | 2012-06-06 23:00:31 +0100 | [diff] [blame] | 826 | : "$8", "$9", "$10", "$11", "$12", "$14", "$15", "$24", "$31", \ |
Maciej W. Rozycki | 619b6e1 | 2007-10-23 12:43:25 +0100 | [diff] [blame] | 827 | DADDI_SCRATCH, "memory"); \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 828 | __cu_len_r; \ |
| 829 | }) |
| 830 | |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 831 | #define __invoke_copy_to_kernel(to, from, n) \ |
| 832 | __invoke_copy_to_user(to, from, n) |
| 833 | |
| 834 | #endif |
| 835 | |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 836 | /* |
| 837 | * __copy_to_user: - Copy a block of data into user space, with less checking. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 838 | * @to: Destination address, in user space. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 839 | * @from: Source address, in kernel space. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 840 | * @n: Number of bytes to copy. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 841 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 842 | * Context: User context only. This function may sleep if pagefaults are |
| 843 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 844 | * |
| 845 | * Copy data from kernel space to user space. Caller must check |
| 846 | * the specified block with access_ok() before calling this function. |
| 847 | * |
| 848 | * Returns number of bytes that could not be copied. |
| 849 | * On success, this will be zero. |
| 850 | */ |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 851 | #define __copy_to_user(to, from, n) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 852 | ({ \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 853 | void __user *__cu_to; \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 854 | const void *__cu_from; \ |
| 855 | long __cu_len; \ |
| 856 | \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 857 | __cu_to = (to); \ |
| 858 | __cu_from = (from); \ |
| 859 | __cu_len = (n); \ |
Ralf Baechle | ef41f46 | 2009-04-28 14:17:54 +0200 | [diff] [blame] | 860 | might_fault(); \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 861 | if (eva_kernel_access()) \ |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 862 | __cu_len = __invoke_copy_to_kernel(__cu_to, __cu_from, \ |
| 863 | __cu_len); \ |
| 864 | else \ |
| 865 | __cu_len = __invoke_copy_to_user(__cu_to, __cu_from, \ |
| 866 | __cu_len); \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 867 | __cu_len; \ |
| 868 | }) |
| 869 | |
Ralf Baechle | d0c91ae | 2007-03-05 15:54:20 +0000 | [diff] [blame] | 870 | extern size_t __copy_user_inatomic(void *__to, const void *__from, size_t __n); |
| 871 | |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 872 | #define __copy_to_user_inatomic(to, from, n) \ |
Ralf Baechle | e03b526 | 2007-02-19 16:59:24 +0000 | [diff] [blame] | 873 | ({ \ |
| 874 | void __user *__cu_to; \ |
| 875 | const void *__cu_from; \ |
| 876 | long __cu_len; \ |
| 877 | \ |
| 878 | __cu_to = (to); \ |
| 879 | __cu_from = (from); \ |
| 880 | __cu_len = (n); \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 881 | if (eva_kernel_access()) \ |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 882 | __cu_len = __invoke_copy_to_kernel(__cu_to, __cu_from, \ |
| 883 | __cu_len); \ |
| 884 | else \ |
| 885 | __cu_len = __invoke_copy_to_user(__cu_to, __cu_from, \ |
| 886 | __cu_len); \ |
Ralf Baechle | e03b526 | 2007-02-19 16:59:24 +0000 | [diff] [blame] | 887 | __cu_len; \ |
| 888 | }) |
| 889 | |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 890 | #define __copy_from_user_inatomic(to, from, n) \ |
Ralf Baechle | e03b526 | 2007-02-19 16:59:24 +0000 | [diff] [blame] | 891 | ({ \ |
| 892 | void *__cu_to; \ |
| 893 | const void __user *__cu_from; \ |
| 894 | long __cu_len; \ |
| 895 | \ |
| 896 | __cu_to = (to); \ |
| 897 | __cu_from = (from); \ |
| 898 | __cu_len = (n); \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 899 | if (eva_kernel_access()) \ |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 900 | __cu_len = __invoke_copy_from_kernel_inatomic(__cu_to, \ |
| 901 | __cu_from,\ |
| 902 | __cu_len);\ |
| 903 | else \ |
| 904 | __cu_len = __invoke_copy_from_user_inatomic(__cu_to, \ |
| 905 | __cu_from, \ |
| 906 | __cu_len); \ |
Ralf Baechle | e03b526 | 2007-02-19 16:59:24 +0000 | [diff] [blame] | 907 | __cu_len; \ |
| 908 | }) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 909 | |
| 910 | /* |
| 911 | * copy_to_user: - Copy a block of data into user space. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 912 | * @to: Destination address, in user space. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 913 | * @from: Source address, in kernel space. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 914 | * @n: Number of bytes to copy. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 915 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 916 | * Context: User context only. This function may sleep if pagefaults are |
| 917 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 918 | * |
| 919 | * Copy data from kernel space to user space. |
| 920 | * |
| 921 | * Returns number of bytes that could not be copied. |
| 922 | * On success, this will be zero. |
| 923 | */ |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 924 | #define copy_to_user(to, from, n) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 925 | ({ \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 926 | void __user *__cu_to; \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 927 | const void *__cu_from; \ |
| 928 | long __cu_len; \ |
| 929 | \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 930 | __cu_to = (to); \ |
| 931 | __cu_from = (from); \ |
| 932 | __cu_len = (n); \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 933 | if (eva_kernel_access()) { \ |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 934 | __cu_len = __invoke_copy_to_kernel(__cu_to, \ |
| 935 | __cu_from, \ |
| 936 | __cu_len); \ |
| 937 | } else { \ |
| 938 | if (access_ok(VERIFY_WRITE, __cu_to, __cu_len)) { \ |
| 939 | might_fault(); \ |
| 940 | __cu_len = __invoke_copy_to_user(__cu_to, \ |
| 941 | __cu_from, \ |
| 942 | __cu_len); \ |
| 943 | } \ |
Ralf Baechle | ef41f46 | 2009-04-28 14:17:54 +0200 | [diff] [blame] | 944 | } \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 945 | __cu_len; \ |
| 946 | }) |
| 947 | |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 948 | #ifndef CONFIG_EVA |
| 949 | |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 950 | #define __invoke_copy_from_user(to, from, n) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 951 | ({ \ |
Ralf Baechle | 49a89ef | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 952 | register void *__cu_to_r __asm__("$4"); \ |
| 953 | register const void __user *__cu_from_r __asm__("$5"); \ |
| 954 | register long __cu_len_r __asm__("$6"); \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 955 | \ |
| 956 | __cu_to_r = (to); \ |
| 957 | __cu_from_r = (from); \ |
| 958 | __cu_len_r = (n); \ |
| 959 | __asm__ __volatile__( \ |
| 960 | ".set\tnoreorder\n\t" \ |
| 961 | __MODULE_JAL(__copy_user) \ |
| 962 | ".set\tnoat\n\t" \ |
| 963 | __UA_ADDU "\t$1, %1, %2\n\t" \ |
| 964 | ".set\tat\n\t" \ |
| 965 | ".set\treorder" \ |
| 966 | : "+r" (__cu_to_r), "+r" (__cu_from_r), "+r" (__cu_len_r) \ |
| 967 | : \ |
David Daney | bb0757e | 2012-06-06 23:00:31 +0100 | [diff] [blame] | 968 | : "$8", "$9", "$10", "$11", "$12", "$14", "$15", "$24", "$31", \ |
Maciej W. Rozycki | 619b6e1 | 2007-10-23 12:43:25 +0100 | [diff] [blame] | 969 | DADDI_SCRATCH, "memory"); \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 970 | __cu_len_r; \ |
| 971 | }) |
| 972 | |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 973 | #define __invoke_copy_from_kernel(to, from, n) \ |
| 974 | __invoke_copy_from_user(to, from, n) |
| 975 | |
| 976 | /* For userland <-> userland operations */ |
| 977 | #define ___invoke_copy_in_user(to, from, n) \ |
| 978 | __invoke_copy_from_user(to, from, n) |
| 979 | |
| 980 | /* For kernel <-> kernel operations */ |
| 981 | #define ___invoke_copy_in_kernel(to, from, n) \ |
| 982 | __invoke_copy_from_user(to, from, n) |
| 983 | |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 984 | #define __invoke_copy_from_user_inatomic(to, from, n) \ |
Ralf Baechle | e03b526 | 2007-02-19 16:59:24 +0000 | [diff] [blame] | 985 | ({ \ |
Ralf Baechle | 49a89ef | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 986 | register void *__cu_to_r __asm__("$4"); \ |
| 987 | register const void __user *__cu_from_r __asm__("$5"); \ |
| 988 | register long __cu_len_r __asm__("$6"); \ |
Ralf Baechle | e03b526 | 2007-02-19 16:59:24 +0000 | [diff] [blame] | 989 | \ |
| 990 | __cu_to_r = (to); \ |
| 991 | __cu_from_r = (from); \ |
| 992 | __cu_len_r = (n); \ |
| 993 | __asm__ __volatile__( \ |
| 994 | ".set\tnoreorder\n\t" \ |
| 995 | __MODULE_JAL(__copy_user_inatomic) \ |
| 996 | ".set\tnoat\n\t" \ |
| 997 | __UA_ADDU "\t$1, %1, %2\n\t" \ |
| 998 | ".set\tat\n\t" \ |
| 999 | ".set\treorder" \ |
| 1000 | : "+r" (__cu_to_r), "+r" (__cu_from_r), "+r" (__cu_len_r) \ |
| 1001 | : \ |
David Daney | bb0757e | 2012-06-06 23:00:31 +0100 | [diff] [blame] | 1002 | : "$8", "$9", "$10", "$11", "$12", "$14", "$15", "$24", "$31", \ |
Maciej W. Rozycki | 619b6e1 | 2007-10-23 12:43:25 +0100 | [diff] [blame] | 1003 | DADDI_SCRATCH, "memory"); \ |
Ralf Baechle | e03b526 | 2007-02-19 16:59:24 +0000 | [diff] [blame] | 1004 | __cu_len_r; \ |
| 1005 | }) |
| 1006 | |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 1007 | #define __invoke_copy_from_kernel_inatomic(to, from, n) \ |
| 1008 | __invoke_copy_from_user_inatomic(to, from, n) \ |
| 1009 | |
| 1010 | #else |
| 1011 | |
| 1012 | /* EVA specific functions */ |
| 1013 | |
| 1014 | extern size_t __copy_user_inatomic_eva(void *__to, const void *__from, |
| 1015 | size_t __n); |
| 1016 | extern size_t __copy_from_user_eva(void *__to, const void *__from, |
| 1017 | size_t __n); |
| 1018 | extern size_t __copy_to_user_eva(void *__to, const void *__from, |
| 1019 | size_t __n); |
| 1020 | extern size_t __copy_in_user_eva(void *__to, const void *__from, size_t __n); |
| 1021 | |
| 1022 | #define __invoke_copy_from_user_eva_generic(to, from, n, func_ptr) \ |
| 1023 | ({ \ |
| 1024 | register void *__cu_to_r __asm__("$4"); \ |
| 1025 | register const void __user *__cu_from_r __asm__("$5"); \ |
| 1026 | register long __cu_len_r __asm__("$6"); \ |
| 1027 | \ |
| 1028 | __cu_to_r = (to); \ |
| 1029 | __cu_from_r = (from); \ |
| 1030 | __cu_len_r = (n); \ |
| 1031 | __asm__ __volatile__( \ |
| 1032 | ".set\tnoreorder\n\t" \ |
| 1033 | __MODULE_JAL(func_ptr) \ |
| 1034 | ".set\tnoat\n\t" \ |
| 1035 | __UA_ADDU "\t$1, %1, %2\n\t" \ |
| 1036 | ".set\tat\n\t" \ |
| 1037 | ".set\treorder" \ |
| 1038 | : "+r" (__cu_to_r), "+r" (__cu_from_r), "+r" (__cu_len_r) \ |
| 1039 | : \ |
| 1040 | : "$8", "$9", "$10", "$11", "$12", "$14", "$15", "$24", "$31", \ |
| 1041 | DADDI_SCRATCH, "memory"); \ |
| 1042 | __cu_len_r; \ |
| 1043 | }) |
| 1044 | |
| 1045 | #define __invoke_copy_to_user_eva_generic(to, from, n, func_ptr) \ |
| 1046 | ({ \ |
| 1047 | register void *__cu_to_r __asm__("$4"); \ |
| 1048 | register const void __user *__cu_from_r __asm__("$5"); \ |
| 1049 | register long __cu_len_r __asm__("$6"); \ |
| 1050 | \ |
| 1051 | __cu_to_r = (to); \ |
| 1052 | __cu_from_r = (from); \ |
| 1053 | __cu_len_r = (n); \ |
| 1054 | __asm__ __volatile__( \ |
| 1055 | __MODULE_JAL(func_ptr) \ |
| 1056 | : "+r" (__cu_to_r), "+r" (__cu_from_r), "+r" (__cu_len_r) \ |
| 1057 | : \ |
| 1058 | : "$8", "$9", "$10", "$11", "$12", "$14", "$15", "$24", "$31", \ |
| 1059 | DADDI_SCRATCH, "memory"); \ |
| 1060 | __cu_len_r; \ |
| 1061 | }) |
| 1062 | |
| 1063 | /* |
| 1064 | * Source or destination address is in userland. We need to go through |
| 1065 | * the TLB |
| 1066 | */ |
| 1067 | #define __invoke_copy_from_user(to, from, n) \ |
| 1068 | __invoke_copy_from_user_eva_generic(to, from, n, __copy_from_user_eva) |
| 1069 | |
| 1070 | #define __invoke_copy_from_user_inatomic(to, from, n) \ |
| 1071 | __invoke_copy_from_user_eva_generic(to, from, n, \ |
| 1072 | __copy_user_inatomic_eva) |
| 1073 | |
| 1074 | #define __invoke_copy_to_user(to, from, n) \ |
| 1075 | __invoke_copy_to_user_eva_generic(to, from, n, __copy_to_user_eva) |
| 1076 | |
| 1077 | #define ___invoke_copy_in_user(to, from, n) \ |
| 1078 | __invoke_copy_from_user_eva_generic(to, from, n, __copy_in_user_eva) |
| 1079 | |
| 1080 | /* |
| 1081 | * Source or destination address in the kernel. We are not going through |
| 1082 | * the TLB |
| 1083 | */ |
| 1084 | #define __invoke_copy_from_kernel(to, from, n) \ |
| 1085 | __invoke_copy_from_user_eva_generic(to, from, n, __copy_user) |
| 1086 | |
| 1087 | #define __invoke_copy_from_kernel_inatomic(to, from, n) \ |
| 1088 | __invoke_copy_from_user_eva_generic(to, from, n, __copy_user_inatomic) |
| 1089 | |
| 1090 | #define __invoke_copy_to_kernel(to, from, n) \ |
| 1091 | __invoke_copy_to_user_eva_generic(to, from, n, __copy_user) |
| 1092 | |
| 1093 | #define ___invoke_copy_in_kernel(to, from, n) \ |
| 1094 | __invoke_copy_from_user_eva_generic(to, from, n, __copy_user) |
| 1095 | |
| 1096 | #endif /* CONFIG_EVA */ |
| 1097 | |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1098 | /* |
Chris Dearman | 131c1a2 | 2007-02-01 19:54:13 +0000 | [diff] [blame] | 1099 | * __copy_from_user: - Copy a block of data from user space, with less checking. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 1100 | * @to: Destination address, in kernel space. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1101 | * @from: Source address, in user space. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 1102 | * @n: Number of bytes to copy. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1103 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 1104 | * Context: User context only. This function may sleep if pagefaults are |
| 1105 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1106 | * |
| 1107 | * Copy data from user space to kernel space. Caller must check |
| 1108 | * the specified block with access_ok() before calling this function. |
| 1109 | * |
| 1110 | * Returns number of bytes that could not be copied. |
| 1111 | * On success, this will be zero. |
| 1112 | * |
| 1113 | * If some data could not be copied, this function will pad the copied |
| 1114 | * data to the requested size using zero bytes. |
| 1115 | */ |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 1116 | #define __copy_from_user(to, from, n) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1117 | ({ \ |
| 1118 | void *__cu_to; \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 1119 | const void __user *__cu_from; \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1120 | long __cu_len; \ |
| 1121 | \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1122 | __cu_to = (to); \ |
| 1123 | __cu_from = (from); \ |
| 1124 | __cu_len = (n); \ |
James Hogan | 6f06a2c | 2015-08-05 16:41:38 +0100 | [diff] [blame] | 1125 | if (eva_kernel_access()) { \ |
| 1126 | __cu_len = __invoke_copy_from_kernel(__cu_to, \ |
| 1127 | __cu_from, \ |
| 1128 | __cu_len); \ |
| 1129 | } else { \ |
| 1130 | might_fault(); \ |
| 1131 | __cu_len = __invoke_copy_from_user(__cu_to, __cu_from, \ |
| 1132 | __cu_len); \ |
| 1133 | } \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1134 | __cu_len; \ |
| 1135 | }) |
| 1136 | |
| 1137 | /* |
| 1138 | * copy_from_user: - Copy a block of data from user space. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 1139 | * @to: Destination address, in kernel space. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1140 | * @from: Source address, in user space. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 1141 | * @n: Number of bytes to copy. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1142 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 1143 | * Context: User context only. This function may sleep if pagefaults are |
| 1144 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1145 | * |
| 1146 | * Copy data from user space to kernel space. |
| 1147 | * |
| 1148 | * Returns number of bytes that could not be copied. |
| 1149 | * On success, this will be zero. |
| 1150 | * |
| 1151 | * If some data could not be copied, this function will pad the copied |
| 1152 | * data to the requested size using zero bytes. |
| 1153 | */ |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 1154 | #define copy_from_user(to, from, n) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1155 | ({ \ |
| 1156 | void *__cu_to; \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 1157 | const void __user *__cu_from; \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1158 | long __cu_len; \ |
| 1159 | \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1160 | __cu_to = (to); \ |
| 1161 | __cu_from = (from); \ |
| 1162 | __cu_len = (n); \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 1163 | if (eva_kernel_access()) { \ |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 1164 | __cu_len = __invoke_copy_from_kernel(__cu_to, \ |
| 1165 | __cu_from, \ |
| 1166 | __cu_len); \ |
| 1167 | } else { \ |
| 1168 | if (access_ok(VERIFY_READ, __cu_from, __cu_len)) { \ |
| 1169 | might_fault(); \ |
| 1170 | __cu_len = __invoke_copy_from_user(__cu_to, \ |
| 1171 | __cu_from, \ |
| 1172 | __cu_len); \ |
| 1173 | } \ |
Ralf Baechle | ef41f46 | 2009-04-28 14:17:54 +0200 | [diff] [blame] | 1174 | } \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1175 | __cu_len; \ |
| 1176 | }) |
| 1177 | |
Ralf Baechle | ed01b3d | 2009-04-27 16:46:21 +0200 | [diff] [blame] | 1178 | #define __copy_in_user(to, from, n) \ |
| 1179 | ({ \ |
| 1180 | void __user *__cu_to; \ |
| 1181 | const void __user *__cu_from; \ |
| 1182 | long __cu_len; \ |
| 1183 | \ |
Ralf Baechle | ed01b3d | 2009-04-27 16:46:21 +0200 | [diff] [blame] | 1184 | __cu_to = (to); \ |
| 1185 | __cu_from = (from); \ |
| 1186 | __cu_len = (n); \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 1187 | if (eva_kernel_access()) { \ |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 1188 | __cu_len = ___invoke_copy_in_kernel(__cu_to, __cu_from, \ |
| 1189 | __cu_len); \ |
| 1190 | } else { \ |
| 1191 | might_fault(); \ |
| 1192 | __cu_len = ___invoke_copy_in_user(__cu_to, __cu_from, \ |
| 1193 | __cu_len); \ |
| 1194 | } \ |
Ralf Baechle | ed01b3d | 2009-04-27 16:46:21 +0200 | [diff] [blame] | 1195 | __cu_len; \ |
| 1196 | }) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1197 | |
Ralf Baechle | 21a151d | 2007-10-11 23:46:15 +0100 | [diff] [blame] | 1198 | #define copy_in_user(to, from, n) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1199 | ({ \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 1200 | void __user *__cu_to; \ |
| 1201 | const void __user *__cu_from; \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1202 | long __cu_len; \ |
| 1203 | \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1204 | __cu_to = (to); \ |
| 1205 | __cu_from = (from); \ |
| 1206 | __cu_len = (n); \ |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 1207 | if (eva_kernel_access()) { \ |
Markos Chandras | 05c6516 | 2013-12-11 16:47:10 +0000 | [diff] [blame] | 1208 | __cu_len = ___invoke_copy_in_kernel(__cu_to,__cu_from, \ |
| 1209 | __cu_len); \ |
| 1210 | } else { \ |
| 1211 | if (likely(access_ok(VERIFY_READ, __cu_from, __cu_len) &&\ |
| 1212 | access_ok(VERIFY_WRITE, __cu_to, __cu_len))) {\ |
| 1213 | might_fault(); \ |
| 1214 | __cu_len = ___invoke_copy_in_user(__cu_to, \ |
| 1215 | __cu_from, \ |
| 1216 | __cu_len); \ |
| 1217 | } \ |
Ralf Baechle | ef41f46 | 2009-04-28 14:17:54 +0200 | [diff] [blame] | 1218 | } \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1219 | __cu_len; \ |
| 1220 | }) |
| 1221 | |
| 1222 | /* |
| 1223 | * __clear_user: - Zero a block of memory in user space, with less checking. |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 1224 | * @to: Destination address, in user space. |
| 1225 | * @n: Number of bytes to zero. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1226 | * |
| 1227 | * Zero a block of memory in user space. Caller must check |
| 1228 | * the specified block with access_ok() before calling this function. |
| 1229 | * |
| 1230 | * Returns number of bytes that could not be cleared. |
| 1231 | * On success, this will be zero. |
| 1232 | */ |
| 1233 | static inline __kernel_size_t |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 1234 | __clear_user(void __user *addr, __kernel_size_t size) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1235 | { |
| 1236 | __kernel_size_t res; |
| 1237 | |
James Hogan | d6a428f | 2015-08-05 16:41:39 +0100 | [diff] [blame] | 1238 | if (eva_kernel_access()) { |
| 1239 | __asm__ __volatile__( |
| 1240 | "move\t$4, %1\n\t" |
| 1241 | "move\t$5, $0\n\t" |
| 1242 | "move\t$6, %2\n\t" |
| 1243 | __MODULE_JAL(__bzero_kernel) |
| 1244 | "move\t%0, $6" |
| 1245 | : "=r" (res) |
| 1246 | : "r" (addr), "r" (size) |
| 1247 | : "$4", "$5", "$6", __UA_t0, __UA_t1, "$31"); |
| 1248 | } else { |
| 1249 | might_fault(); |
| 1250 | __asm__ __volatile__( |
| 1251 | "move\t$4, %1\n\t" |
| 1252 | "move\t$5, $0\n\t" |
| 1253 | "move\t$6, %2\n\t" |
| 1254 | __MODULE_JAL(__bzero) |
| 1255 | "move\t%0, $6" |
| 1256 | : "=r" (res) |
| 1257 | : "r" (addr), "r" (size) |
| 1258 | : "$4", "$5", "$6", __UA_t0, __UA_t1, "$31"); |
| 1259 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1260 | |
| 1261 | return res; |
| 1262 | } |
| 1263 | |
| 1264 | #define clear_user(addr,n) \ |
| 1265 | ({ \ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 1266 | void __user * __cl_addr = (addr); \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1267 | unsigned long __cl_size = (n); \ |
| 1268 | if (__cl_size && access_ok(VERIFY_WRITE, \ |
Wu Zhangjin | 63d3892 | 2009-05-21 05:50:01 +0800 | [diff] [blame] | 1269 | __cl_addr, __cl_size)) \ |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1270 | __cl_size = __clear_user(__cl_addr, __cl_size); \ |
| 1271 | __cl_size; \ |
| 1272 | }) |
| 1273 | |
| 1274 | /* |
| 1275 | * __strncpy_from_user: - Copy a NUL terminated string from userspace, with less checking. |
| 1276 | * @dst: Destination address, in kernel space. This buffer must be at |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 1277 | * least @count bytes long. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1278 | * @src: Source address, in user space. |
| 1279 | * @count: Maximum number of bytes to copy, including the trailing NUL. |
| 1280 | * |
| 1281 | * Copies a NUL-terminated string from userspace to kernel space. |
| 1282 | * Caller must check the specified block with access_ok() before calling |
| 1283 | * this function. |
| 1284 | * |
| 1285 | * On success, returns the length of the string (not including the trailing |
| 1286 | * NUL). |
| 1287 | * |
| 1288 | * If access to userspace fails, returns -EFAULT (some data may have been |
| 1289 | * copied). |
| 1290 | * |
| 1291 | * If @count is smaller than the length of the string, copies @count bytes |
| 1292 | * and returns @count. |
| 1293 | */ |
| 1294 | static inline long |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 1295 | __strncpy_from_user(char *__to, const char __user *__from, long __len) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1296 | { |
| 1297 | long res; |
| 1298 | |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 1299 | if (eva_kernel_access()) { |
Markos Chandras | e3a9b07 | 2014-01-03 14:55:02 +0000 | [diff] [blame] | 1300 | __asm__ __volatile__( |
| 1301 | "move\t$4, %1\n\t" |
| 1302 | "move\t$5, %2\n\t" |
| 1303 | "move\t$6, %3\n\t" |
| 1304 | __MODULE_JAL(__strncpy_from_kernel_nocheck_asm) |
| 1305 | "move\t%0, $2" |
| 1306 | : "=r" (res) |
| 1307 | : "r" (__to), "r" (__from), "r" (__len) |
| 1308 | : "$2", "$3", "$4", "$5", "$6", __UA_t0, "$31", "memory"); |
| 1309 | } else { |
| 1310 | might_fault(); |
| 1311 | __asm__ __volatile__( |
| 1312 | "move\t$4, %1\n\t" |
| 1313 | "move\t$5, %2\n\t" |
| 1314 | "move\t$6, %3\n\t" |
| 1315 | __MODULE_JAL(__strncpy_from_user_nocheck_asm) |
| 1316 | "move\t%0, $2" |
| 1317 | : "=r" (res) |
| 1318 | : "r" (__to), "r" (__from), "r" (__len) |
| 1319 | : "$2", "$3", "$4", "$5", "$6", __UA_t0, "$31", "memory"); |
| 1320 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1321 | |
| 1322 | return res; |
| 1323 | } |
| 1324 | |
| 1325 | /* |
| 1326 | * strncpy_from_user: - Copy a NUL terminated string from userspace. |
| 1327 | * @dst: Destination address, in kernel space. This buffer must be at |
Ralf Baechle | 7034228 | 2013-01-22 12:59:30 +0100 | [diff] [blame] | 1328 | * least @count bytes long. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1329 | * @src: Source address, in user space. |
| 1330 | * @count: Maximum number of bytes to copy, including the trailing NUL. |
| 1331 | * |
| 1332 | * Copies a NUL-terminated string from userspace to kernel space. |
| 1333 | * |
| 1334 | * On success, returns the length of the string (not including the trailing |
| 1335 | * NUL). |
| 1336 | * |
| 1337 | * If access to userspace fails, returns -EFAULT (some data may have been |
| 1338 | * copied). |
| 1339 | * |
| 1340 | * If @count is smaller than the length of the string, copies @count bytes |
| 1341 | * and returns @count. |
| 1342 | */ |
| 1343 | static inline long |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 1344 | strncpy_from_user(char *__to, const char __user *__from, long __len) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1345 | { |
| 1346 | long res; |
| 1347 | |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 1348 | if (eva_kernel_access()) { |
Markos Chandras | e3a9b07 | 2014-01-03 14:55:02 +0000 | [diff] [blame] | 1349 | __asm__ __volatile__( |
| 1350 | "move\t$4, %1\n\t" |
| 1351 | "move\t$5, %2\n\t" |
| 1352 | "move\t$6, %3\n\t" |
| 1353 | __MODULE_JAL(__strncpy_from_kernel_asm) |
| 1354 | "move\t%0, $2" |
| 1355 | : "=r" (res) |
| 1356 | : "r" (__to), "r" (__from), "r" (__len) |
| 1357 | : "$2", "$3", "$4", "$5", "$6", __UA_t0, "$31", "memory"); |
| 1358 | } else { |
| 1359 | might_fault(); |
| 1360 | __asm__ __volatile__( |
| 1361 | "move\t$4, %1\n\t" |
| 1362 | "move\t$5, %2\n\t" |
| 1363 | "move\t$6, %3\n\t" |
| 1364 | __MODULE_JAL(__strncpy_from_user_asm) |
| 1365 | "move\t%0, $2" |
| 1366 | : "=r" (res) |
| 1367 | : "r" (__to), "r" (__from), "r" (__len) |
| 1368 | : "$2", "$3", "$4", "$5", "$6", __UA_t0, "$31", "memory"); |
| 1369 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1370 | |
| 1371 | return res; |
| 1372 | } |
| 1373 | |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1374 | /* |
| 1375 | * strlen_user: - Get the size of a string in user space. |
| 1376 | * @str: The string to measure. |
| 1377 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 1378 | * Context: User context only. This function may sleep if pagefaults are |
| 1379 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1380 | * |
| 1381 | * Get the size of a NUL-terminated string in user space. |
| 1382 | * |
| 1383 | * Returns the size of the string INCLUDING the terminating NUL. |
| 1384 | * On exception, returns 0. |
| 1385 | * |
| 1386 | * If there is a limit on the length of a valid string, you may wish to |
| 1387 | * consider using strnlen_user() instead. |
| 1388 | */ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 1389 | static inline long strlen_user(const char __user *s) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1390 | { |
| 1391 | long res; |
| 1392 | |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 1393 | if (eva_kernel_access()) { |
Markos Chandras | e3a9b07 | 2014-01-03 14:55:02 +0000 | [diff] [blame] | 1394 | __asm__ __volatile__( |
| 1395 | "move\t$4, %1\n\t" |
| 1396 | __MODULE_JAL(__strlen_kernel_asm) |
| 1397 | "move\t%0, $2" |
| 1398 | : "=r" (res) |
| 1399 | : "r" (s) |
| 1400 | : "$2", "$4", __UA_t0, "$31"); |
| 1401 | } else { |
| 1402 | might_fault(); |
| 1403 | __asm__ __volatile__( |
| 1404 | "move\t$4, %1\n\t" |
James Hogan | 5dc62fd | 2015-08-05 16:41:37 +0100 | [diff] [blame] | 1405 | __MODULE_JAL(__strlen_user_asm) |
Markos Chandras | e3a9b07 | 2014-01-03 14:55:02 +0000 | [diff] [blame] | 1406 | "move\t%0, $2" |
| 1407 | : "=r" (res) |
| 1408 | : "r" (s) |
| 1409 | : "$2", "$4", __UA_t0, "$31"); |
| 1410 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1411 | |
| 1412 | return res; |
| 1413 | } |
| 1414 | |
| 1415 | /* Returns: 0 if bad, string length+1 (memory size) of string if ok */ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 1416 | static inline long __strnlen_user(const char __user *s, long n) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1417 | { |
| 1418 | long res; |
| 1419 | |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 1420 | if (eva_kernel_access()) { |
Markos Chandras | e3a9b07 | 2014-01-03 14:55:02 +0000 | [diff] [blame] | 1421 | __asm__ __volatile__( |
| 1422 | "move\t$4, %1\n\t" |
| 1423 | "move\t$5, %2\n\t" |
| 1424 | __MODULE_JAL(__strnlen_kernel_nocheck_asm) |
| 1425 | "move\t%0, $2" |
| 1426 | : "=r" (res) |
| 1427 | : "r" (s), "r" (n) |
| 1428 | : "$2", "$4", "$5", __UA_t0, "$31"); |
| 1429 | } else { |
| 1430 | might_fault(); |
| 1431 | __asm__ __volatile__( |
| 1432 | "move\t$4, %1\n\t" |
| 1433 | "move\t$5, %2\n\t" |
| 1434 | __MODULE_JAL(__strnlen_user_nocheck_asm) |
| 1435 | "move\t%0, $2" |
| 1436 | : "=r" (res) |
| 1437 | : "r" (s), "r" (n) |
| 1438 | : "$2", "$4", "$5", __UA_t0, "$31"); |
| 1439 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1440 | |
| 1441 | return res; |
| 1442 | } |
| 1443 | |
| 1444 | /* |
Ralf Baechle | 80219c6 | 2014-11-04 02:23:45 +0100 | [diff] [blame] | 1445 | * strnlen_user: - Get the size of a string in user space. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1446 | * @str: The string to measure. |
| 1447 | * |
David Hildenbrand | b3c395ef | 2015-05-11 17:52:08 +0200 | [diff] [blame] | 1448 | * Context: User context only. This function may sleep if pagefaults are |
| 1449 | * enabled. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1450 | * |
| 1451 | * Get the size of a NUL-terminated string in user space. |
| 1452 | * |
| 1453 | * Returns the size of the string INCLUDING the terminating NUL. |
| 1454 | * On exception, returns 0. |
Ralf Baechle | 80219c6 | 2014-11-04 02:23:45 +0100 | [diff] [blame] | 1455 | * If the string is too long, returns a value greater than @n. |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1456 | */ |
Ralf Baechle | fe00f94 | 2005-03-01 19:22:29 +0000 | [diff] [blame] | 1457 | static inline long strnlen_user(const char __user *s, long n) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1458 | { |
| 1459 | long res; |
| 1460 | |
Ralf Baechle | ef41f46 | 2009-04-28 14:17:54 +0200 | [diff] [blame] | 1461 | might_fault(); |
Paul Burton | 1206066 | 2015-05-24 16:31:44 +0100 | [diff] [blame] | 1462 | if (eva_kernel_access()) { |
Markos Chandras | e3a9b07 | 2014-01-03 14:55:02 +0000 | [diff] [blame] | 1463 | __asm__ __volatile__( |
| 1464 | "move\t$4, %1\n\t" |
| 1465 | "move\t$5, %2\n\t" |
| 1466 | __MODULE_JAL(__strnlen_kernel_asm) |
| 1467 | "move\t%0, $2" |
| 1468 | : "=r" (res) |
| 1469 | : "r" (s), "r" (n) |
| 1470 | : "$2", "$4", "$5", __UA_t0, "$31"); |
| 1471 | } else { |
| 1472 | __asm__ __volatile__( |
| 1473 | "move\t$4, %1\n\t" |
| 1474 | "move\t$5, %2\n\t" |
| 1475 | __MODULE_JAL(__strnlen_user_asm) |
| 1476 | "move\t%0, $2" |
| 1477 | : "=r" (res) |
| 1478 | : "r" (s), "r" (n) |
| 1479 | : "$2", "$4", "$5", __UA_t0, "$31"); |
| 1480 | } |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1481 | |
| 1482 | return res; |
| 1483 | } |
| 1484 | |
| 1485 | struct exception_table_entry |
| 1486 | { |
| 1487 | unsigned long insn; |
| 1488 | unsigned long nextinsn; |
| 1489 | }; |
| 1490 | |
| 1491 | extern int fixup_exception(struct pt_regs *regs); |
| 1492 | |
| 1493 | #endif /* _ASM_UACCESS_H */ |