/* SPDX-License-Identifier: GPL-2.0 */ /* * S390 version * Copyright IBM Corp. 1999, 2000 * Author(s): Hartmut Penner (hp@de.ibm.com), * Martin Schwidefsky (schwidefsky@de.ibm.com) * * Derived from "include/asm-i386/uaccess.h" */ #ifndef __S390_UACCESS_H #define __S390_UACCESS_H /* * User space memory access functions */ #include #include #include #include #include #include void debug_user_asce(int exit); unsigned long __must_check raw_copy_from_user(void *to, const void __user *from, unsigned long n); unsigned long __must_check raw_copy_to_user(void __user *to, const void *from, unsigned long n); #ifndef CONFIG_KASAN #define INLINE_COPY_FROM_USER #define INLINE_COPY_TO_USER #endif unsigned long __must_check _copy_from_user_key(void *to, const void __user *from, unsigned long n, unsigned long key); static __always_inline unsigned long __must_check copy_from_user_key(void *to, const void __user *from, unsigned long n, unsigned long key) { if (check_copy_size(to, n, false)) n = _copy_from_user_key(to, from, n, key); return n; } unsigned long __must_check _copy_to_user_key(void __user *to, const void *from, unsigned long n, unsigned long key); static __always_inline unsigned long __must_check copy_to_user_key(void __user *to, const void *from, unsigned long n, unsigned long key) { if (check_copy_size(from, n, true)) n = _copy_to_user_key(to, from, n, key); return n; } union oac { unsigned int val; struct { struct { unsigned short key : 4; unsigned short : 4; unsigned short as : 2; unsigned short : 4; unsigned short k : 1; unsigned short a : 1; } oac1; struct { unsigned short key : 4; unsigned short : 4; unsigned short as : 2; unsigned short : 4; unsigned short k : 1; unsigned short a : 1; } oac2; }; }; int __noreturn __put_user_bad(void); #define __put_user_asm(to, from, size) \ ({ \ union oac __oac_spec = { \ .oac1.as = PSW_BITS_AS_SECONDARY, \ .oac1.a = 1, \ }; \ int __rc; \ \ asm volatile( \ " lr 0,%[spec]\n" \ "0: mvcos %[_to],%[_from],%[_size]\n" \ "1: xr %[rc],%[rc]\n" \ "2:\n" \ EX_TABLE_UA_STORE(0b, 2b, %[rc]) \ EX_TABLE_UA_STORE(1b, 2b, %[rc]) \ : [rc] "=&d" (__rc), [_to] "+Q" (*(to)) \ : [_size] "d" (size), [_from] "Q" (*(from)), \ [spec] "d" (__oac_spec.val) \ : "cc", "0"); \ __rc; \ }) static __always_inline int __put_user_fn(void *x, void __user *ptr, unsigned long size) { int rc; switch (size) { case 1: rc = __put_user_asm((unsigned char __user *)ptr, (unsigned char *)x, size); break; case 2: rc = __put_user_asm((unsigned short __user *)ptr, (unsigned short *)x, size); break; case 4: rc = __put_user_asm((unsigned int __user *)ptr, (unsigned int *)x, size); break; case 8: rc = __put_user_asm((unsigned long __user *)ptr, (unsigned long *)x, size); break; default: __put_user_bad(); break; } return rc; } int __noreturn __get_user_bad(void); #define __get_user_asm(to, from, size) \ ({ \ union oac __oac_spec = { \ .oac2.as = PSW_BITS_AS_SECONDARY, \ .oac2.a = 1, \ }; \ int __rc; \ \ asm volatile( \ " lr 0,%[spec]\n" \ "0: mvcos 0(%[_to]),%[_from],%[_size]\n" \ "1: xr %[rc],%[rc]\n" \ "2:\n" \ EX_TABLE_UA_LOAD_MEM(0b, 2b, %[rc], %[_to], %[_ksize]) \ EX_TABLE_UA_LOAD_MEM(1b, 2b, %[rc], %[_to], %[_ksize]) \ : [rc] "=&d" (__rc), "=Q" (*(to)) \ : [_size] "d" (size), [_from] "Q" (*(from)), \ [spec] "d" (__oac_spec.val), [_to] "a" (to), \ [_ksize] "K" (size) \ : "cc", "0"); \ __rc; \ }) static __always_inline int __get_user_fn(void *x, const void __user *ptr, unsigned long size) { int rc; switch (size) { case 1: rc = __get_user_asm((unsigned char *)x, (unsigned char __user *)ptr, size); break; case 2: rc = __get_user_asm((unsigned short *)x, (unsigned short __user *)ptr, size); break; case 4: rc = __get_user_asm((unsigned int *)x, (unsigned int __user *)ptr, size); break; case 8: rc = __get_user_asm((unsigned long *)x, (unsigned long __user *)ptr, size); break; default: __get_user_bad(); break; } return rc; } /* * These are the main single-value transfer routines. They automatically * use the right size if we just have the right pointer type. */ #define __put_user(x, ptr) \ ({ \ __typeof__(*(ptr)) __x = (x); \ int __pu_err = -EFAULT; \ \ __chk_user_ptr(ptr); \ switch (sizeof(*(ptr))) { \ case 1: \ case 2: \ case 4: \ case 8: \ __pu_err = __put_user_fn(&__x, ptr, sizeof(*(ptr))); \ break; \ default: \ __put_user_bad(); \ break; \ } \ __builtin_expect(__pu_err, 0); \ }) #define put_user(x, ptr) \ ({ \ might_fault(); \ __put_user(x, ptr); \ }) #define __get_user(x, ptr) \ ({ \ int __gu_err = -EFAULT; \ \ __chk_user_ptr(ptr); \ switch (sizeof(*(ptr))) { \ case 1: { \ unsigned char __x; \ \ __gu_err = __get_user_fn(&__x, ptr, sizeof(*(ptr))); \ (x) = *(__force __typeof__(*(ptr)) *)&__x; \ break; \ }; \ case 2: { \ unsigned short __x; \ \ __gu_err = __get_user_fn(&__x, ptr, sizeof(*(ptr))); \ (x) = *(__force __typeof__(*(ptr)) *)&__x; \ break; \ }; \ case 4: { \ unsigned int __x; \ \ __gu_err = __get_user_fn(&__x, ptr, sizeof(*(ptr))); \ (x) = *(__force __typeof__(*(ptr)) *)&__x; \ break; \ }; \ case 8: { \ unsigned long __x; \ \ __gu_err = __get_user_fn(&__x, ptr, sizeof(*(ptr))); \ (x) = *(__force __typeof__(*(ptr)) *)&__x; \ break; \ }; \ default: \ __get_user_bad(); \ break; \ } \ __builtin_expect(__gu_err, 0); \ }) #define get_user(x, ptr) \ ({ \ might_fault(); \ __get_user(x, ptr); \ }) /* * Copy a null terminated string from userspace. */ long __must_check strncpy_from_user(char *dst, const char __user *src, long count); long __must_check strnlen_user(const char __user *src, long count); /* * Zero Userspace */ unsigned long __must_check __clear_user(void __user *to, unsigned long size); static inline unsigned long __must_check clear_user(void __user *to, unsigned long n) { might_fault(); return __clear_user(to, n); } void *s390_kernel_write(void *dst, const void *src, size_t size); int __noreturn __put_kernel_bad(void); #define __put_kernel_asm(val, to, insn) \ ({ \ int __rc; \ \ asm volatile( \ "0: " insn " %[_val],%[_to]\n" \ "1: xr %[rc],%[rc]\n" \ "2:\n" \ EX_TABLE_UA_STORE(0b, 2b, %[rc]) \ EX_TABLE_UA_STORE(1b, 2b, %[rc]) \ : [rc] "=d" (__rc), [_to] "+Q" (*(to)) \ : [_val] "d" (val) \ : "cc"); \ __rc; \ }) #define __put_kernel_nofault(dst, src, type, err_label) \ do { \ unsigned long __x = (unsigned long)(*((type *)(src))); \ int __pk_err; \ \ switch (sizeof(type)) { \ case 1: \ __pk_err = __put_kernel_asm(__x, (type *)(dst), "stc"); \ break; \ case 2: \ __pk_err = __put_kernel_asm(__x, (type *)(dst), "sth"); \ break; \ case 4: \ __pk_err = __put_kernel_asm(__x, (type *)(dst), "st"); \ break; \ case 8: \ __pk_err = __put_kernel_asm(__x, (type *)(dst), "stg"); \ break; \ default: \ __pk_err = __put_kernel_bad(); \ break; \ } \ if (unlikely(__pk_err)) \ goto err_label; \ } while (0) int __noreturn __get_kernel_bad(void); #define __get_kernel_asm(val, from, insn) \ ({ \ int __rc; \ \ asm volatile( \ "0: " insn " %[_val],%[_from]\n" \ "1: xr %[rc],%[rc]\n" \ "2:\n" \ EX_TABLE_UA_LOAD_REG(0b, 2b, %[rc], %[_val]) \ EX_TABLE_UA_LOAD_REG(1b, 2b, %[rc], %[_val]) \ : [rc] "=d" (__rc), [_val] "=d" (val) \ : [_from] "Q" (*(from)) \ : "cc"); \ __rc; \ }) #define __get_kernel_nofault(dst, src, type, err_label) \ do { \ int __gk_err; \ \ switch (sizeof(type)) { \ case 1: { \ unsigned char __x; \ \ __gk_err = __get_kernel_asm(__x, (type *)(src), "ic"); \ *((type *)(dst)) = (type)__x; \ break; \ }; \ case 2: { \ unsigned short __x; \ \ __gk_err = __get_kernel_asm(__x, (type *)(src), "lh"); \ *((type *)(dst)) = (type)__x; \ break; \ }; \ case 4: { \ unsigned int __x; \ \ __gk_err = __get_kernel_asm(__x, (type *)(src), "l"); \ *((type *)(dst)) = (type)__x; \ break; \ }; \ case 8: { \ unsigned long __x; \ \ __gk_err = __get_kernel_asm(__x, (type *)(src), "lg"); \ *((type *)(dst)) = (type)__x; \ break; \ }; \ default: \ __gk_err = __get_kernel_bad(); \ break; \ } \ if (unlikely(__gk_err)) \ goto err_label; \ } while (0) void __cmpxchg_user_key_called_with_bad_pointer(void); #define CMPXCHG_USER_KEY_MAX_LOOPS 128 static __always_inline int __cmpxchg_user_key(unsigned long address, void *uval, __uint128_t old, __uint128_t new, unsigned long key, int size) { int rc = 0; switch (size) { case 1: { unsigned int prev, shift, mask, _old, _new; unsigned long count; shift = (3 ^ (address & 3)) << 3; address ^= address & 3; _old = ((unsigned int)old & 0xff) << shift; _new = ((unsigned int)new & 0xff) << shift; mask = ~(0xff << shift); asm volatile( " spka 0(%[key])\n" " sacf 256\n" " llill %[count],%[max_loops]\n" "0: l %[prev],%[address]\n" "1: nr %[prev],%[mask]\n" " xilf %[mask],0xffffffff\n" " or %[new],%[prev]\n" " or %[prev],%[tmp]\n" "2: lr %[tmp],%[prev]\n" "3: cs %[prev],%[new],%[address]\n" "4: jnl 5f\n" " xr %[tmp],%[prev]\n" " xr %[new],%[tmp]\n" " nr %[tmp],%[mask]\n" " jnz 5f\n" " brct %[count],2b\n" "5: sacf 768\n" " spka %[default_key]\n" EX_TABLE_UA_LOAD_REG(0b, 5b, %[rc], %[prev]) EX_TABLE_UA_LOAD_REG(1b, 5b, %[rc], %[prev]) EX_TABLE_UA_LOAD_REG(3b, 5b, %[rc], %[prev]) EX_TABLE_UA_LOAD_REG(4b, 5b, %[rc], %[prev]) : [rc] "+&d" (rc), [prev] "=&d" (prev), [address] "+Q" (*(int *)address), [tmp] "+&d" (_old), [new] "+&d" (_new), [mask] "+&d" (mask), [count] "=a" (count) : [key] "%[count]" (key << 4), [default_key] "J" (PAGE_DEFAULT_KEY), [max_loops] "J" (CMPXCHG_USER_KEY_MAX_LOOPS) : "memory", "cc"); *(unsigned char *)uval = prev >> shift; if (!count) rc = -EAGAIN; return rc; } case 2: { unsigned int prev, shift, mask, _old, _new; unsigned long count; shift = (2 ^ (address & 2)) << 3; address ^= address & 2; _old = ((unsigned int)old & 0xffff) << shift; _new = ((unsigned int)new & 0xffff) << shift; mask = ~(0xffff << shift); asm volatile( " spka 0(%[key])\n" " sacf 256\n" " llill %[count],%[max_loops]\n" "0: l %[prev],%[address]\n" "1: nr %[prev],%[mask]\n" " xilf %[mask],0xffffffff\n" " or %[new],%[prev]\n" " or %[prev],%[tmp]\n" "2: lr %[tmp],%[prev]\n" "3: cs %[prev],%[new],%[address]\n" "4: jnl 5f\n" " xr %[tmp],%[prev]\n" " xr %[new],%[tmp]\n" " nr %[tmp],%[mask]\n" " jnz 5f\n" " brct %[count],2b\n" "5: sacf 768\n" " spka %[default_key]\n" EX_TABLE_UA_LOAD_REG(0b, 5b, %[rc], %[prev]) EX_TABLE_UA_LOAD_REG(1b, 5b, %[rc], %[prev]) EX_TABLE_UA_LOAD_REG(3b, 5b, %[rc], %[prev]) EX_TABLE_UA_LOAD_REG(4b, 5b, %[rc], %[prev]) : [rc] "+&d" (rc), [prev] "=&d" (prev), [address] "+Q" (*(int *)address), [tmp] "+&d" (_old), [new] "+&d" (_new), [mask] "+&d" (mask), [count] "=a" (count) : [key] "%[count]" (key << 4), [default_key] "J" (PAGE_DEFAULT_KEY), [max_loops] "J" (CMPXCHG_USER_KEY_MAX_LOOPS) : "memory", "cc"); *(unsigned short *)uval = prev >> shift; if (!count) rc = -EAGAIN; return rc; } case 4: { unsigned int prev = old; asm volatile( " spka 0(%[key])\n" " sacf 256\n" "0: cs %[prev],%[new],%[address]\n" "1: sacf 768\n" " spka %[default_key]\n" EX_TABLE_UA_LOAD_REG(0b, 1b, %[rc], %[prev]) EX_TABLE_UA_LOAD_REG(1b, 1b, %[rc], %[prev]) : [rc] "+&d" (rc), [prev] "+&d" (prev), [address] "+Q" (*(int *)address) : [new] "d" ((unsigned int)new), [key] "a" (key << 4), [default_key] "J" (PAGE_DEFAULT_KEY) : "memory", "cc"); *(unsigned int *)uval = prev; return rc; } case 8: { unsigned long prev = old; asm volatile( " spka 0(%[key])\n" " sacf 256\n" "0: csg %[prev],%[new],%[address]\n" "1: sacf 768\n" " spka %[default_key]\n" EX_TABLE_UA_LOAD_REG(0b, 1b, %[rc], %[prev]) EX_TABLE_UA_LOAD_REG(1b, 1b, %[rc], %[prev]) : [rc] "+&d" (rc), [prev] "+&d" (prev), [address] "+QS" (*(long *)address) : [new] "d" ((unsigned long)new), [key] "a" (key << 4), [default_key] "J" (PAGE_DEFAULT_KEY) : "memory", "cc"); *(unsigned long *)uval = prev; return rc; } case 16: { __uint128_t prev = old; asm volatile( " spka 0(%[key])\n" " sacf 256\n" "0: cdsg %[prev],%[new],%[address]\n" "1: sacf 768\n" " spka %[default_key]\n" EX_TABLE_UA_LOAD_REGPAIR(0b, 1b, %[rc], %[prev]) EX_TABLE_UA_LOAD_REGPAIR(1b, 1b, %[rc], %[prev]) : [rc] "+&d" (rc), [prev] "+&d" (prev), [address] "+QS" (*(__int128_t *)address) : [new] "d" (new), [key] "a" (key << 4), [default_key] "J" (PAGE_DEFAULT_KEY) : "memory", "cc"); *(__uint128_t *)uval = prev; return rc; } } __cmpxchg_user_key_called_with_bad_pointer(); return rc; } /** * cmpxchg_user_key() - cmpxchg with user space target, honoring storage keys * @ptr: User space address of value to compare to @old and exchange with * @new. Must be aligned to sizeof(*@ptr). * @uval: Address where the old value of *@ptr is written to. * @old: Old value. Compared to the content pointed to by @ptr in order to * determine if the exchange occurs. The old value read from *@ptr is * written to *@uval. * @new: New value to place at *@ptr. * @key: Access key to use for checking storage key protection. * * Perform a cmpxchg on a user space target, honoring storage key protection. * @key alone determines how key checking is performed, neither * storage-protection-override nor fetch-protection-override apply. * The caller must compare *@uval and @old to determine if values have been * exchanged. In case of an exception *@uval is set to zero. * * Return: 0: cmpxchg executed * -EFAULT: an exception happened when trying to access *@ptr * -EAGAIN: maxed out number of retries (byte and short only) */ #define cmpxchg_user_key(ptr, uval, old, new, key) \ ({ \ __typeof__(ptr) __ptr = (ptr); \ __typeof__(uval) __uval = (uval); \ \ BUILD_BUG_ON(sizeof(*(__ptr)) != sizeof(*(__uval))); \ might_fault(); \ __chk_user_ptr(__ptr); \ __cmpxchg_user_key((unsigned long)(__ptr), (void *)(__uval), \ (old), (new), (key), sizeof(*(__ptr))); \ }) #endif /* __S390_UACCESS_H */