| // SPDX-License-Identifier: GPL-2.0 |
| /* Copyright(c) 2016-2020 Intel Corporation. All rights reserved. */ |
| |
| #include <linux/jump_label.h> |
| #include <linux/uaccess.h> |
| #include <linux/export.h> |
| #include <linux/instrumented.h> |
| #include <linux/string.h> |
| #include <linux/types.h> |
| |
| #include <asm/mce.h> |
| |
| #ifdef CONFIG_X86_MCE |
| static DEFINE_STATIC_KEY_FALSE(copy_mc_fragile_key); |
| |
| void enable_copy_mc_fragile(void) |
| { |
| static_branch_inc(©_mc_fragile_key); |
| } |
| #define copy_mc_fragile_enabled (static_branch_unlikely(©_mc_fragile_key)) |
| |
| /* |
| * Similar to copy_user_handle_tail, probe for the write fault point, or |
| * source exception point. |
| */ |
| __visible notrace unsigned long |
| copy_mc_fragile_handle_tail(char *to, char *from, unsigned len) |
| { |
| for (; len; --len, to++, from++) |
| if (copy_mc_fragile(to, from, 1)) |
| break; |
| return len; |
| } |
| #else |
| /* |
| * No point in doing careful copying, or consulting a static key when |
| * there is no #MC handler in the CONFIG_X86_MCE=n case. |
| */ |
| void enable_copy_mc_fragile(void) |
| { |
| } |
| #define copy_mc_fragile_enabled (0) |
| #endif |
| |
| unsigned long copy_mc_enhanced_fast_string(void *dst, const void *src, unsigned len); |
| |
| /** |
| * copy_mc_to_kernel - memory copy that handles source exceptions |
| * |
| * @dst: destination address |
| * @src: source address |
| * @len: number of bytes to copy |
| * |
| * Call into the 'fragile' version on systems that benefit from avoiding |
| * corner case poison consumption scenarios, For example, accessing |
| * poison across 2 cachelines with a single instruction. Almost all |
| * other uses case can use copy_mc_enhanced_fast_string() for a fast |
| * recoverable copy, or fallback to plain memcpy. |
| * |
| * Return 0 for success, or number of bytes not copied if there was an |
| * exception. |
| */ |
| unsigned long __must_check copy_mc_to_kernel(void *dst, const void *src, unsigned len) |
| { |
| unsigned long ret; |
| |
| if (copy_mc_fragile_enabled) { |
| instrument_memcpy_before(dst, src, len); |
| ret = copy_mc_fragile(dst, src, len); |
| instrument_memcpy_after(dst, src, len, ret); |
| return ret; |
| } |
| if (static_cpu_has(X86_FEATURE_ERMS)) { |
| instrument_memcpy_before(dst, src, len); |
| ret = copy_mc_enhanced_fast_string(dst, src, len); |
| instrument_memcpy_after(dst, src, len, ret); |
| return ret; |
| } |
| memcpy(dst, src, len); |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(copy_mc_to_kernel); |
| |
| unsigned long __must_check copy_mc_to_user(void __user *dst, const void *src, unsigned len) |
| { |
| unsigned long ret; |
| |
| if (copy_mc_fragile_enabled) { |
| instrument_copy_to_user(dst, src, len); |
| __uaccess_begin(); |
| ret = copy_mc_fragile((__force void *)dst, src, len); |
| __uaccess_end(); |
| return ret; |
| } |
| |
| if (static_cpu_has(X86_FEATURE_ERMS)) { |
| instrument_copy_to_user(dst, src, len); |
| __uaccess_begin(); |
| ret = copy_mc_enhanced_fast_string((__force void *)dst, src, len); |
| __uaccess_end(); |
| return ret; |
| } |
| |
| return copy_user_generic((__force void *)dst, src, len); |
| } |