| /* SPDX-License-Identifier: GPL-2.0 */ |
| |
| #ifndef _ASM_S390_KFENCE_H |
| #define _ASM_S390_KFENCE_H |
| |
| #include <linux/mm.h> |
| #include <linux/kfence.h> |
| #include <asm/set_memory.h> |
| #include <asm/page.h> |
| |
| void __kernel_map_pages(struct page *page, int numpages, int enable); |
| |
| static __always_inline bool arch_kfence_init_pool(void) |
| { |
| return true; |
| } |
| |
| #define arch_kfence_test_address(addr) ((addr) & PAGE_MASK) |
| |
| /* |
| * Do not split kfence pool to 4k mapping with arch_kfence_init_pool(), |
| * but earlier where page table allocations still happen with memblock. |
| * Reason is that arch_kfence_init_pool() gets called when the system |
| * is still in a limbo state - disabling and enabling bottom halves is |
| * not yet allowed, but that is what our page_table_alloc() would do. |
| */ |
| static __always_inline void kfence_split_mapping(void) |
| { |
| #ifdef CONFIG_KFENCE |
| unsigned long pool_pages = KFENCE_POOL_SIZE >> PAGE_SHIFT; |
| |
| set_memory_4k((unsigned long)__kfence_pool, pool_pages); |
| #endif |
| } |
| |
| static inline bool kfence_protect_page(unsigned long addr, bool protect) |
| { |
| __kernel_map_pages(virt_to_page((void *)addr), 1, !protect); |
| return true; |
| } |
| |
| #endif /* _ASM_S390_KFENCE_H */ |