1/* SPDX-License-Identifier: GPL-2.0 */
2
3#ifndef _ASM_S390_KFENCE_H
4#define _ASM_S390_KFENCE_H
5
6#include <linux/mm.h>
7#include <linux/kfence.h>
8#include <asm/set_memory.h>
9#include <asm/page.h>
10
11void __kernel_map_pages(struct page *page, int numpages, int enable);
12
13static __always_inline bool arch_kfence_init_pool(void)
14{
15	return true;
16}
17
18#define arch_kfence_test_address(addr) ((addr) & PAGE_MASK)
19
20/*
21 * Do not split kfence pool to 4k mapping with arch_kfence_init_pool(),
22 * but earlier where page table allocations still happen with memblock.
23 * Reason is that arch_kfence_init_pool() gets called when the system
24 * is still in a limbo state - disabling and enabling bottom halves is
25 * not yet allowed, but that is what our page_table_alloc() would do.
26 */
27static __always_inline void kfence_split_mapping(void)
28{
29#ifdef CONFIG_KFENCE
30	unsigned long pool_pages = KFENCE_POOL_SIZE >> PAGE_SHIFT;
31
32	set_memory_4k((unsigned long)__kfence_pool, pool_pages);
33#endif
34}
35
36static inline bool kfence_protect_page(unsigned long addr, bool protect)
37{
38	__kernel_map_pages(virt_to_page((void *)addr), 1, !protect);
39	return true;
40}
41
42#endif /* _ASM_S390_KFENCE_H */
43