1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 *  linux/arch/arm/mm/pgd.c
4 *
5 *  Copyright (C) 1998-2005 Russell King
6 */
7#include <linux/mm.h>
8#include <linux/gfp.h>
9#include <linux/highmem.h>
10#include <linux/slab.h>
11
12#include <asm/cp15.h>
13#include <asm/pgalloc.h>
14#include <asm/page.h>
15#include <asm/tlbflush.h>
16
17#include "mm.h"
18
19#ifdef CONFIG_ARM_LPAE
20#define __pgd_alloc()	kmalloc_array(PTRS_PER_PGD, sizeof(pgd_t), GFP_KERNEL)
21#define __pgd_free(pgd)	kfree(pgd)
22#else
23#define __pgd_alloc()	(pgd_t *)__get_free_pages(GFP_KERNEL, 2)
24#define __pgd_free(pgd)	free_pages((unsigned long)pgd, 2)
25#endif
26
27/*
28 * need to get a 16k page for level 1
29 */
30pgd_t *pgd_alloc(struct mm_struct *mm)
31{
32	pgd_t *new_pgd, *init_pgd;
33	p4d_t *new_p4d, *init_p4d;
34	pud_t *new_pud, *init_pud;
35	pmd_t *new_pmd, *init_pmd;
36	pte_t *new_pte, *init_pte;
37
38	new_pgd = __pgd_alloc();
39	if (!new_pgd)
40		goto no_pgd;
41
42	memset(new_pgd, 0, USER_PTRS_PER_PGD * sizeof(pgd_t));
43
44	/*
45	 * Copy over the kernel and IO PGD entries
46	 */
47	init_pgd = pgd_offset_k(0);
48	memcpy(new_pgd + USER_PTRS_PER_PGD, init_pgd + USER_PTRS_PER_PGD,
49		       (PTRS_PER_PGD - USER_PTRS_PER_PGD) * sizeof(pgd_t));
50
51	clean_dcache_area(new_pgd, PTRS_PER_PGD * sizeof(pgd_t));
52
53#ifdef CONFIG_ARM_LPAE
54	/*
55	 * Allocate PMD table for modules and pkmap mappings.
56	 */
57	new_p4d = p4d_alloc(mm, new_pgd + pgd_index(MODULES_VADDR),
58			    MODULES_VADDR);
59	if (!new_p4d)
60		goto no_p4d;
61
62	new_pud = pud_alloc(mm, new_p4d, MODULES_VADDR);
63	if (!new_pud)
64		goto no_pud;
65
66	new_pmd = pmd_alloc(mm, new_pud, 0);
67	if (!new_pmd)
68		goto no_pmd;
69#ifdef CONFIG_KASAN
70	/*
71	 * Copy PMD table for KASAN shadow mappings.
72	 */
73	init_pgd = pgd_offset_k(TASK_SIZE);
74	init_p4d = p4d_offset(init_pgd, TASK_SIZE);
75	init_pud = pud_offset(init_p4d, TASK_SIZE);
76	init_pmd = pmd_offset(init_pud, TASK_SIZE);
77	new_pmd = pmd_offset(new_pud, TASK_SIZE);
78	memcpy(new_pmd, init_pmd,
79	       (pmd_index(MODULES_VADDR) - pmd_index(TASK_SIZE))
80	       * sizeof(pmd_t));
81	clean_dcache_area(new_pmd, PTRS_PER_PMD * sizeof(pmd_t));
82#endif /* CONFIG_KASAN */
83#endif /* CONFIG_LPAE */
84
85	if (!vectors_high()) {
86		/*
87		 * On ARM, first page must always be allocated since it
88		 * contains the machine vectors. The vectors are always high
89		 * with LPAE.
90		 */
91		new_p4d = p4d_alloc(mm, new_pgd, 0);
92		if (!new_p4d)
93			goto no_p4d;
94
95		new_pud = pud_alloc(mm, new_p4d, 0);
96		if (!new_pud)
97			goto no_pud;
98
99		new_pmd = pmd_alloc(mm, new_pud, 0);
100		if (!new_pmd)
101			goto no_pmd;
102
103		new_pte = pte_alloc_map(mm, new_pmd, 0);
104		if (!new_pte)
105			goto no_pte;
106
107#ifndef CONFIG_ARM_LPAE
108		/*
109		 * Modify the PTE pointer to have the correct domain.  This
110		 * needs to be the vectors domain to avoid the low vectors
111		 * being unmapped.
112		 */
113		pmd_val(*new_pmd) &= ~PMD_DOMAIN_MASK;
114		pmd_val(*new_pmd) |= PMD_DOMAIN(DOMAIN_VECTORS);
115#endif
116
117		init_p4d = p4d_offset(init_pgd, 0);
118		init_pud = pud_offset(init_p4d, 0);
119		init_pmd = pmd_offset(init_pud, 0);
120		init_pte = pte_offset_map(init_pmd, 0);
121		set_pte_ext(new_pte + 0, init_pte[0], 0);
122		set_pte_ext(new_pte + 1, init_pte[1], 0);
123		pte_unmap(init_pte);
124		pte_unmap(new_pte);
125	}
126
127	return new_pgd;
128
129no_pte:
130	pmd_free(mm, new_pmd);
131	mm_dec_nr_pmds(mm);
132no_pmd:
133	pud_free(mm, new_pud);
134no_pud:
135	p4d_free(mm, new_p4d);
136no_p4d:
137	__pgd_free(new_pgd);
138no_pgd:
139	return NULL;
140}
141
142void pgd_free(struct mm_struct *mm, pgd_t *pgd_base)
143{
144	pgd_t *pgd;
145	p4d_t *p4d;
146	pud_t *pud;
147	pmd_t *pmd;
148	pgtable_t pte;
149
150	if (!pgd_base)
151		return;
152
153	pgd = pgd_base + pgd_index(0);
154	if (pgd_none_or_clear_bad(pgd))
155		goto no_pgd;
156
157	p4d = p4d_offset(pgd, 0);
158	if (p4d_none_or_clear_bad(p4d))
159		goto no_p4d;
160
161	pud = pud_offset(p4d, 0);
162	if (pud_none_or_clear_bad(pud))
163		goto no_pud;
164
165	pmd = pmd_offset(pud, 0);
166	if (pmd_none_or_clear_bad(pmd))
167		goto no_pmd;
168
169	pte = pmd_pgtable(*pmd);
170	pmd_clear(pmd);
171	pte_free(mm, pte);
172	mm_dec_nr_ptes(mm);
173no_pmd:
174	pud_clear(pud);
175	pmd_free(mm, pmd);
176	mm_dec_nr_pmds(mm);
177no_pud:
178	p4d_clear(p4d);
179	pud_free(mm, pud);
180no_p4d:
181	pgd_clear(pgd);
182	p4d_free(mm, p4d);
183no_pgd:
184#ifdef CONFIG_ARM_LPAE
185	/*
186	 * Free modules/pkmap or identity pmd tables.
187	 */
188	for (pgd = pgd_base; pgd < pgd_base + PTRS_PER_PGD; pgd++) {
189		if (pgd_none_or_clear_bad(pgd))
190			continue;
191		if (pgd_val(*pgd) & L_PGD_SWAPPER)
192			continue;
193		p4d = p4d_offset(pgd, 0);
194		if (p4d_none_or_clear_bad(p4d))
195			continue;
196		pud = pud_offset(p4d, 0);
197		if (pud_none_or_clear_bad(pud))
198			continue;
199		pmd = pmd_offset(pud, 0);
200		pud_clear(pud);
201		pmd_free(mm, pmd);
202		mm_dec_nr_pmds(mm);
203		p4d_clear(p4d);
204		pud_free(mm, pud);
205		mm_dec_nr_puds(mm);
206		pgd_clear(pgd);
207		p4d_free(mm, p4d);
208	}
209#endif
210	__pgd_free(pgd_base);
211}
212