1/*-
2 * SPDX-License-Identifier: BSD-4-Clause
3 *
4 * Copyright (c) 1990 The Regents of the University of California.
5 * All rights reserved.
6 * Copyright (c) 1994 John S. Dyson
7 * All rights reserved.
8 * Copyright (c) 2003 Peter Wemm
9 * All rights reserved.
10 *
11 * This code is derived from software contributed to Berkeley by
12 * William Jolitz.
13 *
14 * Redistribution and use in source and binary forms, with or without
15 * modification, are permitted provided that the following conditions
16 * are met:
17 * 1. Redistributions of source code must retain the above copyright
18 *    notice, this list of conditions and the following disclaimer.
19 * 2. Redistributions in binary form must reproduce the above copyright
20 *    notice, this list of conditions and the following disclaimer in the
21 *    documentation and/or other materials provided with the distribution.
22 * 3. All advertising materials mentioning features or use of this software
23 *    must display the following acknowledgement:
24 *	This product includes software developed by the University of
25 *	California, Berkeley and its contributors.
26 * 4. Neither the name of the University nor the names of its contributors
27 *    may be used to endorse or promote products derived from this software
28 *    without specific prior written permission.
29 *
30 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
31 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
32 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
33 * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
34 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
35 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
36 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
37 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
38 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
39 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
40 * SUCH DAMAGE.
41 *
42 *	from: @(#)vmparam.h	5.9 (Berkeley) 5/12/91
43 * $FreeBSD$
44 */
45
46
47#ifndef _MACHINE_VMPARAM_H_
48#define	_MACHINE_VMPARAM_H_ 1
49
50/*
51 * Machine dependent constants for AMD64.
52 */
53
54/*
55 * Virtual memory related constants, all in bytes
56 */
57#define	MAXTSIZ		(32768UL*1024*1024)	/* max text size */
58#ifndef DFLDSIZ
59#define	DFLDSIZ		(32768UL*1024*1024)	/* initial data size limit */
60#endif
61#ifndef MAXDSIZ
62#define	MAXDSIZ		(32768UL*1024*1024)	/* max data size */
63#endif
64#ifndef	DFLSSIZ
65#define	DFLSSIZ		(8UL*1024*1024)		/* initial stack size limit */
66#endif
67#ifndef	MAXSSIZ
68#define	MAXSSIZ		(512UL*1024*1024)	/* max stack size */
69#endif
70#ifndef SGROWSIZ
71#define	SGROWSIZ	(128UL*1024)		/* amount to grow stack */
72#endif
73
74/*
75 * We provide a machine specific single page allocator through the use
76 * of the direct mapped segment.  This uses 2MB pages for reduced
77 * TLB pressure.
78 */
79#define	UMA_MD_SMALL_ALLOC
80
81/*
82 * The physical address space is densely populated.
83 */
84#define	VM_PHYSSEG_DENSE
85
86/*
87 * The number of PHYSSEG entries must be one greater than the number
88 * of phys_avail entries because the phys_avail entry that spans the
89 * largest physical address that is accessible by ISA DMA is split
90 * into two PHYSSEG entries.
91 */
92#define	VM_PHYSSEG_MAX		63
93
94/*
95 * Create two free page pools: VM_FREEPOOL_DEFAULT is the default pool
96 * from which physical pages are allocated and VM_FREEPOOL_DIRECT is
97 * the pool from which physical pages for page tables and small UMA
98 * objects are allocated.
99 */
100#define	VM_NFREEPOOL		2
101#define	VM_FREEPOOL_DEFAULT	0
102#define	VM_FREEPOOL_DIRECT	1
103
104/*
105 * Create up to three free page lists: VM_FREELIST_DMA32 is for physical pages
106 * that have physical addresses below 4G but are not accessible by ISA DMA,
107 * and VM_FREELIST_ISADMA is for physical pages that are accessible by ISA
108 * DMA.
109 */
110#define	VM_NFREELIST		3
111#define	VM_FREELIST_DEFAULT	0
112#define	VM_FREELIST_DMA32	1
113#define	VM_FREELIST_LOWMEM	2
114
115#define VM_LOWMEM_BOUNDARY	(16 << 20)	/* 16MB ISA DMA limit */
116
117/*
118 * Create the DMA32 free list only if the number of physical pages above
119 * physical address 4G is at least 16M, which amounts to 64GB of physical
120 * memory.
121 */
122#define	VM_DMA32_NPAGES_THRESHOLD	16777216
123
124/*
125 * An allocation size of 16MB is supported in order to optimize the
126 * use of the direct map by UMA.  Specifically, a cache line contains
127 * at most 8 PDEs, collectively mapping 16MB of physical memory.  By
128 * reducing the number of distinct 16MB "pages" that are used by UMA,
129 * the physical memory allocator reduces the likelihood of both 2MB
130 * page TLB misses and cache misses caused by 2MB page TLB misses.
131 */
132#define	VM_NFREEORDER		13
133
134/*
135 * Enable superpage reservations: 1 level.
136 */
137#ifndef	VM_NRESERVLEVEL
138#define	VM_NRESERVLEVEL		1
139#endif
140
141/*
142 * Level 0 reservations consist of 512 pages.
143 */
144#ifndef	VM_LEVEL_0_ORDER
145#define	VM_LEVEL_0_ORDER	9
146#endif
147
148#ifdef	SMP
149#define	PA_LOCK_COUNT	256
150#endif
151
152/*
153 * Virtual addresses of things.  Derived from the page directory and
154 * page table indexes from pmap.h for precision.
155 *
156 * 0x0000000000000000 - 0x00007fffffffffff   user map
157 * 0x0000800000000000 - 0xffff7fffffffffff   does not exist (hole)
158 * 0xffff800000000000 - 0xffff804020100fff   recursive page table (512GB slot)
159 * 0xffff804020100fff - 0xffff807fffffffff   unused
160 * 0xffff808000000000 - 0xffff847fffffffff   large map (can be tuned up)
161 * 0xffff848000000000 - 0xfffff7ffffffffff   unused (large map extends there)
162 * 0xfffff80000000000 - 0xfffffbffffffffff   4TB direct map
163 * 0xfffffc0000000000 - 0xfffffdffffffffff   unused
164 * 0xfffffe0000000000 - 0xffffffffffffffff   2TB kernel map
165 *
166 * Within the kernel map:
167 *
168 * 0xffffffff80000000                        KERNBASE
169 */
170
171#define	VM_MIN_KERNEL_ADDRESS	KVADDR(KPML4BASE, 0, 0, 0)
172#define	VM_MAX_KERNEL_ADDRESS	KVADDR(KPML4BASE + NKPML4E - 1, \
173					NPDPEPG-1, NPDEPG-1, NPTEPG-1)
174
175#define	DMAP_MIN_ADDRESS	KVADDR(DMPML4I, 0, 0, 0)
176#define	DMAP_MAX_ADDRESS	KVADDR(DMPML4I + NDMPML4E, 0, 0, 0)
177
178#define	LARGEMAP_MIN_ADDRESS	KVADDR(LMSPML4I, 0, 0, 0)
179#define	LARGEMAP_MAX_ADDRESS	KVADDR(LMEPML4I + 1, 0, 0, 0)
180
181#define	KERNBASE		KVADDR(KPML4I, KPDPI, 0, 0)
182
183#define	UPT_MAX_ADDRESS		KVADDR(PML4PML4I, PML4PML4I, PML4PML4I, PML4PML4I)
184#define	UPT_MIN_ADDRESS		KVADDR(PML4PML4I, 0, 0, 0)
185
186#define	VM_MAXUSER_ADDRESS	UVADDR(NUPML4E, 0, 0, 0)
187
188#define	SHAREDPAGE		(VM_MAXUSER_ADDRESS - PAGE_SIZE)
189#define	USRSTACK		SHAREDPAGE
190
191#define	VM_MAX_ADDRESS		UPT_MAX_ADDRESS
192#define	VM_MIN_ADDRESS		(0)
193
194/*
195 * XXX Allowing dmaplimit == 0 is a temporary workaround for vt(4) efifb's
196 * early use of PHYS_TO_DMAP before the mapping is actually setup. This works
197 * because the result is not actually accessed until later, but the early
198 * vt fb startup needs to be reworked.
199 */
200#define	PMAP_HAS_DMAP	1
201#define	PHYS_TO_DMAP(x)	({						\
202	KASSERT(dmaplimit == 0 || (x) < dmaplimit,			\
203	    ("physical address %#jx not covered by the DMAP",		\
204	    (uintmax_t)x));						\
205	(x) | DMAP_MIN_ADDRESS; })
206
207#define	DMAP_TO_PHYS(x)	({						\
208	KASSERT((x) < (DMAP_MIN_ADDRESS + dmaplimit) &&			\
209	    (x) >= DMAP_MIN_ADDRESS,					\
210	    ("virtual address %#jx not covered by the DMAP",		\
211	    (uintmax_t)x));						\
212	(x) & ~DMAP_MIN_ADDRESS; })
213
214/*
215 * How many physical pages per kmem arena virtual page.
216 */
217#ifndef VM_KMEM_SIZE_SCALE
218#define	VM_KMEM_SIZE_SCALE	(1)
219#endif
220
221/*
222 * Optional ceiling (in bytes) on the size of the kmem arena: 60% of the
223 * kernel map.
224 */
225#ifndef VM_KMEM_SIZE_MAX
226#define	VM_KMEM_SIZE_MAX	((VM_MAX_KERNEL_ADDRESS - \
227    VM_MIN_KERNEL_ADDRESS + 1) * 3 / 5)
228#endif
229
230/* initial pagein size of beginning of executable file */
231#ifndef VM_INITIAL_PAGEIN
232#define	VM_INITIAL_PAGEIN	16
233#endif
234
235#define	ZERO_REGION_SIZE	(2 * 1024 * 1024)	/* 2MB */
236
237/*
238 * Use a fairly large batch size since we expect amd64 systems to have lots of
239 * memory.
240 */
241#define	VM_BATCHQUEUE_SIZE	31
242
243#endif /* _MACHINE_VMPARAM_H_ */
244