uvm_page_array.c revision 1.8
1/*	$NetBSD: uvm_page_array.c,v 1.8 2020/05/25 22:01:26 ad Exp $	*/
2
3/*-
4 * Copyright (c)2011 YAMAMOTO Takashi,
5 * All rights reserved.
6 *
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
9 * are met:
10 * 1. Redistributions of source code must retain the above copyright
11 *    notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 *    notice, this list of conditions and the following disclaimer in the
14 *    documentation and/or other materials provided with the distribution.
15 *
16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
17 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19 * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
20 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
21 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
22 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
23 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
24 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26 * SUCH DAMAGE.
27 */
28
29#include <sys/cdefs.h>
30__KERNEL_RCSID(0, "$NetBSD: uvm_page_array.c,v 1.8 2020/05/25 22:01:26 ad Exp $");
31
32#include <sys/param.h>
33#include <sys/systm.h>
34
35#include <uvm/uvm_extern.h>
36#include <uvm/uvm_object.h>
37#include <uvm/uvm_page.h>
38#include <uvm/uvm_page_array.h>
39
40/*
41 * uvm_page_array_init: initialize the array.
42 */
43
44void
45uvm_page_array_init(struct uvm_page_array *ar, struct uvm_object *uobj,
46    unsigned int flags)
47{
48
49	ar->ar_idx = 0;
50	ar->ar_npages = 0;
51	ar->ar_uobj = uobj;
52	ar->ar_flags = flags;
53}
54
55/*
56 * uvm_page_array_fini: clean up the array.
57 */
58
59void
60uvm_page_array_fini(struct uvm_page_array *ar)
61{
62
63	/*
64	 * currently nothing to do.
65	 */
66#if defined(DIAGNOSTIC)
67	/*
68	 * poison to trigger assertion in uvm_page_array_peek to
69	 * detect usage errors.
70	 */
71	ar->ar_npages = 1;
72	ar->ar_idx = 1000;
73#endif /* defined(DIAGNOSTIC) */
74}
75
76/*
77 * uvm_page_array_clear: forget the cached pages and initialize the array.
78 */
79
80void
81uvm_page_array_clear(struct uvm_page_array *ar)
82{
83
84	KASSERT(ar->ar_idx <= ar->ar_npages);
85	ar->ar_idx = 0;
86	ar->ar_npages = 0;
87}
88
89/*
90 * uvm_page_array_peek: return the next cached page.
91 */
92
93struct vm_page *
94uvm_page_array_peek(struct uvm_page_array *ar)
95{
96
97	KASSERT(ar->ar_idx <= ar->ar_npages);
98	if (ar->ar_idx == ar->ar_npages) {
99		return NULL;
100	}
101	return ar->ar_pages[ar->ar_idx];
102}
103
104/*
105 * uvm_page_array_advance: advance the array to the next cached page
106 */
107
108void
109uvm_page_array_advance(struct uvm_page_array *ar)
110{
111
112	KASSERT(ar->ar_idx <= ar->ar_npages);
113	ar->ar_idx++;
114	KASSERT(ar->ar_idx <= ar->ar_npages);
115}
116
117/*
118 * uvm_page_array_fill: lookup pages and keep them cached.
119 *
120 * return 0 on success.  in that case, cache the result in the array
121 * so that they will be picked by later uvm_page_array_peek.
122 *
123 * nwant is a number of pages to fetch.  a caller should consider it a hint.
124 * nwant == 0 means a caller have no specific idea.
125 *
126 * return ENOENT if no pages are found.
127 *
128 * called with object lock held.
129 */
130
131int
132uvm_page_array_fill(struct uvm_page_array *ar, voff_t off, unsigned int nwant)
133{
134	unsigned int npages;
135#if defined(DEBUG)
136	unsigned int i;
137#endif /* defined(DEBUG) */
138	unsigned int maxpages = __arraycount(ar->ar_pages);
139	struct uvm_object *uobj = ar->ar_uobj;
140	const int flags = ar->ar_flags;
141	const bool dense = (flags & UVM_PAGE_ARRAY_FILL_DENSE) != 0;
142	const bool backward = (flags & UVM_PAGE_ARRAY_FILL_BACKWARD) != 0;
143
144	if (nwant != 0 && nwant < maxpages) {
145		maxpages = nwant;
146	}
147#if 0 /* called from DDB for "show obj/f" without lock */
148	KASSERT(rw_lock_held(uobj->vmobjlock));
149#endif
150	KASSERT(uvm_page_array_peek(ar) == NULL);
151	if ((flags & UVM_PAGE_ARRAY_FILL_DIRTY) != 0) {
152		unsigned int tagmask = UVM_PAGE_DIRTY_TAG;
153
154		if ((flags & UVM_PAGE_ARRAY_FILL_WRITEBACK) != 0) {
155			tagmask |= UVM_PAGE_WRITEBACK_TAG;
156		}
157		npages =
158		    (backward ? radix_tree_gang_lookup_tagged_node_reverse :
159		    radix_tree_gang_lookup_tagged_node)(
160		    &uobj->uo_pages, off >> PAGE_SHIFT, (void **)ar->ar_pages,
161		    maxpages, dense, tagmask);
162	} else {
163		npages =
164		    (backward ? radix_tree_gang_lookup_node_reverse :
165		    radix_tree_gang_lookup_node)(
166		    &uobj->uo_pages, off >> PAGE_SHIFT, (void **)ar->ar_pages,
167		    maxpages, dense);
168	}
169	if (npages == 0) {
170		if (flags != 0) {
171			/*
172			 * if dense or looking for tagged entries (or
173			 * working backwards), fail right away.
174			 */
175			uvm_page_array_clear(ar);
176			return ENOENT;
177		} else {
178			/*
179			 * there's nothing else to be found with the current
180			 * set of arguments, in the current version of the
181			 * tree.
182			 *
183			 * minimize repeated tree lookups by "finding" a
184			 * null pointer, in case the caller keeps looping (a
185			 * common use case).
186			 */
187			npages = 1;
188			ar->ar_pages[0] = NULL;
189		}
190	}
191	KASSERT(npages <= maxpages);
192	ar->ar_npages = npages;
193	ar->ar_idx = 0;
194#if defined(DEBUG)
195	for (i = 0; i < ar->ar_npages; i++) {
196		struct vm_page * const pg = ar->ar_pages[i];
197
198		if (pg == NULL) {
199			continue;
200		}
201		KDASSERT(pg->uobject == uobj);
202		if (backward) {
203			KDASSERT(pg->offset <= off);
204			KDASSERT(i == 0 ||
205			    pg->offset < ar->ar_pages[i - 1]->offset);
206		} else {
207			KDASSERT(pg->offset >= off);
208			KDASSERT(i == 0 ||
209			    pg->offset > ar->ar_pages[i - 1]->offset);
210		}
211	}
212#endif /* defined(DEBUG) */
213	return 0;
214}
215
216/*
217 * uvm_page_array_fill_and_peek:
218 * same as uvm_page_array_peek except that, if the array is empty, try to fill
219 * it first.
220 */
221
222struct vm_page *
223uvm_page_array_fill_and_peek(struct uvm_page_array *ar, voff_t off,
224    unsigned int nwant)
225{
226	int error;
227
228	if (ar->ar_idx != ar->ar_npages) {
229		return ar->ar_pages[ar->ar_idx];
230	}
231	error = uvm_page_array_fill(ar, off, nwant);
232	if (error != 0) {
233		return NULL;
234	}
235	return uvm_page_array_peek(ar);
236}
237