Deleted Added
full compact
subr_pcpu.c (261725) subr_pcpu.c (302372)
1/*-
2 * Copyright (c) 2001 Wind River Systems, Inc.
3 * All rights reserved.
4 * Written by: John Baldwin <jhb@FreeBSD.org>
5 *
6 * Copyright (c) 2009 Jeffrey Roberson <jeff@freebsd.org>
7 * All rights reserved.
8 *
9 * Redistribution and use in source and binary forms, with or without
10 * modification, are permitted provided that the following conditions
11 * are met:
12 * 1. Redistributions of source code must retain the above copyright
13 * notice, this list of conditions and the following disclaimer.
14 * 2. Redistributions in binary form must reproduce the above copyright
15 * notice, this list of conditions and the following disclaimer in the
16 * documentation and/or other materials provided with the distribution.
17 * 4. Neither the name of the author nor the names of any co-contributors
18 * may be used to endorse or promote products derived from this software
19 * without specific prior written permission.
20 *
21 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
22 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
23 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
24 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
25 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
26 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
27 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
28 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
29 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
30 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
31 * SUCH DAMAGE.
32 */
33
34/*
35 * This module provides MI support for per-cpu data.
36 *
37 * Each architecture determines the mapping of logical CPU IDs to physical
38 * CPUs. The requirements of this mapping are as follows:
39 * - Logical CPU IDs must reside in the range 0 ... MAXCPU - 1.
40 * - The mapping is not required to be dense. That is, there may be
41 * gaps in the mappings.
42 * - The platform sets the value of MAXCPU in <machine/param.h>.
43 * - It is suggested, but not required, that in the non-SMP case, the
44 * platform define MAXCPU to be 1 and define the logical ID of the
45 * sole CPU as 0.
46 */
47
48#include <sys/cdefs.h>
1/*-
2 * Copyright (c) 2001 Wind River Systems, Inc.
3 * All rights reserved.
4 * Written by: John Baldwin <jhb@FreeBSD.org>
5 *
6 * Copyright (c) 2009 Jeffrey Roberson <jeff@freebsd.org>
7 * All rights reserved.
8 *
9 * Redistribution and use in source and binary forms, with or without
10 * modification, are permitted provided that the following conditions
11 * are met:
12 * 1. Redistributions of source code must retain the above copyright
13 * notice, this list of conditions and the following disclaimer.
14 * 2. Redistributions in binary form must reproduce the above copyright
15 * notice, this list of conditions and the following disclaimer in the
16 * documentation and/or other materials provided with the distribution.
17 * 4. Neither the name of the author nor the names of any co-contributors
18 * may be used to endorse or promote products derived from this software
19 * without specific prior written permission.
20 *
21 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
22 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
23 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
24 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
25 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
26 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
27 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
28 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
29 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
30 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
31 * SUCH DAMAGE.
32 */
33
34/*
35 * This module provides MI support for per-cpu data.
36 *
37 * Each architecture determines the mapping of logical CPU IDs to physical
38 * CPUs. The requirements of this mapping are as follows:
39 * - Logical CPU IDs must reside in the range 0 ... MAXCPU - 1.
40 * - The mapping is not required to be dense. That is, there may be
41 * gaps in the mappings.
42 * - The platform sets the value of MAXCPU in <machine/param.h>.
43 * - It is suggested, but not required, that in the non-SMP case, the
44 * platform define MAXCPU to be 1 and define the logical ID of the
45 * sole CPU as 0.
46 */
47
48#include <sys/cdefs.h>
49__FBSDID("$FreeBSD: head/sys/kern/subr_pcpu.c 261725 2014-02-10 19:59:46Z glebius $");
49__FBSDID("$FreeBSD: head/sys/kern/subr_pcpu.c 302372 2016-07-06 14:09:49Z nwhitehorn $");
50
51#include "opt_ddb.h"
52
53#include <sys/param.h>
54#include <sys/systm.h>
55#include <sys/sysctl.h>
56#include <sys/lock.h>
57#include <sys/malloc.h>
58#include <sys/pcpu.h>
59#include <sys/proc.h>
60#include <sys/smp.h>
61#include <sys/sx.h>
62#include <vm/uma.h>
63#include <ddb/ddb.h>
64
65static MALLOC_DEFINE(M_PCPU, "Per-cpu", "Per-cpu resource accouting.");
66
67struct dpcpu_free {
68 uintptr_t df_start;
69 int df_len;
70 TAILQ_ENTRY(dpcpu_free) df_link;
71};
72
73static DPCPU_DEFINE(char, modspace[DPCPU_MODMIN]);
74static TAILQ_HEAD(, dpcpu_free) dpcpu_head = TAILQ_HEAD_INITIALIZER(dpcpu_head);
75static struct sx dpcpu_lock;
76uintptr_t dpcpu_off[MAXCPU];
77struct pcpu *cpuid_to_pcpu[MAXCPU];
78struct cpuhead cpuhead = STAILQ_HEAD_INITIALIZER(cpuhead);
79
80/*
81 * Initialize the MI portions of a struct pcpu.
82 */
83void
84pcpu_init(struct pcpu *pcpu, int cpuid, size_t size)
85{
86
87 bzero(pcpu, size);
88 KASSERT(cpuid >= 0 && cpuid < MAXCPU,
89 ("pcpu_init: invalid cpuid %d", cpuid));
90 pcpu->pc_cpuid = cpuid;
91 cpuid_to_pcpu[cpuid] = pcpu;
92 STAILQ_INSERT_TAIL(&cpuhead, pcpu, pc_allcpu);
93 cpu_pcpu_init(pcpu, cpuid, size);
94 pcpu->pc_rm_queue.rmq_next = &pcpu->pc_rm_queue;
95 pcpu->pc_rm_queue.rmq_prev = &pcpu->pc_rm_queue;
96}
97
98void
99dpcpu_init(void *dpcpu, int cpuid)
100{
101 struct pcpu *pcpu;
102
103 pcpu = pcpu_find(cpuid);
104 pcpu->pc_dynamic = (uintptr_t)dpcpu - DPCPU_START;
105
106 /*
107 * Initialize defaults from our linker section.
108 */
109 memcpy(dpcpu, (void *)DPCPU_START, DPCPU_BYTES);
110
111 /*
112 * Place it in the global pcpu offset array.
113 */
114 dpcpu_off[cpuid] = pcpu->pc_dynamic;
115}
116
117static void
118dpcpu_startup(void *dummy __unused)
119{
120 struct dpcpu_free *df;
121
122 df = malloc(sizeof(*df), M_PCPU, M_WAITOK | M_ZERO);
123 df->df_start = (uintptr_t)&DPCPU_NAME(modspace);
124 df->df_len = DPCPU_MODMIN;
125 TAILQ_INSERT_HEAD(&dpcpu_head, df, df_link);
126 sx_init(&dpcpu_lock, "dpcpu alloc lock");
127}
128SYSINIT(dpcpu, SI_SUB_KLD, SI_ORDER_FIRST, dpcpu_startup, 0);
129
130/*
131 * UMA_PCPU_ZONE zones, that are available for all kernel
132 * consumers. Right now 64 bit zone is used for counter(9)
133 * and pointer zone is used by flowtable.
134 */
135
136uma_zone_t pcpu_zone_64;
137uma_zone_t pcpu_zone_ptr;
138
139static void
140pcpu_zones_startup(void)
141{
142
143 pcpu_zone_64 = uma_zcreate("64 pcpu", sizeof(uint64_t),
144 NULL, NULL, NULL, NULL, UMA_ALIGN_PTR, UMA_ZONE_PCPU);
145
146 if (sizeof(uint64_t) == sizeof(void *))
147 pcpu_zone_ptr = pcpu_zone_64;
148 else
149 pcpu_zone_ptr = uma_zcreate("ptr pcpu", sizeof(void *),
150 NULL, NULL, NULL, NULL, UMA_ALIGN_PTR, UMA_ZONE_PCPU);
151}
152SYSINIT(pcpu_zones, SI_SUB_KMEM, SI_ORDER_ANY, pcpu_zones_startup, NULL);
153
154/*
155 * First-fit extent based allocator for allocating space in the per-cpu
156 * region reserved for modules. This is only intended for use by the
157 * kernel linkers to place module linker sets.
158 */
159void *
160dpcpu_alloc(int size)
161{
162 struct dpcpu_free *df;
163 void *s;
164
165 s = NULL;
166 size = roundup2(size, sizeof(void *));
167 sx_xlock(&dpcpu_lock);
168 TAILQ_FOREACH(df, &dpcpu_head, df_link) {
169 if (df->df_len < size)
170 continue;
171 if (df->df_len == size) {
172 s = (void *)df->df_start;
173 TAILQ_REMOVE(&dpcpu_head, df, df_link);
174 free(df, M_PCPU);
175 break;
176 }
177 s = (void *)df->df_start;
178 df->df_len -= size;
179 df->df_start = df->df_start + size;
180 break;
181 }
182 sx_xunlock(&dpcpu_lock);
183
184 return (s);
185}
186
187/*
188 * Free dynamic per-cpu space at module unload time.
189 */
190void
191dpcpu_free(void *s, int size)
192{
193 struct dpcpu_free *df;
194 struct dpcpu_free *dn;
195 uintptr_t start;
196 uintptr_t end;
197
198 size = roundup2(size, sizeof(void *));
199 start = (uintptr_t)s;
200 end = start + size;
201 /*
202 * Free a region of space and merge it with as many neighbors as
203 * possible. Keeping the list sorted simplifies this operation.
204 */
205 sx_xlock(&dpcpu_lock);
206 TAILQ_FOREACH(df, &dpcpu_head, df_link) {
207 if (df->df_start > end)
208 break;
209 /*
210 * If we expand at the end of an entry we may have to
211 * merge it with the one following it as well.
212 */
213 if (df->df_start + df->df_len == start) {
214 df->df_len += size;
215 dn = TAILQ_NEXT(df, df_link);
216 if (df->df_start + df->df_len == dn->df_start) {
217 df->df_len += dn->df_len;
218 TAILQ_REMOVE(&dpcpu_head, dn, df_link);
219 free(dn, M_PCPU);
220 }
221 sx_xunlock(&dpcpu_lock);
222 return;
223 }
224 if (df->df_start == end) {
225 df->df_start = start;
226 df->df_len += size;
227 sx_xunlock(&dpcpu_lock);
228 return;
229 }
230 }
231 dn = malloc(sizeof(*df), M_PCPU, M_WAITOK | M_ZERO);
232 dn->df_start = start;
233 dn->df_len = size;
234 if (df)
235 TAILQ_INSERT_BEFORE(df, dn, df_link);
236 else
237 TAILQ_INSERT_TAIL(&dpcpu_head, dn, df_link);
238 sx_xunlock(&dpcpu_lock);
239}
240
241/*
242 * Initialize the per-cpu storage from an updated linker-set region.
243 */
244void
245dpcpu_copy(void *s, int size)
246{
247#ifdef SMP
248 uintptr_t dpcpu;
249 int i;
250
50
51#include "opt_ddb.h"
52
53#include <sys/param.h>
54#include <sys/systm.h>
55#include <sys/sysctl.h>
56#include <sys/lock.h>
57#include <sys/malloc.h>
58#include <sys/pcpu.h>
59#include <sys/proc.h>
60#include <sys/smp.h>
61#include <sys/sx.h>
62#include <vm/uma.h>
63#include <ddb/ddb.h>
64
65static MALLOC_DEFINE(M_PCPU, "Per-cpu", "Per-cpu resource accouting.");
66
67struct dpcpu_free {
68 uintptr_t df_start;
69 int df_len;
70 TAILQ_ENTRY(dpcpu_free) df_link;
71};
72
73static DPCPU_DEFINE(char, modspace[DPCPU_MODMIN]);
74static TAILQ_HEAD(, dpcpu_free) dpcpu_head = TAILQ_HEAD_INITIALIZER(dpcpu_head);
75static struct sx dpcpu_lock;
76uintptr_t dpcpu_off[MAXCPU];
77struct pcpu *cpuid_to_pcpu[MAXCPU];
78struct cpuhead cpuhead = STAILQ_HEAD_INITIALIZER(cpuhead);
79
80/*
81 * Initialize the MI portions of a struct pcpu.
82 */
83void
84pcpu_init(struct pcpu *pcpu, int cpuid, size_t size)
85{
86
87 bzero(pcpu, size);
88 KASSERT(cpuid >= 0 && cpuid < MAXCPU,
89 ("pcpu_init: invalid cpuid %d", cpuid));
90 pcpu->pc_cpuid = cpuid;
91 cpuid_to_pcpu[cpuid] = pcpu;
92 STAILQ_INSERT_TAIL(&cpuhead, pcpu, pc_allcpu);
93 cpu_pcpu_init(pcpu, cpuid, size);
94 pcpu->pc_rm_queue.rmq_next = &pcpu->pc_rm_queue;
95 pcpu->pc_rm_queue.rmq_prev = &pcpu->pc_rm_queue;
96}
97
98void
99dpcpu_init(void *dpcpu, int cpuid)
100{
101 struct pcpu *pcpu;
102
103 pcpu = pcpu_find(cpuid);
104 pcpu->pc_dynamic = (uintptr_t)dpcpu - DPCPU_START;
105
106 /*
107 * Initialize defaults from our linker section.
108 */
109 memcpy(dpcpu, (void *)DPCPU_START, DPCPU_BYTES);
110
111 /*
112 * Place it in the global pcpu offset array.
113 */
114 dpcpu_off[cpuid] = pcpu->pc_dynamic;
115}
116
117static void
118dpcpu_startup(void *dummy __unused)
119{
120 struct dpcpu_free *df;
121
122 df = malloc(sizeof(*df), M_PCPU, M_WAITOK | M_ZERO);
123 df->df_start = (uintptr_t)&DPCPU_NAME(modspace);
124 df->df_len = DPCPU_MODMIN;
125 TAILQ_INSERT_HEAD(&dpcpu_head, df, df_link);
126 sx_init(&dpcpu_lock, "dpcpu alloc lock");
127}
128SYSINIT(dpcpu, SI_SUB_KLD, SI_ORDER_FIRST, dpcpu_startup, 0);
129
130/*
131 * UMA_PCPU_ZONE zones, that are available for all kernel
132 * consumers. Right now 64 bit zone is used for counter(9)
133 * and pointer zone is used by flowtable.
134 */
135
136uma_zone_t pcpu_zone_64;
137uma_zone_t pcpu_zone_ptr;
138
139static void
140pcpu_zones_startup(void)
141{
142
143 pcpu_zone_64 = uma_zcreate("64 pcpu", sizeof(uint64_t),
144 NULL, NULL, NULL, NULL, UMA_ALIGN_PTR, UMA_ZONE_PCPU);
145
146 if (sizeof(uint64_t) == sizeof(void *))
147 pcpu_zone_ptr = pcpu_zone_64;
148 else
149 pcpu_zone_ptr = uma_zcreate("ptr pcpu", sizeof(void *),
150 NULL, NULL, NULL, NULL, UMA_ALIGN_PTR, UMA_ZONE_PCPU);
151}
152SYSINIT(pcpu_zones, SI_SUB_KMEM, SI_ORDER_ANY, pcpu_zones_startup, NULL);
153
154/*
155 * First-fit extent based allocator for allocating space in the per-cpu
156 * region reserved for modules. This is only intended for use by the
157 * kernel linkers to place module linker sets.
158 */
159void *
160dpcpu_alloc(int size)
161{
162 struct dpcpu_free *df;
163 void *s;
164
165 s = NULL;
166 size = roundup2(size, sizeof(void *));
167 sx_xlock(&dpcpu_lock);
168 TAILQ_FOREACH(df, &dpcpu_head, df_link) {
169 if (df->df_len < size)
170 continue;
171 if (df->df_len == size) {
172 s = (void *)df->df_start;
173 TAILQ_REMOVE(&dpcpu_head, df, df_link);
174 free(df, M_PCPU);
175 break;
176 }
177 s = (void *)df->df_start;
178 df->df_len -= size;
179 df->df_start = df->df_start + size;
180 break;
181 }
182 sx_xunlock(&dpcpu_lock);
183
184 return (s);
185}
186
187/*
188 * Free dynamic per-cpu space at module unload time.
189 */
190void
191dpcpu_free(void *s, int size)
192{
193 struct dpcpu_free *df;
194 struct dpcpu_free *dn;
195 uintptr_t start;
196 uintptr_t end;
197
198 size = roundup2(size, sizeof(void *));
199 start = (uintptr_t)s;
200 end = start + size;
201 /*
202 * Free a region of space and merge it with as many neighbors as
203 * possible. Keeping the list sorted simplifies this operation.
204 */
205 sx_xlock(&dpcpu_lock);
206 TAILQ_FOREACH(df, &dpcpu_head, df_link) {
207 if (df->df_start > end)
208 break;
209 /*
210 * If we expand at the end of an entry we may have to
211 * merge it with the one following it as well.
212 */
213 if (df->df_start + df->df_len == start) {
214 df->df_len += size;
215 dn = TAILQ_NEXT(df, df_link);
216 if (df->df_start + df->df_len == dn->df_start) {
217 df->df_len += dn->df_len;
218 TAILQ_REMOVE(&dpcpu_head, dn, df_link);
219 free(dn, M_PCPU);
220 }
221 sx_xunlock(&dpcpu_lock);
222 return;
223 }
224 if (df->df_start == end) {
225 df->df_start = start;
226 df->df_len += size;
227 sx_xunlock(&dpcpu_lock);
228 return;
229 }
230 }
231 dn = malloc(sizeof(*df), M_PCPU, M_WAITOK | M_ZERO);
232 dn->df_start = start;
233 dn->df_len = size;
234 if (df)
235 TAILQ_INSERT_BEFORE(df, dn, df_link);
236 else
237 TAILQ_INSERT_TAIL(&dpcpu_head, dn, df_link);
238 sx_xunlock(&dpcpu_lock);
239}
240
241/*
242 * Initialize the per-cpu storage from an updated linker-set region.
243 */
244void
245dpcpu_copy(void *s, int size)
246{
247#ifdef SMP
248 uintptr_t dpcpu;
249 int i;
250
251 for (i = 0; i < mp_ncpus; ++i) {
251 CPU_FOREACH(i) {
252 dpcpu = dpcpu_off[i];
253 if (dpcpu == 0)
254 continue;
255 memcpy((void *)(dpcpu + (uintptr_t)s), s, size);
256 }
257#else
258 memcpy((void *)(dpcpu_off[0] + (uintptr_t)s), s, size);
259#endif
260}
261
262/*
263 * Destroy a struct pcpu.
264 */
265void
266pcpu_destroy(struct pcpu *pcpu)
267{
268
269 STAILQ_REMOVE(&cpuhead, pcpu, pcpu, pc_allcpu);
270 cpuid_to_pcpu[pcpu->pc_cpuid] = NULL;
271 dpcpu_off[pcpu->pc_cpuid] = 0;
272}
273
274/*
275 * Locate a struct pcpu by cpu id.
276 */
277struct pcpu *
278pcpu_find(u_int cpuid)
279{
280
281 return (cpuid_to_pcpu[cpuid]);
282}
283
284int
285sysctl_dpcpu_quad(SYSCTL_HANDLER_ARGS)
286{
287 uintptr_t dpcpu;
288 int64_t count;
289 int i;
290
291 count = 0;
252 dpcpu = dpcpu_off[i];
253 if (dpcpu == 0)
254 continue;
255 memcpy((void *)(dpcpu + (uintptr_t)s), s, size);
256 }
257#else
258 memcpy((void *)(dpcpu_off[0] + (uintptr_t)s), s, size);
259#endif
260}
261
262/*
263 * Destroy a struct pcpu.
264 */
265void
266pcpu_destroy(struct pcpu *pcpu)
267{
268
269 STAILQ_REMOVE(&cpuhead, pcpu, pcpu, pc_allcpu);
270 cpuid_to_pcpu[pcpu->pc_cpuid] = NULL;
271 dpcpu_off[pcpu->pc_cpuid] = 0;
272}
273
274/*
275 * Locate a struct pcpu by cpu id.
276 */
277struct pcpu *
278pcpu_find(u_int cpuid)
279{
280
281 return (cpuid_to_pcpu[cpuid]);
282}
283
284int
285sysctl_dpcpu_quad(SYSCTL_HANDLER_ARGS)
286{
287 uintptr_t dpcpu;
288 int64_t count;
289 int i;
290
291 count = 0;
292 for (i = 0; i < mp_ncpus; ++i) {
292 CPU_FOREACH(i) {
293 dpcpu = dpcpu_off[i];
294 if (dpcpu == 0)
295 continue;
296 count += *(int64_t *)(dpcpu + (uintptr_t)arg1);
297 }
298 return (SYSCTL_OUT(req, &count, sizeof(count)));
299}
300
301int
302sysctl_dpcpu_long(SYSCTL_HANDLER_ARGS)
303{
304 uintptr_t dpcpu;
305 long count;
306 int i;
307
308 count = 0;
293 dpcpu = dpcpu_off[i];
294 if (dpcpu == 0)
295 continue;
296 count += *(int64_t *)(dpcpu + (uintptr_t)arg1);
297 }
298 return (SYSCTL_OUT(req, &count, sizeof(count)));
299}
300
301int
302sysctl_dpcpu_long(SYSCTL_HANDLER_ARGS)
303{
304 uintptr_t dpcpu;
305 long count;
306 int i;
307
308 count = 0;
309 for (i = 0; i < mp_ncpus; ++i) {
309 CPU_FOREACH(i) {
310 dpcpu = dpcpu_off[i];
311 if (dpcpu == 0)
312 continue;
313 count += *(long *)(dpcpu + (uintptr_t)arg1);
314 }
315 return (SYSCTL_OUT(req, &count, sizeof(count)));
316}
317
318int
319sysctl_dpcpu_int(SYSCTL_HANDLER_ARGS)
320{
321 uintptr_t dpcpu;
322 int count;
323 int i;
324
325 count = 0;
310 dpcpu = dpcpu_off[i];
311 if (dpcpu == 0)
312 continue;
313 count += *(long *)(dpcpu + (uintptr_t)arg1);
314 }
315 return (SYSCTL_OUT(req, &count, sizeof(count)));
316}
317
318int
319sysctl_dpcpu_int(SYSCTL_HANDLER_ARGS)
320{
321 uintptr_t dpcpu;
322 int count;
323 int i;
324
325 count = 0;
326 for (i = 0; i < mp_ncpus; ++i) {
326 CPU_FOREACH(i) {
327 dpcpu = dpcpu_off[i];
328 if (dpcpu == 0)
329 continue;
330 count += *(int *)(dpcpu + (uintptr_t)arg1);
331 }
332 return (SYSCTL_OUT(req, &count, sizeof(count)));
333}
334
335#ifdef DDB
336DB_SHOW_COMMAND(dpcpu_off, db_show_dpcpu_off)
337{
338 int id;
339
340 CPU_FOREACH(id) {
341 db_printf("dpcpu_off[%2d] = 0x%jx (+ DPCPU_START = %p)\n",
342 id, (uintmax_t)dpcpu_off[id],
343 (void *)(uintptr_t)(dpcpu_off[id] + DPCPU_START));
344 }
345}
346
347static void
348show_pcpu(struct pcpu *pc)
349{
350 struct thread *td;
351
352 db_printf("cpuid = %d\n", pc->pc_cpuid);
353 db_printf("dynamic pcpu = %p\n", (void *)pc->pc_dynamic);
354 db_printf("curthread = ");
355 td = pc->pc_curthread;
356 if (td != NULL)
357 db_printf("%p: pid %d \"%s\"\n", td, td->td_proc->p_pid,
358 td->td_name);
359 else
360 db_printf("none\n");
361 db_printf("curpcb = %p\n", pc->pc_curpcb);
362 db_printf("fpcurthread = ");
363 td = pc->pc_fpcurthread;
364 if (td != NULL)
365 db_printf("%p: pid %d \"%s\"\n", td, td->td_proc->p_pid,
366 td->td_name);
367 else
368 db_printf("none\n");
369 db_printf("idlethread = ");
370 td = pc->pc_idlethread;
371 if (td != NULL)
372 db_printf("%p: tid %d \"%s\"\n", td, td->td_tid, td->td_name);
373 else
374 db_printf("none\n");
375 db_show_mdpcpu(pc);
376
377#ifdef VIMAGE
378 db_printf("curvnet = %p\n", pc->pc_curthread->td_vnet);
379#endif
380
381#ifdef WITNESS
382 db_printf("spin locks held:\n");
383 witness_list_locks(&pc->pc_spinlocks, db_printf);
384#endif
385}
386
387DB_SHOW_COMMAND(pcpu, db_show_pcpu)
388{
389 struct pcpu *pc;
390 int id;
391
392 if (have_addr)
393 id = ((addr >> 4) % 16) * 10 + (addr % 16);
394 else
395 id = PCPU_GET(cpuid);
396 pc = pcpu_find(id);
397 if (pc == NULL) {
398 db_printf("CPU %d not found\n", id);
399 return;
400 }
401 show_pcpu(pc);
402}
403
404DB_SHOW_ALL_COMMAND(pcpu, db_show_cpu_all)
405{
406 struct pcpu *pc;
407 int id;
408
409 db_printf("Current CPU: %d\n\n", PCPU_GET(cpuid));
410 for (id = 0; id <= mp_maxid; id++) {
411 pc = pcpu_find(id);
412 if (pc != NULL) {
413 show_pcpu(pc);
414 db_printf("\n");
415 }
416 }
417}
418DB_SHOW_ALIAS(allpcpu, db_show_cpu_all);
419#endif
327 dpcpu = dpcpu_off[i];
328 if (dpcpu == 0)
329 continue;
330 count += *(int *)(dpcpu + (uintptr_t)arg1);
331 }
332 return (SYSCTL_OUT(req, &count, sizeof(count)));
333}
334
335#ifdef DDB
336DB_SHOW_COMMAND(dpcpu_off, db_show_dpcpu_off)
337{
338 int id;
339
340 CPU_FOREACH(id) {
341 db_printf("dpcpu_off[%2d] = 0x%jx (+ DPCPU_START = %p)\n",
342 id, (uintmax_t)dpcpu_off[id],
343 (void *)(uintptr_t)(dpcpu_off[id] + DPCPU_START));
344 }
345}
346
347static void
348show_pcpu(struct pcpu *pc)
349{
350 struct thread *td;
351
352 db_printf("cpuid = %d\n", pc->pc_cpuid);
353 db_printf("dynamic pcpu = %p\n", (void *)pc->pc_dynamic);
354 db_printf("curthread = ");
355 td = pc->pc_curthread;
356 if (td != NULL)
357 db_printf("%p: pid %d \"%s\"\n", td, td->td_proc->p_pid,
358 td->td_name);
359 else
360 db_printf("none\n");
361 db_printf("curpcb = %p\n", pc->pc_curpcb);
362 db_printf("fpcurthread = ");
363 td = pc->pc_fpcurthread;
364 if (td != NULL)
365 db_printf("%p: pid %d \"%s\"\n", td, td->td_proc->p_pid,
366 td->td_name);
367 else
368 db_printf("none\n");
369 db_printf("idlethread = ");
370 td = pc->pc_idlethread;
371 if (td != NULL)
372 db_printf("%p: tid %d \"%s\"\n", td, td->td_tid, td->td_name);
373 else
374 db_printf("none\n");
375 db_show_mdpcpu(pc);
376
377#ifdef VIMAGE
378 db_printf("curvnet = %p\n", pc->pc_curthread->td_vnet);
379#endif
380
381#ifdef WITNESS
382 db_printf("spin locks held:\n");
383 witness_list_locks(&pc->pc_spinlocks, db_printf);
384#endif
385}
386
387DB_SHOW_COMMAND(pcpu, db_show_pcpu)
388{
389 struct pcpu *pc;
390 int id;
391
392 if (have_addr)
393 id = ((addr >> 4) % 16) * 10 + (addr % 16);
394 else
395 id = PCPU_GET(cpuid);
396 pc = pcpu_find(id);
397 if (pc == NULL) {
398 db_printf("CPU %d not found\n", id);
399 return;
400 }
401 show_pcpu(pc);
402}
403
404DB_SHOW_ALL_COMMAND(pcpu, db_show_cpu_all)
405{
406 struct pcpu *pc;
407 int id;
408
409 db_printf("Current CPU: %d\n\n", PCPU_GET(cpuid));
410 for (id = 0; id <= mp_maxid; id++) {
411 pc = pcpu_find(id);
412 if (pc != NULL) {
413 show_pcpu(pc);
414 db_printf("\n");
415 }
416 }
417}
418DB_SHOW_ALIAS(allpcpu, db_show_cpu_all);
419#endif