subr_prof.c revision 1.47
1/*	$NetBSD: subr_prof.c,v 1.47 2014/07/10 21:13:52 christos Exp $	*/
2
3/*-
4 * Copyright (c) 1982, 1986, 1993
5 *	The Regents of the University of California.  All rights reserved.
6 *
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
9 * are met:
10 * 1. Redistributions of source code must retain the above copyright
11 *    notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 *    notice, this list of conditions and the following disclaimer in the
14 *    documentation and/or other materials provided with the distribution.
15 * 3. Neither the name of the University nor the names of its contributors
16 *    may be used to endorse or promote products derived from this software
17 *    without specific prior written permission.
18 *
19 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
20 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22 * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
23 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
24 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
25 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
26 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
27 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
28 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29 * SUCH DAMAGE.
30 *
31 *	@(#)subr_prof.c	8.4 (Berkeley) 2/14/95
32 */
33
34#include <sys/cdefs.h>
35__KERNEL_RCSID(0, "$NetBSD: subr_prof.c,v 1.47 2014/07/10 21:13:52 christos Exp $");
36
37#include <sys/param.h>
38#include <sys/systm.h>
39#include <sys/kernel.h>
40#include <sys/proc.h>
41#include <sys/mount.h>
42#include <sys/syscallargs.h>
43#include <sys/sysctl.h>
44
45#include <sys/cpu.h>
46
47#ifdef GPROF
48#include <sys/malloc.h>
49#include <sys/gmon.h>
50
51MALLOC_DEFINE(M_GPROF, "gprof", "kernel profiling buffer");
52
53/*
54 * Froms is actually a bunch of unsigned shorts indexing tos
55 */
56struct gmonparam _gmonparam = { .state = GMON_PROF_OFF };
57
58/* Actual start of the kernel text segment. */
59extern char kernel_text[];
60
61extern char etext[];
62
63
64void
65kmstartup(void)
66{
67	char *cp;
68	struct gmonparam *p = &_gmonparam;
69	/*
70	 * Round lowpc and highpc to multiples of the density we're using
71	 * so the rest of the scaling (here and in gprof) stays in ints.
72	 */
73	p->lowpc = rounddown(((u_long)kernel_text),
74		HISTFRACTION * sizeof(HISTCOUNTER));
75	p->highpc = roundup((u_long)etext,
76		HISTFRACTION * sizeof(HISTCOUNTER));
77	p->textsize = p->highpc - p->lowpc;
78	printf("Profiling kernel, textsize=%ld [%lx..%lx]\n",
79	       p->textsize, p->lowpc, p->highpc);
80	p->kcountsize = p->textsize / HISTFRACTION;
81	p->hashfraction = HASHFRACTION;
82	p->fromssize = p->textsize / HASHFRACTION;
83	p->tolimit = p->textsize * ARCDENSITY / 100;
84	if (p->tolimit < MINARCS)
85		p->tolimit = MINARCS;
86	else if (p->tolimit > MAXARCS)
87		p->tolimit = MAXARCS;
88	p->tossize = p->tolimit * sizeof(struct tostruct);
89	cp = malloc(p->kcountsize + p->fromssize + p->tossize,
90	    M_GPROF, M_NOWAIT | M_ZERO);
91	if (cp == 0) {
92		printf("No memory for profiling.\n");
93		return;
94	}
95	p->tos = (struct tostruct *)cp;
96	cp += p->tossize;
97	p->kcount = (u_short *)cp;
98	cp += p->kcountsize;
99	p->froms = (u_short *)cp;
100}
101
102/*
103 * Return kernel profiling information.
104 */
105/*
106 * sysctl helper routine for kern.profiling subtree.  enables/disables
107 * kernel profiling and gives out copies of the profiling data.
108 */
109static int
110sysctl_kern_profiling(SYSCTLFN_ARGS)
111{
112	struct gmonparam *gp = &_gmonparam;
113	int error;
114	struct sysctlnode node;
115
116	node = *rnode;
117
118	switch (node.sysctl_num) {
119	case GPROF_STATE:
120		node.sysctl_data = &gp->state;
121		break;
122	case GPROF_COUNT:
123		node.sysctl_data = gp->kcount;
124		node.sysctl_size = gp->kcountsize;
125		break;
126	case GPROF_FROMS:
127		node.sysctl_data = gp->froms;
128		node.sysctl_size = gp->fromssize;
129		break;
130	case GPROF_TOS:
131		node.sysctl_data = gp->tos;
132		node.sysctl_size = gp->tossize;
133		break;
134	case GPROF_GMONPARAM:
135		node.sysctl_data = gp;
136		node.sysctl_size = sizeof(*gp);
137		break;
138	default:
139		return (EOPNOTSUPP);
140	}
141
142	error = sysctl_lookup(SYSCTLFN_CALL(&node));
143	if (error || newp == NULL)
144		return (error);
145
146	if (node.sysctl_num == GPROF_STATE) {
147		mutex_spin_enter(&proc0.p_stmutex);
148		if (gp->state == GMON_PROF_OFF)
149			stopprofclock(&proc0);
150		else
151			startprofclock(&proc0);
152		mutex_spin_exit(&proc0.p_stmutex);
153	}
154
155	return (0);
156}
157
158SYSCTL_SETUP(sysctl_kern_gprof_setup, "sysctl kern.profiling subtree setup")
159{
160
161	sysctl_createv(clog, 0, NULL, NULL,
162		       CTLFLAG_PERMANENT,
163		       CTLTYPE_NODE, "profiling",
164		       SYSCTL_DESCR("Profiling information (available)"),
165		       NULL, 0, NULL, 0,
166		       CTL_KERN, KERN_PROF, CTL_EOL);
167
168	sysctl_createv(clog, 0, NULL, NULL,
169		       CTLFLAG_PERMANENT|CTLFLAG_READWRITE,
170		       CTLTYPE_INT, "state",
171		       SYSCTL_DESCR("Profiling state"),
172		       sysctl_kern_profiling, 0, NULL, 0,
173		       CTL_KERN, KERN_PROF, GPROF_STATE, CTL_EOL);
174	sysctl_createv(clog, 0, NULL, NULL,
175		       CTLFLAG_PERMANENT|CTLFLAG_READWRITE,
176		       CTLTYPE_STRUCT, "count",
177		       SYSCTL_DESCR("Array of statistical program counters"),
178		       sysctl_kern_profiling, 0, NULL, 0,
179		       CTL_KERN, KERN_PROF, GPROF_COUNT, CTL_EOL);
180	sysctl_createv(clog, 0, NULL, NULL,
181		       CTLFLAG_PERMANENT|CTLFLAG_READWRITE,
182		       CTLTYPE_STRUCT, "froms",
183		       SYSCTL_DESCR("Array indexed by program counter of "
184				    "call-from points"),
185		       sysctl_kern_profiling, 0, NULL, 0,
186		       CTL_KERN, KERN_PROF, GPROF_FROMS, CTL_EOL);
187	sysctl_createv(clog, 0, NULL, NULL,
188		       CTLFLAG_PERMANENT|CTLFLAG_READWRITE,
189		       CTLTYPE_STRUCT, "tos",
190		       SYSCTL_DESCR("Array of structures describing "
191				    "destination of calls and their counts"),
192		       sysctl_kern_profiling, 0, NULL, 0,
193		       CTL_KERN, KERN_PROF, GPROF_TOS, CTL_EOL);
194	sysctl_createv(clog, 0, NULL, NULL,
195		       CTLFLAG_PERMANENT,
196		       CTLTYPE_STRUCT, "gmonparam",
197		       SYSCTL_DESCR("Structure giving the sizes of the above "
198				    "arrays"),
199		       sysctl_kern_profiling, 0, NULL, 0,
200		       CTL_KERN, KERN_PROF, GPROF_GMONPARAM, CTL_EOL);
201}
202#endif /* GPROF */
203
204/*
205 * Profiling system call.
206 *
207 * The scale factor is a fixed point number with 16 bits of fraction, so that
208 * 1.0 is represented as 0x10000.  A scale factor of 0 turns off profiling.
209 */
210/* ARGSUSED */
211int
212sys_profil(struct lwp *l, const struct sys_profil_args *uap, register_t *retval)
213{
214	/* {
215		syscallarg(char *) samples;
216		syscallarg(size_t) size;
217		syscallarg(u_long) offset;
218		syscallarg(u_int) scale;
219	} */
220	struct proc *p = l->l_proc;
221	struct uprof *upp;
222
223	if (SCARG(uap, scale) > (1 << 16))
224		return (EINVAL);
225	if (SCARG(uap, scale) == 0) {
226		mutex_spin_enter(&p->p_stmutex);
227		stopprofclock(p);
228		mutex_spin_exit(&p->p_stmutex);
229		return (0);
230	}
231	upp = &p->p_stats->p_prof;
232
233	/* Block profile interrupts while changing state. */
234	mutex_spin_enter(&p->p_stmutex);
235	upp->pr_off = SCARG(uap, offset);
236	upp->pr_scale = SCARG(uap, scale);
237	upp->pr_base = SCARG(uap, samples);
238	upp->pr_size = SCARG(uap, size);
239	startprofclock(p);
240	mutex_spin_exit(&p->p_stmutex);
241
242	return (0);
243}
244
245/*
246 * Scale is a fixed-point number with the binary point 16 bits
247 * into the value, and is <= 1.0.  pc is at most 32 bits, so the
248 * intermediate result is at most 48 bits.
249 */
250#define	PC_TO_INDEX(pc, prof) \
251	((int)(((u_quad_t)((pc) - (prof)->pr_off) * \
252	    (u_quad_t)((prof)->pr_scale)) >> 16) & ~1)
253
254/*
255 * Collect user-level profiling statistics; called on a profiling tick,
256 * when a process is running in user-mode.  This routine may be called
257 * from an interrupt context.  We try to update the user profiling buffers
258 * cheaply with fuswintr() and suswintr().  If that fails, we revert to
259 * an AST that will vector us to trap() with a context in which copyin
260 * and copyout will work.  Trap will then call addupc_task().
261 *
262 * Note that we may (rarely) not get around to the AST soon enough, and
263 * lose profile ticks when the next tick overwrites this one, but in this
264 * case the system is overloaded and the profile is probably already
265 * inaccurate.
266 */
267void
268addupc_intr(struct lwp *l, u_long pc)
269{
270	struct uprof *prof;
271	struct proc *p;
272	void *addr;
273	u_int i;
274	int v;
275
276	p = l->l_proc;
277
278	KASSERT(mutex_owned(&p->p_stmutex));
279
280	prof = &p->p_stats->p_prof;
281	if (pc < prof->pr_off ||
282	    (i = PC_TO_INDEX(pc, prof)) >= prof->pr_size)
283		return;			/* out of range; ignore */
284
285	addr = prof->pr_base + i;
286	mutex_spin_exit(&p->p_stmutex);
287	if ((v = fuswintr(addr)) == -1 || suswintr(addr, v + 1) == -1) {
288		/* XXXSMP */
289		prof->pr_addr = pc;
290		prof->pr_ticks++;
291		cpu_need_proftick(l);
292	}
293	mutex_spin_enter(&p->p_stmutex);
294}
295
296/*
297 * Much like before, but we can afford to take faults here.  If the
298 * update fails, we simply turn off profiling.
299 */
300void
301addupc_task(struct lwp *l, u_long pc, u_int ticks)
302{
303	struct uprof *prof;
304	struct proc *p;
305	void *addr;
306	int error;
307	u_int i;
308	u_short v;
309
310	p = l->l_proc;
311
312	if (ticks == 0)
313		return;
314
315	mutex_spin_enter(&p->p_stmutex);
316	prof = &p->p_stats->p_prof;
317
318	/* Testing P_PROFIL may be unnecessary, but is certainly safe. */
319	if ((p->p_stflag & PST_PROFIL) == 0 || pc < prof->pr_off ||
320	    (i = PC_TO_INDEX(pc, prof)) >= prof->pr_size) {
321		mutex_spin_exit(&p->p_stmutex);
322		return;
323	}
324
325	addr = prof->pr_base + i;
326	mutex_spin_exit(&p->p_stmutex);
327	if ((error = copyin(addr, (void *)&v, sizeof(v))) == 0) {
328		v += ticks;
329		error = copyout((void *)&v, addr, sizeof(v));
330	}
331	if (error != 0) {
332		mutex_spin_enter(&p->p_stmutex);
333		stopprofclock(p);
334		mutex_spin_exit(&p->p_stmutex);
335	}
336}
337