vm_machdep.c revision 90643
136865Sdfr/*-
236865Sdfr * Copyright (c) 1982, 1986 The Regents of the University of California.
336865Sdfr * Copyright (c) 1989, 1990 William Jolitz
436865Sdfr * Copyright (c) 1994 John Dyson
536865Sdfr * All rights reserved.
636865Sdfr *
736865Sdfr * This code is derived from software contributed to Berkeley by
836865Sdfr * the Systems Programming Group of the University of Utah Computer
936865Sdfr * Science Department, and William Jolitz.
1036865Sdfr *
1136865Sdfr * Redistribution and use in source and binary forms, with or without
1236865Sdfr * modification, are permitted provided that the following conditions
1336865Sdfr * are met:
1436865Sdfr * 1. Redistributions of source code must retain the above copyright
1536865Sdfr *    notice, this list of conditions and the following disclaimer.
1636865Sdfr * 2. Redistributions in binary form must reproduce the above copyright
1736865Sdfr *    notice, this list of conditions and the following disclaimer in the
1836865Sdfr *    documentation and/or other materials provided with the distribution.
1936865Sdfr * 3. All advertising materials mentioning features or use of this software
2036865Sdfr *    must display the following acknowledgement:
2136865Sdfr *	This product includes software developed by the University of
2236865Sdfr *	California, Berkeley and its contributors.
2336865Sdfr * 4. Neither the name of the University nor the names of its contributors
2436865Sdfr *    may be used to endorse or promote products derived from this software
2536865Sdfr *    without specific prior written permission.
2636865Sdfr *
2736865Sdfr * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
2836865Sdfr * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
2936865Sdfr * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
3036865Sdfr * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
3136865Sdfr * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
3236865Sdfr * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
3336865Sdfr * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
3436865Sdfr * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
3536865Sdfr * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
3636865Sdfr * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
3736865Sdfr * SUCH DAMAGE.
3836865Sdfr *
3936865Sdfr *	from: @(#)vm_machdep.c	7.3 (Berkeley) 5/13/91
4036865Sdfr *	Utah $Hdr: vm_machdep.c 1.16.1.1 89/06/23$
4150477Speter * $FreeBSD: head/sys/powerpc/aim/vm_machdep.c 90643 2002-02-14 01:39:11Z benno $
4236865Sdfr */
4336865Sdfr/*
4436865Sdfr * Copyright (c) 1994, 1995, 1996 Carnegie-Mellon University.
4536865Sdfr * All rights reserved.
4636865Sdfr *
4736865Sdfr * Author: Chris G. Demetriou
4836865Sdfr *
4936865Sdfr * Permission to use, copy, modify and distribute this software and
5036865Sdfr * its documentation is hereby granted, provided that both the copyright
5136865Sdfr * notice and this permission notice appear in all copies of the
5236865Sdfr * software, derivative works or modified versions, and any portions
5336865Sdfr * thereof, and that both notices appear in supporting documentation.
5436865Sdfr *
5536865Sdfr * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
5636865Sdfr * CONDITION.  CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND
5736865Sdfr * FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
5836865Sdfr *
5936865Sdfr * Carnegie Mellon requests users of this software to return to
6036865Sdfr *
6136865Sdfr *  Software Distribution Coordinator  or  Software.Distribution@CS.CMU.EDU
6236865Sdfr *  School of Computer Science
6336865Sdfr *  Carnegie Mellon University
6436865Sdfr *  Pittsburgh PA 15213-3890
6536865Sdfr *
6636865Sdfr * any improvements or extensions that they make and grant Carnegie the
6736865Sdfr * rights to redistribute these changes.
6836865Sdfr */
6936865Sdfr
7036865Sdfr#include <sys/param.h>
7136865Sdfr#include <sys/systm.h>
7236865Sdfr#include <sys/proc.h>
7336865Sdfr#include <sys/malloc.h>
7460041Sphk#include <sys/bio.h>
7536865Sdfr#include <sys/buf.h>
7685201Smp#include <sys/ktr.h>
7785201Smp#include <sys/lock.h>
7867365Sjhb#include <sys/mutex.h>
7936865Sdfr#include <sys/vnode.h>
8036865Sdfr#include <sys/vmmeter.h>
8136865Sdfr#include <sys/kernel.h>
8236865Sdfr#include <sys/sysctl.h>
8354207Speter#include <sys/unistd.h>
8436865Sdfr
8536865Sdfr#include <machine/clock.h>
8636865Sdfr#include <machine/cpu.h>
8741499Sdfr#include <machine/fpu.h>
8836865Sdfr#include <machine/md_var.h>
8936865Sdfr
9078342Sbenno#include <dev/ofw/openfirm.h>
9178342Sbenno
9236865Sdfr#include <vm/vm.h>
9336865Sdfr#include <vm/vm_param.h>
9436865Sdfr#include <vm/vm_kern.h>
9536865Sdfr#include <vm/vm_page.h>
9636865Sdfr#include <vm/vm_map.h>
9736865Sdfr#include <vm/vm_extern.h>
9836865Sdfr
9936865Sdfr#include <sys/user.h>
10036865Sdfr
10136865Sdfr/*
10236865Sdfr * quick version of vm_fault
10336865Sdfr */
10451474Sdillonint
10536865Sdfrvm_fault_quick(v, prot)
10636865Sdfr	caddr_t v;
10736865Sdfr	int prot;
10836865Sdfr{
10951474Sdillon	int r;
11036865Sdfr	if (prot & VM_PROT_WRITE)
11151474Sdillon		r = subyte(v, fubyte(v));
11236865Sdfr	else
11351474Sdillon		r = fubyte(v);
11451474Sdillon	return(r);
11536865Sdfr}
11636865Sdfr
11736865Sdfr/*
11836865Sdfr * Finish a fork operation, with process p2 nearly set up.
11936865Sdfr * Copy and update the pcb, set up the stack so that the child
12036865Sdfr * ready to run and return to user mode.
12136865Sdfr */
12236865Sdfrvoid
12390361Sjuliancpu_fork(struct thread *td1, struct proc *p2, struct thread *td2, int flags)
12436865Sdfr{
12584977Sbenno	struct	proc *p1;
12684977Sbenno	struct	trapframe *tf;
12784977Sbenno	struct	callframe *cf;
12884977Sbenno	struct	switchframe *sf;
12990643Sbenno	struct	pcb *pcb;
13084977Sbenno
13190643Sbenno	KASSERT(td1 == curthread || td1 == thread0,
13290643Sbenno	    ("cpu_fork: p1 not curproc and not proc0"));
13384977Sbenno	CTR3(KTR_PROC, "cpu_fork: called td1=%08x p2=%08x flags=%x", (u_int)td1, (u_int)p2, flags);
13484977Sbenno
13584977Sbenno	if ((flags & RFPROC) == 0)
13684977Sbenno		return;
13784977Sbenno
13884977Sbenno	p1 = td1->td_proc;
13984977Sbenno
14090643Sbenno	pcb = (struct pcb *)(td2->td_kstack + KSTACK_PAGES * PAGE_SIZE -
14190643Sbenno	    sizeof(struct pcb));
14290643Sbenno	td2->td_pcb = pcb;
14384977Sbenno
14490643Sbenno	/* Copy the pcb */
14590643Sbenno	bcopy(td1->td_pcb, pcb, sizeof(struct pcb));
14684977Sbenno
14784977Sbenno	/*
14890643Sbenno	 * Create a fresh stack for the new process.
14984977Sbenno	 * Copy the trap frame for the return to user mode as if from a
15084977Sbenno	 * syscall.  This copies most of the user mode register values.
15184977Sbenno	 */
15284977Sbenno
15390643Sbenno	tf = (struct trapframe *)pcb - 1;
15490643Sbenno	bcopy(td1->td_frame, tf, sizeof(*tf));
15584977Sbenno
15690643Sbenno	/* XXX: Set up trap frame? */
15790643Sbenno
15890643Sbenno	td2->td_frame = tf;
15990643Sbenno
16090643Sbenno	/*
16190643Sbenno	 * There happens to be a callframe, too.
16290643Sbenno	 */
16390643Sbenno	cf = (struct callframe *)tf - 1;
16484977Sbenno	cf->lr = (int)fork_trampoline;
16584977Sbenno
16690643Sbenno	/*
16790643Sbenno	 * Below that, we allocate the switch frame.
16890643Sbenno	 */
16990643Sbenno	sf = (struct switchframe *)cf - 1;
17090643Sbenno	sf->sp = (int)cf;
17190643Sbenno	pcb->pcb_sp = (int)sf;
17284977Sbenno
17390643Sbenno	/*
17490643Sbenno 	 * Now cpu_switch() can schedule the new process.
17590643Sbenno	 */
17636865Sdfr}
17736865Sdfr
17836865Sdfr/*
17936865Sdfr * Intercept the return address from a freshly forked process that has NOT
18036865Sdfr * been scheduled yet.
18136865Sdfr *
18236865Sdfr * This is needed to make kernel threads stay in kernel mode.
18336865Sdfr */
18436865Sdfrvoid
18583682Smpcpu_set_fork_handler(td, func, arg)
18683682Smp	struct thread *td;
18748391Speter	void (*func) __P((void *));
18848391Speter	void *arg;
18936865Sdfr{
19084977Sbenno	struct	switchframe *sf;
19184977Sbenno	struct	callframe *cf;
19284977Sbenno
19384977Sbenno	CTR3(KTR_PROC, "cpu_set_fork_handler: called with td=%08x func=%08x arg=%08x",
19484977Sbenno	    (u_int)td, (u_int)func, (u_int)arg);
19584977Sbenno
19684977Sbenno	sf = (struct switchframe *)td->td_pcb->pcb_sp;
19784977Sbenno	cf = (struct callframe *)sf->sp;
19884977Sbenno
19984977Sbenno	cf->r31 = (register_t)func;
20084977Sbenno	cf->r30 = (register_t)arg;
20136865Sdfr}
20236865Sdfr
20336865Sdfr/*
20436865Sdfr * cpu_exit is called as the last action during exit.
20536865Sdfr * We release the address space of the process, block interrupts,
20636865Sdfr * and call switch_exit.  switch_exit switches to proc0's PCB and stack,
20736865Sdfr * then jumps into the middle of cpu_switch, as if it were switching
20836865Sdfr * from proc0.
20936865Sdfr */
21036865Sdfrvoid
21183682Smpcpu_exit(td)
21283682Smp	register struct thread *td;
21336865Sdfr{
21436865Sdfr}
21536865Sdfr
21636865Sdfrvoid
21783682Smpcpu_wait(td)
21883682Smp	struct proc *td;
21936865Sdfr{
22083276Speter}
22136865Sdfr
22283276Speter/* Temporary helper */
22383276Spetervoid
22483276Spetercpu_throw(void)
22583276Speter{
22683276Speter
22783276Speter	cpu_switch();
22883276Speter	panic("cpu_throw() didn't");
22936865Sdfr}
23036865Sdfr
23136865Sdfr/*
23236865Sdfr * Dump the machine specific header information at the start of a core dump.
23336865Sdfr */
23436865Sdfrint
23583682Smpcpu_coredump(td, vp, cred)
23683682Smp	struct thread *td;
23736865Sdfr	struct vnode *vp;
23836865Sdfr	struct ucred *cred;
23936865Sdfr{
24036865Sdfr
24190643Sbenno	return (vn_rdwr(UIO_WRITE, vp, (caddr_t)td->td_proc->p_uarea,
24290643Sbenno	    ctob(UAREA_PAGES), (off_t)0, UIO_SYSSPACE, IO_UNIT, cred,
24390643Sbenno	    (int *)NULL, td));
24436865Sdfr}
24536865Sdfr
24636865Sdfr/*
24736865Sdfr * Map an IO request into kernel virtual address space.
24836865Sdfr *
24936865Sdfr * All requests are (re)mapped into kernel VA space.
25036865Sdfr * Notice that we use b_bufsize for the size of the buffer
25136865Sdfr * to be mapped.  b_bcount might be modified by the driver.
25236865Sdfr */
25336865Sdfrvoid
25436865Sdfrvmapbuf(bp)
25536865Sdfr	register struct buf *bp;
25636865Sdfr{
25736865Sdfr	register caddr_t addr, v, kva;
25836865Sdfr	vm_offset_t pa;
25936865Sdfr
26079224Sdillon	GIANT_REQUIRED;
26179224Sdillon
26236865Sdfr	if ((bp->b_flags & B_PHYS) == 0)
26336865Sdfr		panic("vmapbuf");
26436865Sdfr
26536865Sdfr	for (v = bp->b_saveaddr, addr = (caddr_t)trunc_page(bp->b_data);
26636865Sdfr	    addr < bp->b_data + bp->b_bufsize;
26736865Sdfr	    addr += PAGE_SIZE, v += PAGE_SIZE) {
26836865Sdfr		/*
26936865Sdfr		 * Do the vm_fault if needed; do the copy-on-write thing
27036865Sdfr		 * when reading stuff off device into memory.
27136865Sdfr		 */
27284381Smjacob		vm_fault_quick((addr >= bp->b_data) ? addr : bp->b_data,
27358345Sphk			(bp->b_iocmd == BIO_READ)?(VM_PROT_READ|VM_PROT_WRITE):VM_PROT_READ);
27436865Sdfr		pa = trunc_page(pmap_kextract((vm_offset_t) addr));
27536865Sdfr		if (pa == 0)
27636865Sdfr			panic("vmapbuf: page not present");
27736865Sdfr		vm_page_hold(PHYS_TO_VM_PAGE(pa));
27836865Sdfr		pmap_kenter((vm_offset_t) v, pa);
27936865Sdfr	}
28036865Sdfr
28136865Sdfr	kva = bp->b_saveaddr;
28236865Sdfr	bp->b_saveaddr = bp->b_data;
28336865Sdfr	bp->b_data = kva + (((vm_offset_t) bp->b_data) & PAGE_MASK);
28436865Sdfr}
28536865Sdfr
28636865Sdfr/*
28736865Sdfr * Free the io map PTEs associated with this IO operation.
28836865Sdfr * We also invalidate the TLB entries and restore the original b_addr.
28936865Sdfr */
29036865Sdfrvoid
29136865Sdfrvunmapbuf(bp)
29236865Sdfr	register struct buf *bp;
29336865Sdfr{
29436865Sdfr	register caddr_t addr;
29536865Sdfr	vm_offset_t pa;
29636865Sdfr
29779224Sdillon	GIANT_REQUIRED;
29879224Sdillon
29936865Sdfr	if ((bp->b_flags & B_PHYS) == 0)
30036865Sdfr		panic("vunmapbuf");
30136865Sdfr
30236865Sdfr	for (addr = (caddr_t)trunc_page(bp->b_data);
30336865Sdfr	    addr < bp->b_data + bp->b_bufsize;
30436865Sdfr	    addr += PAGE_SIZE) {
30536865Sdfr		pa = trunc_page(pmap_kextract((vm_offset_t) addr));
30636865Sdfr		pmap_kremove((vm_offset_t) addr);
30736865Sdfr		vm_page_unhold(PHYS_TO_VM_PAGE(pa));
30836865Sdfr	}
30936865Sdfr
31036865Sdfr	bp->b_data = bp->b_saveaddr;
31136865Sdfr}
31236865Sdfr
31336865Sdfr/*
31465557Sjasone * Reset back to firmware.
31536865Sdfr */
31636865Sdfrvoid
31736865Sdfrcpu_reset()
31836865Sdfr{
31978342Sbenno	OF_exit();
32036865Sdfr}
32136865Sdfr
32236865Sdfrint
32343209Sjuliangrow_stack(p, sp)
32443209Sjulian	struct proc *p;
32543209Sjulian	size_t sp;
32643209Sjulian{
32743209Sjulian	int rv;
32836865Sdfr
32943209Sjulian	rv = vm_map_growstack (p, sp);
33043209Sjulian	if (rv != KERN_SUCCESS)
33143209Sjulian		return (0);
33243209Sjulian
33343209Sjulian	return (1);
33443209Sjulian}
33543209Sjulian
33636865Sdfr/*
33736865Sdfr * Software interrupt handler for queued VM system processing.
33836865Sdfr */
33936865Sdfrvoid
34067551Sjhbswi_vm(void *dummy)
34136865Sdfr{
34278342Sbenno#if 0 /* XXX: Don't have busdma stuff yet */
34336865Sdfr	if (busdma_swi_pending != 0)
34436865Sdfr		busdma_swi();
34578342Sbenno#endif
34636865Sdfr}
34736865Sdfr
34836865Sdfr/*
34936865Sdfr * Tell whether this address is in some physical memory region.
35036865Sdfr * Currently used by the kernel coredump code in order to avoid
35136865Sdfr * dumping the ``ISA memory hole'' which could cause indefinite hangs,
35236865Sdfr * or other unpredictable behaviour.
35336865Sdfr */
35436865Sdfr
35536865Sdfr
35636865Sdfrint
35736865Sdfris_physical_memory(addr)
35836865Sdfr	vm_offset_t addr;
35936865Sdfr{
36036865Sdfr	/*
36136865Sdfr	 * stuff other tests for known memory-mapped devices (PCI?)
36236865Sdfr	 * here
36336865Sdfr	 */
36436865Sdfr
36536865Sdfr	return 1;
36636865Sdfr}
367