locore.s revision 1.26
1/*	$NetBSD: locore.s,v 1.26 1998/02/05 04:57:58 gwr Exp $	*/
2
3/*
4 * Copyright (c) 1988 University of Utah.
5 * Copyright (c) 1980, 1990, 1993
6 *	The Regents of the University of California.  All rights reserved.
7 *
8 * This code is derived from software contributed to Berkeley by
9 * the Systems Programming Group of the University of Utah Computer
10 * Science Department.
11 *
12 * Redistribution and use in source and binary forms, with or without
13 * modification, are permitted provided that the following conditions
14 * are met:
15 * 1. Redistributions of source code must retain the above copyright
16 *    notice, this list of conditions and the following disclaimer.
17 * 2. Redistributions in binary form must reproduce the above copyright
18 *    notice, this list of conditions and the following disclaimer in the
19 *    documentation and/or other materials provided with the distribution.
20 * 3. All advertising materials mentioning features or use of this software
21 *    must display the following acknowledgement:
22 *	This product includes software developed by the University of
23 *	California, Berkeley and its contributors.
24 * 4. Neither the name of the University nor the names of its contributors
25 *    may be used to endorse or promote products derived from this software
26 *    without specific prior written permission.
27 *
28 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
29 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
30 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
31 * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
32 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
33 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
34 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
35 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
36 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
37 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
38 * SUCH DAMAGE.
39 *
40 *	from: Utah $Hdr: locore.s 1.66 92/12/22$
41 *	@(#)locore.s	8.6 (Berkeley) 5/27/94
42 */
43
44#include "assym.h"
45#include <machine/asm.h>
46#include <machine/trap.h>
47
48| Remember this is a fun project!
49
50	.data
51GLOBAL(mon_crp)
52	.long	0,0
53
54| This is for kvm_mkdb, and should be the address of the beginning
55| of the kernel text segment (not necessarily the same as kernbase).
56	.text
57GLOBAL(kernel_text)
58
59| This is the entry point, as well as the end of the temporary stack
60| used during process switch (one 8K page ending at start)
61ASGLOBAL(tmpstk)
62ASGLOBAL(start)
63
64| The first step, after disabling interrupts, is to map enough of the kernel
65| into high virtual address space so that we can use position dependent code.
66| This is a tricky task on the sun3x because the MMU is already enabled and
67| the ROM monitor provides no indication of where the root MMU table is mapped.
68| Therefore we must use one of the 68030's 'transparent translation' registers
69| to define a range in the address space where the MMU translation is
70| turned off.  Once this is complete we can modify the MMU table directly
71| without the need for it to be mapped into virtual memory.
72| All code must be position independent until otherwise noted, as the
73| boot loader has loaded us into low memory but all the symbols in this
74| code have been linked high.
75	movw	#PSL_HIGHIPL, sr	| no interrupts
76	movl	#KERNBASE, a5		| for vtop conversion
77	lea	_C_LABEL(mon_crp), a0	| where to store the CRP
78	subl	a5, a0
79	| Note: borrowing mon_crp for tt0 setup...
80	movl	#0x3F8107, a0@		| map the low 1GB v=p with the
81	.long	0xf0100800		| transparent translation reg0
82					| [ pmove a0@, tt0 ]
83| In order to map the kernel into high memory we will copy the root table
84| entry which maps the 16 megabytes of memory starting at 0x0 into the
85| entry which maps the 16 megabytes starting at KERNBASE.
86	pmove	crp, a0@		| Get monitor CPU root pointer
87	movl	a0@(4), a1		| 2nd word is PA of level A table
88
89	movl	a1, a0			| compute the descriptor address
90	addl	#0x3e0, a1		| for VA starting at KERNBASE
91	movl	a0@, a1@		| copy descriptor type
92	movl	a0@(4), a1@(4)		| copy physical address
93
94| Kernel is now double mapped at zero and KERNBASE.
95| Force a long jump to the relocated code (high VA).
96	movl	#IC_CLEAR, d0		| Flush the I-cache
97	movc	d0, cacr
98	jmp L_high_code:l		| long jump
99
100L_high_code:
101| We are now running in the correctly relocated kernel, so
102| we are no longer restricted to position-independent code.
103| It is handy to leave transparent translation enabled while
104| for the low 1GB while _bootstrap() is doing its thing.
105
106| Do bootstrap stuff needed before main() gets called.
107| Our boot loader leaves a copy of the kernel's exec header
108| just before the start of the kernel text segment, so the
109| kernel can sanity-check the DDB symbols at [end...esym].
110| Pass the struct exec at tmpstk-32 to _bootstrap().
111| Also, make sure the initial frame pointer is zero so that
112| the backtrace algorithm used by KGDB terminates nicely.
113	lea	_ASM_LABEL(tmpstk)-32, sp
114	movl	#0,a6
115	jsr	_C_LABEL(_bootstrap)	| See locore2.c
116
117| Now turn off the transparent translation of the low 1GB.
118| (this also flushes the ATC)
119	clrl	sp@-
120	.long	0xf0170800		| pmove	sp@,tt0
121	addql	#4,sp
122
123| Now that _bootstrap() is done using the PROM functions,
124| we can safely set the sfc/dfc to something != FC_CONTROL
125	moveq	#FC_USERD, d0		| make movs access "user data"
126	movc	d0, sfc			| space for copyin/copyout
127	movc	d0, dfc
128
129| Setup process zero user/kernel stacks.
130	movl	_C_LABEL(proc0paddr),a1	| get proc0 pcb addr
131	lea	a1@(USPACE-4),sp	| set SSP to last word
132	movl	#USRSTACK-4,a2
133	movl	a2,usp			| init user SP
134
135| Note curpcb was already set in _bootstrap().
136| Will do fpu initialization during autoconfig (see fpu.c)
137| The interrupt vector table and stack are now ready.
138| Interrupts will be enabled later, AFTER  autoconfiguration
139| is finished, to avoid spurrious interrupts.
140
141/*
142 * Final preparation for calling main.
143 *
144 * Create a fake exception frame that returns to user mode,
145 * and save its address in p->p_md.md_regs for cpu_fork().
146 * The new frames for process 1 and 2 will be adjusted by
147 * cpu_set_kpc() to arrange for a call to a kernel function
148 * before the new process does its rte out to user mode.
149 */
150	clrw	sp@-			| tf_format,tf_vector
151	clrl	sp@-			| tf_pc (filled in later)
152	movw	#PSL_USER,sp@-		| tf_sr for user mode
153	clrl	sp@-			| tf_stackadj
154	lea	sp@(-64),sp		| tf_regs[16]
155	movl	sp,a1			| a1=trapframe
156	lea	_C_LABEL(proc0),a0	| proc0.p_md.md_regs =
157	movl	a1,a0@(P_MDREGS)	|   trapframe
158	movl	a2,a1@(FR_SP)		| a2 == usp (from above)
159	pea	a1@			| push &trapframe
160	jbsr	_C_LABEL(main)		| main(&trapframe)
161	addql	#4,sp			| help DDB backtrace
162	trap	#15			| should not get here
163
164| This is used by cpu_fork() to return to user mode.
165| It is called with SP pointing to a struct trapframe.
166GLOBAL(proc_do_uret)
167	movl	sp@(FR_SP),a0		| grab and load
168	movl	a0,usp			|   user SP
169	moveml	sp@+,#0x7FFF		| load most registers (all but SSP)
170	addql	#8,sp			| pop SSP and stack adjust count
171	rte
172
173/*
174 * proc_trampoline:
175 * This is used by cpu_set_kpc() to "push" a function call onto the
176 * kernel stack of some process, very much like a signal delivery.
177 * When we get here, the stack has:
178 *
179 * SP+8:	switchframe from before cpu_set_kpc
180 * SP+4:	void *proc;
181 * SP:  	u_long func;
182 *
183 * On entry, the switchframe pushed by cpu_set_kpc has already been
184 * popped off the stack, so all this needs to do is pop the function
185 * pointer into a register, call it, then pop the arg, and finally
186 * return using the switchframe that remains on the stack.
187 */
188GLOBAL(proc_trampoline)
189	movl	sp@+,a0			| function pointer
190	jbsr	a0@			| (*func)(procp)
191	addql	#4,sp			| toss the arg
192	rts				| as cpu_switch would do
193
194| That is all the assembly startup code we need on the sun3x!
195| The rest of this is like the hp300/locore.s where possible.
196
197/*
198 * Trap/interrupt vector routines
199 */
200#include <m68k/m68k/trap_subr.s>
201
202GLOBAL(buserr)
203	tstl	_C_LABEL(nofault)	| device probe?
204	jeq	_C_LABEL(addrerr)	| no, handle as usual
205	movl	_C_LABEL(nofault),sp@-	| yes,
206	jbsr	_C_LABEL(longjmp)	|  longjmp(nofault)
207GLOBAL(addrerr)
208	clrl	sp@-			| stack adjust count
209	moveml	#0xFFFF,sp@-		| save user registers
210	movl	usp,a0			| save the user SP
211	movl	a0,sp@(FR_SP)		|   in the savearea
212	lea	sp@(FR_HW),a1		| grab base of HW berr frame
213	moveq	#0,d0
214	movw	a1@(10),d0		| grab SSW for fault processing
215	btst	#12,d0			| RB set?
216	jeq	LbeX0			| no, test RC
217	bset	#14,d0			| yes, must set FB
218	movw	d0,a1@(10)		| for hardware too
219LbeX0:
220	btst	#13,d0			| RC set?
221	jeq	LbeX1			| no, skip
222	bset	#15,d0			| yes, must set FC
223	movw	d0,a1@(10)		| for hardware too
224LbeX1:
225	btst	#8,d0			| data fault?
226	jeq	Lbe0			| no, check for hard cases
227	movl	a1@(16),d1		| fault address is as given in frame
228	jra	Lbe10			| thats it
229Lbe0:
230	btst	#4,a1@(6)		| long (type B) stack frame?
231	jne	Lbe4			| yes, go handle
232	movl	a1@(2),d1		| no, can use save PC
233	btst	#14,d0			| FB set?
234	jeq	Lbe3			| no, try FC
235	addql	#4,d1			| yes, adjust address
236	jra	Lbe10			| done
237Lbe3:
238	btst	#15,d0			| FC set?
239	jeq	Lbe10			| no, done
240	addql	#2,d1			| yes, adjust address
241	jra	Lbe10			| done
242Lbe4:
243	movl	a1@(36),d1		| long format, use stage B address
244	btst	#15,d0			| FC set?
245	jeq	Lbe10			| no, all done
246	subql	#2,d1			| yes, adjust address
247Lbe10:
248	movl	d1,sp@-			| push fault VA
249	movl	d0,sp@-			| and padded SSW
250	movw	a1@(6),d0		| get frame format/vector offset
251	andw	#0x0FFF,d0		| clear out frame format
252	cmpw	#12,d0			| address error vector?
253	jeq	Lisaerr			| yes, go to it
254
255/* MMU-specific code to determine reason for bus error. */
256	movl	d1,a0			| fault address
257	movl	sp@,d0			| function code from ssw
258	btst	#8,d0			| data fault?
259	jne	Lbe10a
260	movql	#1,d0			| user program access FC
261					| (we dont separate data/program)
262	btst	#5,a1@			| supervisor mode?
263	jeq	Lbe10a			| if no, done
264	movql	#5,d0			| else supervisor program access
265Lbe10a:
266	ptestr	d0,a0@,#7		| do a table search
267	pmove	psr,sp@			| save result
268	movb	sp@,d1
269	btst	#2,d1			| invalid? (incl. limit viol and berr)
270	jeq	Lmightnotbemerr		| no -> wp check
271	btst	#7,d1			| is it MMU table berr?
272	jeq	Lismerr			| no, must be fast
273	jra	Lisberr1		| real bus err needs not be fast
274Lmightnotbemerr:
275	btst	#3,d1			| write protect bit set?
276	jeq	Lisberr1		| no, must be bus error
277	movl	sp@,d0			| ssw into low word of d0
278	andw	#0xc0,d0		| write protect is set on page:
279	cmpw	#0x40,d0		| was it read cycle?
280	jeq	Lisberr1		| yes, was not WPE, must be bus err
281/* End of MMU-specific bus error code. */
282
283Lismerr:
284	movl	#T_MMUFLT,sp@-		| show that we are an MMU fault
285	jra	_ASM_LABEL(faultstkadj)	| and deal with it
286Lisaerr:
287	movl	#T_ADDRERR,sp@-		| mark address error
288	jra	_ASM_LABEL(faultstkadj)	| and deal with it
289Lisberr1:
290	clrw	sp@			| re-clear pad word
291Lisberr:
292	movl	#T_BUSERR,sp@-		| mark bus error
293	jra	_ASM_LABEL(faultstkadj)	| and deal with it
294
295/*
296 * FP exceptions.
297 */
298GLOBAL(fpfline)
299	clrl	sp@-			| stack adjust count
300	moveml	#0xFFFF,sp@-		| save registers
301	moveq	#T_FPEMULI,d0		| denote as FP emulation trap
302	jra	_ASM_LABEL(fault)	| do it
303
304GLOBAL(fpunsupp)
305	clrl	sp@-			| stack adjust count
306	moveml	#0xFFFF,sp@-		| save registers
307	moveq	#T_FPEMULD,d0		| denote as FP emulation trap
308	jra	_ASM_LABEL(fault)	| do it
309
310/*
311 * Handles all other FP coprocessor exceptions.
312 * Note that since some FP exceptions generate mid-instruction frames
313 * and may cause signal delivery, we need to test for stack adjustment
314 * after the trap call.
315 */
316GLOBAL(fpfault)
317	clrl	sp@-		| stack adjust count
318	moveml	#0xFFFF,sp@-	| save user registers
319	movl	usp,a0		| and save
320	movl	a0,sp@(FR_SP)	|   the user stack pointer
321	clrl	sp@-		| no VA arg
322	movl	_C_LABEL(curpcb),a0	| current pcb
323	lea	a0@(PCB_FPCTX),a0 | address of FP savearea
324	fsave	a0@		| save state
325	tstb	a0@		| null state frame?
326	jeq	Lfptnull	| yes, safe
327	clrw	d0		| no, need to tweak BIU
328	movb	a0@(1),d0	| get frame size
329	bset	#3,a0@(0,d0:w)	| set exc_pend bit of BIU
330Lfptnull:
331	fmovem	fpsr,sp@-	| push fpsr as code argument
332	frestore a0@		| restore state
333	movl	#T_FPERR,sp@-	| push type arg
334	jra	_ASM_LABEL(faultstkadj) | call trap and deal with stack cleanup
335
336/*
337 * Other exceptions only cause four and six word stack frame and require
338 * no post-trap stack adjustment.
339 */
340GLOBAL(badtrap)
341	clrl	sp@-			| stack adjust count
342	moveml	#0xFFFF,sp@-		| save std frame regs
343	jbsr	_C_LABEL(straytrap)	| report
344	moveml	sp@+,#0xFFFF		| restore regs
345	addql	#4, sp			| stack adjust count
346	jra	_ASM_LABEL(rei)		| all done
347
348/*
349 * Trap 0 is for system calls
350 */
351GLOBAL(trap0)
352	clrl	sp@-			| stack adjust count
353	moveml	#0xFFFF,sp@-		| save user registers
354	movl	usp,a0			| save the user SP
355	movl	a0,sp@(FR_SP)		|   in the savearea
356	movl	d0,sp@-			| push syscall number
357	jbsr	_C_LABEL(syscall)	| handle it
358	addql	#4,sp			| pop syscall arg
359	movl	sp@(FR_SP),a0		| grab and restore
360	movl	a0,usp			|   user SP
361	moveml	sp@+,#0x7FFF		| restore most registers
362	addql	#8,sp			| pop SP and stack adjust
363	jra	_ASM_LABEL(rei)		| all done
364
365/*
366 * Trap 1 action depends on the emulation type:
367 * NetBSD: sigreturn "syscall"
368 *   HPUX: user breakpoint
369 */
370GLOBAL(trap1)
371#if 0 /* COMPAT_HPUX */
372	/* If process is HPUX, this is a user breakpoint. */
373	jne	_C_LABEL(trap15)	| HPUX user breakpoint
374#endif
375	jra	_ASM_LABEL(sigreturn)	| NetBSD
376
377/*
378 * Trap 2 action depends on the emulation type:
379 * NetBSD: user breakpoint -- See XXX below...
380 *  SunOS: cache flush
381 *   HPUX: sigreturn
382 */
383GLOBAL(trap2)
384#if 0 /* COMPAT_HPUX */
385	/* If process is HPUX, this is a sigreturn call */
386	jne	_ASM_LABEL(sigreturn)
387#endif
388	jra	_C_LABEL(trap15)	| NetBSD user breakpoint
389| XXX - Make NetBSD use trap 15 for breakpoints?
390| XXX - That way, we can allow this cache flush...
391| XXX SunOS trap #2 (and NetBSD?)
392| Flush on-chip cache (leave it enabled)
393|	movl	#CACHE_CLR,d0
394|	movc	d0,cacr
395|	rte
396
397/*
398 * Trap 12 is the entry point for the cachectl "syscall"
399 *	cachectl(command, addr, length)
400 * command in d0, addr in a1, length in d1
401 */
402GLOBAL(trap12)
403	movl	d1,sp@-			| push length
404	movl	a1,sp@-			| push addr
405	movl	d0,sp@-			| push command
406	jbsr	_C_LABEL(cachectl)	| do it
407	lea	sp@(12),sp		| pop args
408	jra	_ASM_LABEL(rei)		| all done
409
410/*
411 * Trace (single-step) trap.  Kernel-mode is special.
412 * User mode traps are simply passed on to trap().
413 */
414GLOBAL(trace)
415	clrl	sp@-			| stack adjust count
416	moveml	#0xFFFF,sp@-
417	moveq	#T_TRACE,d0
418	btst	#5,sp@(FR_HW)		| was supervisor mode?
419	jne	_ASM_LABEL(kbrkpt)	|  yes, kernel brkpt
420	jra	_ASM_LABEL(fault)	| no, user-mode fault
421
422/*
423 * Trap 15 is used for:
424 *	- GDB breakpoints (in user programs)
425 *	- KGDB breakpoints (in the kernel)
426 *	- trace traps for SUN binaries (not fully supported yet)
427 * User mode traps are simply passed to trap().
428 */
429GLOBAL(trap15)
430	clrl	sp@-			| stack adjust count
431	moveml	#0xFFFF,sp@-
432	moveq	#T_TRAP15,d0
433	btst	#5,sp@(FR_HW)		| was supervisor mode?
434	jne	_ASM_LABEL(kbrkpt)	|  yes, kernel brkpt
435	jra	_ASM_LABEL(fault)	| no, user-mode fault
436
437ASLOCAL(kbrkpt)
438	| Kernel-mode breakpoint or trace trap. (d0=trap_type)
439	| Save the system sp rather than the user sp.
440	movw	#PSL_HIGHIPL,sr		| lock out interrupts
441	lea	sp@(FR_SIZE),a6		| Save stack pointer
442	movl	a6,sp@(FR_SP)		|  from before trap
443
444	| If we are not on tmpstk switch to it.
445	| (so debugger can change the stack pointer)
446	movl	a6,d1
447	cmpl	#_ASM_LABEL(tmpstk),d1
448	jls	Lbrkpt2 		| already on tmpstk
449	| Copy frame to the temporary stack
450	movl	sp,a0			| a0=src
451	lea	_ASM_LABEL(tmpstk)-96,a1	| a1=dst
452	movl	a1,sp			| sp=new frame
453	moveq	#FR_SIZE,d1
454Lbrkpt1:
455	movl	a0@+,a1@+
456	subql	#4,d1
457	bgt	Lbrkpt1
458
459Lbrkpt2:
460	| Call the trap handler for the kernel debugger.
461	| Do not call trap() to handle it, so that we can
462	| set breakpoints in trap() if we want.  We know
463	| the trap type is either T_TRACE or T_BREAKPOINT.
464	movl	d0,sp@-			| push trap type
465	jbsr	_C_LABEL(trap_kdebug)
466	addql	#4,sp			| pop args
467
468	| The stack pointer may have been modified, or
469	| data below it modified (by kgdb push call),
470	| so push the hardware frame at the current sp
471	| before restoring registers and returning.
472	movl	sp@(FR_SP),a0		| modified sp
473	lea	sp@(FR_SIZE),a1		| end of our frame
474	movl	a1@-,a0@-		| copy 2 longs with
475	movl	a1@-,a0@-		| ... predecrement
476	movl	a0,sp@(FR_SP)		| sp = h/w frame
477	moveml	sp@+,#0x7FFF		| restore all but sp
478	movl	sp@,sp			| ... and sp
479	rte				| all done
480
481/* Use common m68k sigreturn */
482#include <m68k/m68k/sigreturn.s>
483
484/*
485 * Interrupt handlers.  Most are auto-vectored,
486 * and hard-wired the same way on all sun3 models.
487 * Format in the stack is:
488 *   d0,d1,a0,a1, sr, pc, vo
489 */
490
491#define INTERRUPT_SAVEREG \
492	moveml	#0xC0C0,sp@-
493
494#define INTERRUPT_RESTORE \
495	moveml	sp@+,#0x0303
496
497/*
498 * This is the common auto-vector interrupt handler,
499 * for which the CPU provides the vector=0x18+level.
500 * These are installed in the interrupt vector table.
501 */
502	.align	2
503GLOBAL(_isr_autovec)
504	INTERRUPT_SAVEREG
505	jbsr	_C_LABEL(isr_autovec)
506	INTERRUPT_RESTORE
507	jra	_ASM_LABEL(rei)
508
509/* clock: see clock.c */
510	.align	2
511GLOBAL(_isr_clock)
512	INTERRUPT_SAVEREG
513	jbsr	_C_LABEL(clock_intr)
514	INTERRUPT_RESTORE
515	jra	_ASM_LABEL(rei)
516
517| Handler for all vectored interrupts (i.e. VME interrupts)
518	.align	2
519GLOBAL(_isr_vectored)
520	INTERRUPT_SAVEREG
521	jbsr	_C_LABEL(isr_vectored)
522	INTERRUPT_RESTORE
523	jra	_ASM_LABEL(rei)
524
525#undef	INTERRUPT_SAVEREG
526#undef	INTERRUPT_RESTORE
527
528/* interrupt counters (needed by vmstat) */
529GLOBAL(intrnames)
530	.asciz	"spur"	| 0
531	.asciz	"lev1"	| 1
532	.asciz	"lev2"	| 2
533	.asciz	"lev3"	| 3
534	.asciz	"lev4"	| 4
535	.asciz	"clock"	| 5
536	.asciz	"lev6"	| 6
537	.asciz	"nmi"	| 7
538GLOBAL(eintrnames)
539
540	.data
541	.even
542GLOBAL(intrcnt)
543	.long	0,0,0,0,0,0,0,0,0,0
544GLOBAL(eintrcnt)
545	.text
546
547/*
548 * Emulation of VAX REI instruction.
549 *
550 * This code is (mostly) un-altered from the hp300 code,
551 * except that sun machines do not need a simulated SIR
552 * because they have a real software interrupt register.
553 *
554 * This code deals with checking for and servicing ASTs
555 * (profiling, scheduling) and software interrupts (network, softclock).
556 * We check for ASTs first, just like the VAX.  To avoid excess overhead
557 * the T_ASTFLT handling code will also check for software interrupts so we
558 * do not have to do it here.  After identifying that we need an AST we
559 * drop the IPL to allow device interrupts.
560 *
561 * This code is complicated by the fact that sendsig may have been called
562 * necessitating a stack cleanup.
563 */
564
565ASGLOBAL(rei)
566#ifdef	DIAGNOSTIC
567	tstl	_C_LABEL(panicstr)	| have we paniced?
568	jne	Ldorte			| yes, do not make matters worse
569#endif
570	tstl	_C_LABEL(astpending)	| AST pending?
571	jeq	Ldorte			| no, done
572Lrei1:
573	btst	#5,sp@			| yes, are we returning to user mode?
574	jne	Ldorte			| no, done
575	movw	#PSL_LOWIPL,sr		| lower SPL
576	clrl	sp@-			| stack adjust
577	moveml	#0xFFFF,sp@-		| save all registers
578	movl	usp,a1			| including
579	movl	a1,sp@(FR_SP)		|    the users SP
580	clrl	sp@-			| VA == none
581	clrl	sp@-			| code == none
582	movl	#T_ASTFLT,sp@-		| type == async system trap
583	jbsr	_C_LABEL(trap)		| go handle it
584	lea	sp@(12),sp		| pop value args
585	movl	sp@(FR_SP),a0		| restore user SP
586	movl	a0,usp			|   from save area
587	movw	sp@(FR_ADJ),d0		| need to adjust stack?
588	jne	Laststkadj		| yes, go to it
589	moveml	sp@+,#0x7FFF		| no, restore most user regs
590	addql	#8,sp			| toss SP and stack adjust
591	rte				| and do real RTE
592Laststkadj:
593	lea	sp@(FR_HW),a1		| pointer to HW frame
594	addql	#8,a1			| source pointer
595	movl	a1,a0			| source
596	addw	d0,a0			|  + hole size = dest pointer
597	movl	a1@-,a0@-		| copy
598	movl	a1@-,a0@-		|  8 bytes
599	movl	a0,sp@(FR_SP)		| new SSP
600	moveml	sp@+,#0x7FFF		| restore user registers
601	movl	sp@,sp			| and our SP
602Ldorte:
603	rte				| real return
604
605/*
606 * Initialization is at the beginning of this file, because the
607 * kernel entry point needs to be at zero for compatibility with
608 * the Sun boot loader.  This works on Sun machines because the
609 * interrupt vector table for reset is NOT at address zero.
610 * (The MMU has a "boot" bit that forces access to the PROM)
611 */
612
613/*
614 * Use common m68k sigcode.
615 */
616#include <m68k/m68k/sigcode.s>
617
618	.text
619
620/*
621 * Primitives
622 */
623
624/*
625 * Use common m68k support routines.
626 */
627#include <m68k/m68k/support.s>
628
629BSS(want_resched,4)
630
631/*
632 * Use common m68k process manipulation routines.
633 */
634#include <m68k/m68k/proc_subr.s>
635
636| Message for Lbadsw panic
637Lsw0:
638	.asciz	"cpu_switch"
639	.even
640
641	.data
642GLOBAL(masterpaddr)		| XXX compatibility (debuggers)
643GLOBAL(curpcb)
644	.long	0
645ASBSS(nullpcb,SIZEOF_PCB)
646	.text
647
648/*
649 * At exit of a process, do a cpu_switch for the last time.
650 * Switch to a safe stack and PCB, and deallocate the process's resources.
651 * The ipl is high enough to prevent the memory from being reallocated.
652 */
653ENTRY(switch_exit)
654	movl	sp@(4),a0		| struct proc *p
655					| save state into garbage pcb
656	movl	#_ASM_LABEL(nullpcb),_C_LABEL(curpcb)
657	lea	_ASM_LABEL(tmpstk),sp	| goto a tmp stack
658	movl	a0,sp@-			| pass proc ptr down
659
660	/* Free old process's u-area. */
661	movl	#USPACE,sp@-		| size of u-area
662	movl	a0@(P_ADDR),sp@-	| address of process's u-area
663	movl	_C_LABEL(kernel_map),sp@-	| map it was allocated in
664	jbsr	_C_LABEL(kmem_free)		| deallocate it
665	lea	sp@(12),sp		| pop args
666
667	jra	_C_LABEL(cpu_switch)
668
669/*
670 * When no processes are on the runq, cpu_switch() branches to idle
671 * to wait for something to come ready.
672 */
673	.data
674GLOBAL(Idle_count)
675	.long	0
676	.text
677
678Lidle:
679	stop	#PSL_LOWIPL
680GLOBAL(_Idle)				| See clock.c
681	movw	#PSL_HIGHIPL,sr
682	addql	#1, _C_LABEL(Idle_count)
683	tstl	_C_LABEL(whichqs)
684	jeq	Lidle
685	movw	#PSL_LOWIPL,sr
686	jra	Lsw1
687
688Lbadsw:
689	movl	#Lsw0,sp@-
690	jbsr	_C_LABEL(panic)
691	/*NOTREACHED*/
692
693/*
694 * cpu_switch()
695 * Hacked for sun3
696 * XXX - Arg 1 is a proc pointer (curproc) but this doesn't use it.
697 * XXX - Sould we use p->p_addr instead of curpcb? -gwr
698 */
699ENTRY(cpu_switch)
700	movl	_C_LABEL(curpcb),a1	| current pcb
701	movw	sr,a1@(PCB_PS)		| save sr before changing ipl
702#ifdef notyet
703	movl	_C_LABEL(curproc),sp@-	| remember last proc running
704#endif
705	clrl	_C_LABEL(curproc)
706
707Lsw1:
708	/*
709	 * Find the highest-priority queue that isn't empty,
710	 * then take the first proc from that queue.
711	 */
712	clrl	d0
713	lea	_C_LABEL(whichqs),a0
714	movl	a0@,d1
715Lswchk:
716	btst	d0,d1
717	jne	Lswfnd
718	addqb	#1,d0
719	cmpb	#32,d0
720	jne	Lswchk
721	jra	_C_LABEL(_Idle)
722Lswfnd:
723	movw	#PSL_HIGHIPL,sr		| lock out interrupts
724	movl	a0@,d1			| and check again...
725	bclr	d0,d1
726	jeq	Lsw1			| proc moved, rescan
727	movl	d1,a0@			| update whichqs
728	moveq	#1,d1			| double check for higher priority
729	lsll	d0,d1			| process (which may have snuck in
730	subql	#1,d1			| while we were finding this one)
731	andl	a0@,d1
732	jeq	Lswok			| no one got in, continue
733	movl	a0@,d1
734	bset	d0,d1			| otherwise put this one back
735	movl	d1,a0@
736	jra	Lsw1			| and rescan
737Lswok:
738	movl	d0,d1
739	lslb	#3,d1			| convert queue number to index
740	addl	#_qs,d1			| locate queue (q)
741	movl	d1,a1
742	cmpl	a1@(P_FORW),a1		| anyone on queue?
743	jeq	Lbadsw			| no, panic
744	movl	a1@(P_FORW),a0		| p = q->p_forw
745	movl	a0@(P_FORW),a1@(P_FORW)	| q->p_forw = p->p_forw
746	movl	a0@(P_FORW),a1		| q = p->p_forw
747	movl	a0@(P_BACK),a1@(P_BACK)	| q->p_back = p->p_back
748	cmpl	a0@(P_FORW),d1		| anyone left on queue?
749	jeq	Lsw2			| no, skip
750	movl	_C_LABEL(whichqs),d1
751	bset	d0,d1			| yes, reset bit
752	movl	d1,_C_LABEL(whichqs)
753Lsw2:
754	movl	a0,_C_LABEL(curproc)
755	clrl	_C_LABEL(want_resched)
756#ifdef notyet
757	movl	sp@+,a1			| XXX - Make this work!
758	cmpl	a0,a1			| switching to same proc?
759	jeq	Lswdone			| yes, skip save and restore
760#endif
761	/*
762	 * Save state of previous process in its pcb.
763	 */
764	movl	_C_LABEL(curpcb),a1
765	moveml	#0xFCFC,a1@(PCB_REGS)	| save non-scratch registers
766	movl	usp,a2			| grab USP (a2 has been saved)
767	movl	a2,a1@(PCB_USP)		| and save it
768
769	tstl	_C_LABEL(fputype)	| Do we have an fpu?
770	jeq	Lswnofpsave		| No?  Then don't try save.
771	lea	a1@(PCB_FPCTX),a2	| pointer to FP save area
772	fsave	a2@			| save FP state
773	tstb	a2@			| null state frame?
774	jeq	Lswnofpsave		| yes, all done
775	fmovem	fp0-fp7,a2@(FPF_REGS)		| save FP general regs
776	fmovem	fpcr/fpsr/fpi,a2@(FPF_FPCR)	| save FP control regs
777Lswnofpsave:
778
779	/*
780	 * Now that we have saved all the registers that must be
781	 * preserved, we are free to use those registers until
782	 * we load the registers for the switched-to process.
783	 * In this section, keep:  a0=curproc, a1=curpcb
784	 */
785
786#ifdef DIAGNOSTIC
787	tstl	a0@(P_WCHAN)
788	jne	Lbadsw
789	cmpb	#SRUN,a0@(P_STAT)
790	jne	Lbadsw
791#endif
792	clrl	a0@(P_BACK)		| clear back link
793	movl	a0@(P_ADDR),a1		| get p_addr
794	movl	a1,_C_LABEL(curpcb)
795
796	/*
797	 * Load the new VM context (new MMU root pointer)
798	 */
799	movl	a0@(P_VMSPACE),a2	| vm = p->p_vmspace
800#ifdef DIAGNOSTIC
801	tstl	a2			| vm == VM_MAP_NULL?
802	jeq	Lbadsw			| panic
803#endif
804#ifdef PMAP_DEBUG
805	/* When debugging just call _pmap_switch(). */
806	movl	a2@(VM_PMAP),a2 	| pmap = vm->vm_map.pmap
807	pea	a2@			| push pmap
808	jbsr	_C_LABEL(_pmap_switch)	| _pmap_switch(pmap)
809	addql	#4,sp
810	movl	_C_LABEL(curpcb),a1	| restore p_addr
811#else
812	/* Otherwise, use this inline version. */
813	lea	_C_LABEL(kernel_crp), a3 | our CPU Root Ptr. (CRP)
814	movl	a2@(VM_PMAP),a2 	| pmap = vm->vm_map.pmap
815	movl	a2@(PM_A_PHYS),d0	| phys = pmap->pm_a_phys
816	cmpl	a3@(4),d0		|  == kernel_crp.rp_addr ?
817	jeq	Lsame_mmuctx		| skip loadcrp/flush
818	/* OK, it is a new MMU context.  Load it up. */
819	movl	d0,a3@(4)
820	movl	#CACHE_CLR,d0
821	movc	d0,cacr			| invalidate cache(s)
822	pflusha				| flush entire TLB
823	pmove	a3@,crp			| load new user root pointer
824Lsame_mmuctx:
825#endif
826
827	/*
828	 * Reload the registers for the new process.
829	 * After this point we can only use d0,d1,a0,a1
830	 */
831	moveml	a1@(PCB_REGS),#0xFCFC	| reload registers
832	movl	a1@(PCB_USP),a0
833	movl	a0,usp			| and USP
834
835	tstl	_C_LABEL(fputype)	| If we don't have an fpu,
836	jeq	Lres_skip		|  don't try to restore it.
837	lea	a1@(PCB_FPCTX),a0	| pointer to FP save area
838	tstb	a0@			| null state frame?
839	jeq	Lresfprest		| yes, easy
840	fmovem	a0@(FPF_FPCR),fpcr/fpsr/fpi	| restore FP control regs
841	fmovem	a0@(FPF_REGS),fp0-fp7		| restore FP general regs
842Lresfprest:
843	frestore a0@			| restore state
844Lres_skip:
845	movw	a1@(PCB_PS),d0		| no, restore PS
846#ifdef DIAGNOSTIC
847	btst	#13,d0			| supervisor mode?
848	jeq	Lbadsw			| no? panic!
849#endif
850	movw	d0,sr			| OK, restore PS
851	moveq	#1,d0			| return 1 (for alternate returns)
852	rts
853
854/*
855 * savectx(pcb)
856 * Update pcb, saving current processor state.
857 */
858ENTRY(savectx)
859	movl	sp@(4),a1
860	movw	sr,a1@(PCB_PS)
861	movl	usp,a0			| grab USP
862	movl	a0,a1@(PCB_USP)		| and save it
863	moveml	#0xFCFC,a1@(PCB_REGS)	| save non-scratch registers
864
865	tstl	_C_LABEL(fputype)	| Do we have FPU?
866	jeq	Lsavedone		| No?  Then don't save state.
867	lea	a1@(PCB_FPCTX),a0	| pointer to FP save area
868	fsave	a0@			| save FP state
869	tstb	a0@			| null state frame?
870	jeq	Lsavedone		| yes, all done
871	fmovem	fp0-fp7,a0@(FPF_REGS)		| save FP general regs
872	fmovem	fpcr/fpsr/fpi,a0@(FPF_FPCR)	| save FP control regs
873Lsavedone:
874	moveq	#0,d0			| return 0
875	rts
876
877/* suline() */
878
879#ifdef DEBUG
880	.data
881ASGLOBAL(fulltflush)
882	.long	0
883ASGLOBAL(fullcflush)
884	.long	0
885	.text
886#endif
887
888/*
889 * Invalidate entire TLB.
890 */
891ENTRY(TBIA)
892_C_LABEL(_TBIA):
893	pflusha
894	movl	#DC_CLEAR,d0
895	movc	d0,cacr			| invalidate on-chip d-cache
896	rts
897
898/*
899 * Invalidate any TLB entry for given VA (TB Invalidate Single)
900 */
901ENTRY(TBIS)
902#ifdef DEBUG
903	tstl	_ASM_LABEL(fulltflush)	| being conservative?
904	jne	_C_LABEL(_TBIA)		| yes, flush entire TLB
905#endif
906	movl	sp@(4),a0
907	pflush	#0,#0,a0@		| flush address from both sides
908	movl	#DC_CLEAR,d0
909	movc	d0,cacr			| invalidate on-chip data cache
910	rts
911
912/*
913 * Invalidate supervisor side of TLB
914 */
915ENTRY(TBIAS)
916#ifdef DEBUG
917	tstl	_ASM_LABEL(fulltflush)	| being conservative?
918	jne	_C_LABEL(_TBIA)		| yes, flush everything
919#endif
920	pflush	#4,#4			| flush supervisor TLB entries
921	movl	#DC_CLEAR,d0
922	movc	d0,cacr			| invalidate on-chip d-cache
923	rts
924
925/*
926 * Invalidate user side of TLB
927 */
928ENTRY(TBIAU)
929#ifdef DEBUG
930	tstl	_ASM_LABEL(fulltflush)	| being conservative?
931	jne	_C_LABEL(_TBIA)		| yes, flush everything
932#endif
933	pflush	#0,#4			| flush user TLB entries
934	movl	#DC_CLEAR,d0
935	movc	d0,cacr			| invalidate on-chip d-cache
936	rts
937
938/*
939 * Invalidate instruction cache
940 */
941ENTRY(ICIA)
942	movl	#IC_CLEAR,d0
943	movc	d0,cacr			| invalidate i-cache
944	rts
945
946/*
947 * Invalidate data cache.
948 * NOTE: we do not flush 68030 on-chip cache as there are no aliasing
949 * problems with DC_WA.  The only cases we have to worry about are context
950 * switch and TLB changes, both of which are handled "in-line" in resume
951 * and TBI*.
952 */
953ENTRY(DCIA)
954__DCIA:
955	rts
956
957ENTRY(DCIS)
958__DCIS:
959	rts
960
961/*
962 * Invalidate data cache.
963 */
964ENTRY(DCIU)
965	movl	#DC_CLEAR,d0
966	movc	d0,cacr			| invalidate on-chip d-cache
967	rts
968
969/* ICPL, ICPP, DCPL, DCPP, DCPA, DCFL, DCFP */
970
971ENTRY(PCIA)
972	movl	#DC_CLEAR,d0
973	movc	d0,cacr			| invalidate on-chip d-cache
974	rts
975
976ENTRY(ecacheon)
977	rts
978
979ENTRY(ecacheoff)
980	rts
981
982/*
983 * Get callers current SP value.
984 * Note that simply taking the address of a local variable in a C function
985 * doesn't work because callee saved registers may be outside the stack frame
986 * defined by A6 (e.g. GCC generated code).
987 *
988 * [I don't think the ENTRY() macro will do the right thing with this -- glass]
989 */
990GLOBAL(getsp)
991	movl	sp,d0			| get current SP
992	addql	#4,d0			| compensate for return address
993	rts
994
995ENTRY(getsfc)
996	movc	sfc,d0
997	rts
998
999ENTRY(getdfc)
1000	movc	dfc,d0
1001	rts
1002
1003ENTRY(getvbr)
1004	movc vbr, d0
1005	rts
1006
1007ENTRY(setvbr)
1008	movl sp@(4), d0
1009	movc d0, vbr
1010	rts
1011
1012/*
1013 * Load a new CPU Root Pointer (CRP) into the MMU.
1014 *	void	loadcrp(struct mmu_rootptr *);
1015 */
1016ENTRY(loadcrp)
1017	movl	sp@(4),a0		| arg1: &CRP
1018	movl	#CACHE_CLR,d0
1019	movc	d0,cacr			| invalidate cache(s)
1020	pflusha				| flush entire TLB
1021	pmove	a0@,crp			| load new user root pointer
1022	rts
1023
1024/*
1025 * Get the physical address of the PTE for a given VA.
1026 */
1027ENTRY(ptest_addr)
1028	movl	sp@(4),a0		| VA
1029	ptestr	#5,a0@,#7,a1		| a1 = addr of PTE
1030	movl	a1,d0
1031	rts
1032
1033/*
1034 * Set processor priority level calls.  Most are implemented with
1035 * inline asm expansions.  However, we need one instantiation here
1036 * in case some non-optimized code makes external references.
1037 * Most places will use the inlined functions param.h supplies.
1038 */
1039
1040ENTRY(_getsr)
1041	clrl	d0
1042	movw	sr,d0
1043	rts
1044
1045ENTRY(_spl)
1046	clrl	d0
1047	movw	sr,d0
1048	movl	sp@(4),d1
1049	movw	d1,sr
1050	rts
1051
1052ENTRY(_splraise)
1053	clrl	d0
1054	movw	sr,d0
1055	movl	d0,d1
1056	andl	#PSL_HIGHIPL,d1 	| old &= PSL_HIGHIPL
1057	cmpl	sp@(4),d1		| (old - new)
1058	bge	Lsplr
1059	movl	sp@(4),d1
1060	movw	d1,sr
1061Lsplr:
1062	rts
1063
1064/*
1065 * Save and restore 68881 state.
1066 */
1067ENTRY(m68881_save)
1068	movl	sp@(4),a0		| save area pointer
1069	fsave	a0@			| save state
1070	tstb	a0@			| null state frame?
1071	jeq	Lm68881sdone		| yes, all done
1072	fmovem fp0-fp7,a0@(FPF_REGS)		| save FP general regs
1073	fmovem fpcr/fpsr/fpi,a0@(FPF_FPCR)	| save FP control regs
1074Lm68881sdone:
1075	rts
1076
1077ENTRY(m68881_restore)
1078	movl	sp@(4),a0		| save area pointer
1079	tstb	a0@			| null state frame?
1080	jeq	Lm68881rdone		| yes, easy
1081	fmovem	a0@(FPF_FPCR),fpcr/fpsr/fpi	| restore FP control regs
1082	fmovem	a0@(FPF_REGS),fp0-fp7		| restore FP general regs
1083Lm68881rdone:
1084	frestore a0@			| restore state
1085	rts
1086
1087/*
1088 * _delay(unsigned N)
1089 * Delay for at least (N/256) microseconds.
1090 * This routine depends on the variable:  delay_divisor
1091 * which should be set based on the CPU clock rate.
1092 * XXX: Currently this is set based on the CPU model,
1093 * XXX: but this should be determined at run time...
1094 */
1095GLOBAL(_delay)
1096	| d0 = arg = (usecs << 8)
1097	movl	sp@(4),d0
1098	| d1 = delay_divisor;
1099	movl	_C_LABEL(delay_divisor),d1
1100L_delay:
1101	subl	d1,d0
1102	jgt	L_delay
1103	rts
1104
1105
1106| Define some addresses, mostly so DDB can print useful info.
1107| Not using _C_LABEL() here because these symbols are never
1108| referenced by any C code, and if the leading underscore
1109| ever goes away, these lines turn into syntax errors...
1110	.set	_KERNBASE,KERNBASE
1111	.set	_MONSTART,SUN3X_MONSTART
1112	.set	_PROM_BASE,SUN3X_PROM_BASE
1113	.set	_MONEND,SUN3X_MONEND
1114
1115|The end!
1116