cpu_machdep.c (32464) | cpu_machdep.c (32702) |
---|---|
1/*- 2 * Copyright (c) 1992 Terrence R. Lambert. 3 * Copyright (c) 1982, 1987, 1990 The Regents of the University of California. 4 * All rights reserved. 5 * 6 * This code is derived from software contributed to Berkeley by 7 * William Jolitz. 8 * --- 21 unchanged lines hidden (view full) --- 30 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 31 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 32 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 33 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 34 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 35 * SUCH DAMAGE. 36 * 37 * from: @(#)machdep.c 7.4 (Berkeley) 6/3/91 | 1/*- 2 * Copyright (c) 1992 Terrence R. Lambert. 3 * Copyright (c) 1982, 1987, 1990 The Regents of the University of California. 4 * All rights reserved. 5 * 6 * This code is derived from software contributed to Berkeley by 7 * William Jolitz. 8 * --- 21 unchanged lines hidden (view full) --- 30 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 31 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 32 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 33 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 34 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 35 * SUCH DAMAGE. 36 * 37 * from: @(#)machdep.c 7.4 (Berkeley) 6/3/91 |
38 * $Id: machdep.c,v 1.280 1997/12/27 02:28:27 peter Exp $ | 38 * $Id: machdep.c,v 1.281 1998/01/12 05:16:03 dyson Exp $ |
39 */ 40 41#include "apm.h" 42#include "npx.h" 43#include "opt_bounce.h" 44#include "opt_cpu.h" 45#include "opt_ddb.h" 46#include "opt_maxmem.h" --- 291 unchanged lines hidden (view full) --- 338 * End of second pass, addresses have been assigned 339 */ 340 if ((vm_size_t)(v - firstaddr) != size) 341 panic("startup: table size inconsistency"); 342 343#ifdef BOUNCE_BUFFERS 344 clean_map = kmem_suballoc(kernel_map, &clean_sva, &clean_eva, 345 (nbuf*BKVASIZE) + (nswbuf*MAXPHYS) + | 39 */ 40 41#include "apm.h" 42#include "npx.h" 43#include "opt_bounce.h" 44#include "opt_cpu.h" 45#include "opt_ddb.h" 46#include "opt_maxmem.h" --- 291 unchanged lines hidden (view full) --- 338 * End of second pass, addresses have been assigned 339 */ 340 if ((vm_size_t)(v - firstaddr) != size) 341 panic("startup: table size inconsistency"); 342 343#ifdef BOUNCE_BUFFERS 344 clean_map = kmem_suballoc(kernel_map, &clean_sva, &clean_eva, 345 (nbuf*BKVASIZE) + (nswbuf*MAXPHYS) + |
346 maxbkva + pager_map_size, TRUE); 347 io_map = kmem_suballoc(clean_map, &minaddr, &maxaddr, maxbkva, FALSE); | 346 maxbkva + pager_map_size); 347 io_map = kmem_suballoc(clean_map, &minaddr, &maxaddr, maxbkva); |
348#else 349 clean_map = kmem_suballoc(kernel_map, &clean_sva, &clean_eva, | 348#else 349 clean_map = kmem_suballoc(kernel_map, &clean_sva, &clean_eva, |
350 (nbuf*BKVASIZE) + (nswbuf*MAXPHYS) + pager_map_size, TRUE); | 350 (nbuf*BKVASIZE) + (nswbuf*MAXPHYS) + pager_map_size); |
351#endif 352 buffer_map = kmem_suballoc(clean_map, &buffer_sva, &buffer_eva, | 351#endif 352 buffer_map = kmem_suballoc(clean_map, &buffer_sva, &buffer_eva, |
353 (nbuf*BKVASIZE), TRUE); | 353 (nbuf*BKVASIZE)); |
354 pager_map = kmem_suballoc(clean_map, &pager_sva, &pager_eva, | 354 pager_map = kmem_suballoc(clean_map, &pager_sva, &pager_eva, |
355 (nswbuf*MAXPHYS) + pager_map_size, TRUE); | 355 (nswbuf*MAXPHYS) + pager_map_size); |
356 pager_map->system_map = 1; 357 exec_map = kmem_suballoc(kernel_map, &minaddr, &maxaddr, | 356 pager_map->system_map = 1; 357 exec_map = kmem_suballoc(kernel_map, &minaddr, &maxaddr, |
358 (16*(ARG_MAX+PAGE_SIZE)), TRUE); 359 u_map = kmem_suballoc(kernel_map, &minaddr, &maxaddr, 360 (maxproc*UPAGES*PAGE_SIZE), FALSE); | 358 (16*(ARG_MAX+PAGE_SIZE))); |
361 362 /* 363 * Finally, allocate mbuf pool. Since mclrefcnt is an off-size 364 * we use the more space efficient malloc in place of kmem_alloc. 365 */ 366 { 367 vm_offset_t mb_map_size; 368 369 mb_map_size = nmbufs * MSIZE + nmbclusters * MCLBYTES; 370 mb_map_size = roundup2(mb_map_size, max(MCLBYTES, PAGE_SIZE)); 371 mclrefcnt = malloc(mb_map_size / MCLBYTES, M_MBUF, M_NOWAIT); 372 bzero(mclrefcnt, mb_map_size / MCLBYTES); 373 mb_map = kmem_suballoc(kmem_map, (vm_offset_t *)&mbutl, &maxaddr, | 359 360 /* 361 * Finally, allocate mbuf pool. Since mclrefcnt is an off-size 362 * we use the more space efficient malloc in place of kmem_alloc. 363 */ 364 { 365 vm_offset_t mb_map_size; 366 367 mb_map_size = nmbufs * MSIZE + nmbclusters * MCLBYTES; 368 mb_map_size = roundup2(mb_map_size, max(MCLBYTES, PAGE_SIZE)); 369 mclrefcnt = malloc(mb_map_size / MCLBYTES, M_MBUF, M_NOWAIT); 370 bzero(mclrefcnt, mb_map_size / MCLBYTES); 371 mb_map = kmem_suballoc(kmem_map, (vm_offset_t *)&mbutl, &maxaddr, |
374 mb_map_size, FALSE); | 372 mb_map_size); |
375 mb_map->system_map = 1; 376 } 377 378 /* 379 * Initialize callouts 380 */ 381 SLIST_INIT(&callfree); 382 for (i = 0; i < ncallout; i++) { --- 1425 unchanged lines hidden --- | 373 mb_map->system_map = 1; 374 } 375 376 /* 377 * Initialize callouts 378 */ 379 SLIST_INIT(&callfree); 380 for (i = 0; i < ncallout; i++) { --- 1425 unchanged lines hidden --- |