189 190static int 191sysctl_nmbjumbo9(SYSCTL_HANDLER_ARGS) 192{ 193 int error, newnmbjumbo9; 194 195 newnmbjumbo9 = nmbjumbo9; 196 error = sysctl_handle_int(oidp, &newnmbjumbo9, 0, req); 197 if (error == 0 && req->newptr) { 198 if (newnmbjumbo9 > nmbjumbo9&& 199 nmbufs >= nmbclusters + nmbjumbop + nmbjumbo9 + nmbjumbo16) { 200 nmbjumbo9 = newnmbjumbo9; 201 uma_zone_set_max(zone_jumbo9, nmbjumbo9); 202 nmbjumbo9 = uma_zone_get_max(zone_jumbo9); 203 } else 204 error = EINVAL; 205 } 206 return (error); 207} 208SYSCTL_PROC(_kern_ipc, OID_AUTO, nmbjumbo9, CTLTYPE_INT|CTLFLAG_RW, 209&nmbjumbo9, 0, sysctl_nmbjumbo9, "IU", 210 "Maximum number of mbuf 9k jumbo clusters allowed"); 211 212static int 213sysctl_nmbjumbo16(SYSCTL_HANDLER_ARGS) 214{ 215 int error, newnmbjumbo16; 216 217 newnmbjumbo16 = nmbjumbo16; 218 error = sysctl_handle_int(oidp, &newnmbjumbo16, 0, req); 219 if (error == 0 && req->newptr) { 220 if (newnmbjumbo16 > nmbjumbo16 && 221 nmbufs >= nmbclusters + nmbjumbop + nmbjumbo9 + nmbjumbo16) { 222 nmbjumbo16 = newnmbjumbo16; 223 uma_zone_set_max(zone_jumbo16, nmbjumbo16); 224 nmbjumbo16 = uma_zone_get_max(zone_jumbo16); 225 } else 226 error = EINVAL; 227 } 228 return (error); 229} 230SYSCTL_PROC(_kern_ipc, OID_AUTO, nmbjumbo16, CTLTYPE_INT|CTLFLAG_RW, 231&nmbjumbo16, 0, sysctl_nmbjumbo16, "IU", 232 "Maximum number of mbuf 16k jumbo clusters allowed"); 233 234static int 235sysctl_nmbufs(SYSCTL_HANDLER_ARGS) 236{ 237 int error, newnmbufs; 238 239 newnmbufs = nmbufs; 240 error = sysctl_handle_int(oidp, &newnmbufs, 0, req); 241 if (error == 0 && req->newptr) { 242 if (newnmbufs > nmbufs) { 243 nmbufs = newnmbufs; 244 uma_zone_set_max(zone_mbuf, nmbufs); 245 nmbufs = uma_zone_get_max(zone_mbuf); 246 EVENTHANDLER_INVOKE(nmbufs_change); 247 } else 248 error = EINVAL; 249 } 250 return (error); 251} 252SYSCTL_PROC(_kern_ipc, OID_AUTO, nmbuf, CTLTYPE_INT|CTLFLAG_RW, 253&nmbufs, 0, sysctl_nmbufs, "IU", 254 "Maximum number of mbufs allowed"); 255 256SYSCTL_STRUCT(_kern_ipc, OID_AUTO, mbstat, CTLFLAG_RD, &mbstat, mbstat, 257 "Mbuf general information and statistics"); 258 259/* 260 * Zones from which we allocate. 261 */ 262uma_zone_t zone_mbuf; 263uma_zone_t zone_clust; 264uma_zone_t zone_pack; 265uma_zone_t zone_jumbop; 266uma_zone_t zone_jumbo9; 267uma_zone_t zone_jumbo16; 268uma_zone_t zone_ext_refcnt; 269 270/* 271 * Local prototypes. 272 */ 273static int mb_ctor_mbuf(void *, int, void *, int); 274static int mb_ctor_clust(void *, int, void *, int); 275static int mb_ctor_pack(void *, int, void *, int); 276static void mb_dtor_mbuf(void *, int, void *); 277static void mb_dtor_clust(void *, int, void *); 278static void mb_dtor_pack(void *, int, void *); 279static int mb_zinit_pack(void *, int, int); 280static void mb_zfini_pack(void *, int); 281 282static void mb_reclaim(void *); 283static void mbuf_init(void *); 284static void *mbuf_jumbo_alloc(uma_zone_t, int, uint8_t *, int); 285 286/* Ensure that MSIZE must be a power of 2. */ 287CTASSERT((((MSIZE - 1) ^ MSIZE) + 1) >> 1 == MSIZE); 288 289/* 290 * Initialize FreeBSD Network buffer allocation. 291 */ 292SYSINIT(mbuf, SI_SUB_MBUF, SI_ORDER_FIRST, mbuf_init, NULL); 293static void 294mbuf_init(void *dummy) 295{ 296 297 /* 298 * Configure UMA zones for Mbufs, Clusters, and Packets. 299 */ 300 zone_mbuf = uma_zcreate(MBUF_MEM_NAME, MSIZE, 301 mb_ctor_mbuf, mb_dtor_mbuf, 302#ifdef INVARIANTS 303 trash_init, trash_fini, 304#else 305 NULL, NULL, 306#endif 307 MSIZE - 1, UMA_ZONE_MAXBUCKET); 308 if (nmbufs > 0) { 309 uma_zone_set_max(zone_mbuf, nmbufs); 310 nmbufs = uma_zone_get_max(zone_mbuf); 311 } 312 313 zone_clust = uma_zcreate(MBUF_CLUSTER_MEM_NAME, MCLBYTES, 314 mb_ctor_clust, mb_dtor_clust, 315#ifdef INVARIANTS 316 trash_init, trash_fini, 317#else 318 NULL, NULL, 319#endif 320 UMA_ALIGN_PTR, UMA_ZONE_REFCNT); 321 if (nmbclusters > 0) { 322 uma_zone_set_max(zone_clust, nmbclusters); 323 nmbclusters = uma_zone_get_max(zone_clust); 324 } 325 326 zone_pack = uma_zsecond_create(MBUF_PACKET_MEM_NAME, mb_ctor_pack, 327 mb_dtor_pack, mb_zinit_pack, mb_zfini_pack, zone_mbuf); 328 329 /* Make jumbo frame zone too. Page size, 9k and 16k. */ 330 zone_jumbop = uma_zcreate(MBUF_JUMBOP_MEM_NAME, MJUMPAGESIZE, 331 mb_ctor_clust, mb_dtor_clust, 332#ifdef INVARIANTS 333 trash_init, trash_fini, 334#else 335 NULL, NULL, 336#endif 337 UMA_ALIGN_PTR, UMA_ZONE_REFCNT); 338 if (nmbjumbop > 0) { 339 uma_zone_set_max(zone_jumbop, nmbjumbop); 340 nmbjumbop = uma_zone_get_max(zone_jumbop); 341 } 342 343 zone_jumbo9 = uma_zcreate(MBUF_JUMBO9_MEM_NAME, MJUM9BYTES, 344 mb_ctor_clust, mb_dtor_clust, 345#ifdef INVARIANTS 346 trash_init, trash_fini, 347#else 348 NULL, NULL, 349#endif 350 UMA_ALIGN_PTR, UMA_ZONE_REFCNT); 351 uma_zone_set_allocf(zone_jumbo9, mbuf_jumbo_alloc); 352 if (nmbjumbo9 > 0) { 353 uma_zone_set_max(zone_jumbo9, nmbjumbo9); 354 nmbjumbo9 = uma_zone_get_max(zone_jumbo9); 355 } 356 357 zone_jumbo16 = uma_zcreate(MBUF_JUMBO16_MEM_NAME, MJUM16BYTES, 358 mb_ctor_clust, mb_dtor_clust, 359#ifdef INVARIANTS 360 trash_init, trash_fini, 361#else 362 NULL, NULL, 363#endif 364 UMA_ALIGN_PTR, UMA_ZONE_REFCNT); 365 uma_zone_set_allocf(zone_jumbo16, mbuf_jumbo_alloc); 366 if (nmbjumbo16 > 0) { 367 uma_zone_set_max(zone_jumbo16, nmbjumbo16); 368 nmbjumbo16 = uma_zone_get_max(zone_jumbo16); 369 } 370 371 zone_ext_refcnt = uma_zcreate(MBUF_EXTREFCNT_MEM_NAME, sizeof(u_int), 372 NULL, NULL, 373 NULL, NULL, 374 UMA_ALIGN_PTR, UMA_ZONE_ZINIT); 375 376 /* uma_prealloc() goes here... */ 377 378 /* 379 * Hook event handler for low-memory situation, used to 380 * drain protocols and push data back to the caches (UMA 381 * later pushes it back to VM). 382 */ 383 EVENTHANDLER_REGISTER(vm_lowmem, mb_reclaim, NULL, 384 EVENTHANDLER_PRI_FIRST); 385 386 /* 387 * [Re]set counters and local statistics knobs. 388 * XXX Some of these should go and be replaced, but UMA stat 389 * gathering needs to be revised. 390 */ 391 mbstat.m_mbufs = 0; 392 mbstat.m_mclusts = 0; 393 mbstat.m_drain = 0; 394 mbstat.m_msize = MSIZE; 395 mbstat.m_mclbytes = MCLBYTES; 396 mbstat.m_minclsize = MINCLSIZE; 397 mbstat.m_mlen = MLEN; 398 mbstat.m_mhlen = MHLEN; 399 mbstat.m_numtypes = MT_NTYPES; 400 401 mbstat.m_mcfail = mbstat.m_mpfail = 0; 402 mbstat.sf_iocnt = 0; 403 mbstat.sf_allocwait = mbstat.sf_allocfail = 0; 404} 405 406/* 407 * UMA backend page allocator for the jumbo frame zones. 408 * 409 * Allocates kernel virtual memory that is backed by contiguous physical 410 * pages. 411 */ 412static void * 413mbuf_jumbo_alloc(uma_zone_t zone, int bytes, uint8_t *flags, int wait) 414{ 415 416 /* Inform UMA that this allocator uses kernel_map/object. */ 417 *flags = UMA_SLAB_KERNEL; 418 return ((void *)kmem_alloc_contig(kernel_map, bytes, wait, 419 (vm_paddr_t)0, ~(vm_paddr_t)0, 1, 0, VM_MEMATTR_DEFAULT)); 420} 421 422/* 423 * Constructor for Mbuf master zone. 424 * 425 * The 'arg' pointer points to a mb_args structure which 426 * contains call-specific information required to support the 427 * mbuf allocation API. See mbuf.h. 428 */ 429static int 430mb_ctor_mbuf(void *mem, int size, void *arg, int how) 431{ 432 struct mbuf *m; 433 struct mb_args *args; 434#ifdef MAC 435 int error; 436#endif 437 int flags; 438 short type; 439 440#ifdef INVARIANTS 441 trash_ctor(mem, size, arg, how); 442#endif 443 m = (struct mbuf *)mem; 444 args = (struct mb_args *)arg; 445 flags = args->flags; 446 type = args->type; 447 448 /* 449 * The mbuf is initialized later. The caller has the 450 * responsibility to set up any MAC labels too. 451 */ 452 if (type == MT_NOINIT) 453 return (0); 454 455 m->m_next = NULL; 456 m->m_nextpkt = NULL; 457 m->m_len = 0; 458 m->m_flags = flags; 459 m->m_type = type; 460 if (flags & M_PKTHDR) { 461 m->m_data = m->m_pktdat; 462 m->m_pkthdr.rcvif = NULL; 463 m->m_pkthdr.header = NULL; 464 m->m_pkthdr.len = 0; 465 m->m_pkthdr.csum_flags = 0; 466 m->m_pkthdr.csum_data = 0; 467 m->m_pkthdr.tso_segsz = 0; 468 m->m_pkthdr.ether_vtag = 0; 469 m->m_pkthdr.flowid = 0; 470 SLIST_INIT(&m->m_pkthdr.tags); 471#ifdef MAC 472 /* If the label init fails, fail the alloc */ 473 error = mac_mbuf_init(m, how); 474 if (error) 475 return (error); 476#endif 477 } else 478 m->m_data = m->m_dat; 479 return (0); 480} 481 482/* 483 * The Mbuf master zone destructor. 484 */ 485static void 486mb_dtor_mbuf(void *mem, int size, void *arg) 487{ 488 struct mbuf *m; 489 unsigned long flags; 490 491 m = (struct mbuf *)mem; 492 flags = (unsigned long)arg; 493 494 if ((flags & MB_NOTAGS) == 0 && (m->m_flags & M_PKTHDR) != 0) 495 m_tag_delete_chain(m, NULL); 496 KASSERT((m->m_flags & M_EXT) == 0, ("%s: M_EXT set", __func__)); 497 KASSERT((m->m_flags & M_NOFREE) == 0, ("%s: M_NOFREE set", __func__)); 498#ifdef INVARIANTS 499 trash_dtor(mem, size, arg); 500#endif 501} 502 503/* 504 * The Mbuf Packet zone destructor. 505 */ 506static void 507mb_dtor_pack(void *mem, int size, void *arg) 508{ 509 struct mbuf *m; 510 511 m = (struct mbuf *)mem; 512 if ((m->m_flags & M_PKTHDR) != 0) 513 m_tag_delete_chain(m, NULL); 514 515 /* Make sure we've got a clean cluster back. */ 516 KASSERT((m->m_flags & M_EXT) == M_EXT, ("%s: M_EXT not set", __func__)); 517 KASSERT(m->m_ext.ext_buf != NULL, ("%s: ext_buf == NULL", __func__)); 518 KASSERT(m->m_ext.ext_free == NULL, ("%s: ext_free != NULL", __func__)); 519 KASSERT(m->m_ext.ext_arg1 == NULL, ("%s: ext_arg1 != NULL", __func__)); 520 KASSERT(m->m_ext.ext_arg2 == NULL, ("%s: ext_arg2 != NULL", __func__)); 521 KASSERT(m->m_ext.ext_size == MCLBYTES, ("%s: ext_size != MCLBYTES", __func__)); 522 KASSERT(m->m_ext.ext_type == EXT_PACKET, ("%s: ext_type != EXT_PACKET", __func__)); 523 KASSERT(*m->m_ext.ref_cnt == 1, ("%s: ref_cnt != 1", __func__)); 524#ifdef INVARIANTS 525 trash_dtor(m->m_ext.ext_buf, MCLBYTES, arg); 526#endif 527 /* 528 * If there are processes blocked on zone_clust, waiting for pages 529 * to be freed up, * cause them to be woken up by draining the 530 * packet zone. We are exposed to a race here * (in the check for 531 * the UMA_ZFLAG_FULL) where we might miss the flag set, but that 532 * is deliberate. We don't want to acquire the zone lock for every 533 * mbuf free. 534 */ 535 if (uma_zone_exhausted_nolock(zone_clust)) 536 zone_drain(zone_pack); 537} 538 539/* 540 * The Cluster and Jumbo[PAGESIZE|9|16] zone constructor. 541 * 542 * Here the 'arg' pointer points to the Mbuf which we 543 * are configuring cluster storage for. If 'arg' is 544 * empty we allocate just the cluster without setting 545 * the mbuf to it. See mbuf.h. 546 */ 547static int 548mb_ctor_clust(void *mem, int size, void *arg, int how) 549{ 550 struct mbuf *m; 551 u_int *refcnt; 552 int type; 553 uma_zone_t zone; 554 555#ifdef INVARIANTS 556 trash_ctor(mem, size, arg, how); 557#endif 558 switch (size) { 559 case MCLBYTES: 560 type = EXT_CLUSTER; 561 zone = zone_clust; 562 break; 563#if MJUMPAGESIZE != MCLBYTES 564 case MJUMPAGESIZE: 565 type = EXT_JUMBOP; 566 zone = zone_jumbop; 567 break; 568#endif 569 case MJUM9BYTES: 570 type = EXT_JUMBO9; 571 zone = zone_jumbo9; 572 break; 573 case MJUM16BYTES: 574 type = EXT_JUMBO16; 575 zone = zone_jumbo16; 576 break; 577 default: 578 panic("unknown cluster size"); 579 break; 580 } 581 582 m = (struct mbuf *)arg; 583 refcnt = uma_find_refcnt(zone, mem); 584 *refcnt = 1; 585 if (m != NULL) { 586 m->m_ext.ext_buf = (caddr_t)mem; 587 m->m_data = m->m_ext.ext_buf; 588 m->m_flags |= M_EXT; 589 m->m_ext.ext_free = NULL; 590 m->m_ext.ext_arg1 = NULL; 591 m->m_ext.ext_arg2 = NULL; 592 m->m_ext.ext_size = size; 593 m->m_ext.ext_type = type; 594 m->m_ext.ref_cnt = refcnt; 595 } 596 597 return (0); 598} 599 600/* 601 * The Mbuf Cluster zone destructor. 602 */ 603static void 604mb_dtor_clust(void *mem, int size, void *arg) 605{ 606#ifdef INVARIANTS 607 uma_zone_t zone; 608 609 zone = m_getzone(size); 610 KASSERT(*(uma_find_refcnt(zone, mem)) <= 1, 611 ("%s: refcnt incorrect %u", __func__, 612 *(uma_find_refcnt(zone, mem))) ); 613 614 trash_dtor(mem, size, arg); 615#endif 616} 617 618/* 619 * The Packet secondary zone's init routine, executed on the 620 * object's transition from mbuf keg slab to zone cache. 621 */ 622static int 623mb_zinit_pack(void *mem, int size, int how) 624{ 625 struct mbuf *m; 626 627 m = (struct mbuf *)mem; /* m is virgin. */ 628 if (uma_zalloc_arg(zone_clust, m, how) == NULL || 629 m->m_ext.ext_buf == NULL) 630 return (ENOMEM); 631 m->m_ext.ext_type = EXT_PACKET; /* Override. */ 632#ifdef INVARIANTS 633 trash_init(m->m_ext.ext_buf, MCLBYTES, how); 634#endif 635 return (0); 636} 637 638/* 639 * The Packet secondary zone's fini routine, executed on the 640 * object's transition from zone cache to keg slab. 641 */ 642static void 643mb_zfini_pack(void *mem, int size) 644{ 645 struct mbuf *m; 646 647 m = (struct mbuf *)mem; 648#ifdef INVARIANTS 649 trash_fini(m->m_ext.ext_buf, MCLBYTES); 650#endif 651 uma_zfree_arg(zone_clust, m->m_ext.ext_buf, NULL); 652#ifdef INVARIANTS 653 trash_dtor(mem, size, NULL); 654#endif 655} 656 657/* 658 * The "packet" keg constructor. 659 */ 660static int 661mb_ctor_pack(void *mem, int size, void *arg, int how) 662{ 663 struct mbuf *m; 664 struct mb_args *args; 665#ifdef MAC 666 int error; 667#endif 668 int flags; 669 short type; 670 671 m = (struct mbuf *)mem; 672 args = (struct mb_args *)arg; 673 flags = args->flags; 674 type = args->type; 675 676#ifdef INVARIANTS 677 trash_ctor(m->m_ext.ext_buf, MCLBYTES, arg, how); 678#endif 679 m->m_next = NULL; 680 m->m_nextpkt = NULL; 681 m->m_data = m->m_ext.ext_buf; 682 m->m_len = 0; 683 m->m_flags = (flags | M_EXT); 684 m->m_type = type; 685 686 if (flags & M_PKTHDR) { 687 m->m_pkthdr.rcvif = NULL; 688 m->m_pkthdr.len = 0; 689 m->m_pkthdr.header = NULL; 690 m->m_pkthdr.csum_flags = 0; 691 m->m_pkthdr.csum_data = 0; 692 m->m_pkthdr.tso_segsz = 0; 693 m->m_pkthdr.ether_vtag = 0; 694 m->m_pkthdr.flowid = 0; 695 SLIST_INIT(&m->m_pkthdr.tags); 696#ifdef MAC 697 /* If the label init fails, fail the alloc */ 698 error = mac_mbuf_init(m, how); 699 if (error) 700 return (error); 701#endif 702 } 703 /* m_ext is already initialized. */ 704 705 return (0); 706} 707 708int 709m_pkthdr_init(struct mbuf *m, int how) 710{ 711#ifdef MAC 712 int error; 713#endif 714 m->m_data = m->m_pktdat; 715 SLIST_INIT(&m->m_pkthdr.tags); 716 m->m_pkthdr.rcvif = NULL; 717 m->m_pkthdr.header = NULL; 718 m->m_pkthdr.len = 0; 719 m->m_pkthdr.flowid = 0; 720 m->m_pkthdr.csum_flags = 0; 721 m->m_pkthdr.csum_data = 0; 722 m->m_pkthdr.tso_segsz = 0; 723 m->m_pkthdr.ether_vtag = 0; 724#ifdef MAC 725 /* If the label init fails, fail the alloc */ 726 error = mac_mbuf_init(m, how); 727 if (error) 728 return (error); 729#endif 730 731 return (0); 732} 733 734/* 735 * This is the protocol drain routine. 736 * 737 * No locks should be held when this is called. The drain routines have to 738 * presently acquire some locks which raises the possibility of lock order 739 * reversal. 740 */ 741static void 742mb_reclaim(void *junk) 743{ 744 struct domain *dp; 745 struct protosw *pr; 746 747 WITNESS_WARN(WARN_GIANTOK | WARN_SLEEPOK | WARN_PANIC, NULL, 748 "mb_reclaim()"); 749 750 for (dp = domains; dp != NULL; dp = dp->dom_next) 751 for (pr = dp->dom_protosw; pr < dp->dom_protoswNPROTOSW; pr++) 752 if (pr->pr_drain != NULL) 753 (*pr->pr_drain)(); 754}
| 188 189static int 190sysctl_nmbjumbo9(SYSCTL_HANDLER_ARGS) 191{ 192 int error, newnmbjumbo9; 193 194 newnmbjumbo9 = nmbjumbo9; 195 error = sysctl_handle_int(oidp, &newnmbjumbo9, 0, req); 196 if (error == 0 && req->newptr) { 197 if (newnmbjumbo9 > nmbjumbo9&& 198 nmbufs >= nmbclusters + nmbjumbop + nmbjumbo9 + nmbjumbo16) { 199 nmbjumbo9 = newnmbjumbo9; 200 uma_zone_set_max(zone_jumbo9, nmbjumbo9); 201 nmbjumbo9 = uma_zone_get_max(zone_jumbo9); 202 } else 203 error = EINVAL; 204 } 205 return (error); 206} 207SYSCTL_PROC(_kern_ipc, OID_AUTO, nmbjumbo9, CTLTYPE_INT|CTLFLAG_RW, 208&nmbjumbo9, 0, sysctl_nmbjumbo9, "IU", 209 "Maximum number of mbuf 9k jumbo clusters allowed"); 210 211static int 212sysctl_nmbjumbo16(SYSCTL_HANDLER_ARGS) 213{ 214 int error, newnmbjumbo16; 215 216 newnmbjumbo16 = nmbjumbo16; 217 error = sysctl_handle_int(oidp, &newnmbjumbo16, 0, req); 218 if (error == 0 && req->newptr) { 219 if (newnmbjumbo16 > nmbjumbo16 && 220 nmbufs >= nmbclusters + nmbjumbop + nmbjumbo9 + nmbjumbo16) { 221 nmbjumbo16 = newnmbjumbo16; 222 uma_zone_set_max(zone_jumbo16, nmbjumbo16); 223 nmbjumbo16 = uma_zone_get_max(zone_jumbo16); 224 } else 225 error = EINVAL; 226 } 227 return (error); 228} 229SYSCTL_PROC(_kern_ipc, OID_AUTO, nmbjumbo16, CTLTYPE_INT|CTLFLAG_RW, 230&nmbjumbo16, 0, sysctl_nmbjumbo16, "IU", 231 "Maximum number of mbuf 16k jumbo clusters allowed"); 232 233static int 234sysctl_nmbufs(SYSCTL_HANDLER_ARGS) 235{ 236 int error, newnmbufs; 237 238 newnmbufs = nmbufs; 239 error = sysctl_handle_int(oidp, &newnmbufs, 0, req); 240 if (error == 0 && req->newptr) { 241 if (newnmbufs > nmbufs) { 242 nmbufs = newnmbufs; 243 uma_zone_set_max(zone_mbuf, nmbufs); 244 nmbufs = uma_zone_get_max(zone_mbuf); 245 EVENTHANDLER_INVOKE(nmbufs_change); 246 } else 247 error = EINVAL; 248 } 249 return (error); 250} 251SYSCTL_PROC(_kern_ipc, OID_AUTO, nmbuf, CTLTYPE_INT|CTLFLAG_RW, 252&nmbufs, 0, sysctl_nmbufs, "IU", 253 "Maximum number of mbufs allowed"); 254 255SYSCTL_STRUCT(_kern_ipc, OID_AUTO, mbstat, CTLFLAG_RD, &mbstat, mbstat, 256 "Mbuf general information and statistics"); 257 258/* 259 * Zones from which we allocate. 260 */ 261uma_zone_t zone_mbuf; 262uma_zone_t zone_clust; 263uma_zone_t zone_pack; 264uma_zone_t zone_jumbop; 265uma_zone_t zone_jumbo9; 266uma_zone_t zone_jumbo16; 267uma_zone_t zone_ext_refcnt; 268 269/* 270 * Local prototypes. 271 */ 272static int mb_ctor_mbuf(void *, int, void *, int); 273static int mb_ctor_clust(void *, int, void *, int); 274static int mb_ctor_pack(void *, int, void *, int); 275static void mb_dtor_mbuf(void *, int, void *); 276static void mb_dtor_clust(void *, int, void *); 277static void mb_dtor_pack(void *, int, void *); 278static int mb_zinit_pack(void *, int, int); 279static void mb_zfini_pack(void *, int); 280 281static void mb_reclaim(void *); 282static void mbuf_init(void *); 283static void *mbuf_jumbo_alloc(uma_zone_t, int, uint8_t *, int); 284 285/* Ensure that MSIZE must be a power of 2. */ 286CTASSERT((((MSIZE - 1) ^ MSIZE) + 1) >> 1 == MSIZE); 287 288/* 289 * Initialize FreeBSD Network buffer allocation. 290 */ 291SYSINIT(mbuf, SI_SUB_MBUF, SI_ORDER_FIRST, mbuf_init, NULL); 292static void 293mbuf_init(void *dummy) 294{ 295 296 /* 297 * Configure UMA zones for Mbufs, Clusters, and Packets. 298 */ 299 zone_mbuf = uma_zcreate(MBUF_MEM_NAME, MSIZE, 300 mb_ctor_mbuf, mb_dtor_mbuf, 301#ifdef INVARIANTS 302 trash_init, trash_fini, 303#else 304 NULL, NULL, 305#endif 306 MSIZE - 1, UMA_ZONE_MAXBUCKET); 307 if (nmbufs > 0) { 308 uma_zone_set_max(zone_mbuf, nmbufs); 309 nmbufs = uma_zone_get_max(zone_mbuf); 310 } 311 312 zone_clust = uma_zcreate(MBUF_CLUSTER_MEM_NAME, MCLBYTES, 313 mb_ctor_clust, mb_dtor_clust, 314#ifdef INVARIANTS 315 trash_init, trash_fini, 316#else 317 NULL, NULL, 318#endif 319 UMA_ALIGN_PTR, UMA_ZONE_REFCNT); 320 if (nmbclusters > 0) { 321 uma_zone_set_max(zone_clust, nmbclusters); 322 nmbclusters = uma_zone_get_max(zone_clust); 323 } 324 325 zone_pack = uma_zsecond_create(MBUF_PACKET_MEM_NAME, mb_ctor_pack, 326 mb_dtor_pack, mb_zinit_pack, mb_zfini_pack, zone_mbuf); 327 328 /* Make jumbo frame zone too. Page size, 9k and 16k. */ 329 zone_jumbop = uma_zcreate(MBUF_JUMBOP_MEM_NAME, MJUMPAGESIZE, 330 mb_ctor_clust, mb_dtor_clust, 331#ifdef INVARIANTS 332 trash_init, trash_fini, 333#else 334 NULL, NULL, 335#endif 336 UMA_ALIGN_PTR, UMA_ZONE_REFCNT); 337 if (nmbjumbop > 0) { 338 uma_zone_set_max(zone_jumbop, nmbjumbop); 339 nmbjumbop = uma_zone_get_max(zone_jumbop); 340 } 341 342 zone_jumbo9 = uma_zcreate(MBUF_JUMBO9_MEM_NAME, MJUM9BYTES, 343 mb_ctor_clust, mb_dtor_clust, 344#ifdef INVARIANTS 345 trash_init, trash_fini, 346#else 347 NULL, NULL, 348#endif 349 UMA_ALIGN_PTR, UMA_ZONE_REFCNT); 350 uma_zone_set_allocf(zone_jumbo9, mbuf_jumbo_alloc); 351 if (nmbjumbo9 > 0) { 352 uma_zone_set_max(zone_jumbo9, nmbjumbo9); 353 nmbjumbo9 = uma_zone_get_max(zone_jumbo9); 354 } 355 356 zone_jumbo16 = uma_zcreate(MBUF_JUMBO16_MEM_NAME, MJUM16BYTES, 357 mb_ctor_clust, mb_dtor_clust, 358#ifdef INVARIANTS 359 trash_init, trash_fini, 360#else 361 NULL, NULL, 362#endif 363 UMA_ALIGN_PTR, UMA_ZONE_REFCNT); 364 uma_zone_set_allocf(zone_jumbo16, mbuf_jumbo_alloc); 365 if (nmbjumbo16 > 0) { 366 uma_zone_set_max(zone_jumbo16, nmbjumbo16); 367 nmbjumbo16 = uma_zone_get_max(zone_jumbo16); 368 } 369 370 zone_ext_refcnt = uma_zcreate(MBUF_EXTREFCNT_MEM_NAME, sizeof(u_int), 371 NULL, NULL, 372 NULL, NULL, 373 UMA_ALIGN_PTR, UMA_ZONE_ZINIT); 374 375 /* uma_prealloc() goes here... */ 376 377 /* 378 * Hook event handler for low-memory situation, used to 379 * drain protocols and push data back to the caches (UMA 380 * later pushes it back to VM). 381 */ 382 EVENTHANDLER_REGISTER(vm_lowmem, mb_reclaim, NULL, 383 EVENTHANDLER_PRI_FIRST); 384 385 /* 386 * [Re]set counters and local statistics knobs. 387 * XXX Some of these should go and be replaced, but UMA stat 388 * gathering needs to be revised. 389 */ 390 mbstat.m_mbufs = 0; 391 mbstat.m_mclusts = 0; 392 mbstat.m_drain = 0; 393 mbstat.m_msize = MSIZE; 394 mbstat.m_mclbytes = MCLBYTES; 395 mbstat.m_minclsize = MINCLSIZE; 396 mbstat.m_mlen = MLEN; 397 mbstat.m_mhlen = MHLEN; 398 mbstat.m_numtypes = MT_NTYPES; 399 400 mbstat.m_mcfail = mbstat.m_mpfail = 0; 401 mbstat.sf_iocnt = 0; 402 mbstat.sf_allocwait = mbstat.sf_allocfail = 0; 403} 404 405/* 406 * UMA backend page allocator for the jumbo frame zones. 407 * 408 * Allocates kernel virtual memory that is backed by contiguous physical 409 * pages. 410 */ 411static void * 412mbuf_jumbo_alloc(uma_zone_t zone, int bytes, uint8_t *flags, int wait) 413{ 414 415 /* Inform UMA that this allocator uses kernel_map/object. */ 416 *flags = UMA_SLAB_KERNEL; 417 return ((void *)kmem_alloc_contig(kernel_map, bytes, wait, 418 (vm_paddr_t)0, ~(vm_paddr_t)0, 1, 0, VM_MEMATTR_DEFAULT)); 419} 420 421/* 422 * Constructor for Mbuf master zone. 423 * 424 * The 'arg' pointer points to a mb_args structure which 425 * contains call-specific information required to support the 426 * mbuf allocation API. See mbuf.h. 427 */ 428static int 429mb_ctor_mbuf(void *mem, int size, void *arg, int how) 430{ 431 struct mbuf *m; 432 struct mb_args *args; 433#ifdef MAC 434 int error; 435#endif 436 int flags; 437 short type; 438 439#ifdef INVARIANTS 440 trash_ctor(mem, size, arg, how); 441#endif 442 m = (struct mbuf *)mem; 443 args = (struct mb_args *)arg; 444 flags = args->flags; 445 type = args->type; 446 447 /* 448 * The mbuf is initialized later. The caller has the 449 * responsibility to set up any MAC labels too. 450 */ 451 if (type == MT_NOINIT) 452 return (0); 453 454 m->m_next = NULL; 455 m->m_nextpkt = NULL; 456 m->m_len = 0; 457 m->m_flags = flags; 458 m->m_type = type; 459 if (flags & M_PKTHDR) { 460 m->m_data = m->m_pktdat; 461 m->m_pkthdr.rcvif = NULL; 462 m->m_pkthdr.header = NULL; 463 m->m_pkthdr.len = 0; 464 m->m_pkthdr.csum_flags = 0; 465 m->m_pkthdr.csum_data = 0; 466 m->m_pkthdr.tso_segsz = 0; 467 m->m_pkthdr.ether_vtag = 0; 468 m->m_pkthdr.flowid = 0; 469 SLIST_INIT(&m->m_pkthdr.tags); 470#ifdef MAC 471 /* If the label init fails, fail the alloc */ 472 error = mac_mbuf_init(m, how); 473 if (error) 474 return (error); 475#endif 476 } else 477 m->m_data = m->m_dat; 478 return (0); 479} 480 481/* 482 * The Mbuf master zone destructor. 483 */ 484static void 485mb_dtor_mbuf(void *mem, int size, void *arg) 486{ 487 struct mbuf *m; 488 unsigned long flags; 489 490 m = (struct mbuf *)mem; 491 flags = (unsigned long)arg; 492 493 if ((flags & MB_NOTAGS) == 0 && (m->m_flags & M_PKTHDR) != 0) 494 m_tag_delete_chain(m, NULL); 495 KASSERT((m->m_flags & M_EXT) == 0, ("%s: M_EXT set", __func__)); 496 KASSERT((m->m_flags & M_NOFREE) == 0, ("%s: M_NOFREE set", __func__)); 497#ifdef INVARIANTS 498 trash_dtor(mem, size, arg); 499#endif 500} 501 502/* 503 * The Mbuf Packet zone destructor. 504 */ 505static void 506mb_dtor_pack(void *mem, int size, void *arg) 507{ 508 struct mbuf *m; 509 510 m = (struct mbuf *)mem; 511 if ((m->m_flags & M_PKTHDR) != 0) 512 m_tag_delete_chain(m, NULL); 513 514 /* Make sure we've got a clean cluster back. */ 515 KASSERT((m->m_flags & M_EXT) == M_EXT, ("%s: M_EXT not set", __func__)); 516 KASSERT(m->m_ext.ext_buf != NULL, ("%s: ext_buf == NULL", __func__)); 517 KASSERT(m->m_ext.ext_free == NULL, ("%s: ext_free != NULL", __func__)); 518 KASSERT(m->m_ext.ext_arg1 == NULL, ("%s: ext_arg1 != NULL", __func__)); 519 KASSERT(m->m_ext.ext_arg2 == NULL, ("%s: ext_arg2 != NULL", __func__)); 520 KASSERT(m->m_ext.ext_size == MCLBYTES, ("%s: ext_size != MCLBYTES", __func__)); 521 KASSERT(m->m_ext.ext_type == EXT_PACKET, ("%s: ext_type != EXT_PACKET", __func__)); 522 KASSERT(*m->m_ext.ref_cnt == 1, ("%s: ref_cnt != 1", __func__)); 523#ifdef INVARIANTS 524 trash_dtor(m->m_ext.ext_buf, MCLBYTES, arg); 525#endif 526 /* 527 * If there are processes blocked on zone_clust, waiting for pages 528 * to be freed up, * cause them to be woken up by draining the 529 * packet zone. We are exposed to a race here * (in the check for 530 * the UMA_ZFLAG_FULL) where we might miss the flag set, but that 531 * is deliberate. We don't want to acquire the zone lock for every 532 * mbuf free. 533 */ 534 if (uma_zone_exhausted_nolock(zone_clust)) 535 zone_drain(zone_pack); 536} 537 538/* 539 * The Cluster and Jumbo[PAGESIZE|9|16] zone constructor. 540 * 541 * Here the 'arg' pointer points to the Mbuf which we 542 * are configuring cluster storage for. If 'arg' is 543 * empty we allocate just the cluster without setting 544 * the mbuf to it. See mbuf.h. 545 */ 546static int 547mb_ctor_clust(void *mem, int size, void *arg, int how) 548{ 549 struct mbuf *m; 550 u_int *refcnt; 551 int type; 552 uma_zone_t zone; 553 554#ifdef INVARIANTS 555 trash_ctor(mem, size, arg, how); 556#endif 557 switch (size) { 558 case MCLBYTES: 559 type = EXT_CLUSTER; 560 zone = zone_clust; 561 break; 562#if MJUMPAGESIZE != MCLBYTES 563 case MJUMPAGESIZE: 564 type = EXT_JUMBOP; 565 zone = zone_jumbop; 566 break; 567#endif 568 case MJUM9BYTES: 569 type = EXT_JUMBO9; 570 zone = zone_jumbo9; 571 break; 572 case MJUM16BYTES: 573 type = EXT_JUMBO16; 574 zone = zone_jumbo16; 575 break; 576 default: 577 panic("unknown cluster size"); 578 break; 579 } 580 581 m = (struct mbuf *)arg; 582 refcnt = uma_find_refcnt(zone, mem); 583 *refcnt = 1; 584 if (m != NULL) { 585 m->m_ext.ext_buf = (caddr_t)mem; 586 m->m_data = m->m_ext.ext_buf; 587 m->m_flags |= M_EXT; 588 m->m_ext.ext_free = NULL; 589 m->m_ext.ext_arg1 = NULL; 590 m->m_ext.ext_arg2 = NULL; 591 m->m_ext.ext_size = size; 592 m->m_ext.ext_type = type; 593 m->m_ext.ref_cnt = refcnt; 594 } 595 596 return (0); 597} 598 599/* 600 * The Mbuf Cluster zone destructor. 601 */ 602static void 603mb_dtor_clust(void *mem, int size, void *arg) 604{ 605#ifdef INVARIANTS 606 uma_zone_t zone; 607 608 zone = m_getzone(size); 609 KASSERT(*(uma_find_refcnt(zone, mem)) <= 1, 610 ("%s: refcnt incorrect %u", __func__, 611 *(uma_find_refcnt(zone, mem))) ); 612 613 trash_dtor(mem, size, arg); 614#endif 615} 616 617/* 618 * The Packet secondary zone's init routine, executed on the 619 * object's transition from mbuf keg slab to zone cache. 620 */ 621static int 622mb_zinit_pack(void *mem, int size, int how) 623{ 624 struct mbuf *m; 625 626 m = (struct mbuf *)mem; /* m is virgin. */ 627 if (uma_zalloc_arg(zone_clust, m, how) == NULL || 628 m->m_ext.ext_buf == NULL) 629 return (ENOMEM); 630 m->m_ext.ext_type = EXT_PACKET; /* Override. */ 631#ifdef INVARIANTS 632 trash_init(m->m_ext.ext_buf, MCLBYTES, how); 633#endif 634 return (0); 635} 636 637/* 638 * The Packet secondary zone's fini routine, executed on the 639 * object's transition from zone cache to keg slab. 640 */ 641static void 642mb_zfini_pack(void *mem, int size) 643{ 644 struct mbuf *m; 645 646 m = (struct mbuf *)mem; 647#ifdef INVARIANTS 648 trash_fini(m->m_ext.ext_buf, MCLBYTES); 649#endif 650 uma_zfree_arg(zone_clust, m->m_ext.ext_buf, NULL); 651#ifdef INVARIANTS 652 trash_dtor(mem, size, NULL); 653#endif 654} 655 656/* 657 * The "packet" keg constructor. 658 */ 659static int 660mb_ctor_pack(void *mem, int size, void *arg, int how) 661{ 662 struct mbuf *m; 663 struct mb_args *args; 664#ifdef MAC 665 int error; 666#endif 667 int flags; 668 short type; 669 670 m = (struct mbuf *)mem; 671 args = (struct mb_args *)arg; 672 flags = args->flags; 673 type = args->type; 674 675#ifdef INVARIANTS 676 trash_ctor(m->m_ext.ext_buf, MCLBYTES, arg, how); 677#endif 678 m->m_next = NULL; 679 m->m_nextpkt = NULL; 680 m->m_data = m->m_ext.ext_buf; 681 m->m_len = 0; 682 m->m_flags = (flags | M_EXT); 683 m->m_type = type; 684 685 if (flags & M_PKTHDR) { 686 m->m_pkthdr.rcvif = NULL; 687 m->m_pkthdr.len = 0; 688 m->m_pkthdr.header = NULL; 689 m->m_pkthdr.csum_flags = 0; 690 m->m_pkthdr.csum_data = 0; 691 m->m_pkthdr.tso_segsz = 0; 692 m->m_pkthdr.ether_vtag = 0; 693 m->m_pkthdr.flowid = 0; 694 SLIST_INIT(&m->m_pkthdr.tags); 695#ifdef MAC 696 /* If the label init fails, fail the alloc */ 697 error = mac_mbuf_init(m, how); 698 if (error) 699 return (error); 700#endif 701 } 702 /* m_ext is already initialized. */ 703 704 return (0); 705} 706 707int 708m_pkthdr_init(struct mbuf *m, int how) 709{ 710#ifdef MAC 711 int error; 712#endif 713 m->m_data = m->m_pktdat; 714 SLIST_INIT(&m->m_pkthdr.tags); 715 m->m_pkthdr.rcvif = NULL; 716 m->m_pkthdr.header = NULL; 717 m->m_pkthdr.len = 0; 718 m->m_pkthdr.flowid = 0; 719 m->m_pkthdr.csum_flags = 0; 720 m->m_pkthdr.csum_data = 0; 721 m->m_pkthdr.tso_segsz = 0; 722 m->m_pkthdr.ether_vtag = 0; 723#ifdef MAC 724 /* If the label init fails, fail the alloc */ 725 error = mac_mbuf_init(m, how); 726 if (error) 727 return (error); 728#endif 729 730 return (0); 731} 732 733/* 734 * This is the protocol drain routine. 735 * 736 * No locks should be held when this is called. The drain routines have to 737 * presently acquire some locks which raises the possibility of lock order 738 * reversal. 739 */ 740static void 741mb_reclaim(void *junk) 742{ 743 struct domain *dp; 744 struct protosw *pr; 745 746 WITNESS_WARN(WARN_GIANTOK | WARN_SLEEPOK | WARN_PANIC, NULL, 747 "mb_reclaim()"); 748 749 for (dp = domains; dp != NULL; dp = dp->dom_next) 750 for (pr = dp->dom_protosw; pr < dp->dom_protoswNPROTOSW; pr++) 751 if (pr->pr_drain != NULL) 752 (*pr->pr_drain)(); 753}
|