kern_conf.c revision 171188
1/*- 2 * Copyright (c) 1999-2002 Poul-Henning Kamp 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 */ 26 27#include <sys/cdefs.h> 28__FBSDID("$FreeBSD: head/sys/kern/kern_conf.c 171188 2007-07-03 18:18:30Z kib $"); 29 30#include <sys/param.h> 31#include <sys/kernel.h> 32#include <sys/systm.h> 33#include <sys/bio.h> 34#include <sys/lock.h> 35#include <sys/mutex.h> 36#include <sys/module.h> 37#include <sys/malloc.h> 38#include <sys/conf.h> 39#include <sys/vnode.h> 40#include <sys/queue.h> 41#include <sys/poll.h> 42#include <sys/sx.h> 43#include <sys/ctype.h> 44#include <sys/tty.h> 45#include <sys/ucred.h> 46#include <sys/taskqueue.h> 47#include <machine/stdarg.h> 48 49#include <fs/devfs/devfs_int.h> 50 51static MALLOC_DEFINE(M_DEVT, "cdev", "cdev storage"); 52 53struct mtx devmtx; 54static void destroy_devl(struct cdev *dev); 55static int destroy_dev_sched_cbl(struct cdev *dev, 56 void (*cb)(void *), void *arg); 57static struct cdev *make_dev_credv(int flags, 58 struct cdevsw *devsw, int minornr, 59 struct ucred *cr, uid_t uid, gid_t gid, int mode, const char *fmt, 60 va_list ap); 61 62static struct cdev_priv_list cdevp_free_list = 63 TAILQ_HEAD_INITIALIZER(cdevp_free_list); 64 65void 66dev_lock(void) 67{ 68 69 mtx_lock(&devmtx); 70} 71 72static void 73dev_unlock_and_free(void) 74{ 75 struct cdev_priv *cdp; 76 77 mtx_assert(&devmtx, MA_OWNED); 78 while ((cdp = TAILQ_FIRST(&cdevp_free_list)) != NULL) { 79 TAILQ_REMOVE(&cdevp_free_list, cdp, cdp_list); 80 mtx_unlock(&devmtx); 81 devfs_free(&cdp->cdp_c); 82 mtx_lock(&devmtx); 83 } 84 mtx_unlock(&devmtx); 85} 86 87static void 88dev_free_devlocked(struct cdev *cdev) 89{ 90 struct cdev_priv *cdp; 91 92 mtx_assert(&devmtx, MA_OWNED); 93 cdp = cdev->si_priv; 94 TAILQ_INSERT_HEAD(&cdevp_free_list, cdp, cdp_list); 95} 96 97void 98dev_unlock(void) 99{ 100 101 mtx_unlock(&devmtx); 102} 103 104void 105dev_ref(struct cdev *dev) 106{ 107 108 mtx_assert(&devmtx, MA_NOTOWNED); 109 mtx_lock(&devmtx); 110 dev->si_refcount++; 111 mtx_unlock(&devmtx); 112} 113 114void 115dev_refl(struct cdev *dev) 116{ 117 118 mtx_assert(&devmtx, MA_OWNED); 119 dev->si_refcount++; 120} 121 122void 123dev_rel(struct cdev *dev) 124{ 125 int flag = 0; 126 127 mtx_assert(&devmtx, MA_NOTOWNED); 128 dev_lock(); 129 dev->si_refcount--; 130 KASSERT(dev->si_refcount >= 0, 131 ("dev_rel(%s) gave negative count", devtoname(dev))); 132#if 0 133 if (dev->si_usecount == 0 && 134 (dev->si_flags & SI_CHEAPCLONE) && (dev->si_flags & SI_NAMED)) 135 ; 136 else 137#endif 138 if (dev->si_devsw == NULL && dev->si_refcount == 0) { 139 LIST_REMOVE(dev, si_list); 140 flag = 1; 141 } 142 dev_unlock(); 143 if (flag) 144 devfs_free(dev); 145} 146 147struct cdevsw * 148dev_refthread(struct cdev *dev) 149{ 150 struct cdevsw *csw; 151 struct cdev_priv *cdp; 152 153 mtx_assert(&devmtx, MA_NOTOWNED); 154 dev_lock(); 155 csw = dev->si_devsw; 156 if (csw != NULL) { 157 cdp = dev->si_priv; 158 if ((cdp->cdp_flags & CDP_SCHED_DTR) == 0) 159 dev->si_threadcount++; 160 else 161 csw = NULL; 162 } 163 dev_unlock(); 164 return (csw); 165} 166 167struct cdevsw * 168devvn_refthread(struct vnode *vp, struct cdev **devp) 169{ 170 struct cdevsw *csw; 171 struct cdev_priv *cdp; 172 173 mtx_assert(&devmtx, MA_NOTOWNED); 174 csw = NULL; 175 dev_lock(); 176 *devp = vp->v_rdev; 177 if (*devp != NULL) { 178 cdp = (*devp)->si_priv; 179 if ((cdp->cdp_flags & CDP_SCHED_DTR) == 0) { 180 csw = (*devp)->si_devsw; 181 if (csw != NULL) 182 (*devp)->si_threadcount++; 183 } 184 } 185 dev_unlock(); 186 return (csw); 187} 188 189void 190dev_relthread(struct cdev *dev) 191{ 192 193 mtx_assert(&devmtx, MA_NOTOWNED); 194 dev_lock(); 195 dev->si_threadcount--; 196 dev_unlock(); 197} 198 199int 200nullop(void) 201{ 202 203 return (0); 204} 205 206int 207eopnotsupp(void) 208{ 209 210 return (EOPNOTSUPP); 211} 212 213static int 214enxio(void) 215{ 216 return (ENXIO); 217} 218 219static int 220enodev(void) 221{ 222 return (ENODEV); 223} 224 225/* Define a dead_cdevsw for use when devices leave unexpectedly. */ 226 227#define dead_open (d_open_t *)enxio 228#define dead_close (d_close_t *)enxio 229#define dead_read (d_read_t *)enxio 230#define dead_write (d_write_t *)enxio 231#define dead_ioctl (d_ioctl_t *)enxio 232#define dead_poll (d_poll_t *)enodev 233#define dead_mmap (d_mmap_t *)enodev 234 235static void 236dead_strategy(struct bio *bp) 237{ 238 239 biofinish(bp, NULL, ENXIO); 240} 241 242#define dead_dump (dumper_t *)enxio 243#define dead_kqfilter (d_kqfilter_t *)enxio 244 245static struct cdevsw dead_cdevsw = { 246 .d_version = D_VERSION, 247 .d_flags = D_NEEDGIANT, /* XXX: does dead_strategy need this ? */ 248 .d_open = dead_open, 249 .d_close = dead_close, 250 .d_read = dead_read, 251 .d_write = dead_write, 252 .d_ioctl = dead_ioctl, 253 .d_poll = dead_poll, 254 .d_mmap = dead_mmap, 255 .d_strategy = dead_strategy, 256 .d_name = "dead", 257 .d_dump = dead_dump, 258 .d_kqfilter = dead_kqfilter 259}; 260 261/* Default methods if driver does not specify method */ 262 263#define null_open (d_open_t *)nullop 264#define null_close (d_close_t *)nullop 265#define no_read (d_read_t *)enodev 266#define no_write (d_write_t *)enodev 267#define no_ioctl (d_ioctl_t *)enodev 268#define no_mmap (d_mmap_t *)enodev 269#define no_kqfilter (d_kqfilter_t *)enodev 270 271static void 272no_strategy(struct bio *bp) 273{ 274 275 biofinish(bp, NULL, ENODEV); 276} 277 278static int 279no_poll(struct cdev *dev __unused, int events, struct thread *td __unused) 280{ 281 /* 282 * Return true for read/write. If the user asked for something 283 * special, return POLLNVAL, so that clients have a way of 284 * determining reliably whether or not the extended 285 * functionality is present without hard-coding knowledge 286 * of specific filesystem implementations. 287 * Stay in sync with vop_nopoll(). 288 */ 289 if (events & ~POLLSTANDARD) 290 return (POLLNVAL); 291 292 return (events & (POLLIN | POLLOUT | POLLRDNORM | POLLWRNORM)); 293} 294 295#define no_dump (dumper_t *)enodev 296 297static int 298giant_open(struct cdev *dev, int oflags, int devtype, struct thread *td) 299{ 300 int retval; 301 302 mtx_lock(&Giant); 303 retval = dev->si_devsw->d_gianttrick-> 304 d_open(dev, oflags, devtype, td); 305 mtx_unlock(&Giant); 306 return (retval); 307} 308 309static int 310giant_fdopen(struct cdev *dev, int oflags, struct thread *td, struct file *fp) 311{ 312 int retval; 313 314 mtx_lock(&Giant); 315 retval = dev->si_devsw->d_gianttrick-> 316 d_fdopen(dev, oflags, td, fp); 317 mtx_unlock(&Giant); 318 return (retval); 319} 320 321static int 322giant_close(struct cdev *dev, int fflag, int devtype, struct thread *td) 323{ 324 int retval; 325 326 mtx_lock(&Giant); 327 retval = dev->si_devsw->d_gianttrick-> 328 d_close(dev, fflag, devtype, td); 329 mtx_unlock(&Giant); 330 return (retval); 331} 332 333static void 334giant_strategy(struct bio *bp) 335{ 336 337 mtx_lock(&Giant); 338 bp->bio_dev->si_devsw->d_gianttrick-> 339 d_strategy(bp); 340 mtx_unlock(&Giant); 341} 342 343static int 344giant_ioctl(struct cdev *dev, u_long cmd, caddr_t data, int fflag, struct thread *td) 345{ 346 int retval; 347 348 mtx_lock(&Giant); 349 retval = dev->si_devsw->d_gianttrick-> 350 d_ioctl(dev, cmd, data, fflag, td); 351 mtx_unlock(&Giant); 352 return (retval); 353} 354 355static int 356giant_read(struct cdev *dev, struct uio *uio, int ioflag) 357{ 358 int retval; 359 360 mtx_lock(&Giant); 361 retval = dev->si_devsw->d_gianttrick-> 362 d_read(dev, uio, ioflag); 363 mtx_unlock(&Giant); 364 return (retval); 365} 366 367static int 368giant_write(struct cdev *dev, struct uio *uio, int ioflag) 369{ 370 int retval; 371 372 mtx_lock(&Giant); 373 retval = dev->si_devsw->d_gianttrick-> 374 d_write(dev, uio, ioflag); 375 mtx_unlock(&Giant); 376 return (retval); 377} 378 379static int 380giant_poll(struct cdev *dev, int events, struct thread *td) 381{ 382 int retval; 383 384 mtx_lock(&Giant); 385 retval = dev->si_devsw->d_gianttrick-> 386 d_poll(dev, events, td); 387 mtx_unlock(&Giant); 388 return (retval); 389} 390 391static int 392giant_kqfilter(struct cdev *dev, struct knote *kn) 393{ 394 int retval; 395 396 mtx_lock(&Giant); 397 retval = dev->si_devsw->d_gianttrick-> 398 d_kqfilter(dev, kn); 399 mtx_unlock(&Giant); 400 return (retval); 401} 402 403static int 404giant_mmap(struct cdev *dev, vm_offset_t offset, vm_paddr_t *paddr, int nprot) 405{ 406 int retval; 407 408 mtx_lock(&Giant); 409 retval = dev->si_devsw->d_gianttrick-> 410 d_mmap(dev, offset, paddr, nprot); 411 mtx_unlock(&Giant); 412 return (retval); 413} 414 415 416/* 417 * struct cdev * and u_dev_t primitives 418 */ 419 420int 421minor(struct cdev *x) 422{ 423 if (x == NULL) 424 return NODEV; 425 return(x->si_drv0 & MAXMINOR); 426} 427 428int 429dev2unit(struct cdev *x) 430{ 431 432 if (x == NULL) 433 return NODEV; 434 return (minor2unit(minor(x))); 435} 436 437u_int 438minor2unit(u_int _minor) 439{ 440 441 KASSERT((_minor & ~MAXMINOR) == 0, ("Illegal minor %x", _minor)); 442 return ((_minor & 0xff) | ((_minor >> 8) & 0xffff00)); 443} 444 445int 446unit2minor(int unit) 447{ 448 449 KASSERT(unit <= 0xffffff, ("Invalid unit (%d) in unit2minor", unit)); 450 return ((unit & 0xff) | ((unit << 8) & ~0xffff)); 451} 452 453static struct cdev * 454newdev(struct cdevsw *csw, int y, struct cdev *si) 455{ 456 struct cdev *si2; 457 dev_t udev; 458 459 mtx_assert(&devmtx, MA_OWNED); 460 udev = y; 461 LIST_FOREACH(si2, &csw->d_devs, si_list) { 462 if (si2->si_drv0 == udev) { 463 dev_free_devlocked(si); 464 return (si2); 465 } 466 } 467 si->si_drv0 = udev; 468 si->si_devsw = csw; 469 LIST_INSERT_HEAD(&csw->d_devs, si, si_list); 470 return (si); 471} 472 473int 474uminor(dev_t dev) 475{ 476 return (dev & MAXMINOR); 477} 478 479int 480umajor(dev_t dev) 481{ 482 return ((dev & ~MAXMINOR) >> 8); 483} 484 485static void 486fini_cdevsw(struct cdevsw *devsw) 487{ 488 struct cdevsw *gt; 489 490 if (devsw->d_gianttrick != NULL) { 491 gt = devsw->d_gianttrick; 492 memcpy(devsw, gt, sizeof *devsw); 493 free(gt, M_DEVT); 494 devsw->d_gianttrick = NULL; 495 } 496 devsw->d_flags &= ~D_INIT; 497} 498 499static void 500prep_cdevsw(struct cdevsw *devsw) 501{ 502 struct cdevsw *dsw2; 503 504 if (devsw->d_flags & D_NEEDGIANT) 505 dsw2 = malloc(sizeof *dsw2, M_DEVT, M_WAITOK); 506 else 507 dsw2 = NULL; 508 dev_lock(); 509 510 if (devsw->d_version != D_VERSION_01) { 511 printf( 512 "WARNING: Device driver \"%s\" has wrong version %s\n", 513 devsw->d_name == NULL ? "???" : devsw->d_name, 514 "and is disabled. Recompile KLD module."); 515 devsw->d_open = dead_open; 516 devsw->d_close = dead_close; 517 devsw->d_read = dead_read; 518 devsw->d_write = dead_write; 519 devsw->d_ioctl = dead_ioctl; 520 devsw->d_poll = dead_poll; 521 devsw->d_mmap = dead_mmap; 522 devsw->d_strategy = dead_strategy; 523 devsw->d_dump = dead_dump; 524 devsw->d_kqfilter = dead_kqfilter; 525 } 526 527 if (devsw->d_flags & D_TTY) { 528 if (devsw->d_ioctl == NULL) devsw->d_ioctl = ttyioctl; 529 if (devsw->d_read == NULL) devsw->d_read = ttyread; 530 if (devsw->d_write == NULL) devsw->d_write = ttywrite; 531 if (devsw->d_kqfilter == NULL) devsw->d_kqfilter = ttykqfilter; 532 if (devsw->d_poll == NULL) devsw->d_poll = ttypoll; 533 } 534 535 if (devsw->d_flags & D_NEEDGIANT) { 536 if (devsw->d_gianttrick == NULL) { 537 memcpy(dsw2, devsw, sizeof *dsw2); 538 devsw->d_gianttrick = dsw2; 539 } else 540 free(dsw2, M_DEVT); 541 } 542 543#define FIXUP(member, noop, giant) \ 544 do { \ 545 if (devsw->member == NULL) { \ 546 devsw->member = noop; \ 547 } else if (devsw->d_flags & D_NEEDGIANT) \ 548 devsw->member = giant; \ 549 } \ 550 while (0) 551 552 FIXUP(d_open, null_open, giant_open); 553 FIXUP(d_fdopen, NULL, giant_fdopen); 554 FIXUP(d_close, null_close, giant_close); 555 FIXUP(d_read, no_read, giant_read); 556 FIXUP(d_write, no_write, giant_write); 557 FIXUP(d_ioctl, no_ioctl, giant_ioctl); 558 FIXUP(d_poll, no_poll, giant_poll); 559 FIXUP(d_mmap, no_mmap, giant_mmap); 560 FIXUP(d_strategy, no_strategy, giant_strategy); 561 FIXUP(d_kqfilter, no_kqfilter, giant_kqfilter); 562 563 if (devsw->d_dump == NULL) devsw->d_dump = no_dump; 564 565 LIST_INIT(&devsw->d_devs); 566 567 devsw->d_flags |= D_INIT; 568 569 dev_unlock(); 570} 571 572struct cdev * 573make_dev_credv(int flags, struct cdevsw *devsw, int minornr, 574 struct ucred *cr, uid_t uid, 575 gid_t gid, int mode, const char *fmt, va_list ap) 576{ 577 struct cdev *dev; 578 int i; 579 580 KASSERT((minornr & ~MAXMINOR) == 0, 581 ("Invalid minor (0x%x) in make_dev", minornr)); 582 583 if (!(devsw->d_flags & D_INIT)) 584 prep_cdevsw(devsw); 585 dev = devfs_alloc(); 586 dev_lock(); 587 dev = newdev(devsw, minornr, dev); 588 if (flags & MAKEDEV_REF) 589 dev_refl(dev); 590 if (dev->si_flags & SI_CHEAPCLONE && 591 dev->si_flags & SI_NAMED) { 592 /* 593 * This is allowed as it removes races and generally 594 * simplifies cloning devices. 595 * XXX: still ?? 596 */ 597 dev_unlock_and_free(); 598 return (dev); 599 } 600 KASSERT(!(dev->si_flags & SI_NAMED), 601 ("make_dev() by driver %s on pre-existing device (min=%x, name=%s)", 602 devsw->d_name, minor(dev), devtoname(dev))); 603 604 i = vsnrprintf(dev->__si_namebuf, sizeof dev->__si_namebuf, 32, fmt, ap); 605 if (i > (sizeof dev->__si_namebuf - 1)) { 606 printf("WARNING: Device name truncated! (%s)\n", 607 dev->__si_namebuf); 608 } 609 610 dev->si_flags |= SI_NAMED; 611 if (cr != NULL) 612 dev->si_cred = crhold(cr); 613 else 614 dev->si_cred = NULL; 615 dev->si_uid = uid; 616 dev->si_gid = gid; 617 dev->si_mode = mode; 618 619 devfs_create(dev); 620 dev_unlock(); 621 return (dev); 622} 623 624struct cdev * 625make_dev(struct cdevsw *devsw, int minornr, uid_t uid, gid_t gid, int mode, 626 const char *fmt, ...) 627{ 628 struct cdev *dev; 629 va_list ap; 630 631 va_start(ap, fmt); 632 dev = make_dev_credv(0, devsw, minornr, NULL, uid, gid, mode, fmt, ap); 633 va_end(ap); 634 return (dev); 635} 636 637struct cdev * 638make_dev_cred(struct cdevsw *devsw, int minornr, struct ucred *cr, uid_t uid, 639 gid_t gid, int mode, const char *fmt, ...) 640{ 641 struct cdev *dev; 642 va_list ap; 643 644 va_start(ap, fmt); 645 dev = make_dev_credv(0, devsw, minornr, cr, uid, gid, mode, fmt, ap); 646 va_end(ap); 647 648 return (dev); 649} 650 651struct cdev * 652make_dev_credf(int flags, struct cdevsw *devsw, int minornr, 653 struct ucred *cr, uid_t uid, 654 gid_t gid, int mode, const char *fmt, ...) 655{ 656 struct cdev *dev; 657 va_list ap; 658 659 va_start(ap, fmt); 660 dev = make_dev_credv(flags, devsw, minornr, cr, uid, gid, mode, 661 fmt, ap); 662 va_end(ap); 663 664 return (dev); 665} 666 667static void 668dev_dependsl(struct cdev *pdev, struct cdev *cdev) 669{ 670 671 cdev->si_parent = pdev; 672 cdev->si_flags |= SI_CHILD; 673 LIST_INSERT_HEAD(&pdev->si_children, cdev, si_siblings); 674} 675 676 677void 678dev_depends(struct cdev *pdev, struct cdev *cdev) 679{ 680 681 dev_lock(); 682 dev_dependsl(pdev, cdev); 683 dev_unlock(); 684} 685 686struct cdev * 687make_dev_alias(struct cdev *pdev, const char *fmt, ...) 688{ 689 struct cdev *dev; 690 va_list ap; 691 int i; 692 693 dev = devfs_alloc(); 694 dev_lock(); 695 dev->si_flags |= SI_ALIAS; 696 dev->si_flags |= SI_NAMED; 697 va_start(ap, fmt); 698 i = vsnrprintf(dev->__si_namebuf, sizeof dev->__si_namebuf, 32, fmt, ap); 699 if (i > (sizeof dev->__si_namebuf - 1)) { 700 printf("WARNING: Device name truncated! (%s)\n", 701 dev->__si_namebuf); 702 } 703 va_end(ap); 704 705 devfs_create(dev); 706 dev_unlock(); 707 dev_depends(pdev, dev); 708 return (dev); 709} 710 711static void 712destroy_devl(struct cdev *dev) 713{ 714 struct cdevsw *csw; 715 716 mtx_assert(&devmtx, MA_OWNED); 717 KASSERT(dev->si_flags & SI_NAMED, 718 ("WARNING: Driver mistake: destroy_dev on %d\n", minor(dev))); 719 720 devfs_destroy(dev); 721 722 /* Remove name marking */ 723 dev->si_flags &= ~SI_NAMED; 724 725 /* If we are a child, remove us from the parents list */ 726 if (dev->si_flags & SI_CHILD) { 727 LIST_REMOVE(dev, si_siblings); 728 dev->si_flags &= ~SI_CHILD; 729 } 730 731 /* Kill our children */ 732 while (!LIST_EMPTY(&dev->si_children)) 733 destroy_devl(LIST_FIRST(&dev->si_children)); 734 735 /* Remove from clone list */ 736 if (dev->si_flags & SI_CLONELIST) { 737 LIST_REMOVE(dev, si_clone); 738 dev->si_flags &= ~SI_CLONELIST; 739 } 740 741 dev->si_refcount++; /* Avoid race with dev_rel() */ 742 csw = dev->si_devsw; 743 dev->si_devsw = NULL; /* already NULL for SI_ALIAS */ 744 while (csw != NULL && csw->d_purge != NULL && dev->si_threadcount) { 745 csw->d_purge(dev); 746 msleep(csw, &devmtx, PRIBIO, "devprg", hz/10); 747 if (dev->si_threadcount) 748 printf("Still %lu threads in %s\n", 749 dev->si_threadcount, devtoname(dev)); 750 } 751 while (dev->si_threadcount != 0) { 752 /* Use unique dummy wait ident */ 753 msleep(&csw, &devmtx, PRIBIO, "devdrn", hz / 10); 754 } 755 756 dev->si_drv1 = 0; 757 dev->si_drv2 = 0; 758 bzero(&dev->__si_u, sizeof(dev->__si_u)); 759 760 if (!(dev->si_flags & SI_ALIAS)) { 761 /* Remove from cdevsw list */ 762 LIST_REMOVE(dev, si_list); 763 764 /* If cdevsw has no more struct cdev *'s, clean it */ 765 if (LIST_EMPTY(&csw->d_devs)) { 766 fini_cdevsw(csw); 767 wakeup(&csw->d_devs); 768 } 769 } 770 dev->si_flags &= ~SI_ALIAS; 771 dev->si_refcount--; /* Avoid race with dev_rel() */ 772 773 if (dev->si_refcount > 0) { 774 LIST_INSERT_HEAD(&dead_cdevsw.d_devs, dev, si_list); 775 } else { 776 dev_free_devlocked(dev); 777 } 778} 779 780void 781destroy_dev(struct cdev *dev) 782{ 783 struct cdevsw *csw; 784 785 dev_lock(); 786 csw = dev->si_devsw; 787 if ((csw != NULL && csw->d_purge != NULL) || 788 dev->si_threadcount == 0) { 789 destroy_devl(dev); 790 dev_unlock_and_free(); 791 } else 792 destroy_dev_sched_cbl(dev, NULL, NULL); 793} 794 795const char * 796devtoname(struct cdev *dev) 797{ 798 char *p; 799 struct cdevsw *csw; 800 int mynor; 801 802 if (dev->si_name[0] == '#' || dev->si_name[0] == '\0') { 803 p = dev->si_name; 804 csw = dev_refthread(dev); 805 if (csw != NULL) { 806 sprintf(p, "(%s)", csw->d_name); 807 dev_relthread(dev); 808 } 809 p += strlen(p); 810 mynor = minor(dev); 811 if (mynor < 0 || mynor > 255) 812 sprintf(p, "/%#x", (u_int)mynor); 813 else 814 sprintf(p, "/%d", mynor); 815 } 816 return (dev->si_name); 817} 818 819int 820dev_stdclone(char *name, char **namep, const char *stem, int *unit) 821{ 822 int u, i; 823 824 i = strlen(stem); 825 if (bcmp(stem, name, i) != 0) 826 return (0); 827 if (!isdigit(name[i])) 828 return (0); 829 u = 0; 830 if (name[i] == '0' && isdigit(name[i+1])) 831 return (0); 832 while (isdigit(name[i])) { 833 u *= 10; 834 u += name[i++] - '0'; 835 } 836 if (u > 0xffffff) 837 return (0); 838 *unit = u; 839 if (namep) 840 *namep = &name[i]; 841 if (name[i]) 842 return (2); 843 return (1); 844} 845 846/* 847 * Helper functions for cloning device drivers. 848 * 849 * The objective here is to make it unnecessary for the device drivers to 850 * use rman or similar to manage their unit number space. Due to the way 851 * we do "on-demand" devices, using rman or other "private" methods 852 * will be very tricky to lock down properly once we lock down this file. 853 * 854 * Instead we give the drivers these routines which puts the struct cdev *'s 855 * that are to be managed on their own list, and gives the driver the ability 856 * to ask for the first free unit number or a given specified unit number. 857 * 858 * In addition these routines support paired devices (pty, nmdm and similar) 859 * by respecting a number of "flag" bits in the minor number. 860 * 861 */ 862 863struct clonedevs { 864 LIST_HEAD(,cdev) head; 865}; 866 867void 868clone_setup(struct clonedevs **cdp) 869{ 870 871 *cdp = malloc(sizeof **cdp, M_DEVBUF, M_WAITOK | M_ZERO); 872 LIST_INIT(&(*cdp)->head); 873} 874 875int 876clone_create(struct clonedevs **cdp, struct cdevsw *csw, int *up, struct cdev **dp, int extra) 877{ 878 struct clonedevs *cd; 879 struct cdev *dev, *ndev, *dl, *de; 880 int unit, low, u; 881 882 KASSERT(*cdp != NULL, 883 ("clone_setup() not called in driver \"%s\"", csw->d_name)); 884 KASSERT(!(extra & CLONE_UNITMASK), 885 ("Illegal extra bits (0x%x) in clone_create", extra)); 886 KASSERT(*up <= CLONE_UNITMASK, 887 ("Too high unit (0x%x) in clone_create", *up)); 888 889 if (!(csw->d_flags & D_INIT)) 890 prep_cdevsw(csw); 891 892 /* 893 * Search the list for a lot of things in one go: 894 * A preexisting match is returned immediately. 895 * The lowest free unit number if we are passed -1, and the place 896 * in the list where we should insert that new element. 897 * The place to insert a specified unit number, if applicable 898 * the end of the list. 899 */ 900 unit = *up; 901 ndev = devfs_alloc(); 902 dev_lock(); 903 low = extra; 904 de = dl = NULL; 905 cd = *cdp; 906 LIST_FOREACH(dev, &cd->head, si_clone) { 907 KASSERT(dev->si_flags & SI_CLONELIST, 908 ("Dev %p(%s) should be on clonelist", dev, dev->si_name)); 909 u = dev2unit(dev); 910 if (u == (unit | extra)) { 911 *dp = dev; 912 dev_unlock(); 913 devfs_free(ndev); 914 return (0); 915 } 916 if (unit == -1 && u == low) { 917 low++; 918 de = dev; 919 continue; 920 } else if (u < (unit | extra)) { 921 de = dev; 922 continue; 923 } else if (u > (unit | extra)) { 924 dl = dev; 925 break; 926 } 927 } 928 if (unit == -1) 929 unit = low & CLONE_UNITMASK; 930 dev = newdev(csw, unit2minor(unit | extra), ndev); 931 if (dev->si_flags & SI_CLONELIST) { 932 printf("dev %p (%s) is on clonelist\n", dev, dev->si_name); 933 printf("unit=%d, low=%d, extra=0x%x\n", unit, low, extra); 934 LIST_FOREACH(dev, &cd->head, si_clone) { 935 printf("\t%p %s\n", dev, dev->si_name); 936 } 937 panic("foo"); 938 } 939 KASSERT(!(dev->si_flags & SI_CLONELIST), 940 ("Dev %p(%s) should not be on clonelist", dev, dev->si_name)); 941 if (dl != NULL) 942 LIST_INSERT_BEFORE(dl, dev, si_clone); 943 else if (de != NULL) 944 LIST_INSERT_AFTER(de, dev, si_clone); 945 else 946 LIST_INSERT_HEAD(&cd->head, dev, si_clone); 947 dev->si_flags |= SI_CLONELIST; 948 *up = unit; 949 dev_unlock_and_free(); 950 return (1); 951} 952 953/* 954 * Kill everything still on the list. The driver should already have 955 * disposed of any softc hung of the struct cdev *'s at this time. 956 */ 957void 958clone_cleanup(struct clonedevs **cdp) 959{ 960 struct cdev *dev; 961 struct cdev_priv *cp; 962 struct clonedevs *cd; 963 964 cd = *cdp; 965 if (cd == NULL) 966 return; 967 dev_lock(); 968 while (!LIST_EMPTY(&cd->head)) { 969 dev = LIST_FIRST(&cd->head); 970 LIST_REMOVE(dev, si_clone); 971 KASSERT(dev->si_flags & SI_CLONELIST, 972 ("Dev %p(%s) should be on clonelist", dev, dev->si_name)); 973 dev->si_flags &= ~SI_CLONELIST; 974 cp = dev->si_priv; 975 if (!(cp->cdp_flags & CDP_SCHED_DTR)) { 976 cp->cdp_flags |= CDP_SCHED_DTR; 977 KASSERT(dev->si_flags & SI_NAMED, 978 ("Driver has goofed in cloning underways udev %x", dev->si_drv0)); 979 destroy_devl(dev); 980 } 981 } 982 dev_unlock(); 983 free(cd, M_DEVBUF); 984 *cdp = NULL; 985} 986 987static TAILQ_HEAD(, cdev_priv) dev_ddtr = 988 TAILQ_HEAD_INITIALIZER(dev_ddtr); 989static struct task dev_dtr_task; 990 991static void 992destroy_dev_tq(void *ctx, int pending) 993{ 994 struct cdev_priv *cp; 995 struct cdev *dev; 996 void (*cb)(void *); 997 void *cb_arg; 998 999 dev_lock(); 1000 while (!TAILQ_EMPTY(&dev_ddtr)) { 1001 cp = TAILQ_FIRST(&dev_ddtr); 1002 dev = &cp->cdp_c; 1003 KASSERT(cp->cdp_flags & CDP_SCHED_DTR, 1004 ("cdev %p in dev_destroy_tq without CDP_SCHED_DTR", cp)); 1005 TAILQ_REMOVE(&dev_ddtr, cp, cdp_dtr_list); 1006 cb = cp->cdp_dtr_cb; 1007 cb_arg = cp->cdp_dtr_cb_arg; 1008 destroy_devl(dev); 1009 dev_unlock(); 1010 dev_rel(dev); 1011 if (cb != NULL) 1012 cb(cb_arg); 1013 dev_lock(); 1014 } 1015 dev_unlock(); 1016} 1017 1018/* 1019 * devmtx shall be locked on entry. devmtx will be unlocked after 1020 * function return. 1021 */ 1022static int 1023destroy_dev_sched_cbl(struct cdev *dev, void (*cb)(void *), void *arg) 1024{ 1025 struct cdev_priv *cp; 1026 1027 mtx_assert(&devmtx, MA_OWNED); 1028 cp = dev->si_priv; 1029 if (cp->cdp_flags & CDP_SCHED_DTR) { 1030 dev_unlock(); 1031 return (0); 1032 } 1033 dev_refl(dev); 1034 cp->cdp_flags |= CDP_SCHED_DTR; 1035 cp->cdp_dtr_cb = cb; 1036 cp->cdp_dtr_cb_arg = arg; 1037 TAILQ_INSERT_TAIL(&dev_ddtr, cp, cdp_dtr_list); 1038 dev_unlock(); 1039 taskqueue_enqueue(taskqueue_swi_giant, &dev_dtr_task); 1040 return (1); 1041} 1042 1043int 1044destroy_dev_sched_cb(struct cdev *dev, void (*cb)(void *), void *arg) 1045{ 1046 dev_lock(); 1047 return (destroy_dev_sched_cbl(dev, cb, arg)); 1048} 1049 1050int 1051destroy_dev_sched(struct cdev *dev) 1052{ 1053 return (destroy_dev_sched_cb(dev, NULL, NULL)); 1054} 1055 1056void 1057destroy_dev_drain(struct cdevsw *csw) 1058{ 1059 1060 dev_lock(); 1061 while (!LIST_EMPTY(&csw->d_devs)) { 1062 msleep(&csw->d_devs, &devmtx, PRIBIO, "devscd", hz/10); 1063 } 1064 dev_unlock(); 1065} 1066 1067void 1068drain_dev_clone_events(void) 1069{ 1070 1071 sx_xlock(&clone_drain_lock); 1072 sx_xunlock(&clone_drain_lock); 1073} 1074 1075static void 1076devdtr_init(void *dummy __unused) 1077{ 1078 1079 TASK_INIT(&dev_dtr_task, 0, destroy_dev_tq, NULL); 1080} 1081 1082SYSINIT(devdtr, SI_SUB_DEVFS, SI_ORDER_SECOND, devdtr_init, NULL); 1083