Deleted Added
full compact
1/*
2 * Copyright (c) 1982, 1986, 1989, 1991, 1993
3 * The Regents of the University of California. All rights reserved.
4 * (c) UNIX System Laboratories, Inc.
5 * All or some portions of this file are derived from material licensed
6 * to the University of California by American Telephone and Telegraph
7 * Co. or Unix System Laboratories, Inc. and are reproduced herein with
8 * the permission of UNIX System Laboratories, Inc.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 * 3. All advertising materials mentioning features or use of this software
19 * must display the following acknowledgement:
20 * This product includes software developed by the University of
21 * California, Berkeley and its contributors.
22 * 4. Neither the name of the University nor the names of its contributors
23 * may be used to endorse or promote products derived from this software
24 * without specific prior written permission.
25 *
26 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
27 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
28 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
29 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
30 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
31 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
32 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
33 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
34 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
35 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
36 * SUCH DAMAGE.
37 *
38 * @(#)kern_descrip.c 8.6 (Berkeley) 4/19/94
39 * $FreeBSD: head/sys/kern/kern_descrip.c 92641 2002-03-19 04:30:04Z alfred $
39 * $FreeBSD: head/sys/kern/kern_descrip.c 92654 2002-03-19 09:11:49Z jeff $
40 */
41
42#include "opt_compat.h"
43
44#include <sys/param.h>
45#include <sys/systm.h>
46#include <sys/lock.h>
47#include <sys/malloc.h>
48#include <sys/mutex.h>
49#include <sys/sysproto.h>
50#include <sys/conf.h>
51#include <sys/filedesc.h>
52#include <sys/kernel.h>
53#include <sys/sysctl.h>
54#include <sys/vnode.h>
55#include <sys/proc.h>
56#include <sys/file.h>
57#include <sys/stat.h>
58#include <sys/filio.h>
59#include <sys/fcntl.h>
60#include <sys/unistd.h>
61#include <sys/resourcevar.h>
62#include <sys/event.h>
63#include <sys/sx.h>
64#include <sys/socketvar.h>
65
66#include <machine/limits.h>
67
68#include <vm/vm.h>
69#include <vm/vm_extern.h>
70#include <vm/vm_zone.h>
71
72static MALLOC_DEFINE(M_FILEDESC, "file desc", "Open file descriptor table");
72MALLOC_DEFINE(M_FILE, "file", "Open file structure");
73static MALLOC_DEFINE(M_SIGIO, "sigio", "sigio structures");
74
75uma_zone_t file_zone;
76
77static d_open_t fdopen;
78#define NUMFDESC 64
79
80#define CDEV_MAJOR 22
81static struct cdevsw fildesc_cdevsw = {
82 /* open */ fdopen,
83 /* close */ noclose,
84 /* read */ noread,
85 /* write */ nowrite,
86 /* ioctl */ noioctl,
87 /* poll */ nopoll,
88 /* mmap */ nommap,
89 /* strategy */ nostrategy,
90 /* name */ "FD",
91 /* maj */ CDEV_MAJOR,
92 /* dump */ nodump,
93 /* psize */ nopsize,
94 /* flags */ 0,
95};
96
97static int do_dup __P((struct filedesc *fdp, int old, int new, register_t *retval, struct thread *td));
98static int badfo_readwrite __P((struct file *fp, struct uio *uio,
99 struct ucred *cred, int flags, struct thread *td));
100static int badfo_ioctl __P((struct file *fp, u_long com, caddr_t data,
101 struct thread *td));
102static int badfo_poll __P((struct file *fp, int events,
103 struct ucred *cred, struct thread *td));
104static int badfo_kqfilter __P((struct file *fp, struct knote *kn));
105static int badfo_stat __P((struct file *fp, struct stat *sb, struct thread *td));
106static int badfo_close __P((struct file *fp, struct thread *td));
107
108/*
109 * Descriptor management.
110 */
111struct filelist filehead; /* head of list of open files */
112int nfiles; /* actual number of open files */
113extern int cmask;
114struct sx filelist_lock; /* sx to protect filelist */
115
116/*
117 * System calls on descriptors.
118 */
119#ifndef _SYS_SYSPROTO_H_
120struct getdtablesize_args {
121 int dummy;
122};
123#endif
124/*
125 * MPSAFE
126 */
127/* ARGSUSED */
128int
129getdtablesize(td, uap)
130 struct thread *td;
131 struct getdtablesize_args *uap;
132{
133 struct proc *p = td->td_proc;
134
135 mtx_lock(&Giant);
136 td->td_retval[0] =
137 min((int)p->p_rlimit[RLIMIT_NOFILE].rlim_cur, maxfilesperproc);
138 mtx_unlock(&Giant);
139 return (0);
140}
141
142/*
143 * Duplicate a file descriptor to a particular value.
144 *
145 * note: keep in mind that a potential race condition exists when closing
146 * descriptors from a shared descriptor table (via rfork).
147 */
148#ifndef _SYS_SYSPROTO_H_
149struct dup2_args {
150 u_int from;
151 u_int to;
152};
153#endif
154/*
155 * MPSAFE
156 */
157/* ARGSUSED */
158int
159dup2(td, uap)
160 struct thread *td;
161 struct dup2_args *uap;
162{
163 struct proc *p = td->td_proc;
164 register struct filedesc *fdp = td->td_proc->p_fd;
165 register u_int old = uap->from, new = uap->to;
166 int i, error;
167
168 FILEDESC_LOCK(fdp);
169retry:
170 if (old >= fdp->fd_nfiles ||
171 fdp->fd_ofiles[old] == NULL ||
172 new >= p->p_rlimit[RLIMIT_NOFILE].rlim_cur ||
173 new >= maxfilesperproc) {
174 FILEDESC_UNLOCK(fdp);
175 return (EBADF);
176 }
177 if (old == new) {
178 td->td_retval[0] = new;
179 FILEDESC_UNLOCK(fdp);
180 return (0);
181 }
182 if (new >= fdp->fd_nfiles) {
183 if ((error = fdalloc(td, new, &i))) {
184 FILEDESC_UNLOCK(fdp);
185 return (error);
186 }
187 /*
188 * fdalloc() may block, retest everything.
189 */
190 goto retry;
191 }
192 error = do_dup(fdp, (int)old, (int)new, td->td_retval, td);
193 return(error);
194}
195
196/*
197 * Duplicate a file descriptor.
198 */
199#ifndef _SYS_SYSPROTO_H_
200struct dup_args {
201 u_int fd;
202};
203#endif
204/*
205 * MPSAFE
206 */
207/* ARGSUSED */
208int
209dup(td, uap)
210 struct thread *td;
211 struct dup_args *uap;
212{
213 register struct filedesc *fdp;
214 u_int old;
215 int new, error;
216
217 old = uap->fd;
218 fdp = td->td_proc->p_fd;
219 FILEDESC_LOCK(fdp);
220 if (old >= fdp->fd_nfiles || fdp->fd_ofiles[old] == NULL) {
221 FILEDESC_UNLOCK(fdp);
222 return (EBADF);
223 }
224 if ((error = fdalloc(td, 0, &new))) {
225 FILEDESC_UNLOCK(fdp);
226 return (error);
227 }
228 error = do_dup(fdp, (int)old, new, td->td_retval, td);
229 return (error);
230}
231
232/*
233 * The file control system call.
234 */
235#ifndef _SYS_SYSPROTO_H_
236struct fcntl_args {
237 int fd;
238 int cmd;
239 long arg;
240};
241#endif
242/*
243 * MPSAFE
244 */
245/* ARGSUSED */
246int
247fcntl(td, uap)
248 struct thread *td;
249 register struct fcntl_args *uap;
250{
251 register struct proc *p = td->td_proc;
252 register struct filedesc *fdp;
253 register struct file *fp;
254 register char *pop;
255 struct vnode *vp;
256 int i, tmp, error = 0, flg = F_POSIX;
257 struct flock fl;
258 u_int newmin;
259 struct proc *leaderp;
260
261 mtx_lock(&Giant);
262
263 fdp = p->p_fd;
264 FILEDESC_LOCK(fdp);
265 if ((unsigned)uap->fd >= fdp->fd_nfiles ||
266 (fp = fdp->fd_ofiles[uap->fd]) == NULL) {
267 FILEDESC_UNLOCK(fdp);
268 error = EBADF;
269 goto done2;
270 }
271 pop = &fdp->fd_ofileflags[uap->fd];
272
273 switch (uap->cmd) {
274 case F_DUPFD:
275 newmin = uap->arg;
276 if (newmin >= p->p_rlimit[RLIMIT_NOFILE].rlim_cur ||
277 newmin >= maxfilesperproc) {
278 FILEDESC_UNLOCK(fdp);
279 error = EINVAL;
280 break;
281 }
282 if ((error = fdalloc(td, newmin, &i))) {
283 FILEDESC_UNLOCK(fdp);
284 break;
285 }
286 error = do_dup(fdp, uap->fd, i, td->td_retval, td);
287 break;
288
289 case F_GETFD:
290 td->td_retval[0] = *pop & 1;
291 FILEDESC_UNLOCK(fdp);
292 break;
293
294 case F_SETFD:
295 *pop = (*pop &~ 1) | (uap->arg & 1);
296 FILEDESC_UNLOCK(fdp);
297 break;
298
299 case F_GETFL:
300 FILE_LOCK(fp);
301 FILEDESC_UNLOCK(fdp);
302 td->td_retval[0] = OFLAGS(fp->f_flag);
303 FILE_UNLOCK(fp);
304 break;
305
306 case F_SETFL:
307 fhold(fp);
308 FILEDESC_UNLOCK(fdp);
309 fp->f_flag &= ~FCNTLFLAGS;
310 fp->f_flag |= FFLAGS(uap->arg & ~O_ACCMODE) & FCNTLFLAGS;
311 tmp = fp->f_flag & FNONBLOCK;
312 error = fo_ioctl(fp, FIONBIO, (caddr_t)&tmp, td);
313 if (error) {
314 fdrop(fp, td);
315 break;
316 }
317 tmp = fp->f_flag & FASYNC;
318 error = fo_ioctl(fp, FIOASYNC, (caddr_t)&tmp, td);
319 if (!error) {
320 fdrop(fp, td);
321 break;
322 }
323 fp->f_flag &= ~FNONBLOCK;
324 tmp = 0;
325 (void)fo_ioctl(fp, FIONBIO, (caddr_t)&tmp, td);
326 fdrop(fp, td);
327 break;
328
329 case F_GETOWN:
330 fhold(fp);
331 FILEDESC_UNLOCK(fdp);
332 error = fo_ioctl(fp, FIOGETOWN, (caddr_t)td->td_retval, td);
333 fdrop(fp, td);
334 break;
335
336 case F_SETOWN:
337 fhold(fp);
338 FILEDESC_UNLOCK(fdp);
339 error = fo_ioctl(fp, FIOSETOWN, (caddr_t)&uap->arg, td);
340 fdrop(fp, td);
341 break;
342
343 case F_SETLKW:
344 flg |= F_WAIT;
345 /* Fall into F_SETLK */
346
347 case F_SETLK:
348 if (fp->f_type != DTYPE_VNODE) {
349 FILEDESC_UNLOCK(fdp);
350 error = EBADF;
351 break;
352 }
353 vp = (struct vnode *)fp->f_data;
354 /*
355 * copyin/lockop may block
356 */
357 fhold(fp);
358 FILEDESC_UNLOCK(fdp);
359 vp = (struct vnode *)fp->f_data;
360
361 /* Copy in the lock structure */
362 error = copyin((caddr_t)(intptr_t)uap->arg, (caddr_t)&fl,
363 sizeof(fl));
364 if (error) {
365 fdrop(fp, td);
366 break;
367 }
368 if (fl.l_whence == SEEK_CUR) {
369 if (fp->f_offset < 0 ||
370 (fl.l_start > 0 &&
371 fp->f_offset > OFF_MAX - fl.l_start)) {
372 fdrop(fp, td);
373 error = EOVERFLOW;
374 break;
375 }
376 fl.l_start += fp->f_offset;
377 }
378
379 switch (fl.l_type) {
380 case F_RDLCK:
381 if ((fp->f_flag & FREAD) == 0) {
382 error = EBADF;
383 break;
384 }
385 PROC_LOCK(p);
386 p->p_flag |= P_ADVLOCK;
387 leaderp = p->p_leader;
388 PROC_UNLOCK(p);
389 error = VOP_ADVLOCK(vp, (caddr_t)leaderp, F_SETLK,
390 &fl, flg);
391 break;
392 case F_WRLCK:
393 if ((fp->f_flag & FWRITE) == 0) {
394 error = EBADF;
395 break;
396 }
397 PROC_LOCK(p);
398 p->p_flag |= P_ADVLOCK;
399 leaderp = p->p_leader;
400 PROC_UNLOCK(p);
401 error = VOP_ADVLOCK(vp, (caddr_t)leaderp, F_SETLK,
402 &fl, flg);
403 break;
404 case F_UNLCK:
405 PROC_LOCK(p);
406 leaderp = p->p_leader;
407 PROC_UNLOCK(p);
408 error = VOP_ADVLOCK(vp, (caddr_t)leaderp, F_UNLCK,
409 &fl, F_POSIX);
410 break;
411 default:
412 error = EINVAL;
413 break;
414 }
415 fdrop(fp, td);
416 break;
417
418 case F_GETLK:
419 if (fp->f_type != DTYPE_VNODE) {
420 FILEDESC_UNLOCK(fdp);
421 error = EBADF;
422 break;
423 }
424 vp = (struct vnode *)fp->f_data;
425 /*
426 * copyin/lockop may block
427 */
428 fhold(fp);
429 FILEDESC_UNLOCK(fdp);
430 vp = (struct vnode *)fp->f_data;
431
432 /* Copy in the lock structure */
433 error = copyin((caddr_t)(intptr_t)uap->arg, (caddr_t)&fl,
434 sizeof(fl));
435 if (error) {
436 fdrop(fp, td);
437 break;
438 }
439 if (fl.l_type != F_RDLCK && fl.l_type != F_WRLCK &&
440 fl.l_type != F_UNLCK) {
441 fdrop(fp, td);
442 error = EINVAL;
443 break;
444 }
445 if (fl.l_whence == SEEK_CUR) {
446 if ((fl.l_start > 0 &&
447 fp->f_offset > OFF_MAX - fl.l_start) ||
448 (fl.l_start < 0 &&
449 fp->f_offset < OFF_MIN - fl.l_start)) {
450 fdrop(fp, td);
451 error = EOVERFLOW;
452 break;
453 }
454 fl.l_start += fp->f_offset;
455 }
456 error = VOP_ADVLOCK(vp, (caddr_t)p->p_leader, F_GETLK,
457 &fl, F_POSIX);
458 fdrop(fp, td);
459 if (error == 0) {
460 error = copyout((caddr_t)&fl,
461 (caddr_t)(intptr_t)uap->arg, sizeof(fl));
462 }
463 break;
464 default:
465 FILEDESC_UNLOCK(fdp);
466 error = EINVAL;
467 break;
468 }
469done2:
470 mtx_unlock(&Giant);
471 return (error);
472}
473
474/*
475 * Common code for dup, dup2, and fcntl(F_DUPFD).
476 * filedesc must be locked, but will be unlocked as a side effect.
477 */
478static int
479do_dup(fdp, old, new, retval, td)
480 register struct filedesc *fdp;
481 register int old, new;
482 register_t *retval;
483 struct thread *td;
484{
485 struct file *fp;
486 struct file *delfp;
487
488 FILEDESC_LOCK_ASSERT(fdp, MA_OWNED);
489
490 /*
491 * Save info on the descriptor being overwritten. We have
492 * to do the unmap now, but we cannot close it without
493 * introducing an ownership race for the slot.
494 */
495 delfp = fdp->fd_ofiles[new];
496#if 0
497 if (delfp && (fdp->fd_ofileflags[new] & UF_MAPPED))
498 (void) munmapfd(td, new);
499#endif
500
501 /*
502 * Duplicate the source descriptor, update lastfile
503 */
504 fp = fdp->fd_ofiles[old];
505 fdp->fd_ofiles[new] = fp;
506 fdp->fd_ofileflags[new] = fdp->fd_ofileflags[old] &~ UF_EXCLOSE;
507 fhold(fp);
508 if (new > fdp->fd_lastfile)
509 fdp->fd_lastfile = new;
510 *retval = new;
511
512 FILEDESC_UNLOCK(fdp);
513
514 /*
515 * If we dup'd over a valid file, we now own the reference to it
516 * and must dispose of it using closef() semantics (as if a
517 * close() were performed on it).
518 */
519 if (delfp) {
520 mtx_lock(&Giant);
521 (void) closef(delfp, td);
522 mtx_unlock(&Giant);
523 }
524 return (0);
525}
526
527/*
528 * If sigio is on the list associated with a process or process group,
529 * disable signalling from the device, remove sigio from the list and
530 * free sigio.
531 */
532void
533funsetown(sigio)
534 struct sigio *sigio;
535{
536 int s;
537
538 if (sigio == NULL)
539 return;
540
541 s = splhigh();
542 *(sigio->sio_myref) = NULL;
543 splx(s);
544 if ((sigio)->sio_pgid < 0) {
545 struct pgrp *pg = (sigio)->sio_pgrp;
546 PGRP_LOCK(pg);
547 SLIST_REMOVE(&sigio->sio_pgrp->pg_sigiolst, sigio,
548 sigio, sio_pgsigio);
549 PGRP_UNLOCK(pg);
550 } else {
551 struct proc *p = (sigio)->sio_proc;
552 PROC_LOCK(p);
553 SLIST_REMOVE(&sigio->sio_proc->p_sigiolst, sigio,
554 sigio, sio_pgsigio);
555 PROC_UNLOCK(p);
556 }
557 crfree(sigio->sio_ucred);
558 FREE(sigio, M_SIGIO);
559}
560
561/* Free a list of sigio structures. */
562void
563funsetownlst(sigiolst)
564 struct sigiolst *sigiolst;
565{
566 int s;
567 struct sigio *sigio;
568 struct proc *p;
569 struct pgrp *pg;
570
571 sigio = SLIST_FIRST(sigiolst);
572 if (sigio == NULL)
573 return;
574
575 p = NULL;
576 pg = NULL;
577
578 /*
579 * Every entry of the list should belong
580 * to a single proc or pgrp.
581 */
582 if (sigio->sio_pgid < 0) {
583 pg = sigio->sio_pgrp;
584 PGRP_LOCK_ASSERT(pg, MA_OWNED);
585 } else /* if (sigio->sio_pgid > 0) */ {
586 p = sigio->sio_proc;
587 PROC_LOCK_ASSERT(p, MA_OWNED);
588 }
589
590 while ((sigio = SLIST_FIRST(sigiolst)) != NULL) {
591 s = splhigh();
592 *(sigio->sio_myref) = NULL;
593 splx(s);
594 if (pg != NULL) {
595 KASSERT(sigio->sio_pgid < 0, ("Proc sigio in pgrp sigio list"));
596 KASSERT(sigio->sio_pgrp == pg, ("Bogus pgrp in sigio list"));
597 SLIST_REMOVE(&pg->pg_sigiolst, sigio, sigio, sio_pgsigio);
598 PGRP_UNLOCK(pg);
599 crfree(sigio->sio_ucred);
600 FREE(sigio, M_SIGIO);
601 PGRP_LOCK(pg);
602 } else /* if (p != NULL) */ {
603 KASSERT(sigio->sio_pgid > 0, ("Pgrp sigio in proc sigio list"));
604 KASSERT(sigio->sio_proc == p, ("Bogus proc in sigio list"));
605 SLIST_REMOVE(&p->p_sigiolst, sigio, sigio, sio_pgsigio);
606 PROC_UNLOCK(p);
607 crfree(sigio->sio_ucred);
608 FREE(sigio, M_SIGIO);
609 PROC_LOCK(p);
610 }
611 }
612}
613
614/*
615 * This is common code for FIOSETOWN ioctl called by fcntl(fd, F_SETOWN, arg).
616 *
617 * After permission checking, add a sigio structure to the sigio list for
618 * the process or process group.
619 */
620int
621fsetown(pgid, sigiop)
622 pid_t pgid;
623 struct sigio **sigiop;
624{
625 struct proc *proc;
626 struct pgrp *pgrp;
627 struct sigio *sigio;
628 int s, ret;
629
630 if (pgid == 0) {
631 funsetown(*sigiop);
632 return (0);
633 }
634
635 ret = 0;
636
637 /* Allocate and fill in the new sigio out of locks. */
638 MALLOC(sigio, struct sigio *, sizeof(struct sigio), M_SIGIO, M_WAITOK);
639 sigio->sio_pgid = pgid;
640 sigio->sio_ucred = crhold(curthread->td_ucred);
641 sigio->sio_myref = sigiop;
642
643 PGRPSESS_SLOCK();
644 if (pgid > 0) {
645 proc = pfind(pgid);
646 if (proc == NULL) {
647 ret = ESRCH;
648 goto fail;
649 }
650
651 /*
652 * Policy - Don't allow a process to FSETOWN a process
653 * in another session.
654 *
655 * Remove this test to allow maximum flexibility or
656 * restrict FSETOWN to the current process or process
657 * group for maximum safety.
658 */
659 PROC_UNLOCK(proc);
660 if (proc->p_session != curthread->td_proc->p_session) {
661 ret = EPERM;
662 goto fail;
663 }
664
665 pgrp = NULL;
666 } else /* if (pgid < 0) */ {
667 pgrp = pgfind(-pgid);
668 if (pgrp == NULL) {
669 ret = ESRCH;
670 goto fail;
671 }
672 PGRP_UNLOCK(pgrp);
673
674 /*
675 * Policy - Don't allow a process to FSETOWN a process
676 * in another session.
677 *
678 * Remove this test to allow maximum flexibility or
679 * restrict FSETOWN to the current process or process
680 * group for maximum safety.
681 */
682 if (pgrp->pg_session != curthread->td_proc->p_session) {
683 ret = EPERM;
684 goto fail;
685 }
686
687 proc = NULL;
688 }
689 funsetown(*sigiop);
690 if (pgid > 0) {
691 PROC_LOCK(proc);
692 SLIST_INSERT_HEAD(&proc->p_sigiolst, sigio, sio_pgsigio);
693 sigio->sio_proc = proc;
694 PROC_UNLOCK(proc);
695 } else {
696 PGRP_LOCK(pgrp);
697 SLIST_INSERT_HEAD(&pgrp->pg_sigiolst, sigio, sio_pgsigio);
698 sigio->sio_pgrp = pgrp;
699 PGRP_UNLOCK(pgrp);
700 }
701 PGRPSESS_SUNLOCK();
702 s = splhigh();
703 *sigiop = sigio;
704 splx(s);
705 return (0);
706
707fail:
708 PGRPSESS_SUNLOCK();
709 crfree(sigio->sio_ucred);
710 FREE(sigio, M_SIGIO);
711 return (ret);
712}
713
714/*
715 * This is common code for FIOGETOWN ioctl called by fcntl(fd, F_GETOWN, arg).
716 */
717pid_t
718fgetown(sigio)
719 struct sigio *sigio;
720{
721 return (sigio != NULL ? sigio->sio_pgid : 0);
722}
723
724/*
725 * Close a file descriptor.
726 */
727#ifndef _SYS_SYSPROTO_H_
728struct close_args {
729 int fd;
730};
731#endif
732/*
733 * MPSAFE
734 */
735/* ARGSUSED */
736int
737close(td, uap)
738 struct thread *td;
739 struct close_args *uap;
740{
741 register struct filedesc *fdp;
742 register struct file *fp;
743 register int fd = uap->fd;
744 int error = 0;
745
746 mtx_lock(&Giant);
747 fdp = td->td_proc->p_fd;
748 FILEDESC_LOCK(fdp);
749 if ((unsigned)fd >= fdp->fd_nfiles ||
750 (fp = fdp->fd_ofiles[fd]) == NULL) {
751 FILEDESC_UNLOCK(fdp);
752 error = EBADF;
753 goto done2;
754 }
755#if 0
756 if (fdp->fd_ofileflags[fd] & UF_MAPPED)
757 (void) munmapfd(td, fd);
758#endif
759 fdp->fd_ofiles[fd] = NULL;
760 fdp->fd_ofileflags[fd] = 0;
761
762 /*
763 * we now hold the fp reference that used to be owned by the descriptor
764 * array.
765 */
766 while (fdp->fd_lastfile > 0 && fdp->fd_ofiles[fdp->fd_lastfile] == NULL)
767 fdp->fd_lastfile--;
768 if (fd < fdp->fd_freefile)
769 fdp->fd_freefile = fd;
770 if (fd < fdp->fd_knlistsize) {
771 FILEDESC_UNLOCK(fdp);
772 knote_fdclose(td, fd);
773 } else
774 FILEDESC_UNLOCK(fdp);
775
776 error = closef(fp, td);
777done2:
778 mtx_unlock(&Giant);
779 return(error);
780}
781
782#if defined(COMPAT_43) || defined(COMPAT_SUNOS)
783/*
784 * Return status information about a file descriptor.
785 */
786#ifndef _SYS_SYSPROTO_H_
787struct ofstat_args {
788 int fd;
789 struct ostat *sb;
790};
791#endif
792/*
793 * MPSAFE
794 */
795/* ARGSUSED */
796int
797ofstat(td, uap)
798 struct thread *td;
799 register struct ofstat_args *uap;
800{
801 struct file *fp;
802 struct stat ub;
803 struct ostat oub;
804 int error;
805
806 mtx_lock(&Giant);
807 if ((error = fget(td, uap->fd, &fp)) != 0)
808 goto done2;
809 error = fo_stat(fp, &ub, td);
810 if (error == 0) {
811 cvtstat(&ub, &oub);
812 error = copyout((caddr_t)&oub, (caddr_t)uap->sb, sizeof (oub));
813 }
814 fdrop(fp, td);
815done2:
816 mtx_unlock(&Giant);
817 return (error);
818}
819#endif /* COMPAT_43 || COMPAT_SUNOS */
820
821/*
822 * Return status information about a file descriptor.
823 */
824#ifndef _SYS_SYSPROTO_H_
825struct fstat_args {
826 int fd;
827 struct stat *sb;
828};
829#endif
830/*
831 * MPSAFE
832 */
833/* ARGSUSED */
834int
835fstat(td, uap)
836 struct thread *td;
837 struct fstat_args *uap;
838{
839 struct file *fp;
840 struct stat ub;
841 int error;
842
843 mtx_lock(&Giant);
844 if ((error = fget(td, uap->fd, &fp)) != 0)
845 goto done2;
846 error = fo_stat(fp, &ub, td);
847 if (error == 0)
848 error = copyout((caddr_t)&ub, (caddr_t)uap->sb, sizeof (ub));
849 fdrop(fp, td);
850done2:
851 mtx_unlock(&Giant);
852 return (error);
853}
854
855/*
856 * Return status information about a file descriptor.
857 */
858#ifndef _SYS_SYSPROTO_H_
859struct nfstat_args {
860 int fd;
861 struct nstat *sb;
862};
863#endif
864/*
865 * MPSAFE
866 */
867/* ARGSUSED */
868int
869nfstat(td, uap)
870 struct thread *td;
871 register struct nfstat_args *uap;
872{
873 struct file *fp;
874 struct stat ub;
875 struct nstat nub;
876 int error;
877
878 mtx_lock(&Giant);
879 if ((error = fget(td, uap->fd, &fp)) != 0)
880 goto done2;
881 error = fo_stat(fp, &ub, td);
882 if (error == 0) {
883 cvtnstat(&ub, &nub);
884 error = copyout((caddr_t)&nub, (caddr_t)uap->sb, sizeof (nub));
885 }
886 fdrop(fp, td);
887done2:
888 mtx_unlock(&Giant);
889 return (error);
890}
891
892/*
893 * Return pathconf information about a file descriptor.
894 */
895#ifndef _SYS_SYSPROTO_H_
896struct fpathconf_args {
897 int fd;
898 int name;
899};
900#endif
901/*
902 * MPSAFE
903 */
904/* ARGSUSED */
905int
906fpathconf(td, uap)
907 struct thread *td;
908 register struct fpathconf_args *uap;
909{
910 struct file *fp;
911 struct vnode *vp;
912 int error;
913
914 if ((error = fget(td, uap->fd, &fp)) != 0)
915 return (error);
916
917 switch (fp->f_type) {
918 case DTYPE_PIPE:
919 case DTYPE_SOCKET:
920 if (uap->name != _PC_PIPE_BUF) {
921 error = EINVAL;
922 } else {
923 td->td_retval[0] = PIPE_BUF;
924 error = 0;
925 }
926 break;
927 case DTYPE_FIFO:
928 case DTYPE_VNODE:
929 vp = (struct vnode *)fp->f_data;
930 mtx_lock(&Giant);
931 error = VOP_PATHCONF(vp, uap->name, td->td_retval);
932 mtx_unlock(&Giant);
933 break;
934 default:
935 error = EOPNOTSUPP;
936 break;
937 }
938 fdrop(fp, td);
939 return(error);
940}
941
942/*
943 * Allocate a file descriptor for the process.
944 */
945static int fdexpand;
946SYSCTL_INT(_debug, OID_AUTO, fdexpand, CTLFLAG_RD, &fdexpand, 0, "");
947
948int
949fdalloc(td, want, result)
950 struct thread *td;
951 int want;
952 int *result;
953{
954 struct proc *p = td->td_proc;
955 register struct filedesc *fdp = td->td_proc->p_fd;
956 register int i;
957 int lim, last, nfiles;
958 struct file **newofile, **oldofile;
959 char *newofileflags;
960
961 FILEDESC_LOCK_ASSERT(fdp, MA_OWNED);
962
963 /*
964 * Search for a free descriptor starting at the higher
965 * of want or fd_freefile. If that fails, consider
966 * expanding the ofile array.
967 */
968 lim = min((int)p->p_rlimit[RLIMIT_NOFILE].rlim_cur, maxfilesperproc);
969 for (;;) {
970 last = min(fdp->fd_nfiles, lim);
971 if ((i = want) < fdp->fd_freefile)
972 i = fdp->fd_freefile;
973 for (; i < last; i++) {
974 if (fdp->fd_ofiles[i] == NULL) {
975 fdp->fd_ofileflags[i] = 0;
976 if (i > fdp->fd_lastfile)
977 fdp->fd_lastfile = i;
978 if (want <= fdp->fd_freefile)
979 fdp->fd_freefile = i;
980 *result = i;
981 return (0);
982 }
983 }
984
985 /*
986 * No space in current array. Expand?
987 */
988 if (fdp->fd_nfiles >= lim)
989 return (EMFILE);
990 if (fdp->fd_nfiles < NDEXTENT)
991 nfiles = NDEXTENT;
992 else
993 nfiles = 2 * fdp->fd_nfiles;
994 FILEDESC_UNLOCK(fdp);
995 mtx_lock(&Giant);
996 MALLOC(newofile, struct file **, nfiles * OFILESIZE,
997 M_FILEDESC, M_WAITOK);
998 mtx_unlock(&Giant);
999 FILEDESC_LOCK(fdp);
1000
1001 /*
1002 * deal with file-table extend race that might have occured
1003 * when malloc was blocked.
1004 */
1005 if (fdp->fd_nfiles >= nfiles) {
1006 FILEDESC_UNLOCK(fdp);
1007 mtx_lock(&Giant);
1008 FREE(newofile, M_FILEDESC);
1009 mtx_unlock(&Giant);
1010 FILEDESC_LOCK(fdp);
1011 continue;
1012 }
1013 newofileflags = (char *) &newofile[nfiles];
1014 /*
1015 * Copy the existing ofile and ofileflags arrays
1016 * and zero the new portion of each array.
1017 */
1018 bcopy(fdp->fd_ofiles, newofile,
1019 (i = sizeof(struct file *) * fdp->fd_nfiles));
1020 bzero((char *)newofile + i, nfiles * sizeof(struct file *) - i);
1021 bcopy(fdp->fd_ofileflags, newofileflags,
1022 (i = sizeof(char) * fdp->fd_nfiles));
1023 bzero(newofileflags + i, nfiles * sizeof(char) - i);
1024 if (fdp->fd_nfiles > NDFILE)
1025 oldofile = fdp->fd_ofiles;
1026 else
1027 oldofile = NULL;
1028 fdp->fd_ofiles = newofile;
1029 fdp->fd_ofileflags = newofileflags;
1030 fdp->fd_nfiles = nfiles;
1031 fdexpand++;
1032 if (oldofile != NULL) {
1033 FILEDESC_UNLOCK(fdp);
1034 mtx_lock(&Giant);
1035 FREE(oldofile, M_FILEDESC);
1036 mtx_unlock(&Giant);
1037 FILEDESC_LOCK(fdp);
1038 }
1039 }
1040 return (0);
1041}
1042
1043/*
1044 * Check to see whether n user file descriptors
1045 * are available to the process p.
1046 */
1047int
1048fdavail(td, n)
1049 struct thread *td;
1050 register int n;
1051{
1052 struct proc *p = td->td_proc;
1053 register struct filedesc *fdp = td->td_proc->p_fd;
1054 register struct file **fpp;
1055 register int i, lim, last;
1056
1057 FILEDESC_LOCK_ASSERT(fdp, MA_OWNED);
1058
1059 lim = min((int)p->p_rlimit[RLIMIT_NOFILE].rlim_cur, maxfilesperproc);
1060 if ((i = lim - fdp->fd_nfiles) > 0 && (n -= i) <= 0)
1061 return (1);
1062
1063 last = min(fdp->fd_nfiles, lim);
1064 fpp = &fdp->fd_ofiles[fdp->fd_freefile];
1065 for (i = last - fdp->fd_freefile; --i >= 0; fpp++) {
1066 if (*fpp == NULL && --n <= 0)
1067 return (1);
1068 }
1069 return (0);
1070}
1071
1072/*
1073 * Create a new open file structure and allocate
1074 * a file decriptor for the process that refers to it.
1075 */
1076int
1077falloc(td, resultfp, resultfd)
1078 register struct thread *td;
1079 struct file **resultfp;
1080 int *resultfd;
1081{
1082 struct proc *p = td->td_proc;
1083 register struct file *fp, *fq;
1084 int error, i;
1085
1086 sx_xlock(&filelist_lock);
1087 if (nfiles >= maxfiles) {
1088 sx_xunlock(&filelist_lock);
1089 tablefull("file");
1090 return (ENFILE);
1091 }
1092 nfiles++;
1093 sx_xunlock(&filelist_lock);
1094 /*
1095 * Allocate a new file descriptor.
1096 * If the process has file descriptor zero open, add to the list
1097 * of open files at that point, otherwise put it at the front of
1098 * the list of open files.
1099 */
1098 MALLOC(fp, struct file *, sizeof(struct file), M_FILE, M_WAITOK | M_ZERO);
1100 fp = uma_zalloc(file_zone, M_WAITOK);
1101 bzero(fp, sizeof(*fp));
1102
1103 /*
1104 * wait until after malloc (which may have blocked) returns before
1105 * allocating the slot, else a race might have shrunk it if we had
1106 * allocated it before the malloc.
1107 */
1108 FILEDESC_LOCK(p->p_fd);
1109 if ((error = fdalloc(td, 0, &i))) {
1110 FILEDESC_UNLOCK(p->p_fd);
1111 sx_xlock(&filelist_lock);
1112 nfiles--;
1113 sx_xunlock(&filelist_lock);
1111 FREE(fp, M_FILE);
1114 uma_zfree(file_zone, fp);
1115 return (error);
1116 }
1117 fp->f_mtxp = mtx_pool_alloc();
1118 fp->f_gcflag = 0;
1119 fp->f_count = 1;
1120 fp->f_cred = crhold(td->td_ucred);
1121 fp->f_ops = &badfileops;
1122 fp->f_seqcount = 1;
1123 FILEDESC_UNLOCK(p->p_fd);
1124 sx_xlock(&filelist_lock);
1125 FILEDESC_LOCK(p->p_fd);
1126 if ((fq = p->p_fd->fd_ofiles[0])) {
1127 LIST_INSERT_AFTER(fq, fp, f_list);
1128 } else {
1129 LIST_INSERT_HEAD(&filehead, fp, f_list);
1130 }
1131 p->p_fd->fd_ofiles[i] = fp;
1132 FILEDESC_UNLOCK(p->p_fd);
1133 sx_xunlock(&filelist_lock);
1134 if (resultfp)
1135 *resultfp = fp;
1136 if (resultfd)
1137 *resultfd = i;
1138 return (0);
1139}
1140
1141/*
1142 * Free a file descriptor.
1143 */
1144void
1145ffree(fp)
1146 register struct file *fp;
1147{
1148
1149 KASSERT((fp->f_count == 0), ("ffree: fp_fcount not 0!"));
1150 sx_xlock(&filelist_lock);
1151 LIST_REMOVE(fp, f_list);
1152 nfiles--;
1153 sx_xunlock(&filelist_lock);
1154 crfree(fp->f_cred);
1152 FREE(fp, M_FILE);
1155 uma_zfree(file_zone, fp);
1156}
1157
1158/*
1159 * Build a new filedesc structure.
1160 */
1161struct filedesc *
1162fdinit(td)
1163 struct thread *td;
1164{
1165 register struct filedesc0 *newfdp;
1166 register struct filedesc *fdp = td->td_proc->p_fd;
1167
1168 MALLOC(newfdp, struct filedesc0 *, sizeof(struct filedesc0),
1169 M_FILEDESC, M_WAITOK | M_ZERO);
1170 mtx_init(&newfdp->fd_fd.fd_mtx, "filedesc structure", MTX_DEF);
1171 FILEDESC_LOCK(&newfdp->fd_fd);
1172 newfdp->fd_fd.fd_cdir = fdp->fd_cdir;
1173 if (newfdp->fd_fd.fd_cdir)
1174 VREF(newfdp->fd_fd.fd_cdir);
1175 newfdp->fd_fd.fd_rdir = fdp->fd_rdir;
1176 if (newfdp->fd_fd.fd_rdir)
1177 VREF(newfdp->fd_fd.fd_rdir);
1178 newfdp->fd_fd.fd_jdir = fdp->fd_jdir;
1179 if (newfdp->fd_fd.fd_jdir)
1180 VREF(newfdp->fd_fd.fd_jdir);
1181
1182 /* Create the file descriptor table. */
1183 newfdp->fd_fd.fd_refcnt = 1;
1184 newfdp->fd_fd.fd_cmask = cmask;
1185 newfdp->fd_fd.fd_ofiles = newfdp->fd_dfiles;
1186 newfdp->fd_fd.fd_ofileflags = newfdp->fd_dfileflags;
1187 newfdp->fd_fd.fd_nfiles = NDFILE;
1188 newfdp->fd_fd.fd_knlistsize = -1;
1189 FILEDESC_UNLOCK(&newfdp->fd_fd);
1190
1191 return (&newfdp->fd_fd);
1192}
1193
1194/*
1195 * Share a filedesc structure.
1196 */
1197struct filedesc *
1198fdshare(p)
1199 struct proc *p;
1200{
1201 FILEDESC_LOCK(p->p_fd);
1202 p->p_fd->fd_refcnt++;
1203 FILEDESC_UNLOCK(p->p_fd);
1204 return (p->p_fd);
1205}
1206
1207/*
1208 * Copy a filedesc structure.
1209 */
1210struct filedesc *
1211fdcopy(td)
1212 struct thread *td;
1213{
1214 register struct filedesc *newfdp, *fdp = td->td_proc->p_fd;
1215 register struct file **fpp;
1216 register int i, j;
1217
1218 /* Certain daemons might not have file descriptors. */
1219 if (fdp == NULL)
1220 return (NULL);
1221
1222 FILEDESC_LOCK_ASSERT(fdp, MA_OWNED);
1223
1224 FILEDESC_UNLOCK(fdp);
1225 MALLOC(newfdp, struct filedesc *, sizeof(struct filedesc0),
1226 M_FILEDESC, M_WAITOK);
1227 FILEDESC_LOCK(fdp);
1228 bcopy(fdp, newfdp, sizeof(struct filedesc));
1229 FILEDESC_UNLOCK(fdp);
1230 bzero(&newfdp->fd_mtx, sizeof(newfdp->fd_mtx));
1231 mtx_init(&newfdp->fd_mtx, "filedesc structure", MTX_DEF);
1232 if (newfdp->fd_cdir)
1233 VREF(newfdp->fd_cdir);
1234 if (newfdp->fd_rdir)
1235 VREF(newfdp->fd_rdir);
1236 if (newfdp->fd_jdir)
1237 VREF(newfdp->fd_jdir);
1238 newfdp->fd_refcnt = 1;
1239
1240 /*
1241 * If the number of open files fits in the internal arrays
1242 * of the open file structure, use them, otherwise allocate
1243 * additional memory for the number of descriptors currently
1244 * in use.
1245 */
1246 FILEDESC_LOCK(fdp);
1247 newfdp->fd_lastfile = fdp->fd_lastfile;
1248 newfdp->fd_nfiles = fdp->fd_nfiles;
1249 if (newfdp->fd_lastfile < NDFILE) {
1250 newfdp->fd_ofiles = ((struct filedesc0 *) newfdp)->fd_dfiles;
1251 newfdp->fd_ofileflags =
1252 ((struct filedesc0 *) newfdp)->fd_dfileflags;
1253 i = NDFILE;
1254 } else {
1255 /*
1256 * Compute the smallest multiple of NDEXTENT needed
1257 * for the file descriptors currently in use,
1258 * allowing the table to shrink.
1259 */
1260retry:
1261 i = newfdp->fd_nfiles;
1262 while (i > 2 * NDEXTENT && i > newfdp->fd_lastfile * 2)
1263 i /= 2;
1264 FILEDESC_UNLOCK(fdp);
1265 MALLOC(newfdp->fd_ofiles, struct file **, i * OFILESIZE,
1266 M_FILEDESC, M_WAITOK);
1267 FILEDESC_LOCK(fdp);
1268 newfdp->fd_lastfile = fdp->fd_lastfile;
1269 newfdp->fd_nfiles = fdp->fd_nfiles;
1270 j = newfdp->fd_nfiles;
1271 while (j > 2 * NDEXTENT && j > newfdp->fd_lastfile * 2)
1272 j /= 2;
1273 if (i != j) {
1274 /*
1275 * The size of the original table has changed.
1276 * Go over once again.
1277 */
1278 FILEDESC_UNLOCK(fdp);
1279 FREE(newfdp->fd_ofiles, M_FILEDESC);
1280 FILEDESC_LOCK(fdp);
1281 newfdp->fd_lastfile = fdp->fd_lastfile;
1282 newfdp->fd_nfiles = fdp->fd_nfiles;
1283 goto retry;
1284 }
1285 newfdp->fd_ofileflags = (char *) &newfdp->fd_ofiles[i];
1286 }
1287 newfdp->fd_nfiles = i;
1288 bcopy(fdp->fd_ofiles, newfdp->fd_ofiles, i * sizeof(struct file **));
1289 bcopy(fdp->fd_ofileflags, newfdp->fd_ofileflags, i * sizeof(char));
1290
1291 /*
1292 * kq descriptors cannot be copied.
1293 */
1294 if (newfdp->fd_knlistsize != -1) {
1295 fpp = &newfdp->fd_ofiles[newfdp->fd_lastfile];
1296 for (i = newfdp->fd_lastfile; i >= 0; i--, fpp--) {
1297 if (*fpp != NULL && (*fpp)->f_type == DTYPE_KQUEUE) {
1298 *fpp = NULL;
1299 if (i < newfdp->fd_freefile)
1300 newfdp->fd_freefile = i;
1301 }
1302 if (*fpp == NULL && i == newfdp->fd_lastfile && i > 0)
1303 newfdp->fd_lastfile--;
1304 }
1305 newfdp->fd_knlist = NULL;
1306 newfdp->fd_knlistsize = -1;
1307 newfdp->fd_knhash = NULL;
1308 newfdp->fd_knhashmask = 0;
1309 }
1310
1311 fpp = newfdp->fd_ofiles;
1312 for (i = newfdp->fd_lastfile; i-- >= 0; fpp++) {
1313 if (*fpp != NULL) {
1314 fhold(*fpp);
1315 }
1316 }
1317 return (newfdp);
1318}
1319
1320/*
1321 * Release a filedesc structure.
1322 */
1323void
1324fdfree(td)
1325 struct thread *td;
1326{
1327 register struct filedesc *fdp;
1328 struct file **fpp;
1329 register int i;
1330
1331 fdp = td->td_proc->p_fd;
1332 /* Certain daemons might not have file descriptors. */
1333 if (fdp == NULL)
1334 return;
1335
1336 FILEDESC_LOCK(fdp);
1337 if (--fdp->fd_refcnt > 0) {
1338 FILEDESC_UNLOCK(fdp);
1339 return;
1340 }
1341 /*
1342 * we are the last reference to the structure, we can
1343 * safely assume it will not change out from under us.
1344 */
1345 FILEDESC_UNLOCK(fdp);
1346 fpp = fdp->fd_ofiles;
1347 for (i = fdp->fd_lastfile; i-- >= 0; fpp++) {
1348 if (*fpp)
1349 (void) closef(*fpp, td);
1350 }
1351
1352 PROC_LOCK(td->td_proc);
1353 td->td_proc->p_fd = NULL;
1354 PROC_UNLOCK(td->td_proc);
1355
1356 if (fdp->fd_nfiles > NDFILE)
1357 FREE(fdp->fd_ofiles, M_FILEDESC);
1358 if (fdp->fd_cdir)
1359 vrele(fdp->fd_cdir);
1360 if (fdp->fd_rdir)
1361 vrele(fdp->fd_rdir);
1362 if (fdp->fd_jdir)
1363 vrele(fdp->fd_jdir);
1364 if (fdp->fd_knlist)
1365 FREE(fdp->fd_knlist, M_KQUEUE);
1366 if (fdp->fd_knhash)
1367 FREE(fdp->fd_knhash, M_KQUEUE);
1368 mtx_destroy(&fdp->fd_mtx);
1369 FREE(fdp, M_FILEDESC);
1370}
1371
1372/*
1373 * For setugid programs, we don't want to people to use that setugidness
1374 * to generate error messages which write to a file which otherwise would
1375 * otherwise be off-limits to the process.
1376 *
1377 * This is a gross hack to plug the hole. A better solution would involve
1378 * a special vop or other form of generalized access control mechanism. We
1379 * go ahead and just reject all procfs file systems accesses as dangerous.
1380 *
1381 * Since setugidsafety calls this only for fd 0, 1 and 2, this check is
1382 * sufficient. We also don't for check setugidness since we know we are.
1383 */
1384static int
1385is_unsafe(struct file *fp)
1386{
1387 if (fp->f_type == DTYPE_VNODE &&
1388 ((struct vnode *)(fp->f_data))->v_tag == VT_PROCFS)
1389 return (1);
1390 return (0);
1391}
1392
1393/*
1394 * Make this setguid thing safe, if at all possible.
1395 */
1396void
1397setugidsafety(td)
1398 struct thread *td;
1399{
1400 struct filedesc *fdp = td->td_proc->p_fd;
1401 register int i;
1402
1403 /* Certain daemons might not have file descriptors. */
1404 if (fdp == NULL)
1405 return;
1406
1407 /*
1408 * note: fdp->fd_ofiles may be reallocated out from under us while
1409 * we are blocked in a close. Be careful!
1410 */
1411 FILEDESC_LOCK(fdp);
1412 for (i = 0; i <= fdp->fd_lastfile; i++) {
1413 if (i > 2)
1414 break;
1415 if (fdp->fd_ofiles[i] && is_unsafe(fdp->fd_ofiles[i])) {
1416 struct file *fp;
1417
1418#if 0
1419 if ((fdp->fd_ofileflags[i] & UF_MAPPED) != 0)
1420 (void) munmapfd(td, i);
1421#endif
1422 if (i < fdp->fd_knlistsize) {
1423 FILEDESC_UNLOCK(fdp);
1424 knote_fdclose(td, i);
1425 FILEDESC_LOCK(fdp);
1426 }
1427 /*
1428 * NULL-out descriptor prior to close to avoid
1429 * a race while close blocks.
1430 */
1431 fp = fdp->fd_ofiles[i];
1432 fdp->fd_ofiles[i] = NULL;
1433 fdp->fd_ofileflags[i] = 0;
1434 if (i < fdp->fd_freefile)
1435 fdp->fd_freefile = i;
1436 FILEDESC_UNLOCK(fdp);
1437 (void) closef(fp, td);
1438 FILEDESC_LOCK(fdp);
1439 }
1440 }
1441 while (fdp->fd_lastfile > 0 && fdp->fd_ofiles[fdp->fd_lastfile] == NULL)
1442 fdp->fd_lastfile--;
1443 FILEDESC_UNLOCK(fdp);
1444}
1445
1446/*
1447 * Close any files on exec?
1448 */
1449void
1450fdcloseexec(td)
1451 struct thread *td;
1452{
1453 struct filedesc *fdp = td->td_proc->p_fd;
1454 register int i;
1455
1456 /* Certain daemons might not have file descriptors. */
1457 if (fdp == NULL)
1458 return;
1459
1460 FILEDESC_LOCK(fdp);
1461
1462 /*
1463 * We cannot cache fd_ofiles or fd_ofileflags since operations
1464 * may block and rip them out from under us.
1465 */
1466 for (i = 0; i <= fdp->fd_lastfile; i++) {
1467 if (fdp->fd_ofiles[i] != NULL &&
1468 (fdp->fd_ofileflags[i] & UF_EXCLOSE)) {
1469 struct file *fp;
1470
1471#if 0
1472 if (fdp->fd_ofileflags[i] & UF_MAPPED)
1473 (void) munmapfd(td, i);
1474#endif
1475 if (i < fdp->fd_knlistsize) {
1476 FILEDESC_UNLOCK(fdp);
1477 knote_fdclose(td, i);
1478 FILEDESC_LOCK(fdp);
1479 }
1480 /*
1481 * NULL-out descriptor prior to close to avoid
1482 * a race while close blocks.
1483 */
1484 fp = fdp->fd_ofiles[i];
1485 fdp->fd_ofiles[i] = NULL;
1486 fdp->fd_ofileflags[i] = 0;
1487 if (i < fdp->fd_freefile)
1488 fdp->fd_freefile = i;
1489 FILEDESC_UNLOCK(fdp);
1490 (void) closef(fp, td);
1491 FILEDESC_LOCK(fdp);
1492 }
1493 }
1494 while (fdp->fd_lastfile > 0 && fdp->fd_ofiles[fdp->fd_lastfile] == NULL)
1495 fdp->fd_lastfile--;
1496 FILEDESC_UNLOCK(fdp);
1497}
1498
1499/*
1500 * Internal form of close.
1501 * Decrement reference count on file structure.
1502 * Note: td may be NULL when closing a file
1503 * that was being passed in a message.
1504 */
1505int
1506closef(fp, td)
1507 register struct file *fp;
1508 register struct thread *td;
1509{
1510 struct vnode *vp;
1511 struct flock lf;
1512
1513 if (fp == NULL)
1514 return (0);
1515 /*
1516 * POSIX record locking dictates that any close releases ALL
1517 * locks owned by this process. This is handled by setting
1518 * a flag in the unlock to free ONLY locks obeying POSIX
1519 * semantics, and not to free BSD-style file locks.
1520 * If the descriptor was in a message, POSIX-style locks
1521 * aren't passed with the descriptor.
1522 */
1523 if (td && (td->td_proc->p_flag & P_ADVLOCK) &&
1524 fp->f_type == DTYPE_VNODE) {
1525 lf.l_whence = SEEK_SET;
1526 lf.l_start = 0;
1527 lf.l_len = 0;
1528 lf.l_type = F_UNLCK;
1529 vp = (struct vnode *)fp->f_data;
1530 (void) VOP_ADVLOCK(vp, (caddr_t)td->td_proc->p_leader,
1531 F_UNLCK, &lf, F_POSIX);
1532 }
1533 return (fdrop(fp, td));
1534}
1535
1536/*
1537 * Drop reference on struct file passed in, may call closef if the
1538 * reference hits zero.
1539 */
1540int
1541fdrop(fp, td)
1542 struct file *fp;
1543 struct thread *td;
1544{
1545
1546 FILE_LOCK(fp);
1547 return (fdrop_locked(fp, td));
1548}
1549
1550/*
1551 * Extract the file pointer associated with the specified descriptor for
1552 * the current user process.
1553 *
1554 * If the descriptor doesn't exist, EBADF is returned.
1555 *
1556 * If the descriptor exists but doesn't match 'flags' then
1557 * return EBADF for read attempts and EINVAL for write attempts.
1558 *
1559 * If 'hold' is set (non-zero) the file's refcount will be bumped on return.
1560 * It should be droped with fdrop().
1561 * If it is not set, then the refcount will not be bumped however the
1562 * thread's filedesc struct will be returned locked (for fgetsock).
1563 *
1564 * If an error occured the non-zero error is returned and *fpp is set to NULL.
1565 * Otherwise *fpp is set and zero is returned.
1566 */
1567static __inline
1568int
1569_fget(struct thread *td, int fd, struct file **fpp, int flags, int hold)
1570{
1571 struct filedesc *fdp;
1572 struct file *fp;
1573
1574 *fpp = NULL;
1575 if (td == NULL || (fdp = td->td_proc->p_fd) == NULL)
1576 return(EBADF);
1577 FILEDESC_LOCK(fdp);
1578 if ((fp = fget_locked(fdp, fd)) == NULL || fp->f_ops == &badfileops) {
1579 FILEDESC_UNLOCK(fdp);
1580 return(EBADF);
1581 }
1582
1583 /*
1584 * Note: FREAD failures returns EBADF to maintain backwards
1585 * compatibility with what routines returned before.
1586 *
1587 * Only one flag, or 0, may be specified.
1588 */
1589 if (flags == FREAD && (fp->f_flag & FREAD) == 0) {
1590 FILEDESC_UNLOCK(fdp);
1591 return(EBADF);
1592 }
1593 if (flags == FWRITE && (fp->f_flag & FWRITE) == 0) {
1594 FILEDESC_UNLOCK(fdp);
1595 return(EINVAL);
1596 }
1597 if (hold) {
1598 fhold(fp);
1599 FILEDESC_UNLOCK(fdp);
1600 }
1601 *fpp = fp;
1602 return(0);
1603}
1604
1605int
1606fget(struct thread *td, int fd, struct file **fpp)
1607{
1608 return(_fget(td, fd, fpp, 0, 1));
1609}
1610
1611int
1612fget_read(struct thread *td, int fd, struct file **fpp)
1613{
1614 return(_fget(td, fd, fpp, FREAD, 1));
1615}
1616
1617int
1618fget_write(struct thread *td, int fd, struct file **fpp)
1619{
1620 return(_fget(td, fd, fpp, FWRITE, 1));
1621}
1622
1623/*
1624 * Like fget() but loads the underlying vnode, or returns an error if
1625 * the descriptor does not represent a vnode. Note that pipes use vnodes
1626 * but never have VM objects (so VOP_GETVOBJECT() calls will return an
1627 * error). The returned vnode will be vref()d.
1628 */
1629
1630static __inline
1631int
1632_fgetvp(struct thread *td, int fd, struct vnode **vpp, int flags)
1633{
1634 struct file *fp;
1635 int error;
1636
1637 *vpp = NULL;
1638 if ((error = _fget(td, fd, &fp, 0, 0)) != 0)
1639 return (error);
1640 if (fp->f_type != DTYPE_VNODE && fp->f_type != DTYPE_FIFO) {
1641 error = EINVAL;
1642 } else {
1643 *vpp = (struct vnode *)fp->f_data;
1644 vref(*vpp);
1645 }
1646 FILEDESC_UNLOCK(td->td_proc->p_fd);
1647 return (error);
1648}
1649
1650int
1651fgetvp(struct thread *td, int fd, struct vnode **vpp)
1652{
1653 return(_fgetvp(td, fd, vpp, 0));
1654}
1655
1656int
1657fgetvp_read(struct thread *td, int fd, struct vnode **vpp)
1658{
1659 return(_fgetvp(td, fd, vpp, FREAD));
1660}
1661
1662int
1663fgetvp_write(struct thread *td, int fd, struct vnode **vpp)
1664{
1665 return(_fgetvp(td, fd, vpp, FWRITE));
1666}
1667
1668/*
1669 * Like fget() but loads the underlying socket, or returns an error if
1670 * the descriptor does not represent a socket.
1671 *
1672 * We bump the ref count on the returned socket. XXX Also obtain the SX lock in
1673 * the future.
1674 */
1675int
1676fgetsock(struct thread *td, int fd, struct socket **spp, u_int *fflagp)
1677{
1678 struct file *fp;
1679 int error;
1680
1681 *spp = NULL;
1682 if (fflagp)
1683 *fflagp = 0;
1684 if ((error = _fget(td, fd, &fp, 0, 0)) != 0)
1685 return (error);
1686 if (fp->f_type != DTYPE_SOCKET) {
1687 error = ENOTSOCK;
1688 } else {
1689 *spp = (struct socket *)fp->f_data;
1690 if (fflagp)
1691 *fflagp = fp->f_flag;
1692 soref(*spp);
1693 }
1694 FILEDESC_UNLOCK(td->td_proc->p_fd);
1695 return(error);
1696}
1697
1698/*
1699 * Drop the reference count on the the socket and XXX release the SX lock in
1700 * the future. The last reference closes the socket.
1701 */
1702void
1703fputsock(struct socket *so)
1704{
1705 sorele(so);
1706}
1707
1708/*
1709 * Drop reference on struct file passed in, may call closef if the
1710 * reference hits zero.
1711 * Expects struct file locked, and will unlock it.
1712 */
1713int
1714fdrop_locked(fp, td)
1715 struct file *fp;
1716 struct thread *td;
1717{
1718 struct flock lf;
1719 struct vnode *vp;
1720 int error;
1721
1722 FILE_LOCK_ASSERT(fp, MA_OWNED);
1723
1724 if (--fp->f_count > 0) {
1725 FILE_UNLOCK(fp);
1726 return (0);
1727 }
1728 mtx_lock(&Giant);
1729 if (fp->f_count < 0)
1730 panic("fdrop: count < 0");
1731 if ((fp->f_flag & FHASLOCK) && fp->f_type == DTYPE_VNODE) {
1732 lf.l_whence = SEEK_SET;
1733 lf.l_start = 0;
1734 lf.l_len = 0;
1735 lf.l_type = F_UNLCK;
1736 vp = (struct vnode *)fp->f_data;
1737 FILE_UNLOCK(fp);
1738 (void) VOP_ADVLOCK(vp, (caddr_t)fp, F_UNLCK, &lf, F_FLOCK);
1739 } else
1740 FILE_UNLOCK(fp);
1741 if (fp->f_ops != &badfileops)
1742 error = fo_close(fp, td);
1743 else
1744 error = 0;
1745 ffree(fp);
1746 mtx_unlock(&Giant);
1747 return (error);
1748}
1749
1750/*
1751 * Apply an advisory lock on a file descriptor.
1752 *
1753 * Just attempt to get a record lock of the requested type on
1754 * the entire file (l_whence = SEEK_SET, l_start = 0, l_len = 0).
1755 */
1756#ifndef _SYS_SYSPROTO_H_
1757struct flock_args {
1758 int fd;
1759 int how;
1760};
1761#endif
1762/*
1763 * MPSAFE
1764 */
1765/* ARGSUSED */
1766int
1767flock(td, uap)
1768 struct thread *td;
1769 register struct flock_args *uap;
1770{
1771 struct file *fp;
1772 struct vnode *vp;
1773 struct flock lf;
1774 int error;
1775
1776 if ((error = fget(td, uap->fd, &fp)) != 0)
1777 return (error);
1778 if (fp->f_type != DTYPE_VNODE) {
1779 fdrop(fp, td);
1780 return (EOPNOTSUPP);
1781 }
1782
1783 mtx_lock(&Giant);
1784 vp = (struct vnode *)fp->f_data;
1785 lf.l_whence = SEEK_SET;
1786 lf.l_start = 0;
1787 lf.l_len = 0;
1788 if (uap->how & LOCK_UN) {
1789 lf.l_type = F_UNLCK;
1790 FILE_LOCK(fp);
1791 fp->f_flag &= ~FHASLOCK;
1792 FILE_UNLOCK(fp);
1793 error = VOP_ADVLOCK(vp, (caddr_t)fp, F_UNLCK, &lf, F_FLOCK);
1794 goto done2;
1795 }
1796 if (uap->how & LOCK_EX)
1797 lf.l_type = F_WRLCK;
1798 else if (uap->how & LOCK_SH)
1799 lf.l_type = F_RDLCK;
1800 else {
1801 error = EBADF;
1802 goto done2;
1803 }
1804 FILE_LOCK(fp);
1805 fp->f_flag |= FHASLOCK;
1806 FILE_UNLOCK(fp);
1807 error = VOP_ADVLOCK(vp, (caddr_t)fp, F_SETLK, &lf,
1808 (uap->how & LOCK_NB) ? F_FLOCK : F_FLOCK | F_WAIT);
1809done2:
1810 fdrop(fp, td);
1811 mtx_unlock(&Giant);
1812 return (error);
1813}
1814
1815/*
1816 * File Descriptor pseudo-device driver (/dev/fd/).
1817 *
1818 * Opening minor device N dup()s the file (if any) connected to file
1819 * descriptor N belonging to the calling process. Note that this driver
1820 * consists of only the ``open()'' routine, because all subsequent
1821 * references to this file will be direct to the other driver.
1822 */
1823/* ARGSUSED */
1824static int
1825fdopen(dev, mode, type, td)
1826 dev_t dev;
1827 int mode, type;
1828 struct thread *td;
1829{
1830
1831 /*
1832 * XXX Kludge: set curthread->td_dupfd to contain the value of the
1833 * the file descriptor being sought for duplication. The error
1834 * return ensures that the vnode for this device will be released
1835 * by vn_open. Open will detect this special error and take the
1836 * actions in dupfdopen below. Other callers of vn_open or VOP_OPEN
1837 * will simply report the error.
1838 */
1839 td->td_dupfd = dev2unit(dev);
1840 return (ENODEV);
1841}
1842
1843/*
1844 * Duplicate the specified descriptor to a free descriptor.
1845 */
1846int
1847dupfdopen(td, fdp, indx, dfd, mode, error)
1848 struct thread *td;
1849 struct filedesc *fdp;
1850 int indx, dfd;
1851 int mode;
1852 int error;
1853{
1854 register struct file *wfp;
1855 struct file *fp;
1856
1857 /*
1858 * If the to-be-dup'd fd number is greater than the allowed number
1859 * of file descriptors, or the fd to be dup'd has already been
1860 * closed, then reject.
1861 */
1862 FILEDESC_LOCK(fdp);
1863 if ((u_int)dfd >= fdp->fd_nfiles ||
1864 (wfp = fdp->fd_ofiles[dfd]) == NULL) {
1865 FILEDESC_UNLOCK(fdp);
1866 return (EBADF);
1867 }
1868
1869 /*
1870 * There are two cases of interest here.
1871 *
1872 * For ENODEV simply dup (dfd) to file descriptor
1873 * (indx) and return.
1874 *
1875 * For ENXIO steal away the file structure from (dfd) and
1876 * store it in (indx). (dfd) is effectively closed by
1877 * this operation.
1878 *
1879 * Any other error code is just returned.
1880 */
1881 switch (error) {
1882 case ENODEV:
1883 /*
1884 * Check that the mode the file is being opened for is a
1885 * subset of the mode of the existing descriptor.
1886 */
1887 FILE_LOCK(wfp);
1888 if (((mode & (FREAD|FWRITE)) | wfp->f_flag) != wfp->f_flag) {
1889 FILE_UNLOCK(wfp);
1890 FILEDESC_UNLOCK(fdp);
1891 return (EACCES);
1892 }
1893 fp = fdp->fd_ofiles[indx];
1894#if 0
1895 if (fp && fdp->fd_ofileflags[indx] & UF_MAPPED)
1896 (void) munmapfd(td, indx);
1897#endif
1898 fdp->fd_ofiles[indx] = wfp;
1899 fdp->fd_ofileflags[indx] = fdp->fd_ofileflags[dfd];
1900 fhold_locked(wfp);
1901 FILE_UNLOCK(wfp);
1902 if (indx > fdp->fd_lastfile)
1903 fdp->fd_lastfile = indx;
1904 if (fp != NULL)
1905 FILE_LOCK(fp);
1906 FILEDESC_UNLOCK(fdp);
1907 /*
1908 * we now own the reference to fp that the ofiles[] array
1909 * used to own. Release it.
1910 */
1911 if (fp != NULL)
1912 fdrop_locked(fp, td);
1913 return (0);
1914
1915 case ENXIO:
1916 /*
1917 * Steal away the file pointer from dfd, and stuff it into indx.
1918 */
1919 fp = fdp->fd_ofiles[indx];
1920#if 0
1921 if (fp && fdp->fd_ofileflags[indx] & UF_MAPPED)
1922 (void) munmapfd(td, indx);
1923#endif
1924 fdp->fd_ofiles[indx] = fdp->fd_ofiles[dfd];
1925 fdp->fd_ofiles[dfd] = NULL;
1926 fdp->fd_ofileflags[indx] = fdp->fd_ofileflags[dfd];
1927 fdp->fd_ofileflags[dfd] = 0;
1928
1929 /*
1930 * Complete the clean up of the filedesc structure by
1931 * recomputing the various hints.
1932 */
1933 if (indx > fdp->fd_lastfile) {
1934 fdp->fd_lastfile = indx;
1935 } else {
1936 while (fdp->fd_lastfile > 0 &&
1937 fdp->fd_ofiles[fdp->fd_lastfile] == NULL) {
1938 fdp->fd_lastfile--;
1939 }
1940 if (dfd < fdp->fd_freefile)
1941 fdp->fd_freefile = dfd;
1942 }
1943 if (fp != NULL)
1944 FILE_LOCK(fp);
1945 FILEDESC_UNLOCK(fdp);
1946
1947 /*
1948 * we now own the reference to fp that the ofiles[] array
1949 * used to own. Release it.
1950 */
1951 if (fp != NULL)
1952 fdrop_locked(fp, td);
1953 return (0);
1954
1955 default:
1956 FILEDESC_UNLOCK(fdp);
1957 return (error);
1958 }
1959 /* NOTREACHED */
1960}
1961
1962/*
1963 * Get file structures.
1964 */
1965static int
1966sysctl_kern_file(SYSCTL_HANDLER_ARGS)
1967{
1968 int error;
1969 struct file *fp;
1970
1971 sx_slock(&filelist_lock);
1972 if (!req->oldptr) {
1973 /*
1974 * overestimate by 10 files
1975 */
1976 error = SYSCTL_OUT(req, 0, sizeof(filehead) +
1977 (nfiles + 10) * sizeof(struct file));
1978 sx_sunlock(&filelist_lock);
1979 return (error);
1980 }
1981
1982 error = SYSCTL_OUT(req, (caddr_t)&filehead, sizeof(filehead));
1983 if (error) {
1984 sx_sunlock(&filelist_lock);
1985 return (error);
1986 }
1987
1988 /*
1989 * followed by an array of file structures
1990 */
1991 LIST_FOREACH(fp, &filehead, f_list) {
1992 error = SYSCTL_OUT(req, (caddr_t)fp, sizeof (struct file));
1993 if (error) {
1994 sx_sunlock(&filelist_lock);
1995 return (error);
1996 }
1997 }
1998 sx_sunlock(&filelist_lock);
1999 return (0);
2000}
2001
2002SYSCTL_PROC(_kern, KERN_FILE, file, CTLTYPE_OPAQUE|CTLFLAG_RD,
2003 0, 0, sysctl_kern_file, "S,file", "Entire file table");
2004
2005SYSCTL_INT(_kern, KERN_MAXFILESPERPROC, maxfilesperproc, CTLFLAG_RW,
2006 &maxfilesperproc, 0, "Maximum files allowed open per process");
2007
2008SYSCTL_INT(_kern, KERN_MAXFILES, maxfiles, CTLFLAG_RW,
2009 &maxfiles, 0, "Maximum number of files");
2010
2011SYSCTL_INT(_kern, OID_AUTO, openfiles, CTLFLAG_RD,
2012 &nfiles, 0, "System-wide number of open files");
2013
2014static void
2015fildesc_drvinit(void *unused)
2016{
2017 dev_t dev;
2018
2019 dev = make_dev(&fildesc_cdevsw, 0, UID_BIN, GID_BIN, 0666, "fd/0");
2020 make_dev_alias(dev, "stdin");
2021 dev = make_dev(&fildesc_cdevsw, 1, UID_BIN, GID_BIN, 0666, "fd/1");
2022 make_dev_alias(dev, "stdout");
2023 dev = make_dev(&fildesc_cdevsw, 2, UID_BIN, GID_BIN, 0666, "fd/2");
2024 make_dev_alias(dev, "stderr");
2025 if (!devfs_present) {
2026 int fd;
2027
2028 for (fd = 3; fd < NUMFDESC; fd++)
2029 make_dev(&fildesc_cdevsw, fd, UID_BIN, GID_BIN, 0666,
2030 "fd/%d", fd);
2031 }
2032}
2033
2034struct fileops badfileops = {
2035 badfo_readwrite,
2036 badfo_readwrite,
2037 badfo_ioctl,
2038 badfo_poll,
2039 badfo_kqfilter,
2040 badfo_stat,
2041 badfo_close
2042};
2043
2044static int
2045badfo_readwrite(fp, uio, cred, flags, td)
2046 struct file *fp;
2047 struct uio *uio;
2048 struct ucred *cred;
2049 struct thread *td;
2050 int flags;
2051{
2052
2053 return (EBADF);
2054}
2055
2056static int
2057badfo_ioctl(fp, com, data, td)
2058 struct file *fp;
2059 u_long com;
2060 caddr_t data;
2061 struct thread *td;
2062{
2063
2064 return (EBADF);
2065}
2066
2067static int
2068badfo_poll(fp, events, cred, td)
2069 struct file *fp;
2070 int events;
2071 struct ucred *cred;
2072 struct thread *td;
2073{
2074
2075 return (0);
2076}
2077
2078static int
2079badfo_kqfilter(fp, kn)
2080 struct file *fp;
2081 struct knote *kn;
2082{
2083
2084 return (0);
2085}
2086
2087static int
2088badfo_stat(fp, sb, td)
2089 struct file *fp;
2090 struct stat *sb;
2091 struct thread *td;
2092{
2093
2094 return (EBADF);
2095}
2096
2097static int
2098badfo_close(fp, td)
2099 struct file *fp;
2100 struct thread *td;
2101{
2102
2103 return (EBADF);
2104}
2105
2106SYSINIT(fildescdev,SI_SUB_DRIVERS,SI_ORDER_MIDDLE+CDEV_MAJOR,
2107 fildesc_drvinit,NULL)
2108
2109static void filelistinit __P((void *));
2110SYSINIT(select, SI_SUB_LOCK, SI_ORDER_FIRST, filelistinit, NULL)
2111
2112/* ARGSUSED*/
2113static void
2114filelistinit(dummy)
2115 void *dummy;
2116{
2117 file_zone = uma_zcreate("Files", sizeof(struct file), NULL, NULL,
2118 NULL, NULL, UMA_ALIGN_PTR, 0);
2119
2120 sx_init(&filelist_lock, "filelist lock");
2121}