socketvar.h revision 180198
1/*- 2 * Copyright (c) 1982, 1986, 1990, 1993 3 * The Regents of the University of California. All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 4. Neither the name of the University nor the names of its contributors 14 * may be used to endorse or promote products derived from this software 15 * without specific prior written permission. 16 * 17 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 20 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 21 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 27 * SUCH DAMAGE. 28 * 29 * @(#)socketvar.h 8.3 (Berkeley) 2/19/95 30 * $FreeBSD: head/sys/sys/socketvar.h 180198 2008-07-02 23:23:27Z rwatson $ 31 */ 32 33#ifndef _SYS_SOCKETVAR_H_ 34#define _SYS_SOCKETVAR_H_ 35 36#include <sys/queue.h> /* for TAILQ macros */ 37#include <sys/selinfo.h> /* for struct selinfo */ 38#include <sys/_lock.h> 39#include <sys/_mutex.h> 40#include <sys/_sx.h> 41 42/* 43 * Kernel structure per socket. 44 * Contains send and receive buffer queues, 45 * handle on protocol and pointer to protocol 46 * private data and error information. 47 */ 48typedef u_quad_t so_gen_t; 49 50/*- 51 * Locking key to struct socket: 52 * (a) constant after allocation, no locking required. 53 * (b) locked by SOCK_LOCK(so). 54 * (c) locked by SOCKBUF_LOCK(&so->so_rcv). 55 * (d) locked by SOCKBUF_LOCK(&so->so_snd). 56 * (e) locked by ACCEPT_LOCK(). 57 * (f) not locked since integer reads/writes are atomic. 58 * (g) used only as a sleep/wakeup address, no value. 59 * (h) locked by global mutex so_global_mtx. 60 */ 61struct socket { 62 int so_count; /* (b) reference count */ 63 short so_type; /* (a) generic type, see socket.h */ 64 short so_options; /* from socket call, see socket.h */ 65 short so_linger; /* time to linger while closing */ 66 short so_state; /* (b) internal state flags SS_* */ 67 int so_qstate; /* (e) internal state flags SQ_* */ 68 void *so_pcb; /* protocol control block */ 69 struct protosw *so_proto; /* (a) protocol handle */ 70/* 71 * Variables for connection queuing. 72 * Socket where accepts occur is so_head in all subsidiary sockets. 73 * If so_head is 0, socket is not related to an accept. 74 * For head socket so_incomp queues partially completed connections, 75 * while so_comp is a queue of connections ready to be accepted. 76 * If a connection is aborted and it has so_head set, then 77 * it has to be pulled out of either so_incomp or so_comp. 78 * We allow connections to queue up based on current queue lengths 79 * and limit on number of queued connections for this socket. 80 */ 81 struct socket *so_head; /* (e) back pointer to listen socket */ 82 TAILQ_HEAD(, socket) so_incomp; /* (e) queue of partial unaccepted connections */ 83 TAILQ_HEAD(, socket) so_comp; /* (e) queue of complete unaccepted connections */ 84 TAILQ_ENTRY(socket) so_list; /* (e) list of unaccepted connections */ 85 u_short so_qlen; /* (e) number of unaccepted connections */ 86 u_short so_incqlen; /* (e) number of unaccepted incomplete 87 connections */ 88 u_short so_qlimit; /* (e) max number queued connections */ 89 short so_timeo; /* (g) connection timeout */ 90 u_short so_error; /* (f) error affecting connection */ 91 struct sigio *so_sigio; /* [sg] information for async I/O or 92 out of band data (SIGURG) */ 93 u_long so_oobmark; /* (c) chars to oob mark */ 94 TAILQ_HEAD(, aiocblist) so_aiojobq; /* AIO ops waiting on socket */ 95/* 96 * Variables for socket buffering. 97 */ 98 struct sockbuf { 99 struct selinfo sb_sel; /* process selecting read/write */ 100 struct mtx sb_mtx; /* sockbuf lock */ 101 struct sx sb_sx; /* prevent I/O interlacing */ 102 short sb_state; /* (c/d) socket state on sockbuf */ 103#define sb_startzero sb_mb 104 struct mbuf *sb_mb; /* (c/d) the mbuf chain */ 105 struct mbuf *sb_mbtail; /* (c/d) the last mbuf in the chain */ 106 struct mbuf *sb_lastrecord; /* (c/d) first mbuf of last 107 * record in socket buffer */ 108 struct mbuf *sb_sndptr; /* (c/d) pointer into mbuf chain */ 109 u_int sb_sndptroff; /* (c/d) byte offset of ptr into chain */ 110 u_int sb_cc; /* (c/d) actual chars in buffer */ 111 u_int sb_hiwat; /* (c/d) max actual char count */ 112 u_int sb_mbcnt; /* (c/d) chars of mbufs used */ 113 u_int sb_mcnt; /* (c/d) number of mbufs in buffer */ 114 u_int sb_ccnt; /* (c/d) number of clusters in buffer */ 115 u_int sb_mbmax; /* (c/d) max chars of mbufs to use */ 116 u_int sb_ctl; /* (c/d) non-data chars in buffer */ 117 int sb_lowat; /* (c/d) low water mark */ 118 int sb_timeo; /* (c/d) timeout for read/write */ 119 short sb_flags; /* (c/d) flags, see below */ 120 } so_rcv, so_snd; 121/* 122 * Constants for sb_flags field of struct sockbuf. 123 */ 124#define SB_MAX (256*1024) /* default for max chars in sockbuf */ 125/* 126 * Constants for sb_flags field of struct sockbuf. 127 */ 128#define SB_WAIT 0x04 /* someone is waiting for data/space */ 129#define SB_SEL 0x08 /* someone is selecting */ 130#define SB_ASYNC 0x10 /* ASYNC I/O, need signals */ 131#define SB_UPCALL 0x20 /* someone wants an upcall */ 132#define SB_NOINTR 0x40 /* operations not interruptible */ 133#define SB_AIO 0x80 /* AIO operations queued */ 134#define SB_KNOTE 0x100 /* kernel note attached */ 135#define SB_NOCOALESCE 0x200 /* don't coalesce new data into existing mbufs */ 136#define SB_AUTOSIZE 0x800 /* automatically size socket buffer */ 137 138 void (*so_upcall)(struct socket *, void *, int); 139 void *so_upcallarg; 140 struct ucred *so_cred; /* (a) user credentials */ 141 struct label *so_label; /* (b) MAC label for socket */ 142 struct label *so_peerlabel; /* (b) cached MAC label for peer */ 143 /* NB: generation count must not be first. */ 144 so_gen_t so_gencnt; /* (h) generation count */ 145 void *so_emuldata; /* (b) private data for emulators */ 146 struct so_accf { 147 struct accept_filter *so_accept_filter; 148 void *so_accept_filter_arg; /* saved filter args */ 149 char *so_accept_filter_str; /* saved user args */ 150 } *so_accf; 151 int so_fibnum; /* routing domain for this socket */ 152}; 153 154#define SB_EMPTY_FIXUP(sb) do { \ 155 if ((sb)->sb_mb == NULL) { \ 156 (sb)->sb_mbtail = NULL; \ 157 (sb)->sb_lastrecord = NULL; \ 158 } \ 159} while (/*CONSTCOND*/0) 160 161/* 162 * Global accept mutex to serialize access to accept queues and 163 * fields associated with multiple sockets. This allows us to 164 * avoid defining a lock order between listen and accept sockets 165 * until such time as it proves to be a good idea. 166 */ 167extern struct mtx accept_mtx; 168#define ACCEPT_LOCK_ASSERT() mtx_assert(&accept_mtx, MA_OWNED) 169#define ACCEPT_UNLOCK_ASSERT() mtx_assert(&accept_mtx, MA_NOTOWNED) 170#define ACCEPT_LOCK() mtx_lock(&accept_mtx) 171#define ACCEPT_UNLOCK() mtx_unlock(&accept_mtx) 172 173/* 174 * Per-socket buffer mutex used to protect most fields in the socket 175 * buffer. 176 */ 177#define SOCKBUF_MTX(_sb) (&(_sb)->sb_mtx) 178#define SOCKBUF_LOCK_INIT(_sb, _name) \ 179 mtx_init(SOCKBUF_MTX(_sb), _name, NULL, MTX_DEF) 180#define SOCKBUF_LOCK_DESTROY(_sb) mtx_destroy(SOCKBUF_MTX(_sb)) 181#define SOCKBUF_LOCK(_sb) mtx_lock(SOCKBUF_MTX(_sb)) 182#define SOCKBUF_OWNED(_sb) mtx_owned(SOCKBUF_MTX(_sb)) 183#define SOCKBUF_UNLOCK(_sb) mtx_unlock(SOCKBUF_MTX(_sb)) 184#define SOCKBUF_LOCK_ASSERT(_sb) mtx_assert(SOCKBUF_MTX(_sb), MA_OWNED) 185#define SOCKBUF_UNLOCK_ASSERT(_sb) mtx_assert(SOCKBUF_MTX(_sb), MA_NOTOWNED) 186 187/* 188 * Per-socket mutex: we reuse the receive socket buffer mutex for space 189 * efficiency. This decision should probably be revisited as we optimize 190 * locking for the socket code. 191 */ 192#define SOCK_MTX(_so) SOCKBUF_MTX(&(_so)->so_rcv) 193#define SOCK_LOCK(_so) SOCKBUF_LOCK(&(_so)->so_rcv) 194#define SOCK_OWNED(_so) SOCKBUF_OWNED(&(_so)->so_rcv) 195#define SOCK_UNLOCK(_so) SOCKBUF_UNLOCK(&(_so)->so_rcv) 196#define SOCK_LOCK_ASSERT(_so) SOCKBUF_LOCK_ASSERT(&(_so)->so_rcv) 197 198/* 199 * Socket state bits. 200 * 201 * Historically, this bits were all kept in the so_state field. For 202 * locking reasons, they are now in multiple fields, as they are 203 * locked differently. so_state maintains basic socket state protected 204 * by the socket lock. so_qstate holds information about the socket 205 * accept queues. Each socket buffer also has a state field holding 206 * information relevant to that socket buffer (can't send, rcv). Many 207 * fields will be read without locks to improve performance and avoid 208 * lock order issues. However, this approach must be used with caution. 209 */ 210#define SS_NOFDREF 0x0001 /* no file table ref any more */ 211#define SS_ISCONNECTED 0x0002 /* socket connected to a peer */ 212#define SS_ISCONNECTING 0x0004 /* in process of connecting to peer */ 213#define SS_ISDISCONNECTING 0x0008 /* in process of disconnecting */ 214#define SS_NBIO 0x0100 /* non-blocking ops */ 215#define SS_ASYNC 0x0200 /* async i/o notify */ 216#define SS_ISCONFIRMING 0x0400 /* deciding to accept connection req */ 217#define SS_ISDISCONNECTED 0x2000 /* socket disconnected from peer */ 218/* 219 * Protocols can mark a socket as SS_PROTOREF to indicate that, following 220 * pru_detach, they still want the socket to persist, and will free it 221 * themselves when they are done. Protocols should only ever call sofree() 222 * following setting this flag in pru_detach(), and never otherwise, as 223 * sofree() bypasses socket reference counting. 224 */ 225#define SS_PROTOREF 0x4000 /* strong protocol reference */ 226 227/* 228 * Socket state bits now stored in the socket buffer state field. 229 */ 230#define SBS_CANTSENDMORE 0x0010 /* can't send more data to peer */ 231#define SBS_CANTRCVMORE 0x0020 /* can't receive more data from peer */ 232#define SBS_RCVATMARK 0x0040 /* at mark on input */ 233 234/* 235 * Socket state bits stored in so_qstate. 236 */ 237#define SQ_INCOMP 0x0800 /* unaccepted, incomplete connection */ 238#define SQ_COMP 0x1000 /* unaccepted, complete connection */ 239 240/* 241 * Externalized form of struct socket used by the sysctl(3) interface. 242 */ 243struct xsocket { 244 size_t xso_len; /* length of this structure */ 245 struct socket *xso_so; /* makes a convenient handle sometimes */ 246 short so_type; 247 short so_options; 248 short so_linger; 249 short so_state; 250 caddr_t so_pcb; /* another convenient handle */ 251 int xso_protocol; 252 int xso_family; 253 u_short so_qlen; 254 u_short so_incqlen; 255 u_short so_qlimit; 256 short so_timeo; 257 u_short so_error; 258 pid_t so_pgid; 259 u_long so_oobmark; 260 struct xsockbuf { 261 u_int sb_cc; 262 u_int sb_hiwat; 263 u_int sb_mbcnt; 264 u_int sb_mcnt; 265 u_int sb_ccnt; 266 u_int sb_mbmax; 267 int sb_lowat; 268 int sb_timeo; 269 short sb_flags; 270 } so_rcv, so_snd; 271 uid_t so_uid; /* XXX */ 272}; 273 274#ifdef _KERNEL 275 276/* 277 * Macros for sockets and socket buffering. 278 */ 279 280/* 281 * Flags to sblock(). 282 */ 283#define SBL_WAIT 0x00000001 /* Wait if not immediately available. */ 284#define SBL_NOINTR 0x00000002 /* Force non-interruptible sleep. */ 285#define SBL_VALID (SBL_WAIT | SBL_NOINTR) 286 287/* 288 * Do we need to notify the other side when I/O is possible? 289 */ 290#define sb_notify(sb) (((sb)->sb_flags & (SB_WAIT | SB_SEL | SB_ASYNC | \ 291 SB_UPCALL | SB_AIO | SB_KNOTE)) != 0) 292 293/* 294 * How much space is there in a socket buffer (so->so_snd or so->so_rcv)? 295 * This is problematical if the fields are unsigned, as the space might 296 * still be negative (cc > hiwat or mbcnt > mbmax). Should detect 297 * overflow and return 0. Should use "lmin" but it doesn't exist now. 298 */ 299#define sbspace(sb) \ 300 ((long) imin((int)((sb)->sb_hiwat - (sb)->sb_cc), \ 301 (int)((sb)->sb_mbmax - (sb)->sb_mbcnt))) 302 303/* do we have to send all at once on a socket? */ 304#define sosendallatonce(so) \ 305 ((so)->so_proto->pr_flags & PR_ATOMIC) 306 307/* can we read something from so? */ 308#define soreadable(so) \ 309 ((so)->so_rcv.sb_cc >= (so)->so_rcv.sb_lowat || \ 310 ((so)->so_rcv.sb_state & SBS_CANTRCVMORE) || \ 311 !TAILQ_EMPTY(&(so)->so_comp) || (so)->so_error) 312 313/* can we write something to so? */ 314#define sowriteable(so) \ 315 ((sbspace(&(so)->so_snd) >= (so)->so_snd.sb_lowat && \ 316 (((so)->so_state&SS_ISCONNECTED) || \ 317 ((so)->so_proto->pr_flags&PR_CONNREQUIRED)==0)) || \ 318 ((so)->so_snd.sb_state & SBS_CANTSENDMORE) || \ 319 (so)->so_error) 320 321/* adjust counters in sb reflecting allocation of m */ 322#define sballoc(sb, m) { \ 323 (sb)->sb_cc += (m)->m_len; \ 324 if ((m)->m_type != MT_DATA && (m)->m_type != MT_OOBDATA) \ 325 (sb)->sb_ctl += (m)->m_len; \ 326 (sb)->sb_mbcnt += MSIZE; \ 327 (sb)->sb_mcnt += 1; \ 328 if ((m)->m_flags & M_EXT) { \ 329 (sb)->sb_mbcnt += (m)->m_ext.ext_size; \ 330 (sb)->sb_ccnt += 1; \ 331 } \ 332} 333 334/* adjust counters in sb reflecting freeing of m */ 335#define sbfree(sb, m) { \ 336 (sb)->sb_cc -= (m)->m_len; \ 337 if ((m)->m_type != MT_DATA && (m)->m_type != MT_OOBDATA) \ 338 (sb)->sb_ctl -= (m)->m_len; \ 339 (sb)->sb_mbcnt -= MSIZE; \ 340 (sb)->sb_mcnt -= 1; \ 341 if ((m)->m_flags & M_EXT) { \ 342 (sb)->sb_mbcnt -= (m)->m_ext.ext_size; \ 343 (sb)->sb_ccnt -= 1; \ 344 } \ 345 if ((sb)->sb_sndptr == (m)) { \ 346 (sb)->sb_sndptr = NULL; \ 347 (sb)->sb_sndptroff = 0; \ 348 } \ 349 if ((sb)->sb_sndptroff != 0) \ 350 (sb)->sb_sndptroff -= (m)->m_len; \ 351} 352 353/* 354 * soref()/sorele() ref-count the socket structure. Note that you must 355 * still explicitly close the socket, but the last ref count will free 356 * the structure. 357 */ 358#define soref(so) do { \ 359 SOCK_LOCK_ASSERT(so); \ 360 ++(so)->so_count; \ 361} while (0) 362 363#define sorele(so) do { \ 364 ACCEPT_LOCK_ASSERT(); \ 365 SOCK_LOCK_ASSERT(so); \ 366 if ((so)->so_count <= 0) \ 367 panic("sorele"); \ 368 if (--(so)->so_count == 0) \ 369 sofree(so); \ 370 else { \ 371 SOCK_UNLOCK(so); \ 372 ACCEPT_UNLOCK(); \ 373 } \ 374} while (0) 375 376#define sotryfree(so) do { \ 377 ACCEPT_LOCK_ASSERT(); \ 378 SOCK_LOCK_ASSERT(so); \ 379 if ((so)->so_count == 0) \ 380 sofree(so); \ 381 else { \ 382 SOCK_UNLOCK(so); \ 383 ACCEPT_UNLOCK(); \ 384 } \ 385} while(0) 386 387/* 388 * In sorwakeup() and sowwakeup(), acquire the socket buffer lock to 389 * avoid a non-atomic test-and-wakeup. However, sowakeup is 390 * responsible for releasing the lock if it is called. We unlock only 391 * if we don't call into sowakeup. If any code is introduced that 392 * directly invokes the underlying sowakeup() primitives, it must 393 * maintain the same semantics. 394 */ 395#define sorwakeup_locked(so) do { \ 396 SOCKBUF_LOCK_ASSERT(&(so)->so_rcv); \ 397 if (sb_notify(&(so)->so_rcv)) \ 398 sowakeup((so), &(so)->so_rcv); \ 399 else \ 400 SOCKBUF_UNLOCK(&(so)->so_rcv); \ 401} while (0) 402 403#define sorwakeup(so) do { \ 404 SOCKBUF_LOCK(&(so)->so_rcv); \ 405 sorwakeup_locked(so); \ 406} while (0) 407 408#define sowwakeup_locked(so) do { \ 409 SOCKBUF_LOCK_ASSERT(&(so)->so_snd); \ 410 if (sb_notify(&(so)->so_snd)) \ 411 sowakeup((so), &(so)->so_snd); \ 412 else \ 413 SOCKBUF_UNLOCK(&(so)->so_snd); \ 414} while (0) 415 416#define sowwakeup(so) do { \ 417 SOCKBUF_LOCK(&(so)->so_snd); \ 418 sowwakeup_locked(so); \ 419} while (0) 420 421/* 422 * Argument structure for sosetopt et seq. This is in the KERNEL 423 * section because it will never be visible to user code. 424 */ 425enum sopt_dir { SOPT_GET, SOPT_SET }; 426struct sockopt { 427 enum sopt_dir sopt_dir; /* is this a get or a set? */ 428 int sopt_level; /* second arg of [gs]etsockopt */ 429 int sopt_name; /* third arg of [gs]etsockopt */ 430 void *sopt_val; /* fourth arg of [gs]etsockopt */ 431 size_t sopt_valsize; /* (almost) fifth arg of [gs]etsockopt */ 432 struct thread *sopt_td; /* calling thread or null if kernel */ 433}; 434 435struct accept_filter { 436 char accf_name[16]; 437 void (*accf_callback) 438 (struct socket *so, void *arg, int waitflag); 439 void * (*accf_create) 440 (struct socket *so, char *arg); 441 void (*accf_destroy) 442 (struct socket *so); 443 SLIST_ENTRY(accept_filter) accf_next; 444}; 445 446#ifdef MALLOC_DECLARE 447MALLOC_DECLARE(M_ACCF); 448MALLOC_DECLARE(M_PCB); 449MALLOC_DECLARE(M_SONAME); 450#endif 451 452extern int maxsockets; 453extern u_long sb_max; 454extern struct uma_zone *socket_zone; 455extern so_gen_t so_gencnt; 456 457struct mbuf; 458struct sockaddr; 459struct ucred; 460struct uio; 461 462/* 463 * From uipc_socket and friends 464 */ 465int do_getopt_accept_filter(struct socket *so, struct sockopt *sopt); 466int do_setopt_accept_filter(struct socket *so, struct sockopt *sopt); 467int so_setsockopt(struct socket *so, int level, int optname, 468 void *optval, size_t optlen); 469int sockargs(struct mbuf **mp, caddr_t buf, int buflen, int type); 470int getsockaddr(struct sockaddr **namp, caddr_t uaddr, size_t len); 471void sbappend(struct sockbuf *sb, struct mbuf *m); 472void sbappend_locked(struct sockbuf *sb, struct mbuf *m); 473void sbappendstream(struct sockbuf *sb, struct mbuf *m); 474void sbappendstream_locked(struct sockbuf *sb, struct mbuf *m); 475int sbappendaddr(struct sockbuf *sb, const struct sockaddr *asa, 476 struct mbuf *m0, struct mbuf *control); 477int sbappendaddr_locked(struct sockbuf *sb, const struct sockaddr *asa, 478 struct mbuf *m0, struct mbuf *control); 479int sbappendcontrol(struct sockbuf *sb, struct mbuf *m0, 480 struct mbuf *control); 481int sbappendcontrol_locked(struct sockbuf *sb, struct mbuf *m0, 482 struct mbuf *control); 483void sbappendrecord(struct sockbuf *sb, struct mbuf *m0); 484void sbappendrecord_locked(struct sockbuf *sb, struct mbuf *m0); 485void sbcheck(struct sockbuf *sb); 486void sbcompress(struct sockbuf *sb, struct mbuf *m, struct mbuf *n); 487struct mbuf * 488 sbcreatecontrol(caddr_t p, int size, int type, int level); 489void sbdestroy(struct sockbuf *sb, struct socket *so); 490void sbdrop(struct sockbuf *sb, int len); 491void sbdrop_locked(struct sockbuf *sb, int len); 492void sbdroprecord(struct sockbuf *sb); 493void sbdroprecord_locked(struct sockbuf *sb); 494void sbflush(struct sockbuf *sb); 495void sbflush_locked(struct sockbuf *sb); 496void sbrelease(struct sockbuf *sb, struct socket *so); 497void sbrelease_internal(struct sockbuf *sb, struct socket *so); 498void sbrelease_locked(struct sockbuf *sb, struct socket *so); 499int sbreserve(struct sockbuf *sb, u_long cc, struct socket *so, 500 struct thread *td); 501int sbreserve_locked(struct sockbuf *sb, u_long cc, struct socket *so, 502 struct thread *td); 503struct mbuf * 504 sbsndptr(struct sockbuf *sb, u_int off, u_int len, u_int *moff); 505void sbtoxsockbuf(struct sockbuf *sb, struct xsockbuf *xsb); 506int sbwait(struct sockbuf *sb); 507int sblock(struct sockbuf *sb, int flags); 508void sbunlock(struct sockbuf *sb); 509void soabort(struct socket *so); 510int soaccept(struct socket *so, struct sockaddr **nam); 511int socheckuid(struct socket *so, uid_t uid); 512int sobind(struct socket *so, struct sockaddr *nam, struct thread *td); 513void socantrcvmore(struct socket *so); 514void socantrcvmore_locked(struct socket *so); 515void socantsendmore(struct socket *so); 516void socantsendmore_locked(struct socket *so); 517int soclose(struct socket *so); 518int soconnect(struct socket *so, struct sockaddr *nam, struct thread *td); 519int soconnect2(struct socket *so1, struct socket *so2); 520int socow_setup(struct mbuf *m0, struct uio *uio); 521int socreate(int dom, struct socket **aso, int type, int proto, 522 struct ucred *cred, struct thread *td); 523int sodisconnect(struct socket *so); 524struct sockaddr *sodupsockaddr(const struct sockaddr *sa, int mflags); 525void sofree(struct socket *so); 526int sogetopt(struct socket *so, struct sockopt *sopt); 527void sohasoutofband(struct socket *so); 528void soisconnected(struct socket *so); 529void soisconnecting(struct socket *so); 530void soisdisconnected(struct socket *so); 531void soisdisconnecting(struct socket *so); 532int solisten(struct socket *so, int backlog, struct thread *td); 533void solisten_proto(struct socket *so, int backlog); 534int solisten_proto_check(struct socket *so); 535struct socket * 536 sonewconn(struct socket *head, int connstatus); 537int sooptcopyin(struct sockopt *sopt, void *buf, size_t len, size_t minlen); 538int sooptcopyout(struct sockopt *sopt, const void *buf, size_t len); 539 540/* XXX; prepare mbuf for (__FreeBSD__ < 3) routines. */ 541int soopt_getm(struct sockopt *sopt, struct mbuf **mp); 542int soopt_mcopyin(struct sockopt *sopt, struct mbuf *m); 543int soopt_mcopyout(struct sockopt *sopt, struct mbuf *m); 544 545int sopoll(struct socket *so, int events, struct ucred *active_cred, 546 struct thread *td); 547int sopoll_generic(struct socket *so, int events, 548 struct ucred *active_cred, struct thread *td); 549int soreceive(struct socket *so, struct sockaddr **paddr, struct uio *uio, 550 struct mbuf **mp0, struct mbuf **controlp, int *flagsp); 551int soreceive_dgram(struct socket *so, struct sockaddr **paddr, 552 struct uio *uio, struct mbuf **mp0, struct mbuf **controlp, 553 int *flagsp); 554int soreceive_generic(struct socket *so, struct sockaddr **paddr, 555 struct uio *uio, struct mbuf **mp0, struct mbuf **controlp, 556 int *flagsp); 557int soreserve(struct socket *so, u_long sndcc, u_long rcvcc); 558void sorflush(struct socket *so); 559int sosend(struct socket *so, struct sockaddr *addr, struct uio *uio, 560 struct mbuf *top, struct mbuf *control, int flags, 561 struct thread *td); 562int sosend_dgram(struct socket *so, struct sockaddr *addr, 563 struct uio *uio, struct mbuf *top, struct mbuf *control, 564 int flags, struct thread *td); 565int sosend_generic(struct socket *so, struct sockaddr *addr, 566 struct uio *uio, struct mbuf *top, struct mbuf *control, 567 int flags, struct thread *td); 568int sosetopt(struct socket *so, struct sockopt *sopt); 569int soshutdown(struct socket *so, int how); 570void sotoxsocket(struct socket *so, struct xsocket *xso); 571void sowakeup(struct socket *so, struct sockbuf *sb); 572int selsocket(struct socket *so, int events, struct timeval *tv, 573 struct thread *td); 574 575#ifdef SOCKBUF_DEBUG 576void sblastrecordchk(struct sockbuf *, const char *, int); 577#define SBLASTRECORDCHK(sb) sblastrecordchk((sb), __FILE__, __LINE__) 578 579void sblastmbufchk(struct sockbuf *, const char *, int); 580#define SBLASTMBUFCHK(sb) sblastmbufchk((sb), __FILE__, __LINE__) 581#else 582#define SBLASTRECORDCHK(sb) /* nothing */ 583#define SBLASTMBUFCHK(sb) /* nothing */ 584#endif /* SOCKBUF_DEBUG */ 585 586/* 587 * Accept filter functions (duh). 588 */ 589int accept_filt_add(struct accept_filter *filt); 590int accept_filt_del(char *name); 591struct accept_filter *accept_filt_get(char *name); 592#ifdef ACCEPT_FILTER_MOD 593#ifdef SYSCTL_DECL 594SYSCTL_DECL(_net_inet_accf); 595#endif 596int accept_filt_generic_mod_event(module_t mod, int event, void *data); 597#endif 598 599#endif /* _KERNEL */ 600 601#endif /* !_SYS_SOCKETVAR_H_ */ 602