udp.c revision 7862:f8b6a07acfd6
1/*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21/*
22 * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
23 * Use is subject to license terms.
24 */
25/* Copyright (c) 1990 Mentat Inc. */
26
27const char udp_version[] = "@(#)udp.c	1.206	08/10/17 SMI";
28
29#include <sys/types.h>
30#include <sys/stream.h>
31#include <sys/dlpi.h>
32#include <sys/pattr.h>
33#include <sys/stropts.h>
34#include <sys/strlog.h>
35#include <sys/strsun.h>
36#include <sys/time.h>
37#define	_SUN_TPI_VERSION 2
38#include <sys/tihdr.h>
39#include <sys/timod.h>
40#include <sys/ddi.h>
41#include <sys/sunddi.h>
42#include <sys/strsubr.h>
43#include <sys/suntpi.h>
44#include <sys/xti_inet.h>
45#include <sys/cmn_err.h>
46#include <sys/kmem.h>
47#include <sys/policy.h>
48#include <sys/ucred.h>
49#include <sys/zone.h>
50
51#include <sys/socket.h>
52#include <sys/sockio.h>
53#include <sys/vtrace.h>
54#include <sys/sdt.h>
55#include <sys/debug.h>
56#include <sys/isa_defs.h>
57#include <sys/random.h>
58#include <netinet/in.h>
59#include <netinet/ip6.h>
60#include <netinet/icmp6.h>
61#include <netinet/udp.h>
62#include <net/if.h>
63#include <net/route.h>
64
65#include <inet/common.h>
66#include <inet/ip.h>
67#include <inet/ip_impl.h>
68#include <inet/ip6.h>
69#include <inet/ip_ire.h>
70#include <inet/ip_if.h>
71#include <inet/ip_multi.h>
72#include <inet/ip_ndp.h>
73#include <inet/mi.h>
74#include <inet/mib2.h>
75#include <inet/nd.h>
76#include <inet/optcom.h>
77#include <inet/snmpcom.h>
78#include <inet/kstatcom.h>
79#include <inet/udp_impl.h>
80#include <inet/ipclassifier.h>
81#include <inet/ipsec_impl.h>
82#include <inet/ipp_common.h>
83
84/*
85 * The ipsec_info.h header file is here since it has the definition for the
86 * M_CTL message types used by IP to convey information to the ULP. The
87 * ipsec_info.h needs the pfkeyv2.h, hence the latter's presence.
88 */
89#include <net/pfkeyv2.h>
90#include <inet/ipsec_info.h>
91
92#include <sys/tsol/label.h>
93#include <sys/tsol/tnet.h>
94#include <rpc/pmap_prot.h>
95
96/*
97 * Synchronization notes:
98 *
99 * UDP is MT and uses the usual kernel synchronization primitives. There are 2
100 * locks, the fanout lock (uf_lock) and the udp endpoint lock udp_rwlock.
101 * We also use conn_lock when updating things that affect the IP classifier
102 * lookup.
103 * The lock order is udp_rwlock -> uf_lock and is udp_rwlock -> conn_lock.
104 *
105 * The fanout lock uf_lock:
106 * When a UDP endpoint is bound to a local port, it is inserted into
107 * a bind hash list.  The list consists of an array of udp_fanout_t buckets.
108 * The size of the array is controlled by the udp_bind_fanout_size variable.
109 * This variable can be changed in /etc/system if the default value is
110 * not large enough.  Each bind hash bucket is protected by a per bucket
111 * lock.  It protects the udp_bind_hash and udp_ptpbhn fields in the udp_t
112 * structure and a few other fields in the udp_t. A UDP endpoint is removed
113 * from the bind hash list only when it is being unbound or being closed.
114 * The per bucket lock also protects a UDP endpoint's state changes.
115 *
116 * The udp_rwlock:
117 * This protects most of the other fields in the udp_t. The exact list of
118 * fields which are protected by each of the above locks is documented in
119 * the udp_t structure definition.
120 *
121 * Plumbing notes:
122 * UDP is always a device driver. For compatibility with mibopen() code
123 * it is possible to I_PUSH "udp", but that results in pushing a passthrough
124 * dummy module.
125 *
126 * The above implies that we don't support any intermediate module to
127 * reside in between /dev/ip and udp -- in fact, we never supported such
128 * scenario in the past as the inter-layer communication semantics have
129 * always been private.
130 */
131
132/* For /etc/system control */
133uint_t udp_bind_fanout_size = UDP_BIND_FANOUT_SIZE;
134
135#define	NDD_TOO_QUICK_MSG \
136	"ndd get info rate too high for non-privileged users, try again " \
137	"later.\n"
138#define	NDD_OUT_OF_BUF_MSG	"<< Out of buffer >>\n"
139
140/* Option processing attrs */
141typedef struct udpattrs_s {
142	union {
143		ip6_pkt_t	*udpattr_ipp6;	/* For V6 */
144		ip4_pkt_t 	*udpattr_ipp4;	/* For V4 */
145	} udpattr_ippu;
146#define	udpattr_ipp6 udpattr_ippu.udpattr_ipp6
147#define	udpattr_ipp4 udpattr_ippu.udpattr_ipp4
148	mblk_t		*udpattr_mb;
149	boolean_t	udpattr_credset;
150} udpattrs_t;
151
152static void	udp_addr_req(queue_t *q, mblk_t *mp);
153static void	udp_bind(queue_t *q, mblk_t *mp);
154static void	udp_bind_hash_insert(udp_fanout_t *uf, udp_t *udp);
155static void	udp_bind_hash_remove(udp_t *udp, boolean_t caller_holds_lock);
156static void	udp_bind_result(conn_t *, mblk_t *);
157static void	udp_bind_ack(conn_t *, mblk_t *mp);
158static void	udp_bind_error(conn_t *, mblk_t *mp);
159static int	udp_build_hdrs(udp_t *udp);
160static void	udp_capability_req(queue_t *q, mblk_t *mp);
161static int	udp_close(queue_t *q);
162static void	udp_connect(queue_t *q, mblk_t *mp);
163static void	udp_disconnect(queue_t *q, mblk_t *mp);
164static void	udp_err_ack(queue_t *q, mblk_t *mp, t_scalar_t t_error,
165		    int sys_error);
166static void	udp_err_ack_prim(queue_t *q, mblk_t *mp, int primitive,
167		    t_scalar_t tlierr, int unixerr);
168static int	udp_extra_priv_ports_get(queue_t *q, mblk_t *mp, caddr_t cp,
169		    cred_t *cr);
170static int	udp_extra_priv_ports_add(queue_t *q, mblk_t *mp,
171		    char *value, caddr_t cp, cred_t *cr);
172static int	udp_extra_priv_ports_del(queue_t *q, mblk_t *mp,
173		    char *value, caddr_t cp, cred_t *cr);
174static void	udp_icmp_error(queue_t *q, mblk_t *mp);
175static void	udp_icmp_error_ipv6(queue_t *q, mblk_t *mp);
176static void	udp_info_req(queue_t *q, mblk_t *mp);
177static void	udp_input(void *, mblk_t *, void *);
178static mblk_t	*udp_ip_bind_mp(udp_t *udp, t_scalar_t bind_prim,
179		    t_scalar_t addr_length);
180static void	udp_lrput(queue_t *, mblk_t *);
181static void	udp_lwput(queue_t *, mblk_t *);
182static int	udp_open(queue_t *q, dev_t *devp, int flag, int sflag,
183		    cred_t *credp, boolean_t isv6);
184static int	udp_openv4(queue_t *q, dev_t *devp, int flag, int sflag,
185		    cred_t *credp);
186static int	udp_openv6(queue_t *q, dev_t *devp, int flag, int sflag,
187		    cred_t *credp);
188static  int	udp_unitdata_opt_process(queue_t *q, mblk_t *mp,
189		    int *errorp, udpattrs_t *udpattrs);
190static boolean_t udp_opt_allow_udr_set(t_scalar_t level, t_scalar_t name);
191static int	udp_param_get(queue_t *q, mblk_t *mp, caddr_t cp, cred_t *cr);
192static boolean_t udp_param_register(IDP *ndp, udpparam_t *udppa, int cnt);
193static int	udp_param_set(queue_t *q, mblk_t *mp, char *value, caddr_t cp,
194		    cred_t *cr);
195static void	udp_report_item(mblk_t *mp, udp_t *udp);
196static int	udp_rinfop(queue_t *q, infod_t *dp);
197static int	udp_rrw(queue_t *q, struiod_t *dp);
198static int	udp_status_report(queue_t *q, mblk_t *mp, caddr_t cp,
199		    cred_t *cr);
200static void	udp_send_data(udp_t *, queue_t *, mblk_t *, ipha_t *);
201static void	udp_ud_err(queue_t *q, mblk_t *mp, uchar_t *destaddr,
202		    t_scalar_t destlen, t_scalar_t err);
203static void	udp_unbind(queue_t *q, mblk_t *mp);
204static in_port_t udp_update_next_port(udp_t *udp, in_port_t port,
205    boolean_t random);
206static mblk_t	*udp_output_v4(conn_t *, mblk_t *, ipaddr_t, uint16_t, uint_t,
207		    int *, boolean_t);
208static mblk_t	*udp_output_v6(conn_t *connp, mblk_t *mp, sin6_t *sin6,
209		    int *error);
210static void	udp_wput_other(queue_t *q, mblk_t *mp);
211static void	udp_wput_iocdata(queue_t *q, mblk_t *mp);
212static size_t	udp_set_rcv_hiwat(udp_t *udp, size_t size);
213
214static void	*udp_stack_init(netstackid_t stackid, netstack_t *ns);
215static void	udp_stack_fini(netstackid_t stackid, void *arg);
216
217static void	*udp_kstat_init(netstackid_t stackid);
218static void	udp_kstat_fini(netstackid_t stackid, kstat_t *ksp);
219static void	*udp_kstat2_init(netstackid_t, udp_stat_t *);
220static void	udp_kstat2_fini(netstackid_t, kstat_t *);
221static int	udp_kstat_update(kstat_t *kp, int rw);
222
223static void	udp_rcv_enqueue(queue_t *q, udp_t *udp, mblk_t *mp,
224		    uint_t pkt_len);
225static void	udp_rcv_drain(queue_t *q, udp_t *udp, boolean_t closing);
226static void	udp_xmit(queue_t *, mblk_t *, ire_t *ire, conn_t *, zoneid_t);
227
228#define	UDP_RECV_HIWATER	(56 * 1024)
229#define	UDP_RECV_LOWATER	128
230#define	UDP_XMIT_HIWATER	(56 * 1024)
231#define	UDP_XMIT_LOWATER	1024
232
233static struct module_info udp_mod_info =  {
234	UDP_MOD_ID, UDP_MOD_NAME, 1, INFPSZ, UDP_RECV_HIWATER, UDP_RECV_LOWATER
235};
236
237/*
238 * Entry points for UDP as a device.
239 * We have separate open functions for the /dev/udp and /dev/udp6 devices.
240 */
241static struct qinit udp_rinitv4 = {
242	NULL, NULL, udp_openv4, udp_close, NULL,
243	&udp_mod_info, NULL, udp_rrw, udp_rinfop, STRUIOT_STANDARD
244};
245
246static struct qinit udp_rinitv6 = {
247	NULL, NULL, udp_openv6, udp_close, NULL,
248	&udp_mod_info, NULL, udp_rrw, udp_rinfop, STRUIOT_STANDARD
249};
250
251static struct qinit udp_winit = {
252	(pfi_t)udp_wput, (pfi_t)ip_wsrv, NULL, NULL, NULL,
253	&udp_mod_info, NULL, NULL, NULL, STRUIOT_NONE
254};
255
256/*
257 * UDP needs to handle I_LINK and I_PLINK since ifconfig
258 * likes to use it as a place to hang the various streams.
259 */
260static struct qinit udp_lrinit = {
261	(pfi_t)udp_lrput, NULL, udp_openv4, udp_close, NULL,
262	&udp_mod_info
263};
264
265static struct qinit udp_lwinit = {
266	(pfi_t)udp_lwput, NULL, udp_openv4, udp_close, NULL,
267	&udp_mod_info
268};
269
270/* For AF_INET aka /dev/udp */
271struct streamtab udpinfov4 = {
272	&udp_rinitv4, &udp_winit, &udp_lrinit, &udp_lwinit
273};
274
275/* For AF_INET6 aka /dev/udp6 */
276struct streamtab udpinfov6 = {
277	&udp_rinitv6, &udp_winit, &udp_lrinit, &udp_lwinit
278};
279
280static	sin_t	sin_null;	/* Zero address for quick clears */
281static	sin6_t	sin6_null;	/* Zero address for quick clears */
282
283#define	UDP_MAXPACKET_IPV4 (IP_MAXPACKET - UDPH_SIZE - IP_SIMPLE_HDR_LENGTH)
284
285/* Default structure copied into T_INFO_ACK messages */
286static struct T_info_ack udp_g_t_info_ack_ipv4 = {
287	T_INFO_ACK,
288	UDP_MAXPACKET_IPV4,	/* TSDU_size. Excl. headers */
289	T_INVALID,	/* ETSU_size.  udp does not support expedited data. */
290	T_INVALID,	/* CDATA_size. udp does not support connect data. */
291	T_INVALID,	/* DDATA_size. udp does not support disconnect data. */
292	sizeof (sin_t),	/* ADDR_size. */
293	0,		/* OPT_size - not initialized here */
294	UDP_MAXPACKET_IPV4,	/* TIDU_size.  Excl. headers */
295	T_CLTS,		/* SERV_type.  udp supports connection-less. */
296	TS_UNBND,	/* CURRENT_state.  This is set from udp_state. */
297	(XPG4_1|SENDZERO) /* PROVIDER_flag */
298};
299
300#define	UDP_MAXPACKET_IPV6 (IP_MAXPACKET - UDPH_SIZE - IPV6_HDR_LEN)
301
302static	struct T_info_ack udp_g_t_info_ack_ipv6 = {
303	T_INFO_ACK,
304	UDP_MAXPACKET_IPV6,	/* TSDU_size.  Excl. headers */
305	T_INVALID,	/* ETSU_size.  udp does not support expedited data. */
306	T_INVALID,	/* CDATA_size. udp does not support connect data. */
307	T_INVALID,	/* DDATA_size. udp does not support disconnect data. */
308	sizeof (sin6_t), /* ADDR_size. */
309	0,		/* OPT_size - not initialized here */
310	UDP_MAXPACKET_IPV6,	/* TIDU_size. Excl. headers */
311	T_CLTS,		/* SERV_type.  udp supports connection-less. */
312	TS_UNBND,	/* CURRENT_state.  This is set from udp_state. */
313	(XPG4_1|SENDZERO) /* PROVIDER_flag */
314};
315
316/* largest UDP port number */
317#define	UDP_MAX_PORT	65535
318
319/*
320 * Table of ND variables supported by udp.  These are loaded into us_nd
321 * in udp_open.
322 * All of these are alterable, within the min/max values given, at run time.
323 */
324/* BEGIN CSTYLED */
325udpparam_t udp_param_arr[] = {
326 /*min		max		value		name */
327 { 0L,		256,		32,		"udp_wroff_extra" },
328 { 1L,		255,		255,		"udp_ipv4_ttl" },
329 { 0,		IPV6_MAX_HOPS,	IPV6_DEFAULT_HOPS, "udp_ipv6_hoplimit"},
330 { 1024,	(32 * 1024),	1024,		"udp_smallest_nonpriv_port" },
331 { 0,		1,		1,		"udp_do_checksum" },
332 { 1024,	UDP_MAX_PORT,	(32 * 1024),	"udp_smallest_anon_port" },
333 { 1024,	UDP_MAX_PORT,	UDP_MAX_PORT,	"udp_largest_anon_port" },
334 { UDP_XMIT_LOWATER, (1<<30), UDP_XMIT_HIWATER,	"udp_xmit_hiwat"},
335 { 0,		     (1<<30), UDP_XMIT_LOWATER, "udp_xmit_lowat"},
336 { UDP_RECV_LOWATER, (1<<30), UDP_RECV_HIWATER,	"udp_recv_hiwat"},
337 { 65536,	(1<<30),	2*1024*1024,	"udp_max_buf"},
338 { 100,		60000,		1000,		"udp_ndd_get_info_interval"},
339};
340/* END CSTYLED */
341
342/* Setable in /etc/system */
343/* If set to 0, pick ephemeral port sequentially; otherwise randomly. */
344uint32_t udp_random_anon_port = 1;
345
346/*
347 * Hook functions to enable cluster networking.
348 * On non-clustered systems these vectors must always be NULL
349 */
350
351void (*cl_inet_bind)(uchar_t protocol, sa_family_t addr_family,
352    uint8_t *laddrp, in_port_t lport) = NULL;
353void (*cl_inet_unbind)(uint8_t protocol, sa_family_t addr_family,
354    uint8_t *laddrp, in_port_t lport) = NULL;
355
356typedef union T_primitives *t_primp_t;
357
358/*
359 * Return the next anonymous port in the privileged port range for
360 * bind checking.
361 *
362 * Trusted Extension (TX) notes: TX allows administrator to mark or
363 * reserve ports as Multilevel ports (MLP). MLP has special function
364 * on TX systems. Once a port is made MLP, it's not available as
365 * ordinary port. This creates "holes" in the port name space. It
366 * may be necessary to skip the "holes" find a suitable anon port.
367 */
368static in_port_t
369udp_get_next_priv_port(udp_t *udp)
370{
371	static in_port_t next_priv_port = IPPORT_RESERVED - 1;
372	in_port_t nextport;
373	boolean_t restart = B_FALSE;
374	udp_stack_t *us = udp->udp_us;
375
376retry:
377	if (next_priv_port < us->us_min_anonpriv_port ||
378	    next_priv_port >= IPPORT_RESERVED) {
379		next_priv_port = IPPORT_RESERVED - 1;
380		if (restart)
381			return (0);
382		restart = B_TRUE;
383	}
384
385	if (is_system_labeled() &&
386	    (nextport = tsol_next_port(crgetzone(udp->udp_connp->conn_cred),
387	    next_priv_port, IPPROTO_UDP, B_FALSE)) != 0) {
388		next_priv_port = nextport;
389		goto retry;
390	}
391
392	return (next_priv_port--);
393}
394
395/* UDP bind hash report triggered via the Named Dispatch mechanism. */
396/* ARGSUSED */
397static int
398udp_bind_hash_report(queue_t *q, mblk_t *mp, caddr_t cp, cred_t *cr)
399{
400	udp_fanout_t	*udpf;
401	int		i;
402	zoneid_t	zoneid;
403	conn_t		*connp;
404	udp_t		*udp;
405	udp_stack_t	*us;
406
407	connp = Q_TO_CONN(q);
408	udp = connp->conn_udp;
409	us = udp->udp_us;
410
411	/* Refer to comments in udp_status_report(). */
412	if (cr == NULL || secpolicy_ip_config(cr, B_TRUE) != 0) {
413		if (ddi_get_lbolt() - us->us_last_ndd_get_info_time <
414		    drv_usectohz(us->us_ndd_get_info_interval * 1000)) {
415			(void) mi_mpprintf(mp, NDD_TOO_QUICK_MSG);
416			return (0);
417		}
418	}
419	if ((mp->b_cont = allocb(ND_MAX_BUF_LEN, BPRI_HI)) == NULL) {
420		/* The following may work even if we cannot get a large buf. */
421		(void) mi_mpprintf(mp, NDD_OUT_OF_BUF_MSG);
422		return (0);
423	}
424
425	(void) mi_mpprintf(mp,
426	    "UDP     " MI_COL_HDRPAD_STR
427	/*   12345678[89ABCDEF] */
428	    " zone lport src addr        dest addr       port  state");
429	/*    1234 12345 xxx.xxx.xxx.xxx xxx.xxx.xxx.xxx 12345 UNBOUND */
430
431	zoneid = connp->conn_zoneid;
432
433	for (i = 0; i < us->us_bind_fanout_size; i++) {
434		udpf = &us->us_bind_fanout[i];
435		mutex_enter(&udpf->uf_lock);
436
437		/* Print the hash index. */
438		udp = udpf->uf_udp;
439		if (zoneid != GLOBAL_ZONEID) {
440			/* skip to first entry in this zone; might be none */
441			while (udp != NULL &&
442			    udp->udp_connp->conn_zoneid != zoneid)
443				udp = udp->udp_bind_hash;
444		}
445		if (udp != NULL) {
446			uint_t print_len, buf_len;
447
448			buf_len = mp->b_cont->b_datap->db_lim -
449			    mp->b_cont->b_wptr;
450			print_len = snprintf((char *)mp->b_cont->b_wptr,
451			    buf_len, "%d\n", i);
452			if (print_len < buf_len) {
453				mp->b_cont->b_wptr += print_len;
454			} else {
455				mp->b_cont->b_wptr += buf_len;
456			}
457			for (; udp != NULL; udp = udp->udp_bind_hash) {
458				if (zoneid == GLOBAL_ZONEID ||
459				    zoneid == udp->udp_connp->conn_zoneid)
460					udp_report_item(mp->b_cont, udp);
461			}
462		}
463		mutex_exit(&udpf->uf_lock);
464	}
465	us->us_last_ndd_get_info_time = ddi_get_lbolt();
466	return (0);
467}
468
469/*
470 * Hash list removal routine for udp_t structures.
471 */
472static void
473udp_bind_hash_remove(udp_t *udp, boolean_t caller_holds_lock)
474{
475	udp_t	*udpnext;
476	kmutex_t *lockp;
477	udp_stack_t *us = udp->udp_us;
478
479	if (udp->udp_ptpbhn == NULL)
480		return;
481
482	/*
483	 * Extract the lock pointer in case there are concurrent
484	 * hash_remove's for this instance.
485	 */
486	ASSERT(udp->udp_port != 0);
487	if (!caller_holds_lock) {
488		lockp = &us->us_bind_fanout[UDP_BIND_HASH(udp->udp_port,
489		    us->us_bind_fanout_size)].uf_lock;
490		ASSERT(lockp != NULL);
491		mutex_enter(lockp);
492	}
493	if (udp->udp_ptpbhn != NULL) {
494		udpnext = udp->udp_bind_hash;
495		if (udpnext != NULL) {
496			udpnext->udp_ptpbhn = udp->udp_ptpbhn;
497			udp->udp_bind_hash = NULL;
498		}
499		*udp->udp_ptpbhn = udpnext;
500		udp->udp_ptpbhn = NULL;
501	}
502	if (!caller_holds_lock) {
503		mutex_exit(lockp);
504	}
505}
506
507static void
508udp_bind_hash_insert(udp_fanout_t *uf, udp_t *udp)
509{
510	udp_t	**udpp;
511	udp_t	*udpnext;
512
513	ASSERT(MUTEX_HELD(&uf->uf_lock));
514	ASSERT(udp->udp_ptpbhn == NULL);
515	udpp = &uf->uf_udp;
516	udpnext = udpp[0];
517	if (udpnext != NULL) {
518		/*
519		 * If the new udp bound to the INADDR_ANY address
520		 * and the first one in the list is not bound to
521		 * INADDR_ANY we skip all entries until we find the
522		 * first one bound to INADDR_ANY.
523		 * This makes sure that applications binding to a
524		 * specific address get preference over those binding to
525		 * INADDR_ANY.
526		 */
527		if (V6_OR_V4_INADDR_ANY(udp->udp_bound_v6src) &&
528		    !V6_OR_V4_INADDR_ANY(udpnext->udp_bound_v6src)) {
529			while ((udpnext = udpp[0]) != NULL &&
530			    !V6_OR_V4_INADDR_ANY(
531			    udpnext->udp_bound_v6src)) {
532				udpp = &(udpnext->udp_bind_hash);
533			}
534			if (udpnext != NULL)
535				udpnext->udp_ptpbhn = &udp->udp_bind_hash;
536		} else {
537			udpnext->udp_ptpbhn = &udp->udp_bind_hash;
538		}
539	}
540	udp->udp_bind_hash = udpnext;
541	udp->udp_ptpbhn = udpp;
542	udpp[0] = udp;
543}
544
545/*
546 * This routine is called to handle each O_T_BIND_REQ/T_BIND_REQ message
547 * passed to udp_wput.
548 * It associates a port number and local address with the stream.
549 * The O_T_BIND_REQ/T_BIND_REQ is passed downstream to ip with the UDP
550 * protocol type (IPPROTO_UDP) placed in the message following the address.
551 * A T_BIND_ACK message is passed upstream when ip acknowledges the request.
552 * (Called as writer.)
553 *
554 * Note that UDP over IPv4 and IPv6 sockets can use the same port number
555 * without setting SO_REUSEADDR. This is needed so that they
556 * can be viewed as two independent transport protocols.
557 * However, anonymouns ports are allocated from the same range to avoid
558 * duplicating the us->us_next_port_to_try.
559 */
560static void
561udp_bind(queue_t *q, mblk_t *mp)
562{
563	sin_t		*sin;
564	sin6_t		*sin6;
565	mblk_t		*mp1;
566	in_port_t	port;		/* Host byte order */
567	in_port_t	requested_port;	/* Host byte order */
568	struct T_bind_req *tbr;
569	int		count;
570	in6_addr_t	v6src;
571	boolean_t	bind_to_req_port_only;
572	int		loopmax;
573	udp_fanout_t	*udpf;
574	in_port_t	lport;		/* Network byte order */
575	zoneid_t	zoneid;
576	conn_t		*connp;
577	udp_t		*udp;
578	boolean_t	is_inaddr_any;
579	mlp_type_t	addrtype, mlptype;
580	udp_stack_t	*us;
581
582	connp = Q_TO_CONN(q);
583	udp = connp->conn_udp;
584	us = udp->udp_us;
585	if ((mp->b_wptr - mp->b_rptr) < sizeof (*tbr)) {
586		(void) mi_strlog(q, 1, SL_ERROR|SL_TRACE,
587		    "udp_bind: bad req, len %u",
588		    (uint_t)(mp->b_wptr - mp->b_rptr));
589		udp_err_ack(q, mp, TPROTO, 0);
590		return;
591	}
592	if (udp->udp_state != TS_UNBND) {
593		(void) mi_strlog(q, 1, SL_ERROR|SL_TRACE,
594		    "udp_bind: bad state, %u", udp->udp_state);
595		udp_err_ack(q, mp, TOUTSTATE, 0);
596		return;
597	}
598	/*
599	 * Reallocate the message to make sure we have enough room for an
600	 * address and the protocol type.
601	 */
602	mp1 = reallocb(mp, sizeof (struct T_bind_ack) + sizeof (sin6_t) + 1, 1);
603	if (!mp1) {
604		udp_err_ack(q, mp, TSYSERR, ENOMEM);
605		return;
606	}
607
608	mp = mp1;
609	tbr = (struct T_bind_req *)mp->b_rptr;
610	switch (tbr->ADDR_length) {
611	case 0:			/* Request for a generic port */
612		tbr->ADDR_offset = sizeof (struct T_bind_req);
613		if (udp->udp_family == AF_INET) {
614			tbr->ADDR_length = sizeof (sin_t);
615			sin = (sin_t *)&tbr[1];
616			*sin = sin_null;
617			sin->sin_family = AF_INET;
618			mp->b_wptr = (uchar_t *)&sin[1];
619		} else {
620			ASSERT(udp->udp_family == AF_INET6);
621			tbr->ADDR_length = sizeof (sin6_t);
622			sin6 = (sin6_t *)&tbr[1];
623			*sin6 = sin6_null;
624			sin6->sin6_family = AF_INET6;
625			mp->b_wptr = (uchar_t *)&sin6[1];
626		}
627		port = 0;
628		break;
629
630	case sizeof (sin_t):	/* Complete IPv4 address */
631		sin = (sin_t *)mi_offset_param(mp, tbr->ADDR_offset,
632		    sizeof (sin_t));
633		if (sin == NULL || !OK_32PTR((char *)sin)) {
634			udp_err_ack(q, mp, TSYSERR, EINVAL);
635			return;
636		}
637		if (udp->udp_family != AF_INET ||
638		    sin->sin_family != AF_INET) {
639			udp_err_ack(q, mp, TSYSERR, EAFNOSUPPORT);
640			return;
641		}
642		port = ntohs(sin->sin_port);
643		break;
644
645	case sizeof (sin6_t):	/* complete IPv6 address */
646		sin6 = (sin6_t *)mi_offset_param(mp, tbr->ADDR_offset,
647		    sizeof (sin6_t));
648		if (sin6 == NULL || !OK_32PTR((char *)sin6)) {
649			udp_err_ack(q, mp, TSYSERR, EINVAL);
650			return;
651		}
652		if (udp->udp_family != AF_INET6 ||
653		    sin6->sin6_family != AF_INET6) {
654			udp_err_ack(q, mp, TSYSERR, EAFNOSUPPORT);
655			return;
656		}
657		port = ntohs(sin6->sin6_port);
658		break;
659
660	default:		/* Invalid request */
661		(void) mi_strlog(q, 1, SL_ERROR|SL_TRACE,
662		    "udp_bind: bad ADDR_length length %u", tbr->ADDR_length);
663		udp_err_ack(q, mp, TBADADDR, 0);
664		return;
665	}
666
667	requested_port = port;
668
669	if (requested_port == 0 || tbr->PRIM_type == O_T_BIND_REQ)
670		bind_to_req_port_only = B_FALSE;
671	else			/* T_BIND_REQ and requested_port != 0 */
672		bind_to_req_port_only = B_TRUE;
673
674	if (requested_port == 0) {
675		/*
676		 * If the application passed in zero for the port number, it
677		 * doesn't care which port number we bind to. Get one in the
678		 * valid range.
679		 */
680		if (udp->udp_anon_priv_bind) {
681			port = udp_get_next_priv_port(udp);
682		} else {
683			port = udp_update_next_port(udp,
684			    us->us_next_port_to_try, B_TRUE);
685		}
686	} else {
687		/*
688		 * If the port is in the well-known privileged range,
689		 * make sure the caller was privileged.
690		 */
691		int i;
692		boolean_t priv = B_FALSE;
693
694		if (port < us->us_smallest_nonpriv_port) {
695			priv = B_TRUE;
696		} else {
697			for (i = 0; i < us->us_num_epriv_ports; i++) {
698				if (port == us->us_epriv_ports[i]) {
699					priv = B_TRUE;
700					break;
701				}
702			}
703		}
704
705		if (priv) {
706			cred_t *cr = DB_CREDDEF(mp, connp->conn_cred);
707
708			if (secpolicy_net_privaddr(cr, port,
709			    IPPROTO_UDP) != 0) {
710				udp_err_ack(q, mp, TACCES, 0);
711				return;
712			}
713		}
714	}
715
716	if (port == 0) {
717		udp_err_ack(q, mp, TNOADDR, 0);
718		return;
719	}
720
721	/*
722	 * The state must be TS_UNBND. TPI mandates that users must send
723	 * TPI primitives only 1 at a time and wait for the response before
724	 * sending the next primitive.
725	 */
726	rw_enter(&udp->udp_rwlock, RW_WRITER);
727	if (udp->udp_state != TS_UNBND || udp->udp_pending_op != -1) {
728		rw_exit(&udp->udp_rwlock);
729		(void) mi_strlog(q, 1, SL_ERROR|SL_TRACE,
730		    "udp_bind: bad state, %u", udp->udp_state);
731		udp_err_ack(q, mp, TOUTSTATE, 0);
732		return;
733	}
734	udp->udp_pending_op = tbr->PRIM_type;
735	/*
736	 * Copy the source address into our udp structure. This address
737	 * may still be zero; if so, IP will fill in the correct address
738	 * each time an outbound packet is passed to it. Since the udp is
739	 * not yet in the bind hash list, we don't grab the uf_lock to
740	 * change udp_ipversion
741	 */
742	if (udp->udp_family == AF_INET) {
743		ASSERT(sin != NULL);
744		ASSERT(udp->udp_ipversion == IPV4_VERSION);
745		udp->udp_max_hdr_len = IP_SIMPLE_HDR_LENGTH + UDPH_SIZE +
746		    udp->udp_ip_snd_options_len;
747		IN6_IPADDR_TO_V4MAPPED(sin->sin_addr.s_addr, &v6src);
748	} else {
749		ASSERT(sin6 != NULL);
750		v6src = sin6->sin6_addr;
751		if (IN6_IS_ADDR_V4MAPPED(&v6src)) {
752			/*
753			 * no need to hold the uf_lock to set the udp_ipversion
754			 * since we are not yet in the fanout list
755			 */
756			udp->udp_ipversion = IPV4_VERSION;
757			udp->udp_max_hdr_len = IP_SIMPLE_HDR_LENGTH +
758			    UDPH_SIZE + udp->udp_ip_snd_options_len;
759		} else {
760			udp->udp_ipversion = IPV6_VERSION;
761			udp->udp_max_hdr_len = udp->udp_sticky_hdrs_len;
762		}
763	}
764
765	/*
766	 * If udp_reuseaddr is not set, then we have to make sure that
767	 * the IP address and port number the application requested
768	 * (or we selected for the application) is not being used by
769	 * another stream.  If another stream is already using the
770	 * requested IP address and port, the behavior depends on
771	 * "bind_to_req_port_only". If set the bind fails; otherwise we
772	 * search for any an unused port to bind to the the stream.
773	 *
774	 * As per the BSD semantics, as modified by the Deering multicast
775	 * changes, if udp_reuseaddr is set, then we allow multiple binds
776	 * to the same port independent of the local IP address.
777	 *
778	 * This is slightly different than in SunOS 4.X which did not
779	 * support IP multicast. Note that the change implemented by the
780	 * Deering multicast code effects all binds - not only binding
781	 * to IP multicast addresses.
782	 *
783	 * Note that when binding to port zero we ignore SO_REUSEADDR in
784	 * order to guarantee a unique port.
785	 */
786
787	count = 0;
788	if (udp->udp_anon_priv_bind) {
789		/*
790		 * loopmax = (IPPORT_RESERVED-1) -
791		 *    us->us_min_anonpriv_port + 1
792		 */
793		loopmax = IPPORT_RESERVED - us->us_min_anonpriv_port;
794	} else {
795		loopmax = us->us_largest_anon_port -
796		    us->us_smallest_anon_port + 1;
797	}
798
799	is_inaddr_any = V6_OR_V4_INADDR_ANY(v6src);
800	zoneid = connp->conn_zoneid;
801
802	for (;;) {
803		udp_t		*udp1;
804		boolean_t	found_exclbind = B_FALSE;
805
806		/*
807		 * Walk through the list of udp streams bound to
808		 * requested port with the same IP address.
809		 */
810		lport = htons(port);
811		udpf = &us->us_bind_fanout[UDP_BIND_HASH(lport,
812		    us->us_bind_fanout_size)];
813		mutex_enter(&udpf->uf_lock);
814		for (udp1 = udpf->uf_udp; udp1 != NULL;
815		    udp1 = udp1->udp_bind_hash) {
816			if (lport != udp1->udp_port)
817				continue;
818
819			/*
820			 * On a labeled system, we must treat bindings to ports
821			 * on shared IP addresses by sockets with MAC exemption
822			 * privilege as being in all zones, as there's
823			 * otherwise no way to identify the right receiver.
824			 */
825			if (!(IPCL_ZONE_MATCH(udp1->udp_connp, zoneid) ||
826			    IPCL_ZONE_MATCH(connp,
827			    udp1->udp_connp->conn_zoneid)) &&
828			    !connp->conn_mac_exempt && \
829			    !udp1->udp_connp->conn_mac_exempt)
830				continue;
831
832			/*
833			 * If UDP_EXCLBIND is set for either the bound or
834			 * binding endpoint, the semantics of bind
835			 * is changed according to the following chart.
836			 *
837			 * spec = specified address (v4 or v6)
838			 * unspec = unspecified address (v4 or v6)
839			 * A = specified addresses are different for endpoints
840			 *
841			 * bound	bind to		allowed?
842			 * -------------------------------------
843			 * unspec	unspec		no
844			 * unspec	spec		no
845			 * spec		unspec		no
846			 * spec		spec		yes if A
847			 *
848			 * For labeled systems, SO_MAC_EXEMPT behaves the same
849			 * as UDP_EXCLBIND, except that zoneid is ignored.
850			 */
851			if (udp1->udp_exclbind || udp->udp_exclbind ||
852			    udp1->udp_connp->conn_mac_exempt ||
853			    connp->conn_mac_exempt) {
854				if (V6_OR_V4_INADDR_ANY(
855				    udp1->udp_bound_v6src) ||
856				    is_inaddr_any ||
857				    IN6_ARE_ADDR_EQUAL(&udp1->udp_bound_v6src,
858				    &v6src)) {
859					found_exclbind = B_TRUE;
860					break;
861				}
862				continue;
863			}
864
865			/*
866			 * Check ipversion to allow IPv4 and IPv6 sockets to
867			 * have disjoint port number spaces.
868			 */
869			if (udp->udp_ipversion != udp1->udp_ipversion) {
870
871				/*
872				 * On the first time through the loop, if the
873				 * the user intentionally specified a
874				 * particular port number, then ignore any
875				 * bindings of the other protocol that may
876				 * conflict. This allows the user to bind IPv6
877				 * alone and get both v4 and v6, or bind both
878				 * both and get each seperately. On subsequent
879				 * times through the loop, we're checking a
880				 * port that we chose (not the user) and thus
881				 * we do not allow casual duplicate bindings.
882				 */
883				if (count == 0 && requested_port != 0)
884					continue;
885			}
886
887			/*
888			 * No difference depending on SO_REUSEADDR.
889			 *
890			 * If existing port is bound to a
891			 * non-wildcard IP address and
892			 * the requesting stream is bound to
893			 * a distinct different IP addresses
894			 * (non-wildcard, also), keep going.
895			 */
896			if (!is_inaddr_any &&
897			    !V6_OR_V4_INADDR_ANY(udp1->udp_bound_v6src) &&
898			    !IN6_ARE_ADDR_EQUAL(&udp1->udp_bound_v6src,
899			    &v6src)) {
900				continue;
901			}
902			break;
903		}
904
905		if (!found_exclbind &&
906		    (udp->udp_reuseaddr && requested_port != 0)) {
907			break;
908		}
909
910		if (udp1 == NULL) {
911			/*
912			 * No other stream has this IP address
913			 * and port number. We can use it.
914			 */
915			break;
916		}
917		mutex_exit(&udpf->uf_lock);
918		if (bind_to_req_port_only) {
919			/*
920			 * We get here only when requested port
921			 * is bound (and only first  of the for()
922			 * loop iteration).
923			 *
924			 * The semantics of this bind request
925			 * require it to fail so we return from
926			 * the routine (and exit the loop).
927			 *
928			 */
929			udp->udp_pending_op = -1;
930			rw_exit(&udp->udp_rwlock);
931			udp_err_ack(q, mp, TADDRBUSY, 0);
932			return;
933		}
934
935		if (udp->udp_anon_priv_bind) {
936			port = udp_get_next_priv_port(udp);
937		} else {
938			if ((count == 0) && (requested_port != 0)) {
939				/*
940				 * If the application wants us to find
941				 * a port, get one to start with. Set
942				 * requested_port to 0, so that we will
943				 * update us->us_next_port_to_try below.
944				 */
945				port = udp_update_next_port(udp,
946				    us->us_next_port_to_try, B_TRUE);
947				requested_port = 0;
948			} else {
949				port = udp_update_next_port(udp, port + 1,
950				    B_FALSE);
951			}
952		}
953
954		if (port == 0 || ++count >= loopmax) {
955			/*
956			 * We've tried every possible port number and
957			 * there are none available, so send an error
958			 * to the user.
959			 */
960			udp->udp_pending_op = -1;
961			rw_exit(&udp->udp_rwlock);
962			udp_err_ack(q, mp, TNOADDR, 0);
963			return;
964		}
965	}
966
967	/*
968	 * Copy the source address into our udp structure.  This address
969	 * may still be zero; if so, ip will fill in the correct address
970	 * each time an outbound packet is passed to it.
971	 * If we are binding to a broadcast or multicast address then
972	 * udp_bind_ack will clear the source address when it receives
973	 * the T_BIND_ACK.
974	 */
975	udp->udp_v6src = udp->udp_bound_v6src = v6src;
976	udp->udp_port = lport;
977	/*
978	 * Now reset the the next anonymous port if the application requested
979	 * an anonymous port, or we handed out the next anonymous port.
980	 */
981	if ((requested_port == 0) && (!udp->udp_anon_priv_bind)) {
982		us->us_next_port_to_try = port + 1;
983	}
984
985	/* Initialize the O_T_BIND_REQ/T_BIND_REQ for ip. */
986	if (udp->udp_family == AF_INET) {
987		sin->sin_port = udp->udp_port;
988	} else {
989		int error;
990
991		sin6->sin6_port = udp->udp_port;
992		/* Rebuild the header template */
993		error = udp_build_hdrs(udp);
994		if (error != 0) {
995			udp->udp_pending_op = -1;
996			rw_exit(&udp->udp_rwlock);
997			mutex_exit(&udpf->uf_lock);
998			udp_err_ack(q, mp, TSYSERR, error);
999			return;
1000		}
1001	}
1002	udp->udp_state = TS_IDLE;
1003	udp_bind_hash_insert(udpf, udp);
1004	mutex_exit(&udpf->uf_lock);
1005	rw_exit(&udp->udp_rwlock);
1006
1007	if (cl_inet_bind) {
1008		/*
1009		 * Running in cluster mode - register bind information
1010		 */
1011		if (udp->udp_ipversion == IPV4_VERSION) {
1012			(*cl_inet_bind)(IPPROTO_UDP, AF_INET,
1013			    (uint8_t *)(&V4_PART_OF_V6(udp->udp_v6src)),
1014			    (in_port_t)udp->udp_port);
1015		} else {
1016			(*cl_inet_bind)(IPPROTO_UDP, AF_INET6,
1017			    (uint8_t *)&(udp->udp_v6src),
1018			    (in_port_t)udp->udp_port);
1019		}
1020
1021	}
1022
1023	connp->conn_anon_port = (is_system_labeled() && requested_port == 0);
1024	if (is_system_labeled() && (!connp->conn_anon_port ||
1025	    connp->conn_anon_mlp)) {
1026		uint16_t mlpport;
1027		cred_t *cr = connp->conn_cred;
1028		zone_t *zone;
1029
1030		zone = crgetzone(cr);
1031		connp->conn_mlp_type = udp->udp_recvucred ? mlptBoth :
1032		    mlptSingle;
1033		addrtype = tsol_mlp_addr_type(zone->zone_id, IPV6_VERSION,
1034		    &v6src, us->us_netstack->netstack_ip);
1035		if (addrtype == mlptSingle) {
1036			rw_enter(&udp->udp_rwlock, RW_WRITER);
1037			udp->udp_pending_op = -1;
1038			rw_exit(&udp->udp_rwlock);
1039			udp_err_ack(q, mp, TNOADDR, 0);
1040			connp->conn_anon_port = B_FALSE;
1041			connp->conn_mlp_type = mlptSingle;
1042			return;
1043		}
1044		mlpport = connp->conn_anon_port ? PMAPPORT : port;
1045		mlptype = tsol_mlp_port_type(zone, IPPROTO_UDP, mlpport,
1046		    addrtype);
1047		if (mlptype != mlptSingle &&
1048		    (connp->conn_mlp_type == mlptSingle ||
1049		    secpolicy_net_bindmlp(cr) != 0)) {
1050			if (udp->udp_debug) {
1051				(void) strlog(UDP_MOD_ID, 0, 1,
1052				    SL_ERROR|SL_TRACE,
1053				    "udp_bind: no priv for multilevel port %d",
1054				    mlpport);
1055			}
1056			rw_enter(&udp->udp_rwlock, RW_WRITER);
1057			udp->udp_pending_op = -1;
1058			rw_exit(&udp->udp_rwlock);
1059			udp_err_ack(q, mp, TACCES, 0);
1060			connp->conn_anon_port = B_FALSE;
1061			connp->conn_mlp_type = mlptSingle;
1062			return;
1063		}
1064
1065		/*
1066		 * If we're specifically binding a shared IP address and the
1067		 * port is MLP on shared addresses, then check to see if this
1068		 * zone actually owns the MLP.  Reject if not.
1069		 */
1070		if (mlptype == mlptShared && addrtype == mlptShared) {
1071			/*
1072			 * No need to handle exclusive-stack zones since
1073			 * ALL_ZONES only applies to the shared stack.
1074			 */
1075			zoneid_t mlpzone;
1076
1077			mlpzone = tsol_mlp_findzone(IPPROTO_UDP,
1078			    htons(mlpport));
1079			if (connp->conn_zoneid != mlpzone) {
1080				if (udp->udp_debug) {
1081					(void) strlog(UDP_MOD_ID, 0, 1,
1082					    SL_ERROR|SL_TRACE,
1083					    "udp_bind: attempt to bind port "
1084					    "%d on shared addr in zone %d "
1085					    "(should be %d)",
1086					    mlpport, connp->conn_zoneid,
1087					    mlpzone);
1088				}
1089				rw_enter(&udp->udp_rwlock, RW_WRITER);
1090				udp->udp_pending_op = -1;
1091				rw_exit(&udp->udp_rwlock);
1092				udp_err_ack(q, mp, TACCES, 0);
1093				connp->conn_anon_port = B_FALSE;
1094				connp->conn_mlp_type = mlptSingle;
1095				return;
1096			}
1097		}
1098		if (connp->conn_anon_port) {
1099			int error;
1100
1101			error = tsol_mlp_anon(zone, mlptype, connp->conn_ulp,
1102			    port, B_TRUE);
1103			if (error != 0) {
1104				if (udp->udp_debug) {
1105					(void) strlog(UDP_MOD_ID, 0, 1,
1106					    SL_ERROR|SL_TRACE,
1107					    "udp_bind: cannot establish anon "
1108					    "MLP for port %d", port);
1109				}
1110				rw_enter(&udp->udp_rwlock, RW_WRITER);
1111				udp->udp_pending_op = -1;
1112				rw_exit(&udp->udp_rwlock);
1113				udp_err_ack(q, mp, TACCES, 0);
1114				connp->conn_anon_port = B_FALSE;
1115				connp->conn_mlp_type = mlptSingle;
1116				return;
1117			}
1118		}
1119		connp->conn_mlp_type = mlptype;
1120	}
1121
1122	/* Pass the protocol number in the message following the address. */
1123	*mp->b_wptr++ = IPPROTO_UDP;
1124	if (!V6_OR_V4_INADDR_ANY(udp->udp_v6src)) {
1125		/*
1126		 * Append a request for an IRE if udp_v6src not
1127		 * zero (IPv4 - INADDR_ANY, or IPv6 - all-zeroes address).
1128		 */
1129		mp->b_cont = allocb(sizeof (ire_t), BPRI_HI);
1130		if (!mp->b_cont) {
1131			rw_enter(&udp->udp_rwlock, RW_WRITER);
1132			udp->udp_pending_op = -1;
1133			rw_exit(&udp->udp_rwlock);
1134			udp_err_ack(q, mp, TSYSERR, ENOMEM);
1135			return;
1136		}
1137		mp->b_cont->b_wptr += sizeof (ire_t);
1138		mp->b_cont->b_datap->db_type = IRE_DB_REQ_TYPE;
1139	}
1140	if (udp->udp_family == AF_INET6)
1141		mp = ip_bind_v6(q, mp, connp, NULL);
1142	else
1143		mp = ip_bind_v4(q, mp, connp);
1144
1145	/* The above return NULL if the bind needs to be deferred */
1146	if (mp != NULL)
1147		udp_bind_result(connp, mp);
1148	else
1149		CONN_INC_REF(connp);
1150}
1151
1152/*
1153 * This is called from ip_wput_nondata to handle the results of a
1154 * deferred UDP bind. It is called once the bind has been completed.
1155 */
1156void
1157udp_resume_bind(conn_t *connp, mblk_t *mp)
1158{
1159	ASSERT(connp != NULL && IPCL_IS_UDP(connp));
1160
1161	udp_bind_result(connp, mp);
1162
1163	CONN_OPER_PENDING_DONE(connp);
1164}
1165
1166/*
1167 * This routine handles each T_CONN_REQ message passed to udp.  It
1168 * associates a default destination address with the stream.
1169 *
1170 * This routine sends down a T_BIND_REQ to IP with the following mblks:
1171 *	T_BIND_REQ	- specifying local and remote address/port
1172 *	IRE_DB_REQ_TYPE	- to get an IRE back containing ire_type and src
1173 *	T_OK_ACK	- for the T_CONN_REQ
1174 *	T_CONN_CON	- to keep the TPI user happy
1175 *
1176 * The connect completes in udp_bind_result.
1177 * When a T_BIND_ACK is received information is extracted from the IRE
1178 * and the two appended messages are sent to the TPI user.
1179 * Should udp_bind_result receive T_ERROR_ACK for the T_BIND_REQ it will
1180 * convert it to an error ack for the appropriate primitive.
1181 */
1182static void
1183udp_connect(queue_t *q, mblk_t *mp)
1184{
1185	sin6_t	*sin6;
1186	sin_t	*sin;
1187	struct T_conn_req	*tcr;
1188	in6_addr_t v6dst;
1189	ipaddr_t v4dst;
1190	uint16_t dstport;
1191	uint32_t flowinfo;
1192	mblk_t	*mp1, *mp2;
1193	udp_fanout_t	*udpf;
1194	udp_t	*udp, *udp1;
1195	ushort_t	ipversion;
1196	udp_stack_t	*us;
1197	conn_t		*connp = Q_TO_CONN(q);
1198
1199	udp = connp->conn_udp;
1200	tcr = (struct T_conn_req *)mp->b_rptr;
1201	us = udp->udp_us;
1202
1203	/* A bit of sanity checking */
1204	if ((mp->b_wptr - mp->b_rptr) < sizeof (struct T_conn_req)) {
1205		udp_err_ack(q, mp, TPROTO, 0);
1206		return;
1207	}
1208
1209	if (tcr->OPT_length != 0) {
1210		udp_err_ack(q, mp, TBADOPT, 0);
1211		return;
1212	}
1213
1214	/*
1215	 * Determine packet type based on type of address passed in
1216	 * the request should contain an IPv4 or IPv6 address.
1217	 * Make sure that address family matches the type of
1218	 * family of the the address passed down
1219	 */
1220	switch (tcr->DEST_length) {
1221	default:
1222		udp_err_ack(q, mp, TBADADDR, 0);
1223		return;
1224
1225	case sizeof (sin_t):
1226		sin = (sin_t *)mi_offset_param(mp, tcr->DEST_offset,
1227		    sizeof (sin_t));
1228		if (sin == NULL || !OK_32PTR((char *)sin)) {
1229			udp_err_ack(q, mp, TSYSERR, EINVAL);
1230			return;
1231		}
1232		if (udp->udp_family != AF_INET ||
1233		    sin->sin_family != AF_INET) {
1234			udp_err_ack(q, mp, TSYSERR, EAFNOSUPPORT);
1235			return;
1236		}
1237		v4dst = sin->sin_addr.s_addr;
1238		dstport = sin->sin_port;
1239		IN6_IPADDR_TO_V4MAPPED(v4dst, &v6dst);
1240		ASSERT(udp->udp_ipversion == IPV4_VERSION);
1241		ipversion = IPV4_VERSION;
1242		break;
1243
1244	case sizeof (sin6_t):
1245		sin6 = (sin6_t *)mi_offset_param(mp, tcr->DEST_offset,
1246		    sizeof (sin6_t));
1247		if (sin6 == NULL || !OK_32PTR((char *)sin6)) {
1248			udp_err_ack(q, mp, TSYSERR, EINVAL);
1249			return;
1250		}
1251		if (udp->udp_family != AF_INET6 ||
1252		    sin6->sin6_family != AF_INET6) {
1253			udp_err_ack(q, mp, TSYSERR, EAFNOSUPPORT);
1254			return;
1255		}
1256		v6dst = sin6->sin6_addr;
1257		dstport = sin6->sin6_port;
1258		if (IN6_IS_ADDR_V4MAPPED(&v6dst)) {
1259			IN6_V4MAPPED_TO_IPADDR(&v6dst, v4dst);
1260			ipversion = IPV4_VERSION;
1261			flowinfo = 0;
1262		} else {
1263			ipversion = IPV6_VERSION;
1264			flowinfo = sin6->sin6_flowinfo;
1265		}
1266		break;
1267	}
1268	if (dstport == 0) {
1269		udp_err_ack(q, mp, TBADADDR, 0);
1270		return;
1271	}
1272
1273	rw_enter(&udp->udp_rwlock, RW_WRITER);
1274
1275	/*
1276	 * This UDP must have bound to a port already before doing a connect.
1277	 * TPI mandates that users must send TPI primitives only 1 at a time
1278	 * and wait for the response before sending the next primitive.
1279	 */
1280	if (udp->udp_state == TS_UNBND || udp->udp_pending_op != -1) {
1281		rw_exit(&udp->udp_rwlock);
1282		(void) mi_strlog(q, 1, SL_ERROR|SL_TRACE,
1283		    "udp_connect: bad state, %u", udp->udp_state);
1284		udp_err_ack(q, mp, TOUTSTATE, 0);
1285		return;
1286	}
1287	udp->udp_pending_op = T_CONN_REQ;
1288	ASSERT(udp->udp_port != 0 && udp->udp_ptpbhn != NULL);
1289
1290	if (ipversion == IPV4_VERSION) {
1291		udp->udp_max_hdr_len = IP_SIMPLE_HDR_LENGTH + UDPH_SIZE +
1292		    udp->udp_ip_snd_options_len;
1293	} else {
1294		udp->udp_max_hdr_len = udp->udp_sticky_hdrs_len;
1295	}
1296
1297	udpf = &us->us_bind_fanout[UDP_BIND_HASH(udp->udp_port,
1298	    us->us_bind_fanout_size)];
1299
1300	mutex_enter(&udpf->uf_lock);
1301	if (udp->udp_state == TS_DATA_XFER) {
1302		/* Already connected - clear out state */
1303		udp->udp_v6src = udp->udp_bound_v6src;
1304		udp->udp_state = TS_IDLE;
1305	}
1306
1307	/*
1308	 * Create a default IP header with no IP options.
1309	 */
1310	udp->udp_dstport = dstport;
1311	udp->udp_ipversion = ipversion;
1312	if (ipversion == IPV4_VERSION) {
1313		/*
1314		 * Interpret a zero destination to mean loopback.
1315		 * Update the T_CONN_REQ (sin/sin6) since it is used to
1316		 * generate the T_CONN_CON.
1317		 */
1318		if (v4dst == INADDR_ANY) {
1319			v4dst = htonl(INADDR_LOOPBACK);
1320			IN6_IPADDR_TO_V4MAPPED(v4dst, &v6dst);
1321			if (udp->udp_family == AF_INET) {
1322				sin->sin_addr.s_addr = v4dst;
1323			} else {
1324				sin6->sin6_addr = v6dst;
1325			}
1326		}
1327		udp->udp_v6dst = v6dst;
1328		udp->udp_flowinfo = 0;
1329
1330		/*
1331		 * If the destination address is multicast and
1332		 * an outgoing multicast interface has been set,
1333		 * use the address of that interface as our
1334		 * source address if no source address has been set.
1335		 */
1336		if (V4_PART_OF_V6(udp->udp_v6src) == INADDR_ANY &&
1337		    CLASSD(v4dst) &&
1338		    udp->udp_multicast_if_addr != INADDR_ANY) {
1339			IN6_IPADDR_TO_V4MAPPED(udp->udp_multicast_if_addr,
1340			    &udp->udp_v6src);
1341		}
1342	} else {
1343		ASSERT(udp->udp_ipversion == IPV6_VERSION);
1344		/*
1345		 * Interpret a zero destination to mean loopback.
1346		 * Update the T_CONN_REQ (sin/sin6) since it is used to
1347		 * generate the T_CONN_CON.
1348		 */
1349		if (IN6_IS_ADDR_UNSPECIFIED(&v6dst)) {
1350			v6dst = ipv6_loopback;
1351			sin6->sin6_addr = v6dst;
1352		}
1353		udp->udp_v6dst = v6dst;
1354		udp->udp_flowinfo = flowinfo;
1355		/*
1356		 * If the destination address is multicast and
1357		 * an outgoing multicast interface has been set,
1358		 * then the ip bind logic will pick the correct source
1359		 * address (i.e. matching the outgoing multicast interface).
1360		 */
1361	}
1362
1363	/*
1364	 * Verify that the src/port/dst/port is unique for all
1365	 * connections in TS_DATA_XFER
1366	 */
1367	for (udp1 = udpf->uf_udp; udp1 != NULL; udp1 = udp1->udp_bind_hash) {
1368		if (udp1->udp_state != TS_DATA_XFER)
1369			continue;
1370		if (udp->udp_port != udp1->udp_port ||
1371		    udp->udp_ipversion != udp1->udp_ipversion ||
1372		    dstport != udp1->udp_dstport ||
1373		    !IN6_ARE_ADDR_EQUAL(&udp->udp_v6src, &udp1->udp_v6src) ||
1374		    !IN6_ARE_ADDR_EQUAL(&v6dst, &udp1->udp_v6dst) ||
1375		    !(IPCL_ZONE_MATCH(udp->udp_connp,
1376		    udp1->udp_connp->conn_zoneid) ||
1377		    IPCL_ZONE_MATCH(udp1->udp_connp,
1378		    udp->udp_connp->conn_zoneid)))
1379			continue;
1380		mutex_exit(&udpf->uf_lock);
1381		udp->udp_pending_op = -1;
1382		rw_exit(&udp->udp_rwlock);
1383		udp_err_ack(q, mp, TBADADDR, 0);
1384		return;
1385	}
1386	udp->udp_state = TS_DATA_XFER;
1387	mutex_exit(&udpf->uf_lock);
1388
1389	/*
1390	 * Send down bind to IP to verify that there is a route
1391	 * and to determine the source address.
1392	 * This will come back as T_BIND_ACK with an IRE_DB_TYPE in rput.
1393	 */
1394	if (udp->udp_family == AF_INET)
1395		mp1 = udp_ip_bind_mp(udp, O_T_BIND_REQ, sizeof (ipa_conn_t));
1396	else
1397		mp1 = udp_ip_bind_mp(udp, O_T_BIND_REQ, sizeof (ipa6_conn_t));
1398	if (mp1 == NULL) {
1399bind_failed:
1400		mutex_enter(&udpf->uf_lock);
1401		udp->udp_state = TS_IDLE;
1402		udp->udp_pending_op = -1;
1403		mutex_exit(&udpf->uf_lock);
1404		rw_exit(&udp->udp_rwlock);
1405		udp_err_ack(q, mp, TSYSERR, ENOMEM);
1406		return;
1407	}
1408
1409	rw_exit(&udp->udp_rwlock);
1410	/*
1411	 * We also have to send a connection confirmation to
1412	 * keep TLI happy. Prepare it for udp_bind_result.
1413	 */
1414	if (udp->udp_family == AF_INET)
1415		mp2 = mi_tpi_conn_con(NULL, (char *)sin,
1416		    sizeof (*sin), NULL, 0);
1417	else
1418		mp2 = mi_tpi_conn_con(NULL, (char *)sin6,
1419		    sizeof (*sin6), NULL, 0);
1420	if (mp2 == NULL) {
1421		freemsg(mp1);
1422		rw_enter(&udp->udp_rwlock, RW_WRITER);
1423		goto bind_failed;
1424	}
1425
1426	mp = mi_tpi_ok_ack_alloc(mp);
1427	if (mp == NULL) {
1428		/* Unable to reuse the T_CONN_REQ for the ack. */
1429		freemsg(mp2);
1430		rw_enter(&udp->udp_rwlock, RW_WRITER);
1431		mutex_enter(&udpf->uf_lock);
1432		udp->udp_state = TS_IDLE;
1433		udp->udp_pending_op = -1;
1434		mutex_exit(&udpf->uf_lock);
1435		rw_exit(&udp->udp_rwlock);
1436		udp_err_ack_prim(q, mp1, T_CONN_REQ, TSYSERR, ENOMEM);
1437		return;
1438	}
1439
1440	/* Hang onto the T_OK_ACK and T_CONN_CON for later. */
1441	linkb(mp1, mp);
1442	linkb(mp1, mp2);
1443
1444	mblk_setcred(mp1, connp->conn_cred);
1445	if (udp->udp_family == AF_INET)
1446		mp1 = ip_bind_v4(q, mp1, connp);
1447	else
1448		mp1 = ip_bind_v6(q, mp1, connp, NULL);
1449
1450	/* The above return NULL if the bind needs to be deferred */
1451	if (mp1 != NULL)
1452		udp_bind_result(connp, mp1);
1453	else
1454		CONN_INC_REF(connp);
1455}
1456
1457static int
1458udp_close(queue_t *q)
1459{
1460	conn_t	*connp = (conn_t *)q->q_ptr;
1461	udp_t	*udp;
1462
1463	ASSERT(connp != NULL && IPCL_IS_UDP(connp));
1464	udp = connp->conn_udp;
1465
1466	udp_quiesce_conn(connp);
1467	ip_quiesce_conn(connp);
1468	/*
1469	 * Disable read-side synchronous stream
1470	 * interface and drain any queued data.
1471	 */
1472	udp_rcv_drain(q, udp, B_TRUE);
1473	ASSERT(!udp->udp_direct_sockfs);
1474
1475	qprocsoff(q);
1476
1477	ASSERT(udp->udp_rcv_cnt == 0);
1478	ASSERT(udp->udp_rcv_msgcnt == 0);
1479	ASSERT(udp->udp_rcv_list_head == NULL);
1480	ASSERT(udp->udp_rcv_list_tail == NULL);
1481
1482	udp_close_free(connp);
1483
1484	/*
1485	 * Now we are truly single threaded on this stream, and can
1486	 * delete the things hanging off the connp, and finally the connp.
1487	 * We removed this connp from the fanout list, it cannot be
1488	 * accessed thru the fanouts, and we already waited for the
1489	 * conn_ref to drop to 0. We are already in close, so
1490	 * there cannot be any other thread from the top. qprocsoff
1491	 * has completed, and service has completed or won't run in
1492	 * future.
1493	 */
1494	ASSERT(connp->conn_ref == 1);
1495	inet_minor_free(connp->conn_minor_arena, connp->conn_dev);
1496	connp->conn_ref--;
1497	ipcl_conn_destroy(connp);
1498
1499	q->q_ptr = WR(q)->q_ptr = NULL;
1500	return (0);
1501}
1502
1503/*
1504 * Called in the close path to quiesce the conn
1505 */
1506void
1507udp_quiesce_conn(conn_t *connp)
1508{
1509	udp_t	*udp = connp->conn_udp;
1510
1511	if (cl_inet_unbind != NULL && udp->udp_state == TS_IDLE) {
1512		/*
1513		 * Running in cluster mode - register unbind information
1514		 */
1515		if (udp->udp_ipversion == IPV4_VERSION) {
1516			(*cl_inet_unbind)(IPPROTO_UDP, AF_INET,
1517			    (uint8_t *)(&(V4_PART_OF_V6(udp->udp_v6src))),
1518			    (in_port_t)udp->udp_port);
1519		} else {
1520			(*cl_inet_unbind)(IPPROTO_UDP, AF_INET6,
1521			    (uint8_t *)(&(udp->udp_v6src)),
1522			    (in_port_t)udp->udp_port);
1523		}
1524	}
1525
1526	udp_bind_hash_remove(udp, B_FALSE);
1527
1528}
1529
1530void
1531udp_close_free(conn_t *connp)
1532{
1533	udp_t *udp = connp->conn_udp;
1534
1535	/* If there are any options associated with the stream, free them. */
1536	if (udp->udp_ip_snd_options != NULL) {
1537		mi_free((char *)udp->udp_ip_snd_options);
1538		udp->udp_ip_snd_options = NULL;
1539		udp->udp_ip_snd_options_len = 0;
1540	}
1541
1542	if (udp->udp_ip_rcv_options != NULL) {
1543		mi_free((char *)udp->udp_ip_rcv_options);
1544		udp->udp_ip_rcv_options = NULL;
1545		udp->udp_ip_rcv_options_len = 0;
1546	}
1547
1548	/* Free memory associated with sticky options */
1549	if (udp->udp_sticky_hdrs_len != 0) {
1550		kmem_free(udp->udp_sticky_hdrs,
1551		    udp->udp_sticky_hdrs_len);
1552		udp->udp_sticky_hdrs = NULL;
1553		udp->udp_sticky_hdrs_len = 0;
1554	}
1555
1556	ip6_pkt_free(&udp->udp_sticky_ipp);
1557
1558	/*
1559	 * Clear any fields which the kmem_cache constructor clears.
1560	 * Only udp_connp needs to be preserved.
1561	 * TBD: We should make this more efficient to avoid clearing
1562	 * everything.
1563	 */
1564	ASSERT(udp->udp_connp == connp);
1565	bzero(udp, sizeof (udp_t));
1566	udp->udp_connp = connp;
1567}
1568
1569/*
1570 * This routine handles each T_DISCON_REQ message passed to udp
1571 * as an indicating that UDP is no longer connected. This results
1572 * in sending a T_BIND_REQ to IP to restore the binding to just
1573 * the local address/port.
1574 *
1575 * This routine sends down a T_BIND_REQ to IP with the following mblks:
1576 *	T_BIND_REQ	- specifying just the local address/port
1577 *	T_OK_ACK	- for the T_DISCON_REQ
1578 *
1579 * The disconnect completes in udp_bind_result.
1580 * When a T_BIND_ACK is received the appended T_OK_ACK is sent to the TPI user.
1581 * Should udp_bind_result receive T_ERROR_ACK for the T_BIND_REQ it will
1582 * convert it to an error ack for the appropriate primitive.
1583 */
1584static void
1585udp_disconnect(queue_t *q, mblk_t *mp)
1586{
1587	udp_t	*udp;
1588	mblk_t	*mp1;
1589	udp_fanout_t *udpf;
1590	udp_stack_t *us;
1591	conn_t	*connp = Q_TO_CONN(q);
1592
1593	udp = connp->conn_udp;
1594	us = udp->udp_us;
1595	rw_enter(&udp->udp_rwlock, RW_WRITER);
1596	if (udp->udp_state != TS_DATA_XFER || udp->udp_pending_op != -1) {
1597		rw_exit(&udp->udp_rwlock);
1598		(void) mi_strlog(q, 1, SL_ERROR|SL_TRACE,
1599		    "udp_disconnect: bad state, %u", udp->udp_state);
1600		udp_err_ack(q, mp, TOUTSTATE, 0);
1601		return;
1602	}
1603	udp->udp_pending_op = T_DISCON_REQ;
1604	udpf = &us->us_bind_fanout[UDP_BIND_HASH(udp->udp_port,
1605	    us->us_bind_fanout_size)];
1606	mutex_enter(&udpf->uf_lock);
1607	udp->udp_v6src = udp->udp_bound_v6src;
1608	udp->udp_state = TS_IDLE;
1609	mutex_exit(&udpf->uf_lock);
1610
1611	/*
1612	 * Send down bind to IP to remove the full binding and revert
1613	 * to the local address binding.
1614	 */
1615	if (udp->udp_family == AF_INET)
1616		mp1 = udp_ip_bind_mp(udp, O_T_BIND_REQ, sizeof (sin_t));
1617	else
1618		mp1 = udp_ip_bind_mp(udp, O_T_BIND_REQ, sizeof (sin6_t));
1619	if (mp1 == NULL) {
1620		udp->udp_pending_op = -1;
1621		rw_exit(&udp->udp_rwlock);
1622		udp_err_ack(q, mp, TSYSERR, ENOMEM);
1623		return;
1624	}
1625	mp = mi_tpi_ok_ack_alloc(mp);
1626	if (mp == NULL) {
1627		/* Unable to reuse the T_DISCON_REQ for the ack. */
1628		udp->udp_pending_op = -1;
1629		rw_exit(&udp->udp_rwlock);
1630		udp_err_ack_prim(q, mp1, T_DISCON_REQ, TSYSERR, ENOMEM);
1631		return;
1632	}
1633
1634	if (udp->udp_family == AF_INET6) {
1635		int error;
1636
1637		/* Rebuild the header template */
1638		error = udp_build_hdrs(udp);
1639		if (error != 0) {
1640			udp->udp_pending_op = -1;
1641			rw_exit(&udp->udp_rwlock);
1642			udp_err_ack_prim(q, mp, T_DISCON_REQ, TSYSERR, error);
1643			freemsg(mp1);
1644			return;
1645		}
1646	}
1647
1648	rw_exit(&udp->udp_rwlock);
1649	/* Append the T_OK_ACK to the T_BIND_REQ for udp_bind_ack */
1650	linkb(mp1, mp);
1651
1652	if (udp->udp_family == AF_INET6)
1653		mp1 = ip_bind_v6(q, mp1, connp, NULL);
1654	else
1655		mp1 = ip_bind_v4(q, mp1, connp);
1656
1657	/* The above return NULL if the bind needs to be deferred */
1658	if (mp1 != NULL)
1659		udp_bind_result(connp, mp1);
1660	else
1661		CONN_INC_REF(connp);
1662}
1663
1664/* This routine creates a T_ERROR_ACK message and passes it upstream. */
1665static void
1666udp_err_ack(queue_t *q, mblk_t *mp, t_scalar_t t_error, int sys_error)
1667{
1668	if ((mp = mi_tpi_err_ack_alloc(mp, t_error, sys_error)) != NULL)
1669		qreply(q, mp);
1670}
1671
1672/* Shorthand to generate and send TPI error acks to our client */
1673static void
1674udp_err_ack_prim(queue_t *q, mblk_t *mp, int primitive, t_scalar_t t_error,
1675    int sys_error)
1676{
1677	struct T_error_ack	*teackp;
1678
1679	if ((mp = tpi_ack_alloc(mp, sizeof (struct T_error_ack),
1680	    M_PCPROTO, T_ERROR_ACK)) != NULL) {
1681		teackp = (struct T_error_ack *)mp->b_rptr;
1682		teackp->ERROR_prim = primitive;
1683		teackp->TLI_error = t_error;
1684		teackp->UNIX_error = sys_error;
1685		qreply(q, mp);
1686	}
1687}
1688
1689/*ARGSUSED*/
1690static int
1691udp_extra_priv_ports_get(queue_t *q, mblk_t *mp, caddr_t cp, cred_t *cr)
1692{
1693	int i;
1694	udp_t		*udp = Q_TO_UDP(q);
1695	udp_stack_t *us = udp->udp_us;
1696
1697	for (i = 0; i < us->us_num_epriv_ports; i++) {
1698		if (us->us_epriv_ports[i] != 0)
1699			(void) mi_mpprintf(mp, "%d ", us->us_epriv_ports[i]);
1700	}
1701	return (0);
1702}
1703
1704/* ARGSUSED */
1705static int
1706udp_extra_priv_ports_add(queue_t *q, mblk_t *mp, char *value, caddr_t cp,
1707    cred_t *cr)
1708{
1709	long	new_value;
1710	int	i;
1711	udp_t		*udp = Q_TO_UDP(q);
1712	udp_stack_t *us = udp->udp_us;
1713
1714	/*
1715	 * Fail the request if the new value does not lie within the
1716	 * port number limits.
1717	 */
1718	if (ddi_strtol(value, NULL, 10, &new_value) != 0 ||
1719	    new_value <= 0 || new_value >= 65536) {
1720		return (EINVAL);
1721	}
1722
1723	/* Check if the value is already in the list */
1724	for (i = 0; i < us->us_num_epriv_ports; i++) {
1725		if (new_value == us->us_epriv_ports[i]) {
1726			return (EEXIST);
1727		}
1728	}
1729	/* Find an empty slot */
1730	for (i = 0; i < us->us_num_epriv_ports; i++) {
1731		if (us->us_epriv_ports[i] == 0)
1732			break;
1733	}
1734	if (i == us->us_num_epriv_ports) {
1735		return (EOVERFLOW);
1736	}
1737
1738	/* Set the new value */
1739	us->us_epriv_ports[i] = (in_port_t)new_value;
1740	return (0);
1741}
1742
1743/* ARGSUSED */
1744static int
1745udp_extra_priv_ports_del(queue_t *q, mblk_t *mp, char *value, caddr_t cp,
1746    cred_t *cr)
1747{
1748	long	new_value;
1749	int	i;
1750	udp_t		*udp = Q_TO_UDP(q);
1751	udp_stack_t *us = udp->udp_us;
1752
1753	/*
1754	 * Fail the request if the new value does not lie within the
1755	 * port number limits.
1756	 */
1757	if (ddi_strtol(value, NULL, 10, &new_value) != 0 ||
1758	    new_value <= 0 || new_value >= 65536) {
1759		return (EINVAL);
1760	}
1761
1762	/* Check that the value is already in the list */
1763	for (i = 0; i < us->us_num_epriv_ports; i++) {
1764		if (us->us_epriv_ports[i] == new_value)
1765			break;
1766	}
1767	if (i == us->us_num_epriv_ports) {
1768		return (ESRCH);
1769	}
1770
1771	/* Clear the value */
1772	us->us_epriv_ports[i] = 0;
1773	return (0);
1774}
1775
1776/* At minimum we need 4 bytes of UDP header */
1777#define	ICMP_MIN_UDP_HDR	4
1778
1779/*
1780 * udp_icmp_error is called by udp_input to process ICMP msgs. passed up by IP.
1781 * Generates the appropriate T_UDERROR_IND for permanent (non-transient) errors.
1782 * Assumes that IP has pulled up everything up to and including the ICMP header.
1783 */
1784static void
1785udp_icmp_error(queue_t *q, mblk_t *mp)
1786{
1787	icmph_t *icmph;
1788	ipha_t	*ipha;
1789	int	iph_hdr_length;
1790	udpha_t	*udpha;
1791	sin_t	sin;
1792	sin6_t	sin6;
1793	mblk_t	*mp1;
1794	int	error = 0;
1795	udp_t	*udp = Q_TO_UDP(q);
1796
1797	ipha = (ipha_t *)mp->b_rptr;
1798
1799	ASSERT(OK_32PTR(mp->b_rptr));
1800
1801	if (IPH_HDR_VERSION(ipha) != IPV4_VERSION) {
1802		ASSERT(IPH_HDR_VERSION(ipha) == IPV6_VERSION);
1803		udp_icmp_error_ipv6(q, mp);
1804		return;
1805	}
1806	ASSERT(IPH_HDR_VERSION(ipha) == IPV4_VERSION);
1807
1808	/* Skip past the outer IP and ICMP headers */
1809	iph_hdr_length = IPH_HDR_LENGTH(ipha);
1810	icmph = (icmph_t *)&mp->b_rptr[iph_hdr_length];
1811	ipha = (ipha_t *)&icmph[1];
1812
1813	/* Skip past the inner IP and find the ULP header */
1814	iph_hdr_length = IPH_HDR_LENGTH(ipha);
1815	udpha = (udpha_t *)((char *)ipha + iph_hdr_length);
1816
1817	switch (icmph->icmph_type) {
1818	case ICMP_DEST_UNREACHABLE:
1819		switch (icmph->icmph_code) {
1820		case ICMP_FRAGMENTATION_NEEDED:
1821			/*
1822			 * IP has already adjusted the path MTU.
1823			 */
1824			break;
1825		case ICMP_PORT_UNREACHABLE:
1826		case ICMP_PROTOCOL_UNREACHABLE:
1827			error = ECONNREFUSED;
1828			break;
1829		default:
1830			/* Transient errors */
1831			break;
1832		}
1833		break;
1834	default:
1835		/* Transient errors */
1836		break;
1837	}
1838	if (error == 0) {
1839		freemsg(mp);
1840		return;
1841	}
1842
1843	/*
1844	 * Deliver T_UDERROR_IND when the application has asked for it.
1845	 * The socket layer enables this automatically when connected.
1846	 */
1847	if (!udp->udp_dgram_errind) {
1848		freemsg(mp);
1849		return;
1850	}
1851
1852	switch (udp->udp_family) {
1853	case AF_INET:
1854		sin = sin_null;
1855		sin.sin_family = AF_INET;
1856		sin.sin_addr.s_addr = ipha->ipha_dst;
1857		sin.sin_port = udpha->uha_dst_port;
1858		mp1 = mi_tpi_uderror_ind((char *)&sin, sizeof (sin_t), NULL, 0,
1859		    error);
1860		break;
1861	case AF_INET6:
1862		sin6 = sin6_null;
1863		sin6.sin6_family = AF_INET6;
1864		IN6_IPADDR_TO_V4MAPPED(ipha->ipha_dst, &sin6.sin6_addr);
1865		sin6.sin6_port = udpha->uha_dst_port;
1866
1867		mp1 = mi_tpi_uderror_ind((char *)&sin6, sizeof (sin6_t),
1868		    NULL, 0, error);
1869		break;
1870	}
1871	if (mp1)
1872		putnext(q, mp1);
1873	freemsg(mp);
1874}
1875
1876/*
1877 * udp_icmp_error_ipv6 is called by udp_icmp_error to process ICMP for IPv6.
1878 * Generates the appropriate T_UDERROR_IND for permanent (non-transient) errors.
1879 * Assumes that IP has pulled up all the extension headers as well as the
1880 * ICMPv6 header.
1881 */
1882static void
1883udp_icmp_error_ipv6(queue_t *q, mblk_t *mp)
1884{
1885	icmp6_t		*icmp6;
1886	ip6_t		*ip6h, *outer_ip6h;
1887	uint16_t	iph_hdr_length;
1888	uint8_t		*nexthdrp;
1889	udpha_t		*udpha;
1890	sin6_t		sin6;
1891	mblk_t		*mp1;
1892	int		error = 0;
1893	udp_t		*udp = Q_TO_UDP(q);
1894	udp_stack_t	*us = udp->udp_us;
1895
1896	outer_ip6h = (ip6_t *)mp->b_rptr;
1897	if (outer_ip6h->ip6_nxt != IPPROTO_ICMPV6)
1898		iph_hdr_length = ip_hdr_length_v6(mp, outer_ip6h);
1899	else
1900		iph_hdr_length = IPV6_HDR_LEN;
1901	icmp6 = (icmp6_t *)&mp->b_rptr[iph_hdr_length];
1902	ip6h = (ip6_t *)&icmp6[1];
1903	if (!ip_hdr_length_nexthdr_v6(mp, ip6h, &iph_hdr_length, &nexthdrp)) {
1904		freemsg(mp);
1905		return;
1906	}
1907	udpha = (udpha_t *)((char *)ip6h + iph_hdr_length);
1908
1909	switch (icmp6->icmp6_type) {
1910	case ICMP6_DST_UNREACH:
1911		switch (icmp6->icmp6_code) {
1912		case ICMP6_DST_UNREACH_NOPORT:
1913			error = ECONNREFUSED;
1914			break;
1915		case ICMP6_DST_UNREACH_ADMIN:
1916		case ICMP6_DST_UNREACH_NOROUTE:
1917		case ICMP6_DST_UNREACH_BEYONDSCOPE:
1918		case ICMP6_DST_UNREACH_ADDR:
1919			/* Transient errors */
1920			break;
1921		default:
1922			break;
1923		}
1924		break;
1925	case ICMP6_PACKET_TOO_BIG: {
1926		struct T_unitdata_ind	*tudi;
1927		struct T_opthdr		*toh;
1928		size_t			udi_size;
1929		mblk_t			*newmp;
1930		t_scalar_t		opt_length = sizeof (struct T_opthdr) +
1931		    sizeof (struct ip6_mtuinfo);
1932		sin6_t			*sin6;
1933		struct ip6_mtuinfo	*mtuinfo;
1934
1935		/*
1936		 * If the application has requested to receive path mtu
1937		 * information, send up an empty message containing an
1938		 * IPV6_PATHMTU ancillary data item.
1939		 */
1940		if (!udp->udp_ipv6_recvpathmtu)
1941			break;
1942
1943		udi_size = sizeof (struct T_unitdata_ind) + sizeof (sin6_t) +
1944		    opt_length;
1945		if ((newmp = allocb(udi_size, BPRI_MED)) == NULL) {
1946			BUMP_MIB(&us->us_udp_mib, udpInErrors);
1947			break;
1948		}
1949
1950		/*
1951		 * newmp->b_cont is left to NULL on purpose.  This is an
1952		 * empty message containing only ancillary data.
1953		 */
1954		newmp->b_datap->db_type = M_PROTO;
1955		tudi = (struct T_unitdata_ind *)newmp->b_rptr;
1956		newmp->b_wptr = (uchar_t *)tudi + udi_size;
1957		tudi->PRIM_type = T_UNITDATA_IND;
1958		tudi->SRC_length = sizeof (sin6_t);
1959		tudi->SRC_offset = sizeof (struct T_unitdata_ind);
1960		tudi->OPT_offset = tudi->SRC_offset + sizeof (sin6_t);
1961		tudi->OPT_length = opt_length;
1962
1963		sin6 = (sin6_t *)&tudi[1];
1964		bzero(sin6, sizeof (sin6_t));
1965		sin6->sin6_family = AF_INET6;
1966		sin6->sin6_addr = udp->udp_v6dst;
1967
1968		toh = (struct T_opthdr *)&sin6[1];
1969		toh->level = IPPROTO_IPV6;
1970		toh->name = IPV6_PATHMTU;
1971		toh->len = opt_length;
1972		toh->status = 0;
1973
1974		mtuinfo = (struct ip6_mtuinfo *)&toh[1];
1975		bzero(mtuinfo, sizeof (struct ip6_mtuinfo));
1976		mtuinfo->ip6m_addr.sin6_family = AF_INET6;
1977		mtuinfo->ip6m_addr.sin6_addr = ip6h->ip6_dst;
1978		mtuinfo->ip6m_mtu = icmp6->icmp6_mtu;
1979		/*
1980		 * We've consumed everything we need from the original
1981		 * message.  Free it, then send our empty message.
1982		 */
1983		freemsg(mp);
1984		putnext(q, newmp);
1985		return;
1986	}
1987	case ICMP6_TIME_EXCEEDED:
1988		/* Transient errors */
1989		break;
1990	case ICMP6_PARAM_PROB:
1991		/* If this corresponds to an ICMP_PROTOCOL_UNREACHABLE */
1992		if (icmp6->icmp6_code == ICMP6_PARAMPROB_NEXTHEADER &&
1993		    (uchar_t *)ip6h + icmp6->icmp6_pptr ==
1994		    (uchar_t *)nexthdrp) {
1995			error = ECONNREFUSED;
1996			break;
1997		}
1998		break;
1999	}
2000	if (error == 0) {
2001		freemsg(mp);
2002		return;
2003	}
2004
2005	/*
2006	 * Deliver T_UDERROR_IND when the application has asked for it.
2007	 * The socket layer enables this automatically when connected.
2008	 */
2009	if (!udp->udp_dgram_errind) {
2010		freemsg(mp);
2011		return;
2012	}
2013
2014	sin6 = sin6_null;
2015	sin6.sin6_family = AF_INET6;
2016	sin6.sin6_addr = ip6h->ip6_dst;
2017	sin6.sin6_port = udpha->uha_dst_port;
2018	sin6.sin6_flowinfo = ip6h->ip6_vcf & ~IPV6_VERS_AND_FLOW_MASK;
2019
2020	mp1 = mi_tpi_uderror_ind((char *)&sin6, sizeof (sin6_t), NULL, 0,
2021	    error);
2022	if (mp1)
2023		putnext(q, mp1);
2024	freemsg(mp);
2025}
2026
2027/*
2028 * This routine responds to T_ADDR_REQ messages.  It is called by udp_wput.
2029 * The local address is filled in if endpoint is bound. The remote address
2030 * is filled in if remote address has been precified ("connected endpoint")
2031 * (The concept of connected CLTS sockets is alien to published TPI
2032 *  but we support it anyway).
2033 */
2034static void
2035udp_addr_req(queue_t *q, mblk_t *mp)
2036{
2037	sin_t	*sin;
2038	sin6_t	*sin6;
2039	mblk_t	*ackmp;
2040	struct T_addr_ack *taa;
2041	udp_t	*udp = Q_TO_UDP(q);
2042
2043	/* Make it large enough for worst case */
2044	ackmp = reallocb(mp, sizeof (struct T_addr_ack) +
2045	    2 * sizeof (sin6_t), 1);
2046	if (ackmp == NULL) {
2047		udp_err_ack(q, mp, TSYSERR, ENOMEM);
2048		return;
2049	}
2050	taa = (struct T_addr_ack *)ackmp->b_rptr;
2051
2052	bzero(taa, sizeof (struct T_addr_ack));
2053	ackmp->b_wptr = (uchar_t *)&taa[1];
2054
2055	taa->PRIM_type = T_ADDR_ACK;
2056	ackmp->b_datap->db_type = M_PCPROTO;
2057	rw_enter(&udp->udp_rwlock, RW_READER);
2058	/*
2059	 * Note: Following code assumes 32 bit alignment of basic
2060	 * data structures like sin_t and struct T_addr_ack.
2061	 */
2062	if (udp->udp_state != TS_UNBND) {
2063		/*
2064		 * Fill in local address first
2065		 */
2066		taa->LOCADDR_offset = sizeof (*taa);
2067		if (udp->udp_family == AF_INET) {
2068			taa->LOCADDR_length = sizeof (sin_t);
2069			sin = (sin_t *)&taa[1];
2070			/* Fill zeroes and then initialize non-zero fields */
2071			*sin = sin_null;
2072			sin->sin_family = AF_INET;
2073			if (!IN6_IS_ADDR_V4MAPPED_ANY(&udp->udp_v6src) &&
2074			    !IN6_IS_ADDR_UNSPECIFIED(&udp->udp_v6src)) {
2075				IN6_V4MAPPED_TO_IPADDR(&udp->udp_v6src,
2076				    sin->sin_addr.s_addr);
2077			} else {
2078				/*
2079				 * INADDR_ANY
2080				 * udp_v6src is not set, we might be bound to
2081				 * broadcast/multicast. Use udp_bound_v6src as
2082				 * local address instead (that could
2083				 * also still be INADDR_ANY)
2084				 */
2085				IN6_V4MAPPED_TO_IPADDR(&udp->udp_bound_v6src,
2086				    sin->sin_addr.s_addr);
2087			}
2088			sin->sin_port = udp->udp_port;
2089			ackmp->b_wptr = (uchar_t *)&sin[1];
2090			if (udp->udp_state == TS_DATA_XFER) {
2091				/*
2092				 * connected, fill remote address too
2093				 */
2094				taa->REMADDR_length = sizeof (sin_t);
2095				/* assumed 32-bit alignment */
2096				taa->REMADDR_offset = taa->LOCADDR_offset +
2097				    taa->LOCADDR_length;
2098
2099				sin = (sin_t *)(ackmp->b_rptr +
2100				    taa->REMADDR_offset);
2101				/* initialize */
2102				*sin = sin_null;
2103				sin->sin_family = AF_INET;
2104				sin->sin_addr.s_addr =
2105				    V4_PART_OF_V6(udp->udp_v6dst);
2106				sin->sin_port = udp->udp_dstport;
2107				ackmp->b_wptr = (uchar_t *)&sin[1];
2108			}
2109		} else {
2110			taa->LOCADDR_length = sizeof (sin6_t);
2111			sin6 = (sin6_t *)&taa[1];
2112			/* Fill zeroes and then initialize non-zero fields */
2113			*sin6 = sin6_null;
2114			sin6->sin6_family = AF_INET6;
2115			if (!IN6_IS_ADDR_UNSPECIFIED(&udp->udp_v6src)) {
2116				sin6->sin6_addr = udp->udp_v6src;
2117			} else {
2118				/*
2119				 * UNSPECIFIED
2120				 * udp_v6src is not set, we might be bound to
2121				 * broadcast/multicast. Use udp_bound_v6src as
2122				 * local address instead (that could
2123				 * also still be UNSPECIFIED)
2124				 */
2125				sin6->sin6_addr =
2126				    udp->udp_bound_v6src;
2127			}
2128			sin6->sin6_port = udp->udp_port;
2129			ackmp->b_wptr = (uchar_t *)&sin6[1];
2130			if (udp->udp_state == TS_DATA_XFER) {
2131				/*
2132				 * connected, fill remote address too
2133				 */
2134				taa->REMADDR_length = sizeof (sin6_t);
2135				/* assumed 32-bit alignment */
2136				taa->REMADDR_offset = taa->LOCADDR_offset +
2137				    taa->LOCADDR_length;
2138
2139				sin6 = (sin6_t *)(ackmp->b_rptr +
2140				    taa->REMADDR_offset);
2141				/* initialize */
2142				*sin6 = sin6_null;
2143				sin6->sin6_family = AF_INET6;
2144				sin6->sin6_addr = udp->udp_v6dst;
2145				sin6->sin6_port =  udp->udp_dstport;
2146				ackmp->b_wptr = (uchar_t *)&sin6[1];
2147			}
2148			ackmp->b_wptr = (uchar_t *)&sin6[1];
2149		}
2150	}
2151	rw_exit(&udp->udp_rwlock);
2152	ASSERT(ackmp->b_wptr <= ackmp->b_datap->db_lim);
2153	qreply(q, ackmp);
2154}
2155
2156static void
2157udp_copy_info(struct T_info_ack *tap, udp_t *udp)
2158{
2159	if (udp->udp_family == AF_INET) {
2160		*tap = udp_g_t_info_ack_ipv4;
2161	} else {
2162		*tap = udp_g_t_info_ack_ipv6;
2163	}
2164	tap->CURRENT_state = udp->udp_state;
2165	tap->OPT_size = udp_max_optsize;
2166}
2167
2168/*
2169 * This routine responds to T_CAPABILITY_REQ messages.  It is called by
2170 * udp_wput.  Much of the T_CAPABILITY_ACK information is copied from
2171 * udp_g_t_info_ack.  The current state of the stream is copied from
2172 * udp_state.
2173 */
2174static void
2175udp_capability_req(queue_t *q, mblk_t *mp)
2176{
2177	t_uscalar_t		cap_bits1;
2178	struct T_capability_ack	*tcap;
2179	udp_t	*udp = Q_TO_UDP(q);
2180
2181	cap_bits1 = ((struct T_capability_req *)mp->b_rptr)->CAP_bits1;
2182
2183	mp = tpi_ack_alloc(mp, sizeof (struct T_capability_ack),
2184	    mp->b_datap->db_type, T_CAPABILITY_ACK);
2185	if (!mp)
2186		return;
2187
2188	tcap = (struct T_capability_ack *)mp->b_rptr;
2189	tcap->CAP_bits1 = 0;
2190
2191	if (cap_bits1 & TC1_INFO) {
2192		udp_copy_info(&tcap->INFO_ack, udp);
2193		tcap->CAP_bits1 |= TC1_INFO;
2194	}
2195
2196	qreply(q, mp);
2197}
2198
2199/*
2200 * This routine responds to T_INFO_REQ messages.  It is called by udp_wput.
2201 * Most of the T_INFO_ACK information is copied from udp_g_t_info_ack.
2202 * The current state of the stream is copied from udp_state.
2203 */
2204static void
2205udp_info_req(queue_t *q, mblk_t *mp)
2206{
2207	udp_t *udp = Q_TO_UDP(q);
2208
2209	/* Create a T_INFO_ACK message. */
2210	mp = tpi_ack_alloc(mp, sizeof (struct T_info_ack), M_PCPROTO,
2211	    T_INFO_ACK);
2212	if (!mp)
2213		return;
2214	udp_copy_info((struct T_info_ack *)mp->b_rptr, udp);
2215	qreply(q, mp);
2216}
2217
2218/*
2219 * IP recognizes seven kinds of bind requests:
2220 *
2221 * - A zero-length address binds only to the protocol number.
2222 *
2223 * - A 4-byte address is treated as a request to
2224 * validate that the address is a valid local IPv4
2225 * address, appropriate for an application to bind to.
2226 * IP does the verification, but does not make any note
2227 * of the address at this time.
2228 *
2229 * - A 16-byte address contains is treated as a request
2230 * to validate a local IPv6 address, as the 4-byte
2231 * address case above.
2232 *
2233 * - A 16-byte sockaddr_in to validate the local IPv4 address and also
2234 * use it for the inbound fanout of packets.
2235 *
2236 * - A 24-byte sockaddr_in6 to validate the local IPv6 address and also
2237 * use it for the inbound fanout of packets.
2238 *
2239 * - A 12-byte address (ipa_conn_t) containing complete IPv4 fanout
2240 * information consisting of local and remote addresses
2241 * and ports.  In this case, the addresses are both
2242 * validated as appropriate for this operation, and, if
2243 * so, the information is retained for use in the
2244 * inbound fanout.
2245 *
2246 * - A 36-byte address address (ipa6_conn_t) containing complete IPv6
2247 * fanout information, like the 12-byte case above.
2248 *
2249 * IP will also fill in the IRE request mblk with information
2250 * regarding our peer.  In all cases, we notify IP of our protocol
2251 * type by appending a single protocol byte to the bind request.
2252 */
2253static mblk_t *
2254udp_ip_bind_mp(udp_t *udp, t_scalar_t bind_prim, t_scalar_t addr_length)
2255{
2256	char	*cp;
2257	mblk_t	*mp;
2258	struct T_bind_req *tbr;
2259	ipa_conn_t	*ac;
2260	ipa6_conn_t	*ac6;
2261	sin_t		*sin;
2262	sin6_t		*sin6;
2263
2264	ASSERT(bind_prim == O_T_BIND_REQ || bind_prim == T_BIND_REQ);
2265	ASSERT(RW_LOCK_HELD(&udp->udp_rwlock));
2266	mp = allocb(sizeof (*tbr) + addr_length + 1, BPRI_HI);
2267	if (!mp)
2268		return (mp);
2269	mp->b_datap->db_type = M_PROTO;
2270	tbr = (struct T_bind_req *)mp->b_rptr;
2271	tbr->PRIM_type = bind_prim;
2272	tbr->ADDR_offset = sizeof (*tbr);
2273	tbr->CONIND_number = 0;
2274	tbr->ADDR_length = addr_length;
2275	cp = (char *)&tbr[1];
2276	switch (addr_length) {
2277	case sizeof (ipa_conn_t):
2278		ASSERT(udp->udp_family == AF_INET);
2279		/* Append a request for an IRE */
2280		mp->b_cont = allocb(sizeof (ire_t), BPRI_HI);
2281		if (!mp->b_cont) {
2282			freemsg(mp);
2283			return (NULL);
2284		}
2285		mp->b_cont->b_wptr += sizeof (ire_t);
2286		mp->b_cont->b_datap->db_type = IRE_DB_REQ_TYPE;
2287
2288		/* cp known to be 32 bit aligned */
2289		ac = (ipa_conn_t *)cp;
2290		ac->ac_laddr = V4_PART_OF_V6(udp->udp_v6src);
2291		ac->ac_faddr = V4_PART_OF_V6(udp->udp_v6dst);
2292		ac->ac_fport = udp->udp_dstport;
2293		ac->ac_lport = udp->udp_port;
2294		break;
2295
2296	case sizeof (ipa6_conn_t):
2297		ASSERT(udp->udp_family == AF_INET6);
2298		/* Append a request for an IRE */
2299		mp->b_cont = allocb(sizeof (ire_t), BPRI_HI);
2300		if (!mp->b_cont) {
2301			freemsg(mp);
2302			return (NULL);
2303		}
2304		mp->b_cont->b_wptr += sizeof (ire_t);
2305		mp->b_cont->b_datap->db_type = IRE_DB_REQ_TYPE;
2306
2307		/* cp known to be 32 bit aligned */
2308		ac6 = (ipa6_conn_t *)cp;
2309		ac6->ac6_laddr = udp->udp_v6src;
2310		ac6->ac6_faddr = udp->udp_v6dst;
2311		ac6->ac6_fport = udp->udp_dstport;
2312		ac6->ac6_lport = udp->udp_port;
2313		break;
2314
2315	case sizeof (sin_t):
2316		ASSERT(udp->udp_family == AF_INET);
2317		/* Append a request for an IRE */
2318		mp->b_cont = allocb(sizeof (ire_t), BPRI_HI);
2319		if (!mp->b_cont) {
2320			freemsg(mp);
2321			return (NULL);
2322		}
2323		mp->b_cont->b_wptr += sizeof (ire_t);
2324		mp->b_cont->b_datap->db_type = IRE_DB_REQ_TYPE;
2325
2326		sin = (sin_t *)cp;
2327		*sin = sin_null;
2328		sin->sin_family = AF_INET;
2329		sin->sin_addr.s_addr = V4_PART_OF_V6(udp->udp_bound_v6src);
2330		sin->sin_port = udp->udp_port;
2331		break;
2332
2333	case sizeof (sin6_t):
2334		ASSERT(udp->udp_family == AF_INET6);
2335		/* Append a request for an IRE */
2336		mp->b_cont = allocb(sizeof (ire_t), BPRI_HI);
2337		if (!mp->b_cont) {
2338			freemsg(mp);
2339			return (NULL);
2340		}
2341		mp->b_cont->b_wptr += sizeof (ire_t);
2342		mp->b_cont->b_datap->db_type = IRE_DB_REQ_TYPE;
2343
2344		sin6 = (sin6_t *)cp;
2345		*sin6 = sin6_null;
2346		sin6->sin6_family = AF_INET6;
2347		sin6->sin6_addr = udp->udp_bound_v6src;
2348		sin6->sin6_port = udp->udp_port;
2349		break;
2350	}
2351	/* Add protocol number to end */
2352	cp[addr_length] = (char)IPPROTO_UDP;
2353	mp->b_wptr = (uchar_t *)&cp[addr_length + 1];
2354	return (mp);
2355}
2356
2357/* For /dev/udp aka AF_INET open */
2358static int
2359udp_openv4(queue_t *q, dev_t *devp, int flag, int sflag, cred_t *credp)
2360{
2361	return (udp_open(q, devp, flag, sflag, credp, B_FALSE));
2362}
2363
2364/* For /dev/udp6 aka AF_INET6 open */
2365static int
2366udp_openv6(queue_t *q, dev_t *devp, int flag, int sflag, cred_t *credp)
2367{
2368	return (udp_open(q, devp, flag, sflag, credp, B_TRUE));
2369}
2370
2371/*
2372 * This is the open routine for udp.  It allocates a udp_t structure for
2373 * the stream and, on the first open of the module, creates an ND table.
2374 */
2375/*ARGSUSED2*/
2376static int
2377udp_open(queue_t *q, dev_t *devp, int flag, int sflag, cred_t *credp,
2378    boolean_t isv6)
2379{
2380	int		err;
2381	udp_t		*udp;
2382	conn_t		*connp;
2383	dev_t		conn_dev;
2384	zoneid_t	zoneid;
2385	netstack_t	*ns;
2386	udp_stack_t	*us;
2387	vmem_t		*minor_arena;
2388
2389	TRACE_1(TR_FAC_UDP, TR_UDP_OPEN, "udp_open: q %p", q);
2390
2391	/* If the stream is already open, return immediately. */
2392	if (q->q_ptr != NULL)
2393		return (0);
2394
2395	if (sflag == MODOPEN)
2396		return (EINVAL);
2397
2398	ns = netstack_find_by_cred(credp);
2399	ASSERT(ns != NULL);
2400	us = ns->netstack_udp;
2401	ASSERT(us != NULL);
2402
2403	/*
2404	 * For exclusive stacks we set the zoneid to zero
2405	 * to make UDP operate as if in the global zone.
2406	 */
2407	if (ns->netstack_stackid != GLOBAL_NETSTACKID)
2408		zoneid = GLOBAL_ZONEID;
2409	else
2410		zoneid = crgetzoneid(credp);
2411
2412	if ((ip_minor_arena_la != NULL) && (flag & SO_SOCKSTR) &&
2413	    ((conn_dev = inet_minor_alloc(ip_minor_arena_la)) != 0)) {
2414		minor_arena = ip_minor_arena_la;
2415	} else {
2416		/*
2417		 * Either minor numbers in the large arena were exhausted
2418		 * or a non socket application is doing the open.
2419		 * Try to allocate from the small arena.
2420		 */
2421		if ((conn_dev = inet_minor_alloc(ip_minor_arena_sa)) == 0) {
2422			netstack_rele(ns);
2423			return (EBUSY);
2424		}
2425		minor_arena = ip_minor_arena_sa;
2426	}
2427
2428	*devp = makedevice(getemajor(*devp), (minor_t)conn_dev);
2429
2430	connp = ipcl_conn_create(IPCL_UDPCONN, KM_SLEEP, ns);
2431	connp->conn_dev = conn_dev;
2432	connp->conn_minor_arena = minor_arena;
2433	udp = connp->conn_udp;
2434
2435	/*
2436	 * ipcl_conn_create did a netstack_hold. Undo the hold that was
2437	 * done by netstack_find_by_cred()
2438	 */
2439	netstack_rele(ns);
2440
2441	/*
2442	 * Initialize the udp_t structure for this stream.
2443	 */
2444	q->q_ptr = connp;
2445	WR(q)->q_ptr = connp;
2446	connp->conn_rq = q;
2447	connp->conn_wq = WR(q);
2448
2449	rw_enter(&udp->udp_rwlock, RW_WRITER);
2450	ASSERT(connp->conn_ulp == IPPROTO_UDP);
2451	ASSERT(connp->conn_udp == udp);
2452	ASSERT(udp->udp_connp == connp);
2453
2454	/* Set the initial state of the stream and the privilege status. */
2455	udp->udp_state = TS_UNBND;
2456	if (isv6) {
2457		udp->udp_family = AF_INET6;
2458		udp->udp_ipversion = IPV6_VERSION;
2459		udp->udp_max_hdr_len = IPV6_HDR_LEN + UDPH_SIZE;
2460		udp->udp_ttl = us->us_ipv6_hoplimit;
2461		connp->conn_af_isv6 = B_TRUE;
2462		connp->conn_flags |= IPCL_ISV6;
2463	} else {
2464		udp->udp_family = AF_INET;
2465		udp->udp_ipversion = IPV4_VERSION;
2466		udp->udp_max_hdr_len = IP_SIMPLE_HDR_LENGTH + UDPH_SIZE;
2467		udp->udp_ttl = us->us_ipv4_ttl;
2468		connp->conn_af_isv6 = B_FALSE;
2469		connp->conn_flags &= ~IPCL_ISV6;
2470	}
2471
2472	udp->udp_multicast_ttl = IP_DEFAULT_MULTICAST_TTL;
2473	udp->udp_pending_op = -1;
2474	connp->conn_multicast_loop = IP_DEFAULT_MULTICAST_LOOP;
2475	connp->conn_zoneid = zoneid;
2476
2477	udp->udp_open_time = lbolt64;
2478	udp->udp_open_pid = curproc->p_pid;
2479
2480	/*
2481	 * If the caller has the process-wide flag set, then default to MAC
2482	 * exempt mode.  This allows read-down to unlabeled hosts.
2483	 */
2484	if (getpflags(NET_MAC_AWARE, credp) != 0)
2485		connp->conn_mac_exempt = B_TRUE;
2486
2487	if (flag & SO_SOCKSTR) {
2488		connp->conn_flags |= IPCL_SOCKET;
2489		udp->udp_issocket = B_TRUE;
2490		udp->udp_direct_sockfs = B_TRUE;
2491	}
2492
2493	connp->conn_ulp_labeled = is_system_labeled();
2494
2495	udp->udp_us = us;
2496
2497	q->q_hiwat = us->us_recv_hiwat;
2498	WR(q)->q_hiwat = us->us_xmit_hiwat;
2499	WR(q)->q_lowat = us->us_xmit_lowat;
2500
2501	connp->conn_recv = udp_input;
2502	crhold(credp);
2503	connp->conn_cred = credp;
2504
2505	mutex_enter(&connp->conn_lock);
2506	connp->conn_state_flags &= ~CONN_INCIPIENT;
2507	mutex_exit(&connp->conn_lock);
2508
2509	qprocson(q);
2510
2511	if (udp->udp_family == AF_INET6) {
2512		/* Build initial header template for transmit */
2513		if ((err = udp_build_hdrs(udp)) != 0) {
2514			rw_exit(&udp->udp_rwlock);
2515			qprocsoff(q);
2516			ipcl_conn_destroy(connp);
2517			return (err);
2518		}
2519	}
2520	rw_exit(&udp->udp_rwlock);
2521
2522	/* Set the Stream head write offset and high watermark. */
2523	(void) mi_set_sth_wroff(q,
2524	    udp->udp_max_hdr_len + us->us_wroff_extra);
2525	(void) mi_set_sth_hiwat(q, udp_set_rcv_hiwat(udp, q->q_hiwat));
2526
2527	return (0);
2528}
2529
2530/*
2531 * Which UDP options OK to set through T_UNITDATA_REQ...
2532 */
2533/* ARGSUSED */
2534static boolean_t
2535udp_opt_allow_udr_set(t_scalar_t level, t_scalar_t name)
2536{
2537	return (B_TRUE);
2538}
2539
2540/*
2541 * This routine gets default values of certain options whose default
2542 * values are maintained by protcol specific code
2543 */
2544/* ARGSUSED */
2545int
2546udp_opt_default(queue_t *q, t_scalar_t level, t_scalar_t name, uchar_t *ptr)
2547{
2548	udp_t		*udp = Q_TO_UDP(q);
2549	udp_stack_t *us = udp->udp_us;
2550	int *i1 = (int *)ptr;
2551
2552	switch (level) {
2553	case IPPROTO_IP:
2554		switch (name) {
2555		case IP_MULTICAST_TTL:
2556			*ptr = (uchar_t)IP_DEFAULT_MULTICAST_TTL;
2557			return (sizeof (uchar_t));
2558		case IP_MULTICAST_LOOP:
2559			*ptr = (uchar_t)IP_DEFAULT_MULTICAST_LOOP;
2560			return (sizeof (uchar_t));
2561		}
2562		break;
2563	case IPPROTO_IPV6:
2564		switch (name) {
2565		case IPV6_MULTICAST_HOPS:
2566			*i1 = IP_DEFAULT_MULTICAST_TTL;
2567			return (sizeof (int));
2568		case IPV6_MULTICAST_LOOP:
2569			*i1 = IP_DEFAULT_MULTICAST_LOOP;
2570			return (sizeof (int));
2571		case IPV6_UNICAST_HOPS:
2572			*i1 = us->us_ipv6_hoplimit;
2573			return (sizeof (int));
2574		}
2575		break;
2576	}
2577	return (-1);
2578}
2579
2580/*
2581 * This routine retrieves the current status of socket options.
2582 * It returns the size of the option retrieved.
2583 */
2584int
2585udp_opt_get_locked(queue_t *q, t_scalar_t level, t_scalar_t name, uchar_t *ptr)
2586{
2587	int	*i1 = (int *)ptr;
2588	conn_t	*connp;
2589	udp_t	*udp;
2590	ip6_pkt_t *ipp;
2591	int	len;
2592	udp_stack_t	*us;
2593
2594	connp = Q_TO_CONN(q);
2595	udp = connp->conn_udp;
2596	ipp = &udp->udp_sticky_ipp;
2597	us = udp->udp_us;
2598
2599	switch (level) {
2600	case SOL_SOCKET:
2601		switch (name) {
2602		case SO_DEBUG:
2603			*i1 = udp->udp_debug;
2604			break;	/* goto sizeof (int) option return */
2605		case SO_REUSEADDR:
2606			*i1 = udp->udp_reuseaddr;
2607			break;	/* goto sizeof (int) option return */
2608		case SO_TYPE:
2609			*i1 = SOCK_DGRAM;
2610			break;	/* goto sizeof (int) option return */
2611
2612		/*
2613		 * The following three items are available here,
2614		 * but are only meaningful to IP.
2615		 */
2616		case SO_DONTROUTE:
2617			*i1 = udp->udp_dontroute;
2618			break;	/* goto sizeof (int) option return */
2619		case SO_USELOOPBACK:
2620			*i1 = udp->udp_useloopback;
2621			break;	/* goto sizeof (int) option return */
2622		case SO_BROADCAST:
2623			*i1 = udp->udp_broadcast;
2624			break;	/* goto sizeof (int) option return */
2625
2626		case SO_SNDBUF:
2627			*i1 = q->q_hiwat;
2628			break;	/* goto sizeof (int) option return */
2629		case SO_RCVBUF:
2630			*i1 = RD(q)->q_hiwat;
2631			break;	/* goto sizeof (int) option return */
2632		case SO_DGRAM_ERRIND:
2633			*i1 = udp->udp_dgram_errind;
2634			break;	/* goto sizeof (int) option return */
2635		case SO_RECVUCRED:
2636			*i1 = udp->udp_recvucred;
2637			break;	/* goto sizeof (int) option return */
2638		case SO_TIMESTAMP:
2639			*i1 = udp->udp_timestamp;
2640			break;	/* goto sizeof (int) option return */
2641		case SO_ANON_MLP:
2642			*i1 = connp->conn_anon_mlp;
2643			break;	/* goto sizeof (int) option return */
2644		case SO_MAC_EXEMPT:
2645			*i1 = connp->conn_mac_exempt;
2646			break;	/* goto sizeof (int) option return */
2647		case SO_ALLZONES:
2648			*i1 = connp->conn_allzones;
2649			break;	/* goto sizeof (int) option return */
2650		case SO_EXCLBIND:
2651			*i1 = udp->udp_exclbind ? SO_EXCLBIND : 0;
2652			break;
2653		case SO_PROTOTYPE:
2654			*i1 = IPPROTO_UDP;
2655			break;
2656		case SO_DOMAIN:
2657			*i1 = udp->udp_family;
2658			break;
2659		default:
2660			return (-1);
2661		}
2662		break;
2663	case IPPROTO_IP:
2664		if (udp->udp_family != AF_INET)
2665			return (-1);
2666		switch (name) {
2667		case IP_OPTIONS:
2668		case T_IP_OPTIONS:
2669			len = udp->udp_ip_rcv_options_len - udp->udp_label_len;
2670			if (len > 0) {
2671				bcopy(udp->udp_ip_rcv_options +
2672				    udp->udp_label_len, ptr, len);
2673			}
2674			return (len);
2675		case IP_TOS:
2676		case T_IP_TOS:
2677			*i1 = (int)udp->udp_type_of_service;
2678			break;	/* goto sizeof (int) option return */
2679		case IP_TTL:
2680			*i1 = (int)udp->udp_ttl;
2681			break;	/* goto sizeof (int) option return */
2682		case IP_DHCPINIT_IF:
2683			return (-EINVAL);
2684		case IP_NEXTHOP:
2685		case IP_RECVPKTINFO:
2686			/*
2687			 * This also handles IP_PKTINFO.
2688			 * IP_PKTINFO and IP_RECVPKTINFO have the same value.
2689			 * Differentiation is based on the size of the argument
2690			 * passed in.
2691			 * This option is handled in IP which will return an
2692			 * error for IP_PKTINFO as it's not supported as a
2693			 * sticky option.
2694			 */
2695			return (-EINVAL);
2696		case IP_MULTICAST_IF:
2697			/* 0 address if not set */
2698			*(ipaddr_t *)ptr = udp->udp_multicast_if_addr;
2699			return (sizeof (ipaddr_t));
2700		case IP_MULTICAST_TTL:
2701			*(uchar_t *)ptr = udp->udp_multicast_ttl;
2702			return (sizeof (uchar_t));
2703		case IP_MULTICAST_LOOP:
2704			*ptr = connp->conn_multicast_loop;
2705			return (sizeof (uint8_t));
2706		case IP_RECVOPTS:
2707			*i1 = udp->udp_recvopts;
2708			break;	/* goto sizeof (int) option return */
2709		case IP_RECVDSTADDR:
2710			*i1 = udp->udp_recvdstaddr;
2711			break;	/* goto sizeof (int) option return */
2712		case IP_RECVIF:
2713			*i1 = udp->udp_recvif;
2714			break;	/* goto sizeof (int) option return */
2715		case IP_RECVSLLA:
2716			*i1 = udp->udp_recvslla;
2717			break;	/* goto sizeof (int) option return */
2718		case IP_RECVTTL:
2719			*i1 = udp->udp_recvttl;
2720			break;	/* goto sizeof (int) option return */
2721		case IP_ADD_MEMBERSHIP:
2722		case IP_DROP_MEMBERSHIP:
2723		case IP_BLOCK_SOURCE:
2724		case IP_UNBLOCK_SOURCE:
2725		case IP_ADD_SOURCE_MEMBERSHIP:
2726		case IP_DROP_SOURCE_MEMBERSHIP:
2727		case MCAST_JOIN_GROUP:
2728		case MCAST_LEAVE_GROUP:
2729		case MCAST_BLOCK_SOURCE:
2730		case MCAST_UNBLOCK_SOURCE:
2731		case MCAST_JOIN_SOURCE_GROUP:
2732		case MCAST_LEAVE_SOURCE_GROUP:
2733		case IP_DONTFAILOVER_IF:
2734			/* cannot "get" the value for these */
2735			return (-1);
2736		case IP_BOUND_IF:
2737			/* Zero if not set */
2738			*i1 = udp->udp_bound_if;
2739			break;	/* goto sizeof (int) option return */
2740		case IP_UNSPEC_SRC:
2741			*i1 = udp->udp_unspec_source;
2742			break;	/* goto sizeof (int) option return */
2743		case IP_BROADCAST_TTL:
2744			*(uchar_t *)ptr = connp->conn_broadcast_ttl;
2745			return (sizeof (uchar_t));
2746		default:
2747			return (-1);
2748		}
2749		break;
2750	case IPPROTO_IPV6:
2751		if (udp->udp_family != AF_INET6)
2752			return (-1);
2753		switch (name) {
2754		case IPV6_UNICAST_HOPS:
2755			*i1 = (unsigned int)udp->udp_ttl;
2756			break;	/* goto sizeof (int) option return */
2757		case IPV6_MULTICAST_IF:
2758			/* 0 index if not set */
2759			*i1 = udp->udp_multicast_if_index;
2760			break;	/* goto sizeof (int) option return */
2761		case IPV6_MULTICAST_HOPS:
2762			*i1 = udp->udp_multicast_ttl;
2763			break;	/* goto sizeof (int) option return */
2764		case IPV6_MULTICAST_LOOP:
2765			*i1 = connp->conn_multicast_loop;
2766			break;	/* goto sizeof (int) option return */
2767		case IPV6_JOIN_GROUP:
2768		case IPV6_LEAVE_GROUP:
2769		case MCAST_JOIN_GROUP:
2770		case MCAST_LEAVE_GROUP:
2771		case MCAST_BLOCK_SOURCE:
2772		case MCAST_UNBLOCK_SOURCE:
2773		case MCAST_JOIN_SOURCE_GROUP:
2774		case MCAST_LEAVE_SOURCE_GROUP:
2775			/* cannot "get" the value for these */
2776			return (-1);
2777		case IPV6_BOUND_IF:
2778			/* Zero if not set */
2779			*i1 = udp->udp_bound_if;
2780			break;	/* goto sizeof (int) option return */
2781		case IPV6_UNSPEC_SRC:
2782			*i1 = udp->udp_unspec_source;
2783			break;	/* goto sizeof (int) option return */
2784		case IPV6_RECVPKTINFO:
2785			*i1 = udp->udp_ip_recvpktinfo;
2786			break;	/* goto sizeof (int) option return */
2787		case IPV6_RECVTCLASS:
2788			*i1 = udp->udp_ipv6_recvtclass;
2789			break;	/* goto sizeof (int) option return */
2790		case IPV6_RECVPATHMTU:
2791			*i1 = udp->udp_ipv6_recvpathmtu;
2792			break;	/* goto sizeof (int) option return */
2793		case IPV6_RECVHOPLIMIT:
2794			*i1 = udp->udp_ipv6_recvhoplimit;
2795			break;	/* goto sizeof (int) option return */
2796		case IPV6_RECVHOPOPTS:
2797			*i1 = udp->udp_ipv6_recvhopopts;
2798			break;	/* goto sizeof (int) option return */
2799		case IPV6_RECVDSTOPTS:
2800			*i1 = udp->udp_ipv6_recvdstopts;
2801			break;	/* goto sizeof (int) option return */
2802		case _OLD_IPV6_RECVDSTOPTS:
2803			*i1 = udp->udp_old_ipv6_recvdstopts;
2804			break;	/* goto sizeof (int) option return */
2805		case IPV6_RECVRTHDRDSTOPTS:
2806			*i1 = udp->udp_ipv6_recvrthdrdstopts;
2807			break;	/* goto sizeof (int) option return */
2808		case IPV6_RECVRTHDR:
2809			*i1 = udp->udp_ipv6_recvrthdr;
2810			break;	/* goto sizeof (int) option return */
2811		case IPV6_PKTINFO: {
2812			/* XXX assumes that caller has room for max size! */
2813			struct in6_pktinfo *pkti;
2814
2815			pkti = (struct in6_pktinfo *)ptr;
2816			if (ipp->ipp_fields & IPPF_IFINDEX)
2817				pkti->ipi6_ifindex = ipp->ipp_ifindex;
2818			else
2819				pkti->ipi6_ifindex = 0;
2820			if (ipp->ipp_fields & IPPF_ADDR)
2821				pkti->ipi6_addr = ipp->ipp_addr;
2822			else
2823				pkti->ipi6_addr = ipv6_all_zeros;
2824			return (sizeof (struct in6_pktinfo));
2825		}
2826		case IPV6_TCLASS:
2827			if (ipp->ipp_fields & IPPF_TCLASS)
2828				*i1 = ipp->ipp_tclass;
2829			else
2830				*i1 = IPV6_FLOW_TCLASS(
2831				    IPV6_DEFAULT_VERS_AND_FLOW);
2832			break;	/* goto sizeof (int) option return */
2833		case IPV6_NEXTHOP: {
2834			sin6_t *sin6 = (sin6_t *)ptr;
2835
2836			if (!(ipp->ipp_fields & IPPF_NEXTHOP))
2837				return (0);
2838			*sin6 = sin6_null;
2839			sin6->sin6_family = AF_INET6;
2840			sin6->sin6_addr = ipp->ipp_nexthop;
2841			return (sizeof (sin6_t));
2842		}
2843		case IPV6_HOPOPTS:
2844			if (!(ipp->ipp_fields & IPPF_HOPOPTS))
2845				return (0);
2846			if (ipp->ipp_hopoptslen <= udp->udp_label_len_v6)
2847				return (0);
2848			/*
2849			 * The cipso/label option is added by kernel.
2850			 * User is not usually aware of this option.
2851			 * We copy out the hbh opt after the label option.
2852			 */
2853			bcopy((char *)ipp->ipp_hopopts + udp->udp_label_len_v6,
2854			    ptr, ipp->ipp_hopoptslen - udp->udp_label_len_v6);
2855			if (udp->udp_label_len_v6 > 0) {
2856				ptr[0] = ((char *)ipp->ipp_hopopts)[0];
2857				ptr[1] = (ipp->ipp_hopoptslen -
2858				    udp->udp_label_len_v6 + 7) / 8 - 1;
2859			}
2860			return (ipp->ipp_hopoptslen - udp->udp_label_len_v6);
2861		case IPV6_RTHDRDSTOPTS:
2862			if (!(ipp->ipp_fields & IPPF_RTDSTOPTS))
2863				return (0);
2864			bcopy(ipp->ipp_rtdstopts, ptr, ipp->ipp_rtdstoptslen);
2865			return (ipp->ipp_rtdstoptslen);
2866		case IPV6_RTHDR:
2867			if (!(ipp->ipp_fields & IPPF_RTHDR))
2868				return (0);
2869			bcopy(ipp->ipp_rthdr, ptr, ipp->ipp_rthdrlen);
2870			return (ipp->ipp_rthdrlen);
2871		case IPV6_DSTOPTS:
2872			if (!(ipp->ipp_fields & IPPF_DSTOPTS))
2873				return (0);
2874			bcopy(ipp->ipp_dstopts, ptr, ipp->ipp_dstoptslen);
2875			return (ipp->ipp_dstoptslen);
2876		case IPV6_PATHMTU:
2877			return (ip_fill_mtuinfo(&udp->udp_v6dst,
2878			    udp->udp_dstport, (struct ip6_mtuinfo *)ptr,
2879			    us->us_netstack));
2880		default:
2881			return (-1);
2882		}
2883		break;
2884	case IPPROTO_UDP:
2885		switch (name) {
2886		case UDP_ANONPRIVBIND:
2887			*i1 = udp->udp_anon_priv_bind;
2888			break;
2889		case UDP_EXCLBIND:
2890			*i1 = udp->udp_exclbind ? UDP_EXCLBIND : 0;
2891			break;
2892		case UDP_RCVHDR:
2893			*i1 = udp->udp_rcvhdr ? 1 : 0;
2894			break;
2895		case UDP_NAT_T_ENDPOINT:
2896			*i1 = udp->udp_nat_t_endpoint;
2897			break;
2898		default:
2899			return (-1);
2900		}
2901		break;
2902	default:
2903		return (-1);
2904	}
2905	return (sizeof (int));
2906}
2907
2908int
2909udp_opt_get(queue_t *q, t_scalar_t level, t_scalar_t name, uchar_t *ptr)
2910{
2911	udp_t	*udp;
2912	int	err;
2913
2914	udp = Q_TO_UDP(q);
2915
2916	rw_enter(&udp->udp_rwlock, RW_READER);
2917	err = udp_opt_get_locked(q, level, name, ptr);
2918	rw_exit(&udp->udp_rwlock);
2919	return (err);
2920}
2921
2922/*
2923 * This routine sets socket options.
2924 */
2925/* ARGSUSED */
2926int
2927udp_opt_set_locked(queue_t *q, uint_t optset_context, int level,
2928    int name, uint_t inlen, uchar_t *invalp, uint_t *outlenp,
2929    uchar_t *outvalp, void *thisdg_attrs, cred_t *cr, mblk_t *mblk)
2930{
2931	udpattrs_t *attrs = thisdg_attrs;
2932	int	*i1 = (int *)invalp;
2933	boolean_t onoff = (*i1 == 0) ? 0 : 1;
2934	boolean_t checkonly;
2935	int	error;
2936	conn_t	*connp;
2937	udp_t	*udp;
2938	uint_t	newlen;
2939	udp_stack_t *us;
2940	size_t	sth_wroff;
2941
2942	connp = Q_TO_CONN(q);
2943	udp = connp->conn_udp;
2944	us = udp->udp_us;
2945
2946	switch (optset_context) {
2947	case SETFN_OPTCOM_CHECKONLY:
2948		checkonly = B_TRUE;
2949		/*
2950		 * Note: Implies T_CHECK semantics for T_OPTCOM_REQ
2951		 * inlen != 0 implies value supplied and
2952		 * 	we have to "pretend" to set it.
2953		 * inlen == 0 implies that there is no
2954		 * 	value part in T_CHECK request and just validation
2955		 * done elsewhere should be enough, we just return here.
2956		 */
2957		if (inlen == 0) {
2958			*outlenp = 0;
2959			return (0);
2960		}
2961		break;
2962	case SETFN_OPTCOM_NEGOTIATE:
2963		checkonly = B_FALSE;
2964		break;
2965	case SETFN_UD_NEGOTIATE:
2966	case SETFN_CONN_NEGOTIATE:
2967		checkonly = B_FALSE;
2968		/*
2969		 * Negotiating local and "association-related" options
2970		 * through T_UNITDATA_REQ.
2971		 *
2972		 * Following routine can filter out ones we do not
2973		 * want to be "set" this way.
2974		 */
2975		if (!udp_opt_allow_udr_set(level, name)) {
2976			*outlenp = 0;
2977			return (EINVAL);
2978		}
2979		break;
2980	default:
2981		/*
2982		 * We should never get here
2983		 */
2984		*outlenp = 0;
2985		return (EINVAL);
2986	}
2987
2988	ASSERT((optset_context != SETFN_OPTCOM_CHECKONLY) ||
2989	    (optset_context == SETFN_OPTCOM_CHECKONLY && inlen != 0));
2990
2991	/*
2992	 * For fixed length options, no sanity check
2993	 * of passed in length is done. It is assumed *_optcom_req()
2994	 * routines do the right thing.
2995	 */
2996
2997	switch (level) {
2998	case SOL_SOCKET:
2999		switch (name) {
3000		case SO_REUSEADDR:
3001			if (!checkonly)
3002				udp->udp_reuseaddr = onoff;
3003			break;
3004		case SO_DEBUG:
3005			if (!checkonly)
3006				udp->udp_debug = onoff;
3007			break;
3008		/*
3009		 * The following three items are available here,
3010		 * but are only meaningful to IP.
3011		 */
3012		case SO_DONTROUTE:
3013			if (!checkonly)
3014				udp->udp_dontroute = onoff;
3015			break;
3016		case SO_USELOOPBACK:
3017			if (!checkonly)
3018				udp->udp_useloopback = onoff;
3019			break;
3020		case SO_BROADCAST:
3021			if (!checkonly)
3022				udp->udp_broadcast = onoff;
3023			break;
3024
3025		case SO_SNDBUF:
3026			if (*i1 > us->us_max_buf) {
3027				*outlenp = 0;
3028				return (ENOBUFS);
3029			}
3030			if (!checkonly) {
3031				q->q_hiwat = *i1;
3032			}
3033			break;
3034		case SO_RCVBUF:
3035			if (*i1 > us->us_max_buf) {
3036				*outlenp = 0;
3037				return (ENOBUFS);
3038			}
3039			if (!checkonly) {
3040				RD(q)->q_hiwat = *i1;
3041				rw_exit(&udp->udp_rwlock);
3042				(void) mi_set_sth_hiwat(RD(q),
3043				    udp_set_rcv_hiwat(udp, *i1));
3044				rw_enter(&udp->udp_rwlock, RW_WRITER);
3045			}
3046			break;
3047		case SO_DGRAM_ERRIND:
3048			if (!checkonly)
3049				udp->udp_dgram_errind = onoff;
3050			break;
3051		case SO_RECVUCRED:
3052			if (!checkonly)
3053				udp->udp_recvucred = onoff;
3054			break;
3055		case SO_ALLZONES:
3056			/*
3057			 * "soft" error (negative)
3058			 * option not handled at this level
3059			 * Do not modify *outlenp.
3060			 */
3061			return (-EINVAL);
3062		case SO_TIMESTAMP:
3063			if (!checkonly)
3064				udp->udp_timestamp = onoff;
3065			break;
3066		case SO_ANON_MLP:
3067			/* Pass option along to IP level for handling */
3068			return (-EINVAL);
3069		case SO_MAC_EXEMPT:
3070			/* Pass option along to IP level for handling */
3071			return (-EINVAL);
3072		case SCM_UCRED: {
3073			struct ucred_s *ucr;
3074			cred_t *cr, *newcr;
3075			ts_label_t *tsl;
3076
3077			/*
3078			 * Only sockets that have proper privileges and are
3079			 * bound to MLPs will have any other value here, so
3080			 * this implicitly tests for privilege to set label.
3081			 */
3082			if (connp->conn_mlp_type == mlptSingle)
3083				break;
3084			ucr = (struct ucred_s *)invalp;
3085			if (inlen != ucredsize ||
3086			    ucr->uc_labeloff < sizeof (*ucr) ||
3087			    ucr->uc_labeloff + sizeof (bslabel_t) > inlen)
3088				return (EINVAL);
3089			if (!checkonly) {
3090				mblk_t *mb;
3091
3092				if (attrs == NULL ||
3093				    (mb = attrs->udpattr_mb) == NULL)
3094					return (EINVAL);
3095				if ((cr = DB_CRED(mb)) == NULL)
3096					cr = udp->udp_connp->conn_cred;
3097				ASSERT(cr != NULL);
3098				if ((tsl = crgetlabel(cr)) == NULL)
3099					return (EINVAL);
3100				newcr = copycred_from_bslabel(cr, UCLABEL(ucr),
3101				    tsl->tsl_doi, KM_NOSLEEP);
3102				if (newcr == NULL)
3103					return (ENOSR);
3104				mblk_setcred(mb, newcr);
3105				attrs->udpattr_credset = B_TRUE;
3106				crfree(newcr);
3107			}
3108			break;
3109		}
3110		case SO_EXCLBIND:
3111			if (!checkonly)
3112				udp->udp_exclbind = onoff;
3113			break;
3114		default:
3115			*outlenp = 0;
3116			return (EINVAL);
3117		}
3118		break;
3119	case IPPROTO_IP:
3120		if (udp->udp_family != AF_INET) {
3121			*outlenp = 0;
3122			return (ENOPROTOOPT);
3123		}
3124		switch (name) {
3125		case IP_OPTIONS:
3126		case T_IP_OPTIONS:
3127			/* Save options for use by IP. */
3128			newlen = inlen + udp->udp_label_len;
3129			if ((inlen & 0x3) || newlen > IP_MAX_OPT_LENGTH) {
3130				*outlenp = 0;
3131				return (EINVAL);
3132			}
3133			if (checkonly)
3134				break;
3135
3136			/*
3137			 * Update the stored options taking into account
3138			 * any CIPSO option which we should not overwrite.
3139			 */
3140			if (!tsol_option_set(&udp->udp_ip_snd_options,
3141			    &udp->udp_ip_snd_options_len,
3142			    udp->udp_label_len, invalp, inlen)) {
3143				*outlenp = 0;
3144				return (ENOMEM);
3145			}
3146
3147			udp->udp_max_hdr_len = IP_SIMPLE_HDR_LENGTH +
3148			    UDPH_SIZE + udp->udp_ip_snd_options_len;
3149			sth_wroff = udp->udp_max_hdr_len + us->us_wroff_extra;
3150			rw_exit(&udp->udp_rwlock);
3151			(void) mi_set_sth_wroff(RD(q), sth_wroff);
3152			rw_enter(&udp->udp_rwlock, RW_WRITER);
3153			break;
3154
3155		case IP_TTL:
3156			if (!checkonly) {
3157				udp->udp_ttl = (uchar_t)*i1;
3158			}
3159			break;
3160		case IP_TOS:
3161		case T_IP_TOS:
3162			if (!checkonly) {
3163				udp->udp_type_of_service = (uchar_t)*i1;
3164			}
3165			break;
3166		case IP_MULTICAST_IF: {
3167			/*
3168			 * TODO should check OPTMGMT reply and undo this if
3169			 * there is an error.
3170			 */
3171			struct in_addr *inap = (struct in_addr *)invalp;
3172			if (!checkonly) {
3173				udp->udp_multicast_if_addr =
3174				    inap->s_addr;
3175			}
3176			break;
3177		}
3178		case IP_MULTICAST_TTL:
3179			if (!checkonly)
3180				udp->udp_multicast_ttl = *invalp;
3181			break;
3182		case IP_MULTICAST_LOOP:
3183			if (!checkonly)
3184				connp->conn_multicast_loop = *invalp;
3185			break;
3186		case IP_RECVOPTS:
3187			if (!checkonly)
3188				udp->udp_recvopts = onoff;
3189			break;
3190		case IP_RECVDSTADDR:
3191			if (!checkonly)
3192				udp->udp_recvdstaddr = onoff;
3193			break;
3194		case IP_RECVIF:
3195			if (!checkonly)
3196				udp->udp_recvif = onoff;
3197			break;
3198		case IP_RECVSLLA:
3199			if (!checkonly)
3200				udp->udp_recvslla = onoff;
3201			break;
3202		case IP_RECVTTL:
3203			if (!checkonly)
3204				udp->udp_recvttl = onoff;
3205			break;
3206		case IP_PKTINFO: {
3207			/*
3208			 * This also handles IP_RECVPKTINFO.
3209			 * IP_PKTINFO and IP_RECVPKTINFO have same value.
3210			 * Differentiation is based on the size of the
3211			 * argument passed in.
3212			 */
3213			struct in_pktinfo *pktinfop;
3214			ip4_pkt_t *attr_pktinfop;
3215
3216			if (checkonly)
3217				break;
3218
3219			if (inlen == sizeof (int)) {
3220				/*
3221				 * This is IP_RECVPKTINFO option.
3222				 * Keep a local copy of whether this option is
3223				 * set or not and pass it down to IP for
3224				 * processing.
3225				 */
3226
3227				udp->udp_ip_recvpktinfo = onoff;
3228				return (-EINVAL);
3229			}
3230
3231			if (attrs == NULL ||
3232			    (attr_pktinfop = attrs->udpattr_ipp4) == NULL) {
3233				/*
3234				 * sticky option or no buffer to return
3235				 * the results.
3236				 */
3237				return (EINVAL);
3238			}
3239
3240			if (inlen != sizeof (struct in_pktinfo))
3241				return (EINVAL);
3242
3243			pktinfop = (struct in_pktinfo *)invalp;
3244
3245			/*
3246			 * At least one of the values should be specified
3247			 */
3248			if (pktinfop->ipi_ifindex == 0 &&
3249			    pktinfop->ipi_spec_dst.s_addr == INADDR_ANY) {
3250				return (EINVAL);
3251			}
3252
3253			attr_pktinfop->ip4_addr = pktinfop->ipi_spec_dst.s_addr;
3254			attr_pktinfop->ip4_ill_index = pktinfop->ipi_ifindex;
3255
3256			break;
3257		}
3258		case IP_ADD_MEMBERSHIP:
3259		case IP_DROP_MEMBERSHIP:
3260		case IP_BLOCK_SOURCE:
3261		case IP_UNBLOCK_SOURCE:
3262		case IP_ADD_SOURCE_MEMBERSHIP:
3263		case IP_DROP_SOURCE_MEMBERSHIP:
3264		case MCAST_JOIN_GROUP:
3265		case MCAST_LEAVE_GROUP:
3266		case MCAST_BLOCK_SOURCE:
3267		case MCAST_UNBLOCK_SOURCE:
3268		case MCAST_JOIN_SOURCE_GROUP:
3269		case MCAST_LEAVE_SOURCE_GROUP:
3270		case IP_SEC_OPT:
3271		case IP_NEXTHOP:
3272		case IP_DHCPINIT_IF:
3273			/*
3274			 * "soft" error (negative)
3275			 * option not handled at this level
3276			 * Do not modify *outlenp.
3277			 */
3278			return (-EINVAL);
3279		case IP_BOUND_IF:
3280			if (!checkonly)
3281				udp->udp_bound_if = *i1;
3282			break;
3283		case IP_UNSPEC_SRC:
3284			if (!checkonly)
3285				udp->udp_unspec_source = onoff;
3286			break;
3287		case IP_BROADCAST_TTL:
3288			if (!checkonly)
3289				connp->conn_broadcast_ttl = *invalp;
3290			break;
3291		default:
3292			*outlenp = 0;
3293			return (EINVAL);
3294		}
3295		break;
3296	case IPPROTO_IPV6: {
3297		ip6_pkt_t		*ipp;
3298		boolean_t		sticky;
3299
3300		if (udp->udp_family != AF_INET6) {
3301			*outlenp = 0;
3302			return (ENOPROTOOPT);
3303		}
3304		/*
3305		 * Deal with both sticky options and ancillary data
3306		 */
3307		sticky = B_FALSE;
3308		if (attrs == NULL || (ipp = attrs->udpattr_ipp6) ==
3309		    NULL) {
3310			/* sticky options, or none */
3311			ipp = &udp->udp_sticky_ipp;
3312			sticky = B_TRUE;
3313		}
3314
3315		switch (name) {
3316		case IPV6_MULTICAST_IF:
3317			if (!checkonly)
3318				udp->udp_multicast_if_index = *i1;
3319			break;
3320		case IPV6_UNICAST_HOPS:
3321			/* -1 means use default */
3322			if (*i1 < -1 || *i1 > IPV6_MAX_HOPS) {
3323				*outlenp = 0;
3324				return (EINVAL);
3325			}
3326			if (!checkonly) {
3327				if (*i1 == -1) {
3328					udp->udp_ttl = ipp->ipp_unicast_hops =
3329					    us->us_ipv6_hoplimit;
3330					ipp->ipp_fields &= ~IPPF_UNICAST_HOPS;
3331					/* Pass modified value to IP. */
3332					*i1 = udp->udp_ttl;
3333				} else {
3334					udp->udp_ttl = ipp->ipp_unicast_hops =
3335					    (uint8_t)*i1;
3336					ipp->ipp_fields |= IPPF_UNICAST_HOPS;
3337				}
3338				/* Rebuild the header template */
3339				error = udp_build_hdrs(udp);
3340				if (error != 0) {
3341					*outlenp = 0;
3342					return (error);
3343				}
3344			}
3345			break;
3346		case IPV6_MULTICAST_HOPS:
3347			/* -1 means use default */
3348			if (*i1 < -1 || *i1 > IPV6_MAX_HOPS) {
3349				*outlenp = 0;
3350				return (EINVAL);
3351			}
3352			if (!checkonly) {
3353				if (*i1 == -1) {
3354					udp->udp_multicast_ttl =
3355					    ipp->ipp_multicast_hops =
3356					    IP_DEFAULT_MULTICAST_TTL;
3357					ipp->ipp_fields &= ~IPPF_MULTICAST_HOPS;
3358					/* Pass modified value to IP. */
3359					*i1 = udp->udp_multicast_ttl;
3360				} else {
3361					udp->udp_multicast_ttl =
3362					    ipp->ipp_multicast_hops =
3363					    (uint8_t)*i1;
3364					ipp->ipp_fields |= IPPF_MULTICAST_HOPS;
3365				}
3366			}
3367			break;
3368		case IPV6_MULTICAST_LOOP:
3369			if (*i1 != 0 && *i1 != 1) {
3370				*outlenp = 0;
3371				return (EINVAL);
3372			}
3373			if (!checkonly)
3374				connp->conn_multicast_loop = *i1;
3375			break;
3376		case IPV6_JOIN_GROUP:
3377		case IPV6_LEAVE_GROUP:
3378		case MCAST_JOIN_GROUP:
3379		case MCAST_LEAVE_GROUP:
3380		case MCAST_BLOCK_SOURCE:
3381		case MCAST_UNBLOCK_SOURCE:
3382		case MCAST_JOIN_SOURCE_GROUP:
3383		case MCAST_LEAVE_SOURCE_GROUP:
3384			/*
3385			 * "soft" error (negative)
3386			 * option not handled at this level
3387			 * Note: Do not modify *outlenp
3388			 */
3389			return (-EINVAL);
3390		case IPV6_BOUND_IF:
3391			if (!checkonly)
3392				udp->udp_bound_if = *i1;
3393			break;
3394		case IPV6_UNSPEC_SRC:
3395			if (!checkonly)
3396				udp->udp_unspec_source = onoff;
3397			break;
3398		/*
3399		 * Set boolean switches for ancillary data delivery
3400		 */
3401		case IPV6_RECVPKTINFO:
3402			if (!checkonly)
3403				udp->udp_ip_recvpktinfo = onoff;
3404			break;
3405		case IPV6_RECVTCLASS:
3406			if (!checkonly) {
3407				udp->udp_ipv6_recvtclass = onoff;
3408			}
3409			break;
3410		case IPV6_RECVPATHMTU:
3411			if (!checkonly) {
3412				udp->udp_ipv6_recvpathmtu = onoff;
3413			}
3414			break;
3415		case IPV6_RECVHOPLIMIT:
3416			if (!checkonly)
3417				udp->udp_ipv6_recvhoplimit = onoff;
3418			break;
3419		case IPV6_RECVHOPOPTS:
3420			if (!checkonly)
3421				udp->udp_ipv6_recvhopopts = onoff;
3422			break;
3423		case IPV6_RECVDSTOPTS:
3424			if (!checkonly)
3425				udp->udp_ipv6_recvdstopts = onoff;
3426			break;
3427		case _OLD_IPV6_RECVDSTOPTS:
3428			if (!checkonly)
3429				udp->udp_old_ipv6_recvdstopts = onoff;
3430			break;
3431		case IPV6_RECVRTHDRDSTOPTS:
3432			if (!checkonly)
3433				udp->udp_ipv6_recvrthdrdstopts = onoff;
3434			break;
3435		case IPV6_RECVRTHDR:
3436			if (!checkonly)
3437				udp->udp_ipv6_recvrthdr = onoff;
3438			break;
3439		/*
3440		 * Set sticky options or ancillary data.
3441		 * If sticky options, (re)build any extension headers
3442		 * that might be needed as a result.
3443		 */
3444		case IPV6_PKTINFO:
3445			/*
3446			 * The source address and ifindex are verified
3447			 * in ip_opt_set(). For ancillary data the
3448			 * source address is checked in ip_wput_v6.
3449			 */
3450			if (inlen != 0 && inlen != sizeof (struct in6_pktinfo))
3451				return (EINVAL);
3452			if (checkonly)
3453				break;
3454
3455			if (inlen == 0) {
3456				ipp->ipp_fields &= ~(IPPF_IFINDEX|IPPF_ADDR);
3457				ipp->ipp_sticky_ignored |=
3458				    (IPPF_IFINDEX|IPPF_ADDR);
3459			} else {
3460				struct in6_pktinfo *pkti;
3461
3462				pkti = (struct in6_pktinfo *)invalp;
3463				ipp->ipp_ifindex = pkti->ipi6_ifindex;
3464				ipp->ipp_addr = pkti->ipi6_addr;
3465				if (ipp->ipp_ifindex != 0)
3466					ipp->ipp_fields |= IPPF_IFINDEX;
3467				else
3468					ipp->ipp_fields &= ~IPPF_IFINDEX;
3469				if (!IN6_IS_ADDR_UNSPECIFIED(
3470				    &ipp->ipp_addr))
3471					ipp->ipp_fields |= IPPF_ADDR;
3472				else
3473					ipp->ipp_fields &= ~IPPF_ADDR;
3474			}
3475			if (sticky) {
3476				error = udp_build_hdrs(udp);
3477				if (error != 0)
3478					return (error);
3479			}
3480			break;
3481		case IPV6_HOPLIMIT:
3482			if (sticky)
3483				return (EINVAL);
3484			if (inlen != 0 && inlen != sizeof (int))
3485				return (EINVAL);
3486			if (checkonly)
3487				break;
3488
3489			if (inlen == 0) {
3490				ipp->ipp_fields &= ~IPPF_HOPLIMIT;
3491				ipp->ipp_sticky_ignored |= IPPF_HOPLIMIT;
3492			} else {
3493				if (*i1 > 255 || *i1 < -1)
3494					return (EINVAL);
3495				if (*i1 == -1)
3496					ipp->ipp_hoplimit =
3497					    us->us_ipv6_hoplimit;
3498				else
3499					ipp->ipp_hoplimit = *i1;
3500				ipp->ipp_fields |= IPPF_HOPLIMIT;
3501			}
3502			break;
3503		case IPV6_TCLASS:
3504			if (inlen != 0 && inlen != sizeof (int))
3505				return (EINVAL);
3506			if (checkonly)
3507				break;
3508
3509			if (inlen == 0) {
3510				ipp->ipp_fields &= ~IPPF_TCLASS;
3511				ipp->ipp_sticky_ignored |= IPPF_TCLASS;
3512			} else {
3513				if (*i1 > 255 || *i1 < -1)
3514					return (EINVAL);
3515				if (*i1 == -1)
3516					ipp->ipp_tclass = 0;
3517				else
3518					ipp->ipp_tclass = *i1;
3519				ipp->ipp_fields |= IPPF_TCLASS;
3520			}
3521			if (sticky) {
3522				error = udp_build_hdrs(udp);
3523				if (error != 0)
3524					return (error);
3525			}
3526			break;
3527		case IPV6_NEXTHOP:
3528			/*
3529			 * IP will verify that the nexthop is reachable
3530			 * and fail for sticky options.
3531			 */
3532			if (inlen != 0 && inlen != sizeof (sin6_t))
3533				return (EINVAL);
3534			if (checkonly)
3535				break;
3536
3537			if (inlen == 0) {
3538				ipp->ipp_fields &= ~IPPF_NEXTHOP;
3539				ipp->ipp_sticky_ignored |= IPPF_NEXTHOP;
3540			} else {
3541				sin6_t *sin6 = (sin6_t *)invalp;
3542
3543				if (sin6->sin6_family != AF_INET6)
3544					return (EAFNOSUPPORT);
3545				if (IN6_IS_ADDR_V4MAPPED(
3546				    &sin6->sin6_addr))
3547					return (EADDRNOTAVAIL);
3548				ipp->ipp_nexthop = sin6->sin6_addr;
3549				if (!IN6_IS_ADDR_UNSPECIFIED(
3550				    &ipp->ipp_nexthop))
3551					ipp->ipp_fields |= IPPF_NEXTHOP;
3552				else
3553					ipp->ipp_fields &= ~IPPF_NEXTHOP;
3554			}
3555			if (sticky) {
3556				error = udp_build_hdrs(udp);
3557				if (error != 0)
3558					return (error);
3559			}
3560			break;
3561		case IPV6_HOPOPTS: {
3562			ip6_hbh_t *hopts = (ip6_hbh_t *)invalp;
3563			/*
3564			 * Sanity checks - minimum size, size a multiple of
3565			 * eight bytes, and matching size passed in.
3566			 */
3567			if (inlen != 0 &&
3568			    inlen != (8 * (hopts->ip6h_len + 1)))
3569				return (EINVAL);
3570
3571			if (checkonly)
3572				break;
3573
3574			error = optcom_pkt_set(invalp, inlen, sticky,
3575			    (uchar_t **)&ipp->ipp_hopopts,
3576			    &ipp->ipp_hopoptslen,
3577			    sticky ? udp->udp_label_len_v6 : 0);
3578			if (error != 0)
3579				return (error);
3580			if (ipp->ipp_hopoptslen == 0) {
3581				ipp->ipp_fields &= ~IPPF_HOPOPTS;
3582				ipp->ipp_sticky_ignored |= IPPF_HOPOPTS;
3583			} else {
3584				ipp->ipp_fields |= IPPF_HOPOPTS;
3585			}
3586			if (sticky) {
3587				error = udp_build_hdrs(udp);
3588				if (error != 0)
3589					return (error);
3590			}
3591			break;
3592		}
3593		case IPV6_RTHDRDSTOPTS: {
3594			ip6_dest_t *dopts = (ip6_dest_t *)invalp;
3595
3596			/*
3597			 * Sanity checks - minimum size, size a multiple of
3598			 * eight bytes, and matching size passed in.
3599			 */
3600			if (inlen != 0 &&
3601			    inlen != (8 * (dopts->ip6d_len + 1)))
3602				return (EINVAL);
3603
3604			if (checkonly)
3605				break;
3606
3607			if (inlen == 0) {
3608				if (sticky &&
3609				    (ipp->ipp_fields & IPPF_RTDSTOPTS) != 0) {
3610					kmem_free(ipp->ipp_rtdstopts,
3611					    ipp->ipp_rtdstoptslen);
3612					ipp->ipp_rtdstopts = NULL;
3613					ipp->ipp_rtdstoptslen = 0;
3614				}
3615
3616				ipp->ipp_fields &= ~IPPF_RTDSTOPTS;
3617				ipp->ipp_sticky_ignored |= IPPF_RTDSTOPTS;
3618			} else {
3619				error = optcom_pkt_set(invalp, inlen, sticky,
3620				    (uchar_t **)&ipp->ipp_rtdstopts,
3621				    &ipp->ipp_rtdstoptslen, 0);
3622				if (error != 0)
3623					return (error);
3624				ipp->ipp_fields |= IPPF_RTDSTOPTS;
3625			}
3626			if (sticky) {
3627				error = udp_build_hdrs(udp);
3628				if (error != 0)
3629					return (error);
3630			}
3631			break;
3632		}
3633		case IPV6_DSTOPTS: {
3634			ip6_dest_t *dopts = (ip6_dest_t *)invalp;
3635
3636			/*
3637			 * Sanity checks - minimum size, size a multiple of
3638			 * eight bytes, and matching size passed in.
3639			 */
3640			if (inlen != 0 &&
3641			    inlen != (8 * (dopts->ip6d_len + 1)))
3642				return (EINVAL);
3643
3644			if (checkonly)
3645				break;
3646
3647			if (inlen == 0) {
3648				if (sticky &&
3649				    (ipp->ipp_fields & IPPF_DSTOPTS) != 0) {
3650					kmem_free(ipp->ipp_dstopts,
3651					    ipp->ipp_dstoptslen);
3652					ipp->ipp_dstopts = NULL;
3653					ipp->ipp_dstoptslen = 0;
3654				}
3655				ipp->ipp_fields &= ~IPPF_DSTOPTS;
3656				ipp->ipp_sticky_ignored |= IPPF_DSTOPTS;
3657			} else {
3658				error = optcom_pkt_set(invalp, inlen, sticky,
3659				    (uchar_t **)&ipp->ipp_dstopts,
3660				    &ipp->ipp_dstoptslen, 0);
3661				if (error != 0)
3662					return (error);
3663				ipp->ipp_fields |= IPPF_DSTOPTS;
3664			}
3665			if (sticky) {
3666				error = udp_build_hdrs(udp);
3667				if (error != 0)
3668					return (error);
3669			}
3670			break;
3671		}
3672		case IPV6_RTHDR: {
3673			ip6_rthdr_t *rt = (ip6_rthdr_t *)invalp;
3674
3675			/*
3676			 * Sanity checks - minimum size, size a multiple of
3677			 * eight bytes, and matching size passed in.
3678			 */
3679			if (inlen != 0 &&
3680			    inlen != (8 * (rt->ip6r_len + 1)))
3681				return (EINVAL);
3682
3683			if (checkonly)
3684				break;
3685
3686			if (inlen == 0) {
3687				if (sticky &&
3688				    (ipp->ipp_fields & IPPF_RTHDR) != 0) {
3689					kmem_free(ipp->ipp_rthdr,
3690					    ipp->ipp_rthdrlen);
3691					ipp->ipp_rthdr = NULL;
3692					ipp->ipp_rthdrlen = 0;
3693				}
3694				ipp->ipp_fields &= ~IPPF_RTHDR;
3695				ipp->ipp_sticky_ignored |= IPPF_RTHDR;
3696			} else {
3697				error = optcom_pkt_set(invalp, inlen, sticky,
3698				    (uchar_t **)&ipp->ipp_rthdr,
3699				    &ipp->ipp_rthdrlen, 0);
3700				if (error != 0)
3701					return (error);
3702				ipp->ipp_fields |= IPPF_RTHDR;
3703			}
3704			if (sticky) {
3705				error = udp_build_hdrs(udp);
3706				if (error != 0)
3707					return (error);
3708			}
3709			break;
3710		}
3711
3712		case IPV6_DONTFRAG:
3713			if (checkonly)
3714				break;
3715
3716			if (onoff) {
3717				ipp->ipp_fields |= IPPF_DONTFRAG;
3718			} else {
3719				ipp->ipp_fields &= ~IPPF_DONTFRAG;
3720			}
3721			break;
3722
3723		case IPV6_USE_MIN_MTU:
3724			if (inlen != sizeof (int))
3725				return (EINVAL);
3726
3727			if (*i1 < -1 || *i1 > 1)
3728				return (EINVAL);
3729
3730			if (checkonly)
3731				break;
3732
3733			ipp->ipp_fields |= IPPF_USE_MIN_MTU;
3734			ipp->ipp_use_min_mtu = *i1;
3735			break;
3736
3737		case IPV6_BOUND_PIF:
3738		case IPV6_SEC_OPT:
3739		case IPV6_DONTFAILOVER_IF:
3740		case IPV6_SRC_PREFERENCES:
3741		case IPV6_V6ONLY:
3742			/* Handled at the IP level */
3743			return (-EINVAL);
3744		default:
3745			*outlenp = 0;
3746			return (EINVAL);
3747		}
3748		break;
3749		}		/* end IPPROTO_IPV6 */
3750	case IPPROTO_UDP:
3751		switch (name) {
3752		case UDP_ANONPRIVBIND:
3753			if ((error = secpolicy_net_privaddr(cr, 0,
3754			    IPPROTO_UDP)) != 0) {
3755				*outlenp = 0;
3756				return (error);
3757			}
3758			if (!checkonly) {
3759				udp->udp_anon_priv_bind = onoff;
3760			}
3761			break;
3762		case UDP_EXCLBIND:
3763			if (!checkonly)
3764				udp->udp_exclbind = onoff;
3765			break;
3766		case UDP_RCVHDR:
3767			if (!checkonly)
3768				udp->udp_rcvhdr = onoff;
3769			break;
3770		case UDP_NAT_T_ENDPOINT:
3771			if ((error = secpolicy_ip_config(cr, B_FALSE)) != 0) {
3772				*outlenp = 0;
3773				return (error);
3774			}
3775
3776			/*
3777			 * Use udp_family instead so we can avoid ambiguitites
3778			 * with AF_INET6 sockets that may switch from IPv4
3779			 * to IPv6.
3780			 */
3781			if (udp->udp_family != AF_INET) {
3782				*outlenp = 0;
3783				return (EAFNOSUPPORT);
3784			}
3785
3786			if (!checkonly) {
3787				udp->udp_nat_t_endpoint = onoff;
3788
3789				udp->udp_max_hdr_len = IP_SIMPLE_HDR_LENGTH +
3790				    UDPH_SIZE + udp->udp_ip_snd_options_len;
3791
3792				/* Also, adjust wroff */
3793				if (onoff) {
3794					udp->udp_max_hdr_len +=
3795					    sizeof (uint32_t);
3796				}
3797				(void) mi_set_sth_wroff(RD(q),
3798				    udp->udp_max_hdr_len + us->us_wroff_extra);
3799			}
3800			break;
3801		default:
3802			*outlenp = 0;
3803			return (EINVAL);
3804		}
3805		break;
3806	default:
3807		*outlenp = 0;
3808		return (EINVAL);
3809	}
3810	/*
3811	 * Common case of OK return with outval same as inval.
3812	 */
3813	if (invalp != outvalp) {
3814		/* don't trust bcopy for identical src/dst */
3815		(void) bcopy(invalp, outvalp, inlen);
3816	}
3817	*outlenp = inlen;
3818	return (0);
3819}
3820
3821int
3822udp_opt_set(queue_t *q, uint_t optset_context, int level,
3823    int name, uint_t inlen, uchar_t *invalp, uint_t *outlenp,
3824    uchar_t *outvalp, void *thisdg_attrs, cred_t *cr, mblk_t *mblk)
3825{
3826	udp_t	*udp;
3827	int	err;
3828
3829	udp = Q_TO_UDP(q);
3830
3831	rw_enter(&udp->udp_rwlock, RW_WRITER);
3832	err = udp_opt_set_locked(q, optset_context, level, name, inlen, invalp,
3833	    outlenp, outvalp, thisdg_attrs, cr, mblk);
3834	rw_exit(&udp->udp_rwlock);
3835	return (err);
3836}
3837
3838/*
3839 * Update udp_sticky_hdrs based on udp_sticky_ipp, udp_v6src, and udp_ttl.
3840 * The headers include ip6i_t (if needed), ip6_t, any sticky extension
3841 * headers, and the udp header.
3842 * Returns failure if can't allocate memory.
3843 */
3844static int
3845udp_build_hdrs(udp_t *udp)
3846{
3847	udp_stack_t *us = udp->udp_us;
3848	uchar_t	*hdrs;
3849	uint_t	hdrs_len;
3850	ip6_t	*ip6h;
3851	ip6i_t	*ip6i;
3852	udpha_t	*udpha;
3853	ip6_pkt_t *ipp = &udp->udp_sticky_ipp;
3854	size_t	sth_wroff;
3855
3856	ASSERT(RW_WRITE_HELD(&udp->udp_rwlock));
3857	hdrs_len = ip_total_hdrs_len_v6(ipp) + UDPH_SIZE;
3858	ASSERT(hdrs_len != 0);
3859	if (hdrs_len != udp->udp_sticky_hdrs_len) {
3860		/* Need to reallocate */
3861		hdrs = kmem_alloc(hdrs_len, KM_NOSLEEP);
3862		if (hdrs == NULL)
3863			return (ENOMEM);
3864
3865		if (udp->udp_sticky_hdrs_len != 0) {
3866			kmem_free(udp->udp_sticky_hdrs,
3867			    udp->udp_sticky_hdrs_len);
3868		}
3869		udp->udp_sticky_hdrs = hdrs;
3870		udp->udp_sticky_hdrs_len = hdrs_len;
3871	}
3872	ip_build_hdrs_v6(udp->udp_sticky_hdrs,
3873	    udp->udp_sticky_hdrs_len - UDPH_SIZE, ipp, IPPROTO_UDP);
3874
3875	/* Set header fields not in ipp */
3876	if (ipp->ipp_fields & IPPF_HAS_IP6I) {
3877		ip6i = (ip6i_t *)udp->udp_sticky_hdrs;
3878		ip6h = (ip6_t *)&ip6i[1];
3879	} else {
3880		ip6h = (ip6_t *)udp->udp_sticky_hdrs;
3881	}
3882
3883	if (!(ipp->ipp_fields & IPPF_ADDR))
3884		ip6h->ip6_src = udp->udp_v6src;
3885
3886	udpha = (udpha_t *)(udp->udp_sticky_hdrs + hdrs_len - UDPH_SIZE);
3887	udpha->uha_src_port = udp->udp_port;
3888
3889	/* Try to get everything in a single mblk */
3890	if (hdrs_len > udp->udp_max_hdr_len) {
3891		udp->udp_max_hdr_len = hdrs_len;
3892		sth_wroff = udp->udp_max_hdr_len + us->us_wroff_extra;
3893		rw_exit(&udp->udp_rwlock);
3894		(void) mi_set_sth_wroff(udp->udp_connp->conn_rq, sth_wroff);
3895		rw_enter(&udp->udp_rwlock, RW_WRITER);
3896	}
3897	return (0);
3898}
3899
3900/*
3901 * This routine retrieves the value of an ND variable in a udpparam_t
3902 * structure.  It is called through nd_getset when a user reads the
3903 * variable.
3904 */
3905/* ARGSUSED */
3906static int
3907udp_param_get(queue_t *q, mblk_t *mp, caddr_t cp, cred_t *cr)
3908{
3909	udpparam_t *udppa = (udpparam_t *)cp;
3910
3911	(void) mi_mpprintf(mp, "%d", udppa->udp_param_value);
3912	return (0);
3913}
3914
3915/*
3916 * Walk through the param array specified registering each element with the
3917 * named dispatch (ND) handler.
3918 */
3919static boolean_t
3920udp_param_register(IDP *ndp, udpparam_t *udppa, int cnt)
3921{
3922	for (; cnt-- > 0; udppa++) {
3923		if (udppa->udp_param_name && udppa->udp_param_name[0]) {
3924			if (!nd_load(ndp, udppa->udp_param_name,
3925			    udp_param_get, udp_param_set,
3926			    (caddr_t)udppa)) {
3927				nd_free(ndp);
3928				return (B_FALSE);
3929			}
3930		}
3931	}
3932	if (!nd_load(ndp, "udp_extra_priv_ports",
3933	    udp_extra_priv_ports_get, NULL, NULL)) {
3934		nd_free(ndp);
3935		return (B_FALSE);
3936	}
3937	if (!nd_load(ndp, "udp_extra_priv_ports_add",
3938	    NULL, udp_extra_priv_ports_add, NULL)) {
3939		nd_free(ndp);
3940		return (B_FALSE);
3941	}
3942	if (!nd_load(ndp, "udp_extra_priv_ports_del",
3943	    NULL, udp_extra_priv_ports_del, NULL)) {
3944		nd_free(ndp);
3945		return (B_FALSE);
3946	}
3947	if (!nd_load(ndp, "udp_status", udp_status_report, NULL,
3948	    NULL)) {
3949		nd_free(ndp);
3950		return (B_FALSE);
3951	}
3952	if (!nd_load(ndp, "udp_bind_hash", udp_bind_hash_report, NULL,
3953	    NULL)) {
3954		nd_free(ndp);
3955		return (B_FALSE);
3956	}
3957	return (B_TRUE);
3958}
3959
3960/* This routine sets an ND variable in a udpparam_t structure. */
3961/* ARGSUSED */
3962static int
3963udp_param_set(queue_t *q, mblk_t *mp, char *value, caddr_t cp, cred_t *cr)
3964{
3965	long		new_value;
3966	udpparam_t	*udppa = (udpparam_t *)cp;
3967
3968	/*
3969	 * Fail the request if the new value does not lie within the
3970	 * required bounds.
3971	 */
3972	if (ddi_strtol(value, NULL, 10, &new_value) != 0 ||
3973	    new_value < udppa->udp_param_min ||
3974	    new_value > udppa->udp_param_max) {
3975		return (EINVAL);
3976	}
3977
3978	/* Set the new value */
3979	udppa->udp_param_value = new_value;
3980	return (0);
3981}
3982
3983/*
3984 * Copy hop-by-hop option from ipp->ipp_hopopts to the buffer provided (with
3985 * T_opthdr) and return the number of bytes copied.  'dbuf' may be NULL to
3986 * just count the length needed for allocation.  If 'dbuf' is non-NULL,
3987 * then it's assumed to be allocated to be large enough.
3988 *
3989 * Returns zero if trimming of the security option causes all options to go
3990 * away.
3991 */
3992static size_t
3993copy_hop_opts(const ip6_pkt_t *ipp, uchar_t *dbuf)
3994{
3995	struct T_opthdr *toh;
3996	size_t hol = ipp->ipp_hopoptslen;
3997	ip6_hbh_t *dstopt = NULL;
3998	const ip6_hbh_t *srcopt = ipp->ipp_hopopts;
3999	size_t tlen, olen, plen;
4000	boolean_t deleting;
4001	const struct ip6_opt *sopt, *lastpad;
4002	struct ip6_opt *dopt;
4003
4004	if ((toh = (struct T_opthdr *)dbuf) != NULL) {
4005		toh->level = IPPROTO_IPV6;
4006		toh->name = IPV6_HOPOPTS;
4007		toh->status = 0;
4008		dstopt = (ip6_hbh_t *)(toh + 1);
4009	}
4010
4011	/*
4012	 * If labeling is enabled, then skip the label option
4013	 * but get other options if there are any.
4014	 */
4015	if (is_system_labeled()) {
4016		dopt = NULL;
4017		if (dstopt != NULL) {
4018			/* will fill in ip6h_len later */
4019			dstopt->ip6h_nxt = srcopt->ip6h_nxt;
4020			dopt = (struct ip6_opt *)(dstopt + 1);
4021		}
4022		sopt = (const struct ip6_opt *)(srcopt + 1);
4023		hol -= sizeof (*srcopt);
4024		tlen = sizeof (*dstopt);
4025		lastpad = NULL;
4026		deleting = B_FALSE;
4027		/*
4028		 * This loop finds the first (lastpad pointer) of any number of
4029		 * pads that preceeds the security option, then treats the
4030		 * security option as though it were a pad, and then finds the
4031		 * next non-pad option (or end of list).
4032		 *
4033		 * It then treats the entire block as one big pad.  To preserve
4034		 * alignment of any options that follow, or just the end of the
4035		 * list, it computes a minimal new padding size that keeps the
4036		 * same alignment for the next option.
4037		 *
4038		 * If it encounters just a sequence of pads with no security
4039		 * option, those are copied as-is rather than collapsed.
4040		 *
4041		 * Note that to handle the end of list case, the code makes one
4042		 * loop with 'hol' set to zero.
4043		 */
4044		for (;;) {
4045			if (hol > 0) {
4046				if (sopt->ip6o_type == IP6OPT_PAD1) {
4047					if (lastpad == NULL)
4048						lastpad = sopt;
4049					sopt = (const struct ip6_opt *)
4050					    &sopt->ip6o_len;
4051					hol--;
4052					continue;
4053				}
4054				olen = sopt->ip6o_len + sizeof (*sopt);
4055				if (olen > hol)
4056					olen = hol;
4057				if (sopt->ip6o_type == IP6OPT_PADN ||
4058				    sopt->ip6o_type == ip6opt_ls) {
4059					if (sopt->ip6o_type == ip6opt_ls)
4060						deleting = B_TRUE;
4061					if (lastpad == NULL)
4062						lastpad = sopt;
4063					sopt = (const struct ip6_opt *)
4064					    ((const char *)sopt + olen);
4065					hol -= olen;
4066					continue;
4067				}
4068			} else {
4069				/* if nothing was copied at all, then delete */
4070				if (tlen == sizeof (*dstopt))
4071					return (0);
4072				/* last pass; pick up any trailing padding */
4073				olen = 0;
4074			}
4075			if (deleting) {
4076				/*
4077				 * compute aligning effect of deleted material
4078				 * to reproduce with pad.
4079				 */
4080				plen = ((const char *)sopt -
4081				    (const char *)lastpad) & 7;
4082				tlen += plen;
4083				if (dopt != NULL) {
4084					if (plen == 1) {
4085						dopt->ip6o_type = IP6OPT_PAD1;
4086					} else if (plen > 1) {
4087						plen -= sizeof (*dopt);
4088						dopt->ip6o_type = IP6OPT_PADN;
4089						dopt->ip6o_len = plen;
4090						if (plen > 0)
4091							bzero(dopt + 1, plen);
4092					}
4093					dopt = (struct ip6_opt *)
4094					    ((char *)dopt + plen);
4095				}
4096				deleting = B_FALSE;
4097				lastpad = NULL;
4098			}
4099			/* if there's uncopied padding, then copy that now */
4100			if (lastpad != NULL) {
4101				olen += (const char *)sopt -
4102				    (const char *)lastpad;
4103				sopt = lastpad;
4104				lastpad = NULL;
4105			}
4106			if (dopt != NULL && olen > 0) {
4107				bcopy(sopt, dopt, olen);
4108				dopt = (struct ip6_opt *)((char *)dopt + olen);
4109			}
4110			if (hol == 0)
4111				break;
4112			tlen += olen;
4113			sopt = (const struct ip6_opt *)
4114			    ((const char *)sopt + olen);
4115			hol -= olen;
4116		}
4117		/* go back and patch up the length value, rounded upward */
4118		if (dstopt != NULL)
4119			dstopt->ip6h_len = (tlen - 1) >> 3;
4120	} else {
4121		tlen = hol;
4122		if (dstopt != NULL)
4123			bcopy(srcopt, dstopt, hol);
4124	}
4125
4126	tlen += sizeof (*toh);
4127	if (toh != NULL)
4128		toh->len = tlen;
4129
4130	return (tlen);
4131}
4132
4133/*
4134 * Update udp_rcv_opt_len from the packet.
4135 * Called when options received, and when no options received but
4136 * udp_ip_recv_opt_len has previously recorded options.
4137 */
4138static void
4139udp_save_ip_rcv_opt(udp_t *udp, void *opt, int opt_len)
4140{
4141	/* Save the options if any */
4142	if (opt_len > 0) {
4143		if (opt_len > udp->udp_ip_rcv_options_len) {
4144			/* Need to allocate larger buffer */
4145			if (udp->udp_ip_rcv_options_len != 0)
4146				mi_free((char *)udp->udp_ip_rcv_options);
4147			udp->udp_ip_rcv_options_len = 0;
4148			udp->udp_ip_rcv_options =
4149			    (uchar_t *)mi_alloc(opt_len, BPRI_HI);
4150			if (udp->udp_ip_rcv_options != NULL)
4151				udp->udp_ip_rcv_options_len = opt_len;
4152		}
4153		if (udp->udp_ip_rcv_options_len != 0) {
4154			bcopy(opt, udp->udp_ip_rcv_options, opt_len);
4155			/* Adjust length if we are resusing the space */
4156			udp->udp_ip_rcv_options_len = opt_len;
4157		}
4158	} else if (udp->udp_ip_rcv_options_len != 0) {
4159		/* Clear out previously recorded options */
4160		mi_free((char *)udp->udp_ip_rcv_options);
4161		udp->udp_ip_rcv_options = NULL;
4162		udp->udp_ip_rcv_options_len = 0;
4163	}
4164}
4165
4166/* ARGSUSED2 */
4167static void
4168udp_input(void *arg1, mblk_t *mp, void *arg2)
4169{
4170	conn_t *connp = (conn_t *)arg1;
4171	struct T_unitdata_ind	*tudi;
4172	uchar_t			*rptr;		/* Pointer to IP header */
4173	int			hdr_length;	/* Length of IP+UDP headers */
4174	int			opt_len;
4175	int			udi_size;	/* Size of T_unitdata_ind */
4176	int			mp_len;
4177	udp_t			*udp;
4178	udpha_t			*udpha;
4179	int			ipversion;
4180	ip6_pkt_t		ipp;
4181	ip6_t			*ip6h;
4182	ip6i_t			*ip6i;
4183	mblk_t			*mp1;
4184	mblk_t			*options_mp = NULL;
4185	ip_pktinfo_t		*pinfo = NULL;
4186	cred_t			*cr = NULL;
4187	pid_t			cpid;
4188	uint32_t		udp_ip_rcv_options_len;
4189	udp_bits_t		udp_bits;
4190	cred_t			*rcr = connp->conn_cred;
4191	udp_stack_t *us;
4192
4193	ASSERT(connp->conn_flags & IPCL_UDPCONN);
4194
4195	udp = connp->conn_udp;
4196	us = udp->udp_us;
4197	rptr = mp->b_rptr;
4198	ASSERT(DB_TYPE(mp) == M_DATA || DB_TYPE(mp) == M_CTL);
4199	ASSERT(OK_32PTR(rptr));
4200
4201	/*
4202	 * IP should have prepended the options data in an M_CTL
4203	 * Check M_CTL "type" to make sure are not here bcos of
4204	 * a valid ICMP message
4205	 */
4206	if (DB_TYPE(mp) == M_CTL) {
4207		if (MBLKL(mp) == sizeof (ip_pktinfo_t) &&
4208		    ((ip_pktinfo_t *)mp->b_rptr)->ip_pkt_ulp_type ==
4209		    IN_PKTINFO) {
4210			/*
4211			 * IP_RECVIF or IP_RECVSLLA or IPF_RECVADDR information
4212			 * has been prepended to the packet by IP. We need to
4213			 * extract the mblk and adjust the rptr
4214			 */
4215			pinfo = (ip_pktinfo_t *)mp->b_rptr;
4216			options_mp = mp;
4217			mp = mp->b_cont;
4218			rptr = mp->b_rptr;
4219			UDP_STAT(us, udp_in_pktinfo);
4220		} else {
4221			/*
4222			 * ICMP messages.
4223			 */
4224			udp_icmp_error(connp->conn_rq, mp);
4225			return;
4226		}
4227	}
4228
4229	mp_len = msgdsize(mp);
4230	/*
4231	 * This is the inbound data path.
4232	 * First, we check to make sure the IP version number is correct,
4233	 * and then pull the IP and UDP headers into the first mblk.
4234	 */
4235
4236	/* Initialize regardless if ipversion is IPv4 or IPv6 */
4237	ipp.ipp_fields = 0;
4238
4239	ipversion = IPH_HDR_VERSION(rptr);
4240
4241	rw_enter(&udp->udp_rwlock, RW_READER);
4242	udp_ip_rcv_options_len = udp->udp_ip_rcv_options_len;
4243	udp_bits = udp->udp_bits;
4244	rw_exit(&udp->udp_rwlock);
4245
4246	switch (ipversion) {
4247	case IPV4_VERSION:
4248		ASSERT(MBLKL(mp) >= sizeof (ipha_t));
4249		ASSERT(((ipha_t *)rptr)->ipha_protocol == IPPROTO_UDP);
4250		hdr_length = IPH_HDR_LENGTH(rptr) + UDPH_SIZE;
4251		opt_len = hdr_length - (IP_SIMPLE_HDR_LENGTH + UDPH_SIZE);
4252		if ((opt_len > 0 || udp_ip_rcv_options_len > 0) &&
4253		    udp->udp_family == AF_INET) {
4254			/*
4255			 * Record/update udp_ip_rcv_options with the lock
4256			 * held. Not needed for AF_INET6 sockets
4257			 * since they don't support a getsockopt of IP_OPTIONS.
4258			 */
4259			rw_enter(&udp->udp_rwlock, RW_WRITER);
4260			udp_save_ip_rcv_opt(udp, rptr + IP_SIMPLE_HDR_LENGTH,
4261			    opt_len);
4262			rw_exit(&udp->udp_rwlock);
4263		}
4264		/* Handle IPV6_RECVPKTINFO even for IPv4 packet. */
4265		if ((udp->udp_family == AF_INET6) && (pinfo != NULL) &&
4266		    udp->udp_ip_recvpktinfo) {
4267			if (pinfo->ip_pkt_flags & IPF_RECVIF) {
4268				ipp.ipp_fields |= IPPF_IFINDEX;
4269				ipp.ipp_ifindex = pinfo->ip_pkt_ifindex;
4270			}
4271		}
4272		break;
4273	case IPV6_VERSION:
4274		/*
4275		 * IPv6 packets can only be received by applications
4276		 * that are prepared to receive IPv6 addresses.
4277		 * The IP fanout must ensure this.
4278		 */
4279		ASSERT(udp->udp_family == AF_INET6);
4280
4281		ip6h = (ip6_t *)rptr;
4282		ASSERT((uchar_t *)&ip6h[1] <= mp->b_wptr);
4283
4284		if (ip6h->ip6_nxt != IPPROTO_UDP) {
4285			uint8_t nexthdrp;
4286			/* Look for ifindex information */
4287			if (ip6h->ip6_nxt == IPPROTO_RAW) {
4288				ip6i = (ip6i_t *)ip6h;
4289				if ((uchar_t *)&ip6i[1] > mp->b_wptr)
4290					goto tossit;
4291
4292				if (ip6i->ip6i_flags & IP6I_IFINDEX) {
4293					ASSERT(ip6i->ip6i_ifindex != 0);
4294					ipp.ipp_fields |= IPPF_IFINDEX;
4295					ipp.ipp_ifindex = ip6i->ip6i_ifindex;
4296				}
4297				rptr = (uchar_t *)&ip6i[1];
4298				mp->b_rptr = rptr;
4299				if (rptr == mp->b_wptr) {
4300					mp1 = mp->b_cont;
4301					freeb(mp);
4302					mp = mp1;
4303					rptr = mp->b_rptr;
4304				}
4305				if (MBLKL(mp) < (IPV6_HDR_LEN + UDPH_SIZE))
4306					goto tossit;
4307				ip6h = (ip6_t *)rptr;
4308				mp_len = msgdsize(mp);
4309			}
4310			/*
4311			 * Find any potentially interesting extension headers
4312			 * as well as the length of the IPv6 + extension
4313			 * headers.
4314			 */
4315			hdr_length = ip_find_hdr_v6(mp, ip6h, &ipp, &nexthdrp) +
4316			    UDPH_SIZE;
4317			ASSERT(nexthdrp == IPPROTO_UDP);
4318		} else {
4319			hdr_length = IPV6_HDR_LEN + UDPH_SIZE;
4320			ip6i = NULL;
4321		}
4322		break;
4323	default:
4324		ASSERT(0);
4325	}
4326
4327	/*
4328	 * IP inspected the UDP header thus all of it must be in the mblk.
4329	 * UDP length check is performed for IPv6 packets and IPv4 packets
4330	 * to check if the size of the packet as specified
4331	 * by the header is the same as the physical size of the packet.
4332	 * FIXME? Didn't IP already check this?
4333	 */
4334	udpha = (udpha_t *)(rptr + (hdr_length - UDPH_SIZE));
4335	if ((MBLKL(mp) < hdr_length) ||
4336	    (mp_len != (ntohs(udpha->uha_length) + hdr_length - UDPH_SIZE))) {
4337		goto tossit;
4338	}
4339
4340
4341	/* Walk past the headers unless IP_RECVHDR was set. */
4342	if (!udp_bits.udpb_rcvhdr) {
4343		mp->b_rptr = rptr + hdr_length;
4344		mp_len -= hdr_length;
4345	}
4346
4347	/*
4348	 * This is the inbound data path.  Packets are passed upstream as
4349	 * T_UNITDATA_IND messages with full IP headers still attached.
4350	 */
4351	if (udp->udp_family == AF_INET) {
4352		sin_t *sin;
4353
4354		ASSERT(IPH_HDR_VERSION((ipha_t *)rptr) == IPV4_VERSION);
4355
4356		/*
4357		 * Normally only send up the source address.
4358		 * If IP_RECVDSTADDR is set we include the destination IP
4359		 * address as an option. With IP_RECVOPTS we include all
4360		 * the IP options.
4361		 */
4362		udi_size = sizeof (struct T_unitdata_ind) + sizeof (sin_t);
4363		if (udp_bits.udpb_recvdstaddr) {
4364			udi_size += sizeof (struct T_opthdr) +
4365			    sizeof (struct in_addr);
4366			UDP_STAT(us, udp_in_recvdstaddr);
4367		}
4368
4369		if (udp_bits.udpb_ip_recvpktinfo && (pinfo != NULL) &&
4370		    (pinfo->ip_pkt_flags & IPF_RECVADDR)) {
4371			udi_size += sizeof (struct T_opthdr) +
4372			    sizeof (struct in_pktinfo);
4373			UDP_STAT(us, udp_ip_rcvpktinfo);
4374		}
4375
4376		if ((udp_bits.udpb_recvopts) && opt_len > 0) {
4377			udi_size += sizeof (struct T_opthdr) + opt_len;
4378			UDP_STAT(us, udp_in_recvopts);
4379		}
4380
4381		/*
4382		 * If the IP_RECVSLLA or the IP_RECVIF is set then allocate
4383		 * space accordingly
4384		 */
4385		if ((udp_bits.udpb_recvif) && (pinfo != NULL) &&
4386		    (pinfo->ip_pkt_flags & IPF_RECVIF)) {
4387			udi_size += sizeof (struct T_opthdr) + sizeof (uint_t);
4388			UDP_STAT(us, udp_in_recvif);
4389		}
4390
4391		if ((udp_bits.udpb_recvslla) && (pinfo != NULL) &&
4392		    (pinfo->ip_pkt_flags & IPF_RECVSLLA)) {
4393			udi_size += sizeof (struct T_opthdr) +
4394			    sizeof (struct sockaddr_dl);
4395			UDP_STAT(us, udp_in_recvslla);
4396		}
4397
4398		if ((udp_bits.udpb_recvucred) &&
4399		    (cr = DB_CRED(mp)) != NULL) {
4400			udi_size += sizeof (struct T_opthdr) + ucredsize;
4401			cpid = DB_CPID(mp);
4402			UDP_STAT(us, udp_in_recvucred);
4403		}
4404
4405		/*
4406		 * If SO_TIMESTAMP is set allocate the appropriate sized
4407		 * buffer. Since gethrestime() expects a pointer aligned
4408		 * argument, we allocate space necessary for extra
4409		 * alignment (even though it might not be used).
4410		 */
4411		if (udp_bits.udpb_timestamp) {
4412			udi_size += sizeof (struct T_opthdr) +
4413			    sizeof (timestruc_t) + _POINTER_ALIGNMENT;
4414			UDP_STAT(us, udp_in_timestamp);
4415		}
4416
4417		/*
4418		 * If IP_RECVTTL is set allocate the appropriate sized buffer
4419		 */
4420		if (udp_bits.udpb_recvttl) {
4421			udi_size += sizeof (struct T_opthdr) + sizeof (uint8_t);
4422			UDP_STAT(us, udp_in_recvttl);
4423		}
4424
4425		/* Allocate a message block for the T_UNITDATA_IND structure. */
4426		mp1 = allocb(udi_size, BPRI_MED);
4427		if (mp1 == NULL) {
4428			freemsg(mp);
4429			if (options_mp != NULL)
4430				freeb(options_mp);
4431			BUMP_MIB(&us->us_udp_mib, udpInErrors);
4432			return;
4433		}
4434		mp1->b_cont = mp;
4435		mp = mp1;
4436		mp->b_datap->db_type = M_PROTO;
4437		tudi = (struct T_unitdata_ind *)mp->b_rptr;
4438		mp->b_wptr = (uchar_t *)tudi + udi_size;
4439		tudi->PRIM_type = T_UNITDATA_IND;
4440		tudi->SRC_length = sizeof (sin_t);
4441		tudi->SRC_offset = sizeof (struct T_unitdata_ind);
4442		tudi->OPT_offset = sizeof (struct T_unitdata_ind) +
4443		    sizeof (sin_t);
4444		udi_size -= (sizeof (struct T_unitdata_ind) + sizeof (sin_t));
4445		tudi->OPT_length = udi_size;
4446		sin = (sin_t *)&tudi[1];
4447		sin->sin_addr.s_addr = ((ipha_t *)rptr)->ipha_src;
4448		sin->sin_port =	udpha->uha_src_port;
4449		sin->sin_family = udp->udp_family;
4450		*(uint32_t *)&sin->sin_zero[0] = 0;
4451		*(uint32_t *)&sin->sin_zero[4] = 0;
4452
4453		/*
4454		 * Add options if IP_RECVDSTADDR, IP_RECVIF, IP_RECVSLLA or
4455		 * IP_RECVTTL has been set.
4456		 */
4457		if (udi_size != 0) {
4458			/*
4459			 * Copy in destination address before options to avoid
4460			 * any padding issues.
4461			 */
4462			char *dstopt;
4463
4464			dstopt = (char *)&sin[1];
4465			if (udp_bits.udpb_recvdstaddr) {
4466				struct T_opthdr *toh;
4467				ipaddr_t *dstptr;
4468
4469				toh = (struct T_opthdr *)dstopt;
4470				toh->level = IPPROTO_IP;
4471				toh->name = IP_RECVDSTADDR;
4472				toh->len = sizeof (struct T_opthdr) +
4473				    sizeof (ipaddr_t);
4474				toh->status = 0;
4475				dstopt += sizeof (struct T_opthdr);
4476				dstptr = (ipaddr_t *)dstopt;
4477				*dstptr = ((ipha_t *)rptr)->ipha_dst;
4478				dstopt += sizeof (ipaddr_t);
4479				udi_size -= toh->len;
4480			}
4481
4482			if (udp_bits.udpb_recvopts && opt_len > 0) {
4483				struct T_opthdr *toh;
4484
4485				toh = (struct T_opthdr *)dstopt;
4486				toh->level = IPPROTO_IP;
4487				toh->name = IP_RECVOPTS;
4488				toh->len = sizeof (struct T_opthdr) + opt_len;
4489				toh->status = 0;
4490				dstopt += sizeof (struct T_opthdr);
4491				bcopy(rptr + IP_SIMPLE_HDR_LENGTH, dstopt,
4492				    opt_len);
4493				dstopt += opt_len;
4494				udi_size -= toh->len;
4495			}
4496
4497			if ((udp_bits.udpb_ip_recvpktinfo) && (pinfo != NULL) &&
4498			    (pinfo->ip_pkt_flags & IPF_RECVADDR)) {
4499				struct T_opthdr *toh;
4500				struct in_pktinfo *pktinfop;
4501
4502				toh = (struct T_opthdr *)dstopt;
4503				toh->level = IPPROTO_IP;
4504				toh->name = IP_PKTINFO;
4505				toh->len = sizeof (struct T_opthdr) +
4506				    sizeof (*pktinfop);
4507				toh->status = 0;
4508				dstopt += sizeof (struct T_opthdr);
4509				pktinfop = (struct in_pktinfo *)dstopt;
4510				pktinfop->ipi_ifindex = pinfo->ip_pkt_ifindex;
4511				pktinfop->ipi_spec_dst =
4512				    pinfo->ip_pkt_match_addr;
4513				pktinfop->ipi_addr.s_addr =
4514				    ((ipha_t *)rptr)->ipha_dst;
4515
4516				dstopt += sizeof (struct in_pktinfo);
4517				udi_size -= toh->len;
4518			}
4519
4520			if ((udp_bits.udpb_recvslla) && (pinfo != NULL) &&
4521			    (pinfo->ip_pkt_flags & IPF_RECVSLLA)) {
4522
4523				struct T_opthdr *toh;
4524				struct sockaddr_dl	*dstptr;
4525
4526				toh = (struct T_opthdr *)dstopt;
4527				toh->level = IPPROTO_IP;
4528				toh->name = IP_RECVSLLA;
4529				toh->len = sizeof (struct T_opthdr) +
4530				    sizeof (struct sockaddr_dl);
4531				toh->status = 0;
4532				dstopt += sizeof (struct T_opthdr);
4533				dstptr = (struct sockaddr_dl *)dstopt;
4534				bcopy(&pinfo->ip_pkt_slla, dstptr,
4535				    sizeof (struct sockaddr_dl));
4536				dstopt += sizeof (struct sockaddr_dl);
4537				udi_size -= toh->len;
4538			}
4539
4540			if ((udp_bits.udpb_recvif) && (pinfo != NULL) &&
4541			    (pinfo->ip_pkt_flags & IPF_RECVIF)) {
4542
4543				struct T_opthdr *toh;
4544				uint_t		*dstptr;
4545
4546				toh = (struct T_opthdr *)dstopt;
4547				toh->level = IPPROTO_IP;
4548				toh->name = IP_RECVIF;
4549				toh->len = sizeof (struct T_opthdr) +
4550				    sizeof (uint_t);
4551				toh->status = 0;
4552				dstopt += sizeof (struct T_opthdr);
4553				dstptr = (uint_t *)dstopt;
4554				*dstptr = pinfo->ip_pkt_ifindex;
4555				dstopt += sizeof (uint_t);
4556				udi_size -= toh->len;
4557			}
4558
4559			if (cr != NULL) {
4560				struct T_opthdr *toh;
4561
4562				toh = (struct T_opthdr *)dstopt;
4563				toh->level = SOL_SOCKET;
4564				toh->name = SCM_UCRED;
4565				toh->len = sizeof (struct T_opthdr) + ucredsize;
4566				toh->status = 0;
4567				dstopt += sizeof (struct T_opthdr);
4568				(void) cred2ucred(cr, cpid, dstopt, rcr);
4569				dstopt += ucredsize;
4570				udi_size -= toh->len;
4571			}
4572
4573			if (udp_bits.udpb_timestamp) {
4574				struct	T_opthdr *toh;
4575
4576				toh = (struct T_opthdr *)dstopt;
4577				toh->level = SOL_SOCKET;
4578				toh->name = SCM_TIMESTAMP;
4579				toh->len = sizeof (struct T_opthdr) +
4580				    sizeof (timestruc_t) + _POINTER_ALIGNMENT;
4581				toh->status = 0;
4582				dstopt += sizeof (struct T_opthdr);
4583				/* Align for gethrestime() */
4584				dstopt = (char *)P2ROUNDUP((intptr_t)dstopt,
4585				    sizeof (intptr_t));
4586				gethrestime((timestruc_t *)dstopt);
4587				dstopt = (char *)toh + toh->len;
4588				udi_size -= toh->len;
4589			}
4590
4591			/*
4592			 * CAUTION:
4593			 * Due to aligment issues
4594			 * Processing of IP_RECVTTL option
4595			 * should always be the last. Adding
4596			 * any option processing after this will
4597			 * cause alignment panic.
4598			 */
4599			if (udp_bits.udpb_recvttl) {
4600				struct	T_opthdr *toh;
4601				uint8_t	*dstptr;
4602
4603				toh = (struct T_opthdr *)dstopt;
4604				toh->level = IPPROTO_IP;
4605				toh->name = IP_RECVTTL;
4606				toh->len = sizeof (struct T_opthdr) +
4607				    sizeof (uint8_t);
4608				toh->status = 0;
4609				dstopt += sizeof (struct T_opthdr);
4610				dstptr = (uint8_t *)dstopt;
4611				*dstptr = ((ipha_t *)rptr)->ipha_ttl;
4612				dstopt += sizeof (uint8_t);
4613				udi_size -= toh->len;
4614			}
4615
4616			/* Consumed all of allocated space */
4617			ASSERT(udi_size == 0);
4618		}
4619	} else {
4620		sin6_t *sin6;
4621
4622		/*
4623		 * Handle both IPv4 and IPv6 packets for IPv6 sockets.
4624		 *
4625		 * Normally we only send up the address. If receiving of any
4626		 * optional receive side information is enabled, we also send
4627		 * that up as options.
4628		 */
4629		udi_size = sizeof (struct T_unitdata_ind) + sizeof (sin6_t);
4630
4631		if (ipp.ipp_fields & (IPPF_HOPOPTS|IPPF_DSTOPTS|IPPF_RTDSTOPTS|
4632		    IPPF_RTHDR|IPPF_IFINDEX)) {
4633			if ((udp_bits.udpb_ipv6_recvhopopts) &&
4634			    (ipp.ipp_fields & IPPF_HOPOPTS)) {
4635				size_t hlen;
4636
4637				UDP_STAT(us, udp_in_recvhopopts);
4638				hlen = copy_hop_opts(&ipp, NULL);
4639				if (hlen == 0)
4640					ipp.ipp_fields &= ~IPPF_HOPOPTS;
4641				udi_size += hlen;
4642			}
4643			if (((udp_bits.udpb_ipv6_recvdstopts) ||
4644			    udp_bits.udpb_old_ipv6_recvdstopts) &&
4645			    (ipp.ipp_fields & IPPF_DSTOPTS)) {
4646				udi_size += sizeof (struct T_opthdr) +
4647				    ipp.ipp_dstoptslen;
4648				UDP_STAT(us, udp_in_recvdstopts);
4649			}
4650			if ((((udp_bits.udpb_ipv6_recvdstopts) &&
4651			    udp_bits.udpb_ipv6_recvrthdr &&
4652			    (ipp.ipp_fields & IPPF_RTHDR)) ||
4653			    (udp_bits.udpb_ipv6_recvrthdrdstopts)) &&
4654			    (ipp.ipp_fields & IPPF_RTDSTOPTS)) {
4655				udi_size += sizeof (struct T_opthdr) +
4656				    ipp.ipp_rtdstoptslen;
4657				UDP_STAT(us, udp_in_recvrtdstopts);
4658			}
4659			if ((udp_bits.udpb_ipv6_recvrthdr) &&
4660			    (ipp.ipp_fields & IPPF_RTHDR)) {
4661				udi_size += sizeof (struct T_opthdr) +
4662				    ipp.ipp_rthdrlen;
4663				UDP_STAT(us, udp_in_recvrthdr);
4664			}
4665			if ((udp_bits.udpb_ip_recvpktinfo) &&
4666			    (ipp.ipp_fields & IPPF_IFINDEX)) {
4667				udi_size += sizeof (struct T_opthdr) +
4668				    sizeof (struct in6_pktinfo);
4669				UDP_STAT(us, udp_in_recvpktinfo);
4670			}
4671
4672		}
4673		if ((udp_bits.udpb_recvucred) &&
4674		    (cr = DB_CRED(mp)) != NULL) {
4675			udi_size += sizeof (struct T_opthdr) + ucredsize;
4676			cpid = DB_CPID(mp);
4677			UDP_STAT(us, udp_in_recvucred);
4678		}
4679
4680		/*
4681		 * If SO_TIMESTAMP is set allocate the appropriate sized
4682		 * buffer. Since gethrestime() expects a pointer aligned
4683		 * argument, we allocate space necessary for extra
4684		 * alignment (even though it might not be used).
4685		 */
4686		if (udp_bits.udpb_timestamp) {
4687			udi_size += sizeof (struct T_opthdr) +
4688			    sizeof (timestruc_t) + _POINTER_ALIGNMENT;
4689			UDP_STAT(us, udp_in_timestamp);
4690		}
4691
4692		if (udp_bits.udpb_ipv6_recvhoplimit) {
4693			udi_size += sizeof (struct T_opthdr) + sizeof (int);
4694			UDP_STAT(us, udp_in_recvhoplimit);
4695		}
4696
4697		if (udp_bits.udpb_ipv6_recvtclass) {
4698			udi_size += sizeof (struct T_opthdr) + sizeof (int);
4699			UDP_STAT(us, udp_in_recvtclass);
4700		}
4701
4702		mp1 = allocb(udi_size, BPRI_MED);
4703		if (mp1 == NULL) {
4704			freemsg(mp);
4705			if (options_mp != NULL)
4706				freeb(options_mp);
4707			BUMP_MIB(&us->us_udp_mib, udpInErrors);
4708			return;
4709		}
4710		mp1->b_cont = mp;
4711		mp = mp1;
4712		mp->b_datap->db_type = M_PROTO;
4713		tudi = (struct T_unitdata_ind *)mp->b_rptr;
4714		mp->b_wptr = (uchar_t *)tudi + udi_size;
4715		tudi->PRIM_type = T_UNITDATA_IND;
4716		tudi->SRC_length = sizeof (sin6_t);
4717		tudi->SRC_offset = sizeof (struct T_unitdata_ind);
4718		tudi->OPT_offset = sizeof (struct T_unitdata_ind) +
4719		    sizeof (sin6_t);
4720		udi_size -= (sizeof (struct T_unitdata_ind) + sizeof (sin6_t));
4721		tudi->OPT_length = udi_size;
4722		sin6 = (sin6_t *)&tudi[1];
4723		if (ipversion == IPV4_VERSION) {
4724			in6_addr_t v6dst;
4725
4726			IN6_IPADDR_TO_V4MAPPED(((ipha_t *)rptr)->ipha_src,
4727			    &sin6->sin6_addr);
4728			IN6_IPADDR_TO_V4MAPPED(((ipha_t *)rptr)->ipha_dst,
4729			    &v6dst);
4730			sin6->sin6_flowinfo = 0;
4731			sin6->sin6_scope_id = 0;
4732			sin6->__sin6_src_id = ip_srcid_find_addr(&v6dst,
4733			    connp->conn_zoneid, us->us_netstack);
4734		} else {
4735			sin6->sin6_addr = ip6h->ip6_src;
4736			/* No sin6_flowinfo per API */
4737			sin6->sin6_flowinfo = 0;
4738			/* For link-scope source pass up scope id */
4739			if ((ipp.ipp_fields & IPPF_IFINDEX) &&
4740			    IN6_IS_ADDR_LINKSCOPE(&ip6h->ip6_src))
4741				sin6->sin6_scope_id = ipp.ipp_ifindex;
4742			else
4743				sin6->sin6_scope_id = 0;
4744			sin6->__sin6_src_id = ip_srcid_find_addr(
4745			    &ip6h->ip6_dst, connp->conn_zoneid,
4746			    us->us_netstack);
4747		}
4748		sin6->sin6_port = udpha->uha_src_port;
4749		sin6->sin6_family = udp->udp_family;
4750
4751		if (udi_size != 0) {
4752			uchar_t *dstopt;
4753
4754			dstopt = (uchar_t *)&sin6[1];
4755			if ((udp_bits.udpb_ip_recvpktinfo) &&
4756			    (ipp.ipp_fields & IPPF_IFINDEX)) {
4757				struct T_opthdr *toh;
4758				struct in6_pktinfo *pkti;
4759
4760				toh = (struct T_opthdr *)dstopt;
4761				toh->level = IPPROTO_IPV6;
4762				toh->name = IPV6_PKTINFO;
4763				toh->len = sizeof (struct T_opthdr) +
4764				    sizeof (*pkti);
4765				toh->status = 0;
4766				dstopt += sizeof (struct T_opthdr);
4767				pkti = (struct in6_pktinfo *)dstopt;
4768				if (ipversion == IPV6_VERSION)
4769					pkti->ipi6_addr = ip6h->ip6_dst;
4770				else
4771					IN6_IPADDR_TO_V4MAPPED(
4772					    ((ipha_t *)rptr)->ipha_dst,
4773					    &pkti->ipi6_addr);
4774				pkti->ipi6_ifindex = ipp.ipp_ifindex;
4775				dstopt += sizeof (*pkti);
4776				udi_size -= toh->len;
4777			}
4778			if (udp_bits.udpb_ipv6_recvhoplimit) {
4779				struct T_opthdr *toh;
4780
4781				toh = (struct T_opthdr *)dstopt;
4782				toh->level = IPPROTO_IPV6;
4783				toh->name = IPV6_HOPLIMIT;
4784				toh->len = sizeof (struct T_opthdr) +
4785				    sizeof (uint_t);
4786				toh->status = 0;
4787				dstopt += sizeof (struct T_opthdr);
4788				if (ipversion == IPV6_VERSION)
4789					*(uint_t *)dstopt = ip6h->ip6_hops;
4790				else
4791					*(uint_t *)dstopt =
4792					    ((ipha_t *)rptr)->ipha_ttl;
4793				dstopt += sizeof (uint_t);
4794				udi_size -= toh->len;
4795			}
4796			if (udp_bits.udpb_ipv6_recvtclass) {
4797				struct T_opthdr *toh;
4798
4799				toh = (struct T_opthdr *)dstopt;
4800				toh->level = IPPROTO_IPV6;
4801				toh->name = IPV6_TCLASS;
4802				toh->len = sizeof (struct T_opthdr) +
4803				    sizeof (uint_t);
4804				toh->status = 0;
4805				dstopt += sizeof (struct T_opthdr);
4806				if (ipversion == IPV6_VERSION) {
4807					*(uint_t *)dstopt =
4808					    IPV6_FLOW_TCLASS(ip6h->ip6_flow);
4809				} else {
4810					ipha_t *ipha = (ipha_t *)rptr;
4811					*(uint_t *)dstopt =
4812					    ipha->ipha_type_of_service;
4813				}
4814				dstopt += sizeof (uint_t);
4815				udi_size -= toh->len;
4816			}
4817			if ((udp_bits.udpb_ipv6_recvhopopts) &&
4818			    (ipp.ipp_fields & IPPF_HOPOPTS)) {
4819				size_t hlen;
4820
4821				hlen = copy_hop_opts(&ipp, dstopt);
4822				dstopt += hlen;
4823				udi_size -= hlen;
4824			}
4825			if ((udp_bits.udpb_ipv6_recvdstopts) &&
4826			    (udp_bits.udpb_ipv6_recvrthdr) &&
4827			    (ipp.ipp_fields & IPPF_RTHDR) &&
4828			    (ipp.ipp_fields & IPPF_RTDSTOPTS)) {
4829				struct T_opthdr *toh;
4830
4831				toh = (struct T_opthdr *)dstopt;
4832				toh->level = IPPROTO_IPV6;
4833				toh->name = IPV6_DSTOPTS;
4834				toh->len = sizeof (struct T_opthdr) +
4835				    ipp.ipp_rtdstoptslen;
4836				toh->status = 0;
4837				dstopt += sizeof (struct T_opthdr);
4838				bcopy(ipp.ipp_rtdstopts, dstopt,
4839				    ipp.ipp_rtdstoptslen);
4840				dstopt += ipp.ipp_rtdstoptslen;
4841				udi_size -= toh->len;
4842			}
4843			if ((udp_bits.udpb_ipv6_recvrthdr) &&
4844			    (ipp.ipp_fields & IPPF_RTHDR)) {
4845				struct T_opthdr *toh;
4846
4847				toh = (struct T_opthdr *)dstopt;
4848				toh->level = IPPROTO_IPV6;
4849				toh->name = IPV6_RTHDR;
4850				toh->len = sizeof (struct T_opthdr) +
4851				    ipp.ipp_rthdrlen;
4852				toh->status = 0;
4853				dstopt += sizeof (struct T_opthdr);
4854				bcopy(ipp.ipp_rthdr, dstopt, ipp.ipp_rthdrlen);
4855				dstopt += ipp.ipp_rthdrlen;
4856				udi_size -= toh->len;
4857			}
4858			if ((udp_bits.udpb_ipv6_recvdstopts) &&
4859			    (ipp.ipp_fields & IPPF_DSTOPTS)) {
4860				struct T_opthdr *toh;
4861
4862				toh = (struct T_opthdr *)dstopt;
4863				toh->level = IPPROTO_IPV6;
4864				toh->name = IPV6_DSTOPTS;
4865				toh->len = sizeof (struct T_opthdr) +
4866				    ipp.ipp_dstoptslen;
4867				toh->status = 0;
4868				dstopt += sizeof (struct T_opthdr);
4869				bcopy(ipp.ipp_dstopts, dstopt,
4870				    ipp.ipp_dstoptslen);
4871				dstopt += ipp.ipp_dstoptslen;
4872				udi_size -= toh->len;
4873			}
4874
4875			if (cr != NULL) {
4876				struct T_opthdr *toh;
4877
4878				toh = (struct T_opthdr *)dstopt;
4879				toh->level = SOL_SOCKET;
4880				toh->name = SCM_UCRED;
4881				toh->len = sizeof (struct T_opthdr) + ucredsize;
4882				toh->status = 0;
4883				(void) cred2ucred(cr, cpid, &toh[1], rcr);
4884				dstopt += toh->len;
4885				udi_size -= toh->len;
4886			}
4887			if (udp_bits.udpb_timestamp) {
4888				struct	T_opthdr *toh;
4889
4890				toh = (struct T_opthdr *)dstopt;
4891				toh->level = SOL_SOCKET;
4892				toh->name = SCM_TIMESTAMP;
4893				toh->len = sizeof (struct T_opthdr) +
4894				    sizeof (timestruc_t) + _POINTER_ALIGNMENT;
4895				toh->status = 0;
4896				dstopt += sizeof (struct T_opthdr);
4897				/* Align for gethrestime() */
4898				dstopt = (uchar_t *)P2ROUNDUP((intptr_t)dstopt,
4899				    sizeof (intptr_t));
4900				gethrestime((timestruc_t *)dstopt);
4901				dstopt = (uchar_t *)toh + toh->len;
4902				udi_size -= toh->len;
4903			}
4904
4905			/* Consumed all of allocated space */
4906			ASSERT(udi_size == 0);
4907		}
4908#undef	sin6
4909		/* No IP_RECVDSTADDR for IPv6. */
4910	}
4911
4912	BUMP_MIB(&us->us_udp_mib, udpHCInDatagrams);
4913	if (options_mp != NULL)
4914		freeb(options_mp);
4915
4916	if (udp_bits.udpb_direct_sockfs) {
4917		/*
4918		 * There is nothing above us except for the stream head;
4919		 * use the read-side synchronous stream interface in
4920		 * order to reduce the time spent in interrupt thread.
4921		 */
4922		ASSERT(udp->udp_issocket);
4923		udp_rcv_enqueue(connp->conn_rq, udp, mp, mp_len);
4924	} else {
4925		/*
4926		 * Use regular STREAMS interface to pass data upstream
4927		 * if this is not a socket endpoint, or if we have
4928		 * switched over to the slow mode due to sockmod being
4929		 * popped or a module being pushed on top of us.
4930		 */
4931		putnext(connp->conn_rq, mp);
4932	}
4933	return;
4934
4935tossit:
4936	freemsg(mp);
4937	if (options_mp != NULL)
4938		freeb(options_mp);
4939	BUMP_MIB(&us->us_udp_mib, udpInErrors);
4940}
4941
4942/*
4943 * Handle the results of a T_BIND_REQ whether deferred by IP or handled
4944 * immediately.
4945 */
4946static void
4947udp_bind_result(conn_t *connp, mblk_t *mp)
4948{
4949	struct T_error_ack	*tea;
4950
4951	switch (mp->b_datap->db_type) {
4952	case M_PROTO:
4953	case M_PCPROTO:
4954		/* M_PROTO messages contain some type of TPI message. */
4955		ASSERT((uintptr_t)(mp->b_wptr - mp->b_rptr) <=
4956		    (uintptr_t)INT_MAX);
4957		if (mp->b_wptr - mp->b_rptr < sizeof (t_scalar_t)) {
4958			freemsg(mp);
4959			return;
4960		}
4961		tea = (struct T_error_ack *)mp->b_rptr;
4962
4963		switch (tea->PRIM_type) {
4964		case T_ERROR_ACK:
4965			switch (tea->ERROR_prim) {
4966			case O_T_BIND_REQ:
4967			case T_BIND_REQ:
4968				udp_bind_error(connp, mp);
4969				return;
4970			default:
4971				break;
4972			}
4973			ASSERT(0);
4974			freemsg(mp);
4975			return;
4976
4977		case T_BIND_ACK:
4978			udp_bind_ack(connp, mp);
4979			return;
4980
4981		default:
4982			break;
4983		}
4984		freemsg(mp);
4985		return;
4986	default:
4987		/* FIXME: other cases? */
4988		ASSERT(0);
4989		freemsg(mp);
4990		return;
4991	}
4992}
4993
4994/*
4995 * Process a T_BIND_ACK
4996 */
4997static void
4998udp_bind_ack(conn_t *connp, mblk_t *mp)
4999{
5000	udp_t	*udp = connp->conn_udp;
5001	mblk_t	*mp1;
5002	ire_t	*ire;
5003	struct T_bind_ack *tba;
5004	uchar_t *addrp;
5005	ipa_conn_t	*ac;
5006	ipa6_conn_t	*ac6;
5007	udp_fanout_t	*udpf;
5008	udp_stack_t	*us = udp->udp_us;
5009
5010	ASSERT(udp->udp_pending_op != -1);
5011	rw_enter(&udp->udp_rwlock, RW_WRITER);
5012	/*
5013	 * If a broadcast/multicast address was bound set
5014	 * the source address to 0.
5015	 * This ensures no datagrams with broadcast address
5016	 * as source address are emitted (which would violate
5017	 * RFC1122 - Hosts requirements)
5018	 *
5019	 * Note that when connecting the returned IRE is
5020	 * for the destination address and we only perform
5021	 * the broadcast check for the source address (it
5022	 * is OK to connect to a broadcast/multicast address.)
5023	 */
5024	mp1 = mp->b_cont;
5025	if (mp1 != NULL && mp1->b_datap->db_type == IRE_DB_TYPE) {
5026		ire = (ire_t *)mp1->b_rptr;
5027
5028		/*
5029		 * Note: we get IRE_BROADCAST for IPv6 to "mark" a multicast
5030		 * local address.
5031		 */
5032		udpf = &us->us_bind_fanout[UDP_BIND_HASH(udp->udp_port,
5033		    us->us_bind_fanout_size)];
5034		if (ire->ire_type == IRE_BROADCAST &&
5035		    udp->udp_state != TS_DATA_XFER) {
5036			ASSERT(udp->udp_pending_op == T_BIND_REQ ||
5037			    udp->udp_pending_op == O_T_BIND_REQ);
5038			/* This was just a local bind to a broadcast addr */
5039			mutex_enter(&udpf->uf_lock);
5040			V6_SET_ZERO(udp->udp_v6src);
5041			mutex_exit(&udpf->uf_lock);
5042			if (udp->udp_family == AF_INET6)
5043				(void) udp_build_hdrs(udp);
5044		} else if (V6_OR_V4_INADDR_ANY(udp->udp_v6src)) {
5045			/*
5046			 * Local address not yet set - pick it from the
5047			 * T_bind_ack
5048			 */
5049			tba = (struct T_bind_ack *)mp->b_rptr;
5050			addrp = &mp->b_rptr[tba->ADDR_offset];
5051			switch (udp->udp_family) {
5052			case AF_INET:
5053				if (tba->ADDR_length == sizeof (ipa_conn_t)) {
5054					ac = (ipa_conn_t *)addrp;
5055				} else {
5056					ASSERT(tba->ADDR_length ==
5057					    sizeof (ipa_conn_x_t));
5058					ac = &((ipa_conn_x_t *)addrp)->acx_conn;
5059				}
5060				mutex_enter(&udpf->uf_lock);
5061				IN6_IPADDR_TO_V4MAPPED(ac->ac_laddr,
5062				    &udp->udp_v6src);
5063				mutex_exit(&udpf->uf_lock);
5064				break;
5065			case AF_INET6:
5066				if (tba->ADDR_length == sizeof (ipa6_conn_t)) {
5067					ac6 = (ipa6_conn_t *)addrp;
5068				} else {
5069					ASSERT(tba->ADDR_length ==
5070					    sizeof (ipa6_conn_x_t));
5071					ac6 = &((ipa6_conn_x_t *)
5072					    addrp)->ac6x_conn;
5073				}
5074				mutex_enter(&udpf->uf_lock);
5075				udp->udp_v6src = ac6->ac6_laddr;
5076				mutex_exit(&udpf->uf_lock);
5077				(void) udp_build_hdrs(udp);
5078				break;
5079			}
5080		}
5081		mp1 = mp1->b_cont;
5082	}
5083	udp->udp_pending_op = -1;
5084	rw_exit(&udp->udp_rwlock);
5085	/*
5086	 * Look for one or more appended ACK message added by
5087	 * udp_connect or udp_disconnect.
5088	 * If none found just send up the T_BIND_ACK.
5089	 * udp_connect has appended a T_OK_ACK and a T_CONN_CON.
5090	 * udp_disconnect has appended a T_OK_ACK.
5091	 */
5092	if (mp1 != NULL) {
5093		if (mp->b_cont == mp1)
5094			mp->b_cont = NULL;
5095		else {
5096			ASSERT(mp->b_cont->b_cont == mp1);
5097			mp->b_cont->b_cont = NULL;
5098		}
5099		freemsg(mp);
5100		mp = mp1;
5101		while (mp != NULL) {
5102			mp1 = mp->b_cont;
5103			mp->b_cont = NULL;
5104			putnext(connp->conn_rq, mp);
5105			mp = mp1;
5106		}
5107		return;
5108	}
5109	freemsg(mp->b_cont);
5110	mp->b_cont = NULL;
5111	putnext(connp->conn_rq, mp);
5112}
5113
5114static void
5115udp_bind_error(conn_t *connp, mblk_t *mp)
5116{
5117	udp_t	*udp = connp->conn_udp;
5118	struct T_error_ack *tea;
5119	udp_fanout_t	*udpf;
5120	udp_stack_t	*us = udp->udp_us;
5121
5122	tea = (struct T_error_ack *)mp->b_rptr;
5123
5124	/*
5125	 * If our O_T_BIND_REQ/T_BIND_REQ fails,
5126	 * clear out the associated port and source
5127	 * address before passing the message
5128	 * upstream. If this was caused by a T_CONN_REQ
5129	 * revert back to bound state.
5130	 */
5131
5132	rw_enter(&udp->udp_rwlock, RW_WRITER);
5133	ASSERT(udp->udp_pending_op != -1);
5134	tea->ERROR_prim = udp->udp_pending_op;
5135	udp->udp_pending_op = -1;
5136	udpf = &us->us_bind_fanout[
5137	    UDP_BIND_HASH(udp->udp_port,
5138	    us->us_bind_fanout_size)];
5139	mutex_enter(&udpf->uf_lock);
5140
5141	switch (tea->ERROR_prim) {
5142	case T_CONN_REQ:
5143		ASSERT(udp->udp_state == TS_DATA_XFER);
5144		/* Connect failed */
5145		/* Revert back to the bound source */
5146		udp->udp_v6src = udp->udp_bound_v6src;
5147		udp->udp_state = TS_IDLE;
5148		mutex_exit(&udpf->uf_lock);
5149		if (udp->udp_family == AF_INET6)
5150			(void) udp_build_hdrs(udp);
5151		rw_exit(&udp->udp_rwlock);
5152		break;
5153
5154	case T_DISCON_REQ:
5155	case T_BIND_REQ:
5156	case O_T_BIND_REQ:
5157		V6_SET_ZERO(udp->udp_v6src);
5158		V6_SET_ZERO(udp->udp_bound_v6src);
5159		udp->udp_state = TS_UNBND;
5160		udp_bind_hash_remove(udp, B_TRUE);
5161		udp->udp_port = 0;
5162		mutex_exit(&udpf->uf_lock);
5163		if (udp->udp_family == AF_INET6)
5164			(void) udp_build_hdrs(udp);
5165		rw_exit(&udp->udp_rwlock);
5166		break;
5167
5168	default:
5169		mutex_exit(&udpf->uf_lock);
5170		rw_exit(&udp->udp_rwlock);
5171		(void) mi_strlog(connp->conn_rq, 1,
5172		    SL_ERROR|SL_TRACE,
5173		    "udp_input_other: bad ERROR_prim, "
5174		    "len %d", tea->ERROR_prim);
5175	}
5176	putnext(connp->conn_rq, mp);
5177}
5178
5179/*
5180 * return SNMP stuff in buffer in mpdata. We don't hold any lock and report
5181 * information that can be changing beneath us.
5182 */
5183mblk_t *
5184udp_snmp_get(queue_t *q, mblk_t *mpctl)
5185{
5186	mblk_t			*mpdata;
5187	mblk_t			*mp_conn_ctl;
5188	mblk_t			*mp_attr_ctl;
5189	mblk_t			*mp6_conn_ctl;
5190	mblk_t			*mp6_attr_ctl;
5191	mblk_t			*mp_conn_tail;
5192	mblk_t			*mp_attr_tail;
5193	mblk_t			*mp6_conn_tail;
5194	mblk_t			*mp6_attr_tail;
5195	struct opthdr		*optp;
5196	mib2_udpEntry_t		ude;
5197	mib2_udp6Entry_t	ude6;
5198	mib2_transportMLPEntry_t mlp;
5199	int			state;
5200	zoneid_t		zoneid;
5201	int			i;
5202	connf_t			*connfp;
5203	conn_t			*connp = Q_TO_CONN(q);
5204	int			v4_conn_idx;
5205	int			v6_conn_idx;
5206	boolean_t		needattr;
5207	udp_t			*udp;
5208	ip_stack_t		*ipst = connp->conn_netstack->netstack_ip;
5209	udp_stack_t		*us = connp->conn_netstack->netstack_udp;
5210	mblk_t			*mp2ctl;
5211
5212	/*
5213	 * make a copy of the original message
5214	 */
5215	mp2ctl = copymsg(mpctl);
5216
5217	mp_conn_ctl = mp_attr_ctl = mp6_conn_ctl = NULL;
5218	if (mpctl == NULL ||
5219	    (mpdata = mpctl->b_cont) == NULL ||
5220	    (mp_conn_ctl = copymsg(mpctl)) == NULL ||
5221	    (mp_attr_ctl = copymsg(mpctl)) == NULL ||
5222	    (mp6_conn_ctl = copymsg(mpctl)) == NULL ||
5223	    (mp6_attr_ctl = copymsg(mpctl)) == NULL) {
5224		freemsg(mp_conn_ctl);
5225		freemsg(mp_attr_ctl);
5226		freemsg(mp6_conn_ctl);
5227		freemsg(mpctl);
5228		freemsg(mp2ctl);
5229		return (0);
5230	}
5231
5232	zoneid = connp->conn_zoneid;
5233
5234	/* fixed length structure for IPv4 and IPv6 counters */
5235	SET_MIB(us->us_udp_mib.udpEntrySize, sizeof (mib2_udpEntry_t));
5236	SET_MIB(us->us_udp_mib.udp6EntrySize, sizeof (mib2_udp6Entry_t));
5237	/* synchronize 64- and 32-bit counters */
5238	SYNC32_MIB(&us->us_udp_mib, udpInDatagrams, udpHCInDatagrams);
5239	SYNC32_MIB(&us->us_udp_mib, udpOutDatagrams, udpHCOutDatagrams);
5240
5241	optp = (struct opthdr *)&mpctl->b_rptr[sizeof (struct T_optmgmt_ack)];
5242	optp->level = MIB2_UDP;
5243	optp->name = 0;
5244	(void) snmp_append_data(mpdata, (char *)&us->us_udp_mib,
5245	    sizeof (us->us_udp_mib));
5246	optp->len = msgdsize(mpdata);
5247	qreply(q, mpctl);
5248
5249	mp_conn_tail = mp_attr_tail = mp6_conn_tail = mp6_attr_tail = NULL;
5250	v4_conn_idx = v6_conn_idx = 0;
5251
5252	for (i = 0; i < CONN_G_HASH_SIZE; i++) {
5253		connfp = &ipst->ips_ipcl_globalhash_fanout[i];
5254		connp = NULL;
5255
5256		while ((connp = ipcl_get_next_conn(connfp, connp,
5257		    IPCL_UDPCONN))) {
5258			udp = connp->conn_udp;
5259			if (zoneid != connp->conn_zoneid)
5260				continue;
5261
5262			/*
5263			 * Note that the port numbers are sent in
5264			 * host byte order
5265			 */
5266
5267			if (udp->udp_state == TS_UNBND)
5268				state = MIB2_UDP_unbound;
5269			else if (udp->udp_state == TS_IDLE)
5270				state = MIB2_UDP_idle;
5271			else if (udp->udp_state == TS_DATA_XFER)
5272				state = MIB2_UDP_connected;
5273			else
5274				state = MIB2_UDP_unknown;
5275
5276			needattr = B_FALSE;
5277			bzero(&mlp, sizeof (mlp));
5278			if (connp->conn_mlp_type != mlptSingle) {
5279				if (connp->conn_mlp_type == mlptShared ||
5280				    connp->conn_mlp_type == mlptBoth)
5281					mlp.tme_flags |= MIB2_TMEF_SHARED;
5282				if (connp->conn_mlp_type == mlptPrivate ||
5283				    connp->conn_mlp_type == mlptBoth)
5284					mlp.tme_flags |= MIB2_TMEF_PRIVATE;
5285				needattr = B_TRUE;
5286			}
5287
5288			/*
5289			 * Create an IPv4 table entry for IPv4 entries and also
5290			 * any IPv6 entries which are bound to in6addr_any
5291			 * (i.e. anything a IPv4 peer could connect/send to).
5292			 */
5293			if (udp->udp_ipversion == IPV4_VERSION ||
5294			    (udp->udp_state <= TS_IDLE &&
5295			    IN6_IS_ADDR_UNSPECIFIED(&udp->udp_v6src))) {
5296				ude.udpEntryInfo.ue_state = state;
5297				/*
5298				 * If in6addr_any this will set it to
5299				 * INADDR_ANY
5300				 */
5301				ude.udpLocalAddress =
5302				    V4_PART_OF_V6(udp->udp_v6src);
5303				ude.udpLocalPort = ntohs(udp->udp_port);
5304				if (udp->udp_state == TS_DATA_XFER) {
5305					/*
5306					 * Can potentially get here for
5307					 * v6 socket if another process
5308					 * (say, ping) has just done a
5309					 * sendto(), changing the state
5310					 * from the TS_IDLE above to
5311					 * TS_DATA_XFER by the time we hit
5312					 * this part of the code.
5313					 */
5314					ude.udpEntryInfo.ue_RemoteAddress =
5315					    V4_PART_OF_V6(udp->udp_v6dst);
5316					ude.udpEntryInfo.ue_RemotePort =
5317					    ntohs(udp->udp_dstport);
5318				} else {
5319					ude.udpEntryInfo.ue_RemoteAddress = 0;
5320					ude.udpEntryInfo.ue_RemotePort = 0;
5321				}
5322
5323				/*
5324				 * We make the assumption that all udp_t
5325				 * structs will be created within an address
5326				 * region no larger than 32-bits.
5327				 */
5328				ude.udpInstance = (uint32_t)(uintptr_t)udp;
5329				ude.udpCreationProcess =
5330				    (udp->udp_open_pid < 0) ?
5331				    MIB2_UNKNOWN_PROCESS :
5332				    udp->udp_open_pid;
5333				ude.udpCreationTime = udp->udp_open_time;
5334
5335				(void) snmp_append_data2(mp_conn_ctl->b_cont,
5336				    &mp_conn_tail, (char *)&ude, sizeof (ude));
5337				mlp.tme_connidx = v4_conn_idx++;
5338				if (needattr)
5339					(void) snmp_append_data2(
5340					    mp_attr_ctl->b_cont, &mp_attr_tail,
5341					    (char *)&mlp, sizeof (mlp));
5342			}
5343			if (udp->udp_ipversion == IPV6_VERSION) {
5344				ude6.udp6EntryInfo.ue_state  = state;
5345				ude6.udp6LocalAddress = udp->udp_v6src;
5346				ude6.udp6LocalPort = ntohs(udp->udp_port);
5347				ude6.udp6IfIndex = udp->udp_bound_if;
5348				if (udp->udp_state == TS_DATA_XFER) {
5349					ude6.udp6EntryInfo.ue_RemoteAddress =
5350					    udp->udp_v6dst;
5351					ude6.udp6EntryInfo.ue_RemotePort =
5352					    ntohs(udp->udp_dstport);
5353				} else {
5354					ude6.udp6EntryInfo.ue_RemoteAddress =
5355					    sin6_null.sin6_addr;
5356					ude6.udp6EntryInfo.ue_RemotePort = 0;
5357				}
5358				/*
5359				 * We make the assumption that all udp_t
5360				 * structs will be created within an address
5361				 * region no larger than 32-bits.
5362				 */
5363				ude6.udp6Instance = (uint32_t)(uintptr_t)udp;
5364				ude6.udp6CreationProcess =
5365				    (udp->udp_open_pid < 0) ?
5366				    MIB2_UNKNOWN_PROCESS :
5367				    udp->udp_open_pid;
5368				ude6.udp6CreationTime = udp->udp_open_time;
5369
5370				(void) snmp_append_data2(mp6_conn_ctl->b_cont,
5371				    &mp6_conn_tail, (char *)&ude6,
5372				    sizeof (ude6));
5373				mlp.tme_connidx = v6_conn_idx++;
5374				if (needattr)
5375					(void) snmp_append_data2(
5376					    mp6_attr_ctl->b_cont,
5377					    &mp6_attr_tail, (char *)&mlp,
5378					    sizeof (mlp));
5379			}
5380		}
5381	}
5382
5383	/* IPv4 UDP endpoints */
5384	optp = (struct opthdr *)&mp_conn_ctl->b_rptr[
5385	    sizeof (struct T_optmgmt_ack)];
5386	optp->level = MIB2_UDP;
5387	optp->name = MIB2_UDP_ENTRY;
5388	optp->len = msgdsize(mp_conn_ctl->b_cont);
5389	qreply(q, mp_conn_ctl);
5390
5391	/* table of MLP attributes... */
5392	optp = (struct opthdr *)&mp_attr_ctl->b_rptr[
5393	    sizeof (struct T_optmgmt_ack)];
5394	optp->level = MIB2_UDP;
5395	optp->name = EXPER_XPORT_MLP;
5396	optp->len = msgdsize(mp_attr_ctl->b_cont);
5397	if (optp->len == 0)
5398		freemsg(mp_attr_ctl);
5399	else
5400		qreply(q, mp_attr_ctl);
5401
5402	/* IPv6 UDP endpoints */
5403	optp = (struct opthdr *)&mp6_conn_ctl->b_rptr[
5404	    sizeof (struct T_optmgmt_ack)];
5405	optp->level = MIB2_UDP6;
5406	optp->name = MIB2_UDP6_ENTRY;
5407	optp->len = msgdsize(mp6_conn_ctl->b_cont);
5408	qreply(q, mp6_conn_ctl);
5409
5410	/* table of MLP attributes... */
5411	optp = (struct opthdr *)&mp6_attr_ctl->b_rptr[
5412	    sizeof (struct T_optmgmt_ack)];
5413	optp->level = MIB2_UDP6;
5414	optp->name = EXPER_XPORT_MLP;
5415	optp->len = msgdsize(mp6_attr_ctl->b_cont);
5416	if (optp->len == 0)
5417		freemsg(mp6_attr_ctl);
5418	else
5419		qreply(q, mp6_attr_ctl);
5420
5421	return (mp2ctl);
5422}
5423
5424/*
5425 * Return 0 if invalid set request, 1 otherwise, including non-udp requests.
5426 * NOTE: Per MIB-II, UDP has no writable data.
5427 * TODO:  If this ever actually tries to set anything, it needs to be
5428 * to do the appropriate locking.
5429 */
5430/* ARGSUSED */
5431int
5432udp_snmp_set(queue_t *q, t_scalar_t level, t_scalar_t name,
5433    uchar_t *ptr, int len)
5434{
5435	switch (level) {
5436	case MIB2_UDP:
5437		return (0);
5438	default:
5439		return (1);
5440	}
5441}
5442
5443static void
5444udp_report_item(mblk_t *mp, udp_t *udp)
5445{
5446	char *state;
5447	char addrbuf1[INET6_ADDRSTRLEN];
5448	char addrbuf2[INET6_ADDRSTRLEN];
5449	uint_t print_len, buf_len;
5450
5451	buf_len = mp->b_datap->db_lim - mp->b_wptr;
5452	ASSERT(buf_len >= 0);
5453	if (buf_len == 0)
5454		return;
5455
5456	if (udp->udp_state == TS_UNBND)
5457		state = "UNBOUND";
5458	else if (udp->udp_state == TS_IDLE)
5459		state = "IDLE";
5460	else if (udp->udp_state == TS_DATA_XFER)
5461		state = "CONNECTED";
5462	else
5463		state = "UnkState";
5464	print_len = snprintf((char *)mp->b_wptr, buf_len,
5465	    MI_COL_PTRFMT_STR "%4d %5u %s %s %5u %s\n",
5466	    (void *)udp, udp->udp_connp->conn_zoneid, ntohs(udp->udp_port),
5467	    inet_ntop(AF_INET6, &udp->udp_v6src, addrbuf1, sizeof (addrbuf1)),
5468	    inet_ntop(AF_INET6, &udp->udp_v6dst, addrbuf2, sizeof (addrbuf2)),
5469	    ntohs(udp->udp_dstport), state);
5470	if (print_len < buf_len) {
5471		mp->b_wptr += print_len;
5472	} else {
5473		mp->b_wptr += buf_len;
5474	}
5475}
5476
5477/* Report for ndd "udp_status" */
5478/* ARGSUSED */
5479static int
5480udp_status_report(queue_t *q, mblk_t *mp, caddr_t cp, cred_t *cr)
5481{
5482	zoneid_t zoneid;
5483	connf_t	*connfp;
5484	conn_t	*connp = Q_TO_CONN(q);
5485	udp_t	*udp = connp->conn_udp;
5486	int	i;
5487	udp_stack_t *us = udp->udp_us;
5488	ip_stack_t *ipst = connp->conn_netstack->netstack_ip;
5489
5490	/*
5491	 * Because of the ndd constraint, at most we can have 64K buffer
5492	 * to put in all UDP info.  So to be more efficient, just
5493	 * allocate a 64K buffer here, assuming we need that large buffer.
5494	 * This may be a problem as any user can read udp_status.  Therefore
5495	 * we limit the rate of doing this using us_ndd_get_info_interval.
5496	 * This should be OK as normal users should not do this too often.
5497	 */
5498	if (cr == NULL || secpolicy_ip_config(cr, B_TRUE) != 0) {
5499		if (ddi_get_lbolt() - us->us_last_ndd_get_info_time <
5500		    drv_usectohz(us->us_ndd_get_info_interval * 1000)) {
5501			(void) mi_mpprintf(mp, NDD_TOO_QUICK_MSG);
5502			return (0);
5503		}
5504	}
5505	if ((mp->b_cont = allocb(ND_MAX_BUF_LEN, BPRI_HI)) == NULL) {
5506		/* The following may work even if we cannot get a large buf. */
5507		(void) mi_mpprintf(mp, NDD_OUT_OF_BUF_MSG);
5508		return (0);
5509	}
5510	(void) mi_mpprintf(mp,
5511	    "UDP     " MI_COL_HDRPAD_STR
5512	/*   12345678[89ABCDEF] */
5513	    " zone lport src addr        dest addr       port  state");
5514	/*    1234 12345 xxx.xxx.xxx.xxx xxx.xxx.xxx.xxx 12345 UNBOUND */
5515
5516	zoneid = connp->conn_zoneid;
5517
5518	for (i = 0; i < CONN_G_HASH_SIZE; i++) {
5519		connfp = &ipst->ips_ipcl_globalhash_fanout[i];
5520		connp = NULL;
5521
5522		while ((connp = ipcl_get_next_conn(connfp, connp,
5523		    IPCL_UDPCONN))) {
5524			udp = connp->conn_udp;
5525			if (zoneid != GLOBAL_ZONEID &&
5526			    zoneid != connp->conn_zoneid)
5527				continue;
5528
5529			udp_report_item(mp->b_cont, udp);
5530		}
5531	}
5532	us->us_last_ndd_get_info_time = ddi_get_lbolt();
5533	return (0);
5534}
5535
5536/*
5537 * This routine creates a T_UDERROR_IND message and passes it upstream.
5538 * The address and options are copied from the T_UNITDATA_REQ message
5539 * passed in mp.  This message is freed.
5540 */
5541static void
5542udp_ud_err(queue_t *q, mblk_t *mp, uchar_t *destaddr, t_scalar_t destlen,
5543    t_scalar_t err)
5544{
5545	struct T_unitdata_req *tudr;
5546	mblk_t	*mp1;
5547	uchar_t	*optaddr;
5548	t_scalar_t optlen;
5549
5550	if (DB_TYPE(mp) == M_DATA) {
5551		ASSERT(destaddr != NULL && destlen != 0);
5552		optaddr = NULL;
5553		optlen = 0;
5554	} else {
5555		if ((mp->b_wptr < mp->b_rptr) ||
5556		    (MBLKL(mp)) < sizeof (struct T_unitdata_req)) {
5557			goto done;
5558		}
5559		tudr = (struct T_unitdata_req *)mp->b_rptr;
5560		destaddr = mp->b_rptr + tudr->DEST_offset;
5561		if (destaddr < mp->b_rptr || destaddr >= mp->b_wptr ||
5562		    destaddr + tudr->DEST_length < mp->b_rptr ||
5563		    destaddr + tudr->DEST_length > mp->b_wptr) {
5564			goto done;
5565		}
5566		optaddr = mp->b_rptr + tudr->OPT_offset;
5567		if (optaddr < mp->b_rptr || optaddr >= mp->b_wptr ||
5568		    optaddr + tudr->OPT_length < mp->b_rptr ||
5569		    optaddr + tudr->OPT_length > mp->b_wptr) {
5570			goto done;
5571		}
5572		destlen = tudr->DEST_length;
5573		optlen = tudr->OPT_length;
5574	}
5575
5576	mp1 = mi_tpi_uderror_ind((char *)destaddr, destlen,
5577	    (char *)optaddr, optlen, err);
5578	if (mp1 != NULL)
5579		qreply(q, mp1);
5580
5581done:
5582	freemsg(mp);
5583}
5584
5585/*
5586 * This routine removes a port number association from a stream.  It
5587 * is called by udp_wput to handle T_UNBIND_REQ messages.
5588 */
5589static void
5590udp_unbind(queue_t *q, mblk_t *mp)
5591{
5592	udp_t *udp = Q_TO_UDP(q);
5593	udp_fanout_t	*udpf;
5594	udp_stack_t	*us = udp->udp_us;
5595
5596	if (cl_inet_unbind != NULL) {
5597		/*
5598		 * Running in cluster mode - register unbind information
5599		 */
5600		if (udp->udp_ipversion == IPV4_VERSION) {
5601			(*cl_inet_unbind)(IPPROTO_UDP, AF_INET,
5602			    (uint8_t *)(&V4_PART_OF_V6(udp->udp_v6src)),
5603			    (in_port_t)udp->udp_port);
5604		} else {
5605			(*cl_inet_unbind)(IPPROTO_UDP, AF_INET6,
5606			    (uint8_t *)&(udp->udp_v6src),
5607			    (in_port_t)udp->udp_port);
5608		}
5609	}
5610
5611	rw_enter(&udp->udp_rwlock, RW_WRITER);
5612	if (udp->udp_state == TS_UNBND || udp->udp_pending_op != -1) {
5613		rw_exit(&udp->udp_rwlock);
5614		udp_err_ack(q, mp, TOUTSTATE, 0);
5615		return;
5616	}
5617	udp->udp_pending_op = T_UNBIND_REQ;
5618	rw_exit(&udp->udp_rwlock);
5619
5620	/*
5621	 * Pass the unbind to IP; T_UNBIND_REQ is larger than T_OK_ACK
5622	 * and therefore ip_unbind must never return NULL.
5623	 */
5624	mp = ip_unbind(q, mp);
5625	ASSERT(mp != NULL);
5626	ASSERT(((struct T_ok_ack *)mp->b_rptr)->PRIM_type == T_OK_ACK);
5627
5628	/*
5629	 * Once we're unbound from IP, the pending operation may be cleared
5630	 * here.
5631	 */
5632	rw_enter(&udp->udp_rwlock, RW_WRITER);
5633	udpf = &us->us_bind_fanout[UDP_BIND_HASH(udp->udp_port,
5634	    us->us_bind_fanout_size)];
5635	mutex_enter(&udpf->uf_lock);
5636	udp_bind_hash_remove(udp, B_TRUE);
5637	V6_SET_ZERO(udp->udp_v6src);
5638	V6_SET_ZERO(udp->udp_bound_v6src);
5639	udp->udp_port = 0;
5640	mutex_exit(&udpf->uf_lock);
5641
5642	udp->udp_pending_op = -1;
5643	udp->udp_state = TS_UNBND;
5644	if (udp->udp_family == AF_INET6)
5645		(void) udp_build_hdrs(udp);
5646	rw_exit(&udp->udp_rwlock);
5647
5648	qreply(q, mp);
5649}
5650
5651/*
5652 * Don't let port fall into the privileged range.
5653 * Since the extra privileged ports can be arbitrary we also
5654 * ensure that we exclude those from consideration.
5655 * us->us_epriv_ports is not sorted thus we loop over it until
5656 * there are no changes.
5657 */
5658static in_port_t
5659udp_update_next_port(udp_t *udp, in_port_t port, boolean_t random)
5660{
5661	int i;
5662	in_port_t nextport;
5663	boolean_t restart = B_FALSE;
5664	udp_stack_t *us = udp->udp_us;
5665
5666	if (random && udp_random_anon_port != 0) {
5667		(void) random_get_pseudo_bytes((uint8_t *)&port,
5668		    sizeof (in_port_t));
5669		/*
5670		 * Unless changed by a sys admin, the smallest anon port
5671		 * is 32768 and the largest anon port is 65535.  It is
5672		 * very likely (50%) for the random port to be smaller
5673		 * than the smallest anon port.  When that happens,
5674		 * add port % (anon port range) to the smallest anon
5675		 * port to get the random port.  It should fall into the
5676		 * valid anon port range.
5677		 */
5678		if (port < us->us_smallest_anon_port) {
5679			port = us->us_smallest_anon_port +
5680			    port % (us->us_largest_anon_port -
5681			    us->us_smallest_anon_port);
5682		}
5683	}
5684
5685retry:
5686	if (port < us->us_smallest_anon_port)
5687		port = us->us_smallest_anon_port;
5688
5689	if (port > us->us_largest_anon_port) {
5690		port = us->us_smallest_anon_port;
5691		if (restart)
5692			return (0);
5693		restart = B_TRUE;
5694	}
5695
5696	if (port < us->us_smallest_nonpriv_port)
5697		port = us->us_smallest_nonpriv_port;
5698
5699	for (i = 0; i < us->us_num_epriv_ports; i++) {
5700		if (port == us->us_epriv_ports[i]) {
5701			port++;
5702			/*
5703			 * Make sure that the port is in the
5704			 * valid range.
5705			 */
5706			goto retry;
5707		}
5708	}
5709
5710	if (is_system_labeled() &&
5711	    (nextport = tsol_next_port(crgetzone(udp->udp_connp->conn_cred),
5712	    port, IPPROTO_UDP, B_TRUE)) != 0) {
5713		port = nextport;
5714		goto retry;
5715	}
5716
5717	return (port);
5718}
5719
5720static int
5721udp_update_label(queue_t *wq, mblk_t *mp, ipaddr_t dst)
5722{
5723	int err;
5724	uchar_t opt_storage[IP_MAX_OPT_LENGTH];
5725	udp_t *udp = Q_TO_UDP(wq);
5726	udp_stack_t	*us = udp->udp_us;
5727
5728	err = tsol_compute_label(DB_CREDDEF(mp, udp->udp_connp->conn_cred), dst,
5729	    opt_storage, udp->udp_connp->conn_mac_exempt,
5730	    us->us_netstack->netstack_ip);
5731	if (err == 0) {
5732		err = tsol_update_options(&udp->udp_ip_snd_options,
5733		    &udp->udp_ip_snd_options_len, &udp->udp_label_len,
5734		    opt_storage);
5735	}
5736	if (err != 0) {
5737		DTRACE_PROBE4(
5738		    tx__ip__log__info__updatelabel__udp,
5739		    char *, "queue(1) failed to update options(2) on mp(3)",
5740		    queue_t *, wq, char *, opt_storage, mblk_t *, mp);
5741	} else {
5742		IN6_IPADDR_TO_V4MAPPED(dst, &udp->udp_v6lastdst);
5743	}
5744	return (err);
5745}
5746
5747static mblk_t *
5748udp_output_v4(conn_t *connp, mblk_t *mp, ipaddr_t v4dst, uint16_t port,
5749    uint_t srcid, int *error, boolean_t insert_spi)
5750{
5751	udp_t	*udp = connp->conn_udp;
5752	queue_t	*q = connp->conn_wq;
5753	mblk_t	*mp1 = mp;
5754	mblk_t	*mp2;
5755	ipha_t	*ipha;
5756	int	ip_hdr_length;
5757	uint32_t ip_len;
5758	udpha_t	*udpha;
5759	boolean_t lock_held = B_FALSE;
5760	in_port_t	uha_src_port;
5761	udpattrs_t	attrs;
5762	uchar_t	ip_snd_opt[IP_MAX_OPT_LENGTH];
5763	uint32_t	ip_snd_opt_len = 0;
5764	ip4_pkt_t  pktinfo;
5765	ip4_pkt_t  *pktinfop = &pktinfo;
5766	ip_opt_info_t optinfo;
5767	ip_stack_t	*ipst = connp->conn_netstack->netstack_ip;
5768	udp_stack_t	*us = udp->udp_us;
5769	ipsec_stack_t	*ipss = ipst->ips_netstack->netstack_ipsec;
5770
5771
5772	*error = 0;
5773	pktinfop->ip4_ill_index = 0;
5774	pktinfop->ip4_addr = INADDR_ANY;
5775	optinfo.ip_opt_flags = 0;
5776	optinfo.ip_opt_ill_index = 0;
5777
5778	if (v4dst == INADDR_ANY)
5779		v4dst = htonl(INADDR_LOOPBACK);
5780
5781	/*
5782	 * If options passed in, feed it for verification and handling
5783	 */
5784	attrs.udpattr_credset = B_FALSE;
5785	if (DB_TYPE(mp) != M_DATA) {
5786		mp1 = mp->b_cont;
5787		if (((struct T_unitdata_req *)mp->b_rptr)->OPT_length != 0) {
5788			attrs.udpattr_ipp4 = pktinfop;
5789			attrs.udpattr_mb = mp;
5790			if (udp_unitdata_opt_process(q, mp, error, &attrs) < 0)
5791				goto done;
5792			/*
5793			 * Note: success in processing options.
5794			 * mp option buffer represented by
5795			 * OPT_length/offset now potentially modified
5796			 * and contain option setting results
5797			 */
5798			ASSERT(*error == 0);
5799		}
5800	}
5801
5802	/* mp1 points to the M_DATA mblk carrying the packet */
5803	ASSERT(mp1 != NULL && DB_TYPE(mp1) == M_DATA);
5804
5805	rw_enter(&udp->udp_rwlock, RW_READER);
5806	lock_held = B_TRUE;
5807	/*
5808	 * Check if our saved options are valid; update if not.
5809	 * TSOL Note: Since we are not in WRITER mode, UDP packets
5810	 * to different destination may require different labels,
5811	 * or worse, UDP packets to same IP address may require
5812	 * different labels due to use of shared all-zones address.
5813	 * We use conn_lock to ensure that lastdst, ip_snd_options,
5814	 * and ip_snd_options_len are consistent for the current
5815	 * destination and are updated atomically.
5816	 */
5817	mutex_enter(&connp->conn_lock);
5818	if (is_system_labeled()) {
5819		/* Using UDP MLP requires SCM_UCRED from user */
5820		if (connp->conn_mlp_type != mlptSingle &&
5821		    !attrs.udpattr_credset) {
5822			mutex_exit(&connp->conn_lock);
5823			DTRACE_PROBE4(
5824			    tx__ip__log__info__output__udp,
5825			    char *, "MLP mp(1) lacks SCM_UCRED attr(2) on q(3)",
5826			    mblk_t *, mp1, udpattrs_t *, &attrs, queue_t *, q);
5827			*error = ECONNREFUSED;
5828			goto done;
5829		}
5830		/*
5831		 * update label option for this UDP socket if
5832		 * - the destination has changed, or
5833		 * - the UDP socket is MLP
5834		 */
5835		if ((!IN6_IS_ADDR_V4MAPPED(&udp->udp_v6lastdst) ||
5836		    V4_PART_OF_V6(udp->udp_v6lastdst) != v4dst ||
5837		    connp->conn_mlp_type != mlptSingle) &&
5838		    (*error = udp_update_label(q, mp, v4dst)) != 0) {
5839			mutex_exit(&connp->conn_lock);
5840			goto done;
5841		}
5842	}
5843	if (udp->udp_ip_snd_options_len > 0) {
5844		ip_snd_opt_len = udp->udp_ip_snd_options_len;
5845		bcopy(udp->udp_ip_snd_options, ip_snd_opt, ip_snd_opt_len);
5846	}
5847	mutex_exit(&connp->conn_lock);
5848
5849	/* Add an IP header */
5850	ip_hdr_length = IP_SIMPLE_HDR_LENGTH + UDPH_SIZE + ip_snd_opt_len +
5851	    (insert_spi ? sizeof (uint32_t) : 0);
5852	ipha = (ipha_t *)&mp1->b_rptr[-ip_hdr_length];
5853	if (DB_REF(mp1) != 1 || (uchar_t *)ipha < DB_BASE(mp1) ||
5854	    !OK_32PTR(ipha)) {
5855		mp2 = allocb(ip_hdr_length + us->us_wroff_extra, BPRI_LO);
5856		if (mp2 == NULL) {
5857			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
5858			    "udp_wput_end: q %p (%S)", q, "allocbfail2");
5859			*error = ENOMEM;
5860			goto done;
5861		}
5862		mp2->b_wptr = DB_LIM(mp2);
5863		mp2->b_cont = mp1;
5864		mp1 = mp2;
5865		if (DB_TYPE(mp) != M_DATA)
5866			mp->b_cont = mp1;
5867		else
5868			mp = mp1;
5869
5870		ipha = (ipha_t *)(mp1->b_wptr - ip_hdr_length);
5871	}
5872	ip_hdr_length -= (UDPH_SIZE + (insert_spi ? sizeof (uint32_t) : 0));
5873#ifdef	_BIG_ENDIAN
5874	/* Set version, header length, and tos */
5875	*(uint16_t *)&ipha->ipha_version_and_hdr_length =
5876	    ((((IP_VERSION << 4) | (ip_hdr_length>>2)) << 8) |
5877	    udp->udp_type_of_service);
5878	/* Set ttl and protocol */
5879	*(uint16_t *)&ipha->ipha_ttl = (udp->udp_ttl << 8) | IPPROTO_UDP;
5880#else
5881	/* Set version, header length, and tos */
5882	*(uint16_t *)&ipha->ipha_version_and_hdr_length =
5883	    ((udp->udp_type_of_service << 8) |
5884	    ((IP_VERSION << 4) | (ip_hdr_length>>2)));
5885	/* Set ttl and protocol */
5886	*(uint16_t *)&ipha->ipha_ttl = (IPPROTO_UDP << 8) | udp->udp_ttl;
5887#endif
5888	if (pktinfop->ip4_addr != INADDR_ANY) {
5889		ipha->ipha_src = pktinfop->ip4_addr;
5890		optinfo.ip_opt_flags = IP_VERIFY_SRC;
5891	} else {
5892		/*
5893		 * Copy our address into the packet.  If this is zero,
5894		 * first look at __sin6_src_id for a hint. If we leave the
5895		 * source as INADDR_ANY then ip will fill in the real source
5896		 * address.
5897		 */
5898		IN6_V4MAPPED_TO_IPADDR(&udp->udp_v6src, ipha->ipha_src);
5899		if (srcid != 0 && ipha->ipha_src == INADDR_ANY) {
5900			in6_addr_t v6src;
5901
5902			ip_srcid_find_id(srcid, &v6src, connp->conn_zoneid,
5903			    us->us_netstack);
5904			IN6_V4MAPPED_TO_IPADDR(&v6src, ipha->ipha_src);
5905		}
5906	}
5907	uha_src_port = udp->udp_port;
5908	if (ip_hdr_length == IP_SIMPLE_HDR_LENGTH) {
5909		rw_exit(&udp->udp_rwlock);
5910		lock_held = B_FALSE;
5911	}
5912
5913	if (pktinfop->ip4_ill_index != 0) {
5914		optinfo.ip_opt_ill_index = pktinfop->ip4_ill_index;
5915	}
5916
5917	ipha->ipha_fragment_offset_and_flags = 0;
5918	ipha->ipha_ident = 0;
5919
5920	mp1->b_rptr = (uchar_t *)ipha;
5921
5922	ASSERT((uintptr_t)(mp1->b_wptr - (uchar_t *)ipha) <=
5923	    (uintptr_t)UINT_MAX);
5924
5925	/* Determine length of packet */
5926	ip_len = (uint32_t)(mp1->b_wptr - (uchar_t *)ipha);
5927	if ((mp2 = mp1->b_cont) != NULL) {
5928		do {
5929			ASSERT((uintptr_t)MBLKL(mp2) <= (uintptr_t)UINT_MAX);
5930			ip_len += (uint32_t)MBLKL(mp2);
5931		} while ((mp2 = mp2->b_cont) != NULL);
5932	}
5933	/*
5934	 * If the size of the packet is greater than the maximum allowed by
5935	 * ip, return an error. Passing this down could cause panics because
5936	 * the size will have wrapped and be inconsistent with the msg size.
5937	 */
5938	if (ip_len > IP_MAXPACKET) {
5939		TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
5940		    "udp_wput_end: q %p (%S)", q, "IP length exceeded");
5941		*error = EMSGSIZE;
5942		goto done;
5943	}
5944	ipha->ipha_length = htons((uint16_t)ip_len);
5945	ip_len -= ip_hdr_length;
5946	ip_len = htons((uint16_t)ip_len);
5947	udpha = (udpha_t *)(((uchar_t *)ipha) + ip_hdr_length);
5948
5949	/* Insert all-0s SPI now. */
5950	if (insert_spi)
5951		*((uint32_t *)(udpha + 1)) = 0;
5952
5953	/*
5954	 * Copy in the destination address
5955	 */
5956	ipha->ipha_dst = v4dst;
5957
5958	/*
5959	 * Set ttl based on IP_MULTICAST_TTL to match IPv6 logic.
5960	 */
5961	if (CLASSD(v4dst))
5962		ipha->ipha_ttl = udp->udp_multicast_ttl;
5963
5964	udpha->uha_dst_port = port;
5965	udpha->uha_src_port = uha_src_port;
5966
5967	if (ip_snd_opt_len > 0) {
5968		uint32_t	cksum;
5969
5970		bcopy(ip_snd_opt, &ipha[1], ip_snd_opt_len);
5971		lock_held = B_FALSE;
5972		rw_exit(&udp->udp_rwlock);
5973		/*
5974		 * Massage source route putting first source route in ipha_dst.
5975		 * Ignore the destination in T_unitdata_req.
5976		 * Create a checksum adjustment for a source route, if any.
5977		 */
5978		cksum = ip_massage_options(ipha, us->us_netstack);
5979		cksum = (cksum & 0xFFFF) + (cksum >> 16);
5980		cksum -= ((ipha->ipha_dst >> 16) & 0xFFFF) +
5981		    (ipha->ipha_dst & 0xFFFF);
5982		if ((int)cksum < 0)
5983			cksum--;
5984		cksum = (cksum & 0xFFFF) + (cksum >> 16);
5985		/*
5986		 * IP does the checksum if uha_checksum is non-zero,
5987		 * We make it easy for IP to include our pseudo header
5988		 * by putting our length in uha_checksum.
5989		 */
5990		cksum += ip_len;
5991		cksum = (cksum & 0xFFFF) + (cksum >> 16);
5992		/* There might be a carry. */
5993		cksum = (cksum & 0xFFFF) + (cksum >> 16);
5994#ifdef _LITTLE_ENDIAN
5995		if (us->us_do_checksum)
5996			ip_len = (cksum << 16) | ip_len;
5997#else
5998		if (us->us_do_checksum)
5999			ip_len = (ip_len << 16) | cksum;
6000		else
6001			ip_len <<= 16;
6002#endif
6003	} else {
6004		/*
6005		 * IP does the checksum if uha_checksum is non-zero,
6006		 * We make it easy for IP to include our pseudo header
6007		 * by putting our length in uha_checksum.
6008		 */
6009		if (us->us_do_checksum)
6010			ip_len |= (ip_len << 16);
6011#ifndef _LITTLE_ENDIAN
6012		else
6013			ip_len <<= 16;
6014#endif
6015	}
6016	ASSERT(!lock_held);
6017	/* Set UDP length and checksum */
6018	*((uint32_t *)&udpha->uha_length) = ip_len;
6019	if (DB_CRED(mp) != NULL)
6020		mblk_setcred(mp1, DB_CRED(mp));
6021
6022	if (DB_TYPE(mp) != M_DATA) {
6023		ASSERT(mp != mp1);
6024		freeb(mp);
6025	}
6026
6027	/* mp has been consumed and we'll return success */
6028	ASSERT(*error == 0);
6029	mp = NULL;
6030
6031	/* We're done.  Pass the packet to ip. */
6032	BUMP_MIB(&us->us_udp_mib, udpHCOutDatagrams);
6033	TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
6034	    "udp_wput_end: q %p (%S)", q, "end");
6035
6036	if ((connp->conn_flags & IPCL_CHECK_POLICY) != 0 ||
6037	    CONN_OUTBOUND_POLICY_PRESENT(connp, ipss) ||
6038	    connp->conn_dontroute ||
6039	    connp->conn_nofailover_ill != NULL ||
6040	    connp->conn_outgoing_ill != NULL || optinfo.ip_opt_flags != 0 ||
6041	    optinfo.ip_opt_ill_index != 0 ||
6042	    ipha->ipha_version_and_hdr_length != IP_SIMPLE_HDR_VERSION ||
6043	    IPP_ENABLED(IPP_LOCAL_OUT, ipst) ||
6044	    ipst->ips_ip_g_mrouter != NULL) {
6045		UDP_STAT(us, udp_ip_send);
6046		ip_output_options(connp, mp1, connp->conn_wq, IP_WPUT,
6047		    &optinfo);
6048	} else {
6049		udp_send_data(udp, connp->conn_wq, mp1, ipha);
6050	}
6051
6052done:
6053	if (lock_held)
6054		rw_exit(&udp->udp_rwlock);
6055	if (*error != 0) {
6056		ASSERT(mp != NULL);
6057		BUMP_MIB(&us->us_udp_mib, udpOutErrors);
6058	}
6059	return (mp);
6060}
6061
6062static void
6063udp_send_data(udp_t *udp, queue_t *q, mblk_t *mp, ipha_t *ipha)
6064{
6065	conn_t	*connp = udp->udp_connp;
6066	ipaddr_t src, dst;
6067	ire_t	*ire;
6068	ipif_t	*ipif = NULL;
6069	mblk_t	*ire_fp_mp;
6070	boolean_t retry_caching;
6071	udp_stack_t *us = udp->udp_us;
6072	ip_stack_t	*ipst = connp->conn_netstack->netstack_ip;
6073
6074	dst = ipha->ipha_dst;
6075	src = ipha->ipha_src;
6076	ASSERT(ipha->ipha_ident == 0);
6077
6078	if (CLASSD(dst)) {
6079		int err;
6080
6081		ipif = conn_get_held_ipif(connp,
6082		    &connp->conn_multicast_ipif, &err);
6083
6084		if (ipif == NULL || ipif->ipif_isv6 ||
6085		    (ipif->ipif_ill->ill_phyint->phyint_flags &
6086		    PHYI_LOOPBACK)) {
6087			if (ipif != NULL)
6088				ipif_refrele(ipif);
6089			UDP_STAT(us, udp_ip_send);
6090			ip_output(connp, mp, q, IP_WPUT);
6091			return;
6092		}
6093	}
6094
6095	retry_caching = B_FALSE;
6096	mutex_enter(&connp->conn_lock);
6097	ire = connp->conn_ire_cache;
6098	ASSERT(!(connp->conn_state_flags & CONN_INCIPIENT));
6099
6100	if (ire == NULL || ire->ire_addr != dst ||
6101	    (ire->ire_marks & IRE_MARK_CONDEMNED)) {
6102		retry_caching = B_TRUE;
6103	} else if (CLASSD(dst) && (ire->ire_type & IRE_CACHE)) {
6104		ill_t *stq_ill = (ill_t *)ire->ire_stq->q_ptr;
6105
6106		ASSERT(ipif != NULL);
6107		if (stq_ill != ipif->ipif_ill && (stq_ill->ill_group == NULL ||
6108		    stq_ill->ill_group != ipif->ipif_ill->ill_group))
6109			retry_caching = B_TRUE;
6110	}
6111
6112	if (!retry_caching) {
6113		ASSERT(ire != NULL);
6114		IRE_REFHOLD(ire);
6115		mutex_exit(&connp->conn_lock);
6116	} else {
6117		boolean_t cached = B_FALSE;
6118
6119		connp->conn_ire_cache = NULL;
6120		mutex_exit(&connp->conn_lock);
6121
6122		/* Release the old ire */
6123		if (ire != NULL) {
6124			IRE_REFRELE_NOTR(ire);
6125			ire = NULL;
6126		}
6127
6128		if (CLASSD(dst)) {
6129			ASSERT(ipif != NULL);
6130			ire = ire_ctable_lookup(dst, 0, 0, ipif,
6131			    connp->conn_zoneid, MBLK_GETLABEL(mp),
6132			    MATCH_IRE_ILL_GROUP, ipst);
6133		} else {
6134			ASSERT(ipif == NULL);
6135			ire = ire_cache_lookup(dst, connp->conn_zoneid,
6136			    MBLK_GETLABEL(mp), ipst);
6137		}
6138
6139		if (ire == NULL) {
6140			if (ipif != NULL)
6141				ipif_refrele(ipif);
6142			UDP_STAT(us, udp_ire_null);
6143			ip_output(connp, mp, q, IP_WPUT);
6144			return;
6145		}
6146		IRE_REFHOLD_NOTR(ire);
6147
6148		mutex_enter(&connp->conn_lock);
6149		if (CONN_CACHE_IRE(connp) && connp->conn_ire_cache == NULL &&
6150		    !(ire->ire_marks & IRE_MARK_CONDEMNED)) {
6151			irb_t		*irb = ire->ire_bucket;
6152
6153			/*
6154			 * IRE's created for non-connection oriented transports
6155			 * are normally initialized with IRE_MARK_TEMPORARY set
6156			 * in the ire_marks. These IRE's are preferentially
6157			 * reaped when the hash chain length in the cache
6158			 * bucket exceeds the maximum value specified in
6159			 * ip[6]_ire_max_bucket_cnt. This can severely affect
6160			 * UDP performance if IRE cache entries that we need
6161			 * to reuse are continually removed. To remedy this,
6162			 * when we cache the IRE in the conn_t, we remove the
6163			 * IRE_MARK_TEMPORARY bit from the ire_marks if it was
6164			 * set.
6165			 */
6166			if (ire->ire_marks & IRE_MARK_TEMPORARY) {
6167				rw_enter(&irb->irb_lock, RW_WRITER);
6168				if (ire->ire_marks & IRE_MARK_TEMPORARY) {
6169					ire->ire_marks &= ~IRE_MARK_TEMPORARY;
6170					irb->irb_tmp_ire_cnt--;
6171				}
6172				rw_exit(&irb->irb_lock);
6173			}
6174			connp->conn_ire_cache = ire;
6175			cached = B_TRUE;
6176		}
6177		mutex_exit(&connp->conn_lock);
6178
6179		/*
6180		 * We can continue to use the ire but since it was not
6181		 * cached, we should drop the extra reference.
6182		 */
6183		if (!cached)
6184			IRE_REFRELE_NOTR(ire);
6185	}
6186	ASSERT(ire != NULL && ire->ire_ipversion == IPV4_VERSION);
6187	ASSERT(!CLASSD(dst) || ipif != NULL);
6188
6189	/*
6190	 * Check if we can take the fast-path.
6191	 * Note that "incomplete" ire's (where the link-layer for next hop
6192	 * is not resolved, or where the fast-path header in nce_fp_mp is not
6193	 * available yet) are sent down the legacy (slow) path
6194	 */
6195	if ((ire->ire_type & (IRE_BROADCAST|IRE_LOCAL|IRE_LOOPBACK)) ||
6196	    (ire->ire_flags & RTF_MULTIRT) || (ire->ire_stq == NULL) ||
6197	    (ire->ire_max_frag < ntohs(ipha->ipha_length)) ||
6198	    ((ire->ire_nce == NULL) ||
6199	    ((ire_fp_mp = ire->ire_nce->nce_fp_mp) == NULL)) ||
6200	    connp->conn_nexthop_set || (MBLKL(ire_fp_mp) > MBLKHEAD(mp))) {
6201		if (ipif != NULL)
6202			ipif_refrele(ipif);
6203		UDP_STAT(us, udp_ip_ire_send);
6204		IRE_REFRELE(ire);
6205		ip_output(connp, mp, q, IP_WPUT);
6206		return;
6207	}
6208
6209	if (src == INADDR_ANY && !connp->conn_unspec_src) {
6210		if (CLASSD(dst) && !(ire->ire_flags & RTF_SETSRC))
6211			ipha->ipha_src = ipif->ipif_src_addr;
6212		else
6213			ipha->ipha_src = ire->ire_src_addr;
6214	}
6215
6216	if (ipif != NULL)
6217		ipif_refrele(ipif);
6218
6219	udp_xmit(connp->conn_wq, mp, ire, connp, connp->conn_zoneid);
6220}
6221
6222static void
6223udp_xmit(queue_t *q, mblk_t *mp, ire_t *ire, conn_t *connp, zoneid_t zoneid)
6224{
6225	ipaddr_t src, dst;
6226	ill_t	*ill;
6227	mblk_t	*ire_fp_mp;
6228	uint_t	ire_fp_mp_len;
6229	uint16_t *up;
6230	uint32_t cksum, hcksum_txflags;
6231	queue_t	*dev_q;
6232	udp_t	*udp = connp->conn_udp;
6233	ipha_t	*ipha = (ipha_t *)mp->b_rptr;
6234	udp_stack_t	*us = udp->udp_us;
6235	ip_stack_t	*ipst = connp->conn_netstack->netstack_ip;
6236	boolean_t ll_multicast = B_FALSE;
6237
6238	dev_q = ire->ire_stq->q_next;
6239	ASSERT(dev_q != NULL);
6240
6241
6242	if (DEV_Q_IS_FLOW_CTLED(dev_q)) {
6243		BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsHCOutRequests);
6244		BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsOutDiscards);
6245		if (ipst->ips_ip_output_queue)
6246			(void) putq(connp->conn_wq, mp);
6247		else
6248			freemsg(mp);
6249		ire_refrele(ire);
6250		return;
6251	}
6252
6253	ire_fp_mp = ire->ire_nce->nce_fp_mp;
6254	ire_fp_mp_len = MBLKL(ire_fp_mp);
6255	ASSERT(MBLKHEAD(mp) >= ire_fp_mp_len);
6256
6257	dst = ipha->ipha_dst;
6258	src = ipha->ipha_src;
6259
6260	ill = ire_to_ill(ire);
6261	ASSERT(ill != NULL);
6262
6263	BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCOutRequests);
6264
6265	ipha->ipha_ident = (uint16_t)atomic_add_32_nv(&ire->ire_ident, 1);
6266#ifndef _BIG_ENDIAN
6267	ipha->ipha_ident = (ipha->ipha_ident << 8) | (ipha->ipha_ident >> 8);
6268#endif
6269
6270	if (ILL_HCKSUM_CAPABLE(ill) && dohwcksum) {
6271		ASSERT(ill->ill_hcksum_capab != NULL);
6272		hcksum_txflags = ill->ill_hcksum_capab->ill_hcksum_txflags;
6273	} else {
6274		hcksum_txflags = 0;
6275	}
6276
6277	/* pseudo-header checksum (do it in parts for IP header checksum) */
6278	cksum = (dst >> 16) + (dst & 0xFFFF) + (src >> 16) + (src & 0xFFFF);
6279
6280	ASSERT(ipha->ipha_version_and_hdr_length == IP_SIMPLE_HDR_VERSION);
6281	up = IPH_UDPH_CHECKSUMP(ipha, IP_SIMPLE_HDR_LENGTH);
6282	if (*up != 0) {
6283		IP_CKSUM_XMIT_FAST(ire->ire_ipversion, hcksum_txflags,
6284		    mp, ipha, up, IPPROTO_UDP, IP_SIMPLE_HDR_LENGTH,
6285		    ntohs(ipha->ipha_length), cksum);
6286
6287		/* Software checksum? */
6288		if (DB_CKSUMFLAGS(mp) == 0) {
6289			UDP_STAT(us, udp_out_sw_cksum);
6290			UDP_STAT_UPDATE(us, udp_out_sw_cksum_bytes,
6291			    ntohs(ipha->ipha_length) - IP_SIMPLE_HDR_LENGTH);
6292		}
6293	}
6294
6295	if (!CLASSD(dst)) {
6296		ipha->ipha_fragment_offset_and_flags |=
6297		    (uint32_t)htons(ire->ire_frag_flag);
6298	}
6299
6300	/* Calculate IP header checksum if hardware isn't capable */
6301	if (!(DB_CKSUMFLAGS(mp) & HCK_IPV4_HDRCKSUM)) {
6302		IP_HDR_CKSUM(ipha, cksum, ((uint32_t *)ipha)[0],
6303		    ((uint16_t *)ipha)[4]);
6304	}
6305
6306	if (CLASSD(dst)) {
6307		boolean_t ilm_exists;
6308
6309		ILM_WALKER_HOLD(ill);
6310		ilm_exists = (ilm_lookup_ill(ill, dst, ALL_ZONES) != NULL);
6311		ILM_WALKER_RELE(ill);
6312		if (ilm_exists) {
6313			ip_multicast_loopback(q, ill, mp,
6314			    connp->conn_multicast_loop ? 0 :
6315			    IP_FF_NO_MCAST_LOOP, zoneid);
6316		}
6317
6318		/* If multicast TTL is 0 then we are done */
6319		if (ipha->ipha_ttl == 0) {
6320			freemsg(mp);
6321			ire_refrele(ire);
6322			return;
6323		}
6324		ll_multicast = B_TRUE;
6325	}
6326
6327	ASSERT(DB_TYPE(ire_fp_mp) == M_DATA);
6328	mp->b_rptr = (uchar_t *)ipha - ire_fp_mp_len;
6329	bcopy(ire_fp_mp->b_rptr, mp->b_rptr, ire_fp_mp_len);
6330
6331	UPDATE_OB_PKT_COUNT(ire);
6332	ire->ire_last_used_time = lbolt;
6333
6334	BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCOutTransmits);
6335	UPDATE_MIB(ill->ill_ip_mib, ipIfStatsHCOutOctets,
6336	    ntohs(ipha->ipha_length));
6337
6338	if (ILL_DLS_CAPABLE(ill)) {
6339		/*
6340		 * Send the packet directly to DLD, where it may be queued
6341		 * depending on the availability of transmit resources at
6342		 * the media layer.
6343		 */
6344		IP_DLS_ILL_TX(ill, ipha, mp, ipst);
6345	} else {
6346		DTRACE_PROBE4(ip4__physical__out__start,
6347		    ill_t *, NULL, ill_t *, ill,
6348		    ipha_t *, ipha, mblk_t *, mp);
6349		FW_HOOKS(ipst->ips_ip4_physical_out_event,
6350		    ipst->ips_ipv4firewall_physical_out,
6351		    NULL, ill, ipha, mp, mp, ll_multicast, ipst);
6352		DTRACE_PROBE1(ip4__physical__out__end, mblk_t *, mp);
6353		if (mp != NULL) {
6354			DTRACE_IP7(send, mblk_t *, mp, conn_t *, NULL,
6355			    void_ip_t *, ipha, __dtrace_ipsr_ill_t *, ill,
6356			    ipha_t *, ipha, ip6_t *, NULL, int, 0);
6357			putnext(ire->ire_stq, mp);
6358		}
6359	}
6360
6361	IRE_REFRELE(ire);
6362}
6363
6364static boolean_t
6365udp_update_label_v6(queue_t *wq, mblk_t *mp, in6_addr_t *dst)
6366{
6367	udp_t *udp = Q_TO_UDP(wq);
6368	int err;
6369	uchar_t opt_storage[TSOL_MAX_IPV6_OPTION];
6370	udp_stack_t		*us = udp->udp_us;
6371
6372	err = tsol_compute_label_v6(DB_CREDDEF(mp, udp->udp_connp->conn_cred),
6373	    dst, opt_storage, udp->udp_connp->conn_mac_exempt,
6374	    us->us_netstack->netstack_ip);
6375	if (err == 0) {
6376		err = tsol_update_sticky(&udp->udp_sticky_ipp,
6377		    &udp->udp_label_len_v6, opt_storage);
6378	}
6379	if (err != 0) {
6380		DTRACE_PROBE4(
6381		    tx__ip__log__drop__updatelabel__udp6,
6382		    char *, "queue(1) failed to update options(2) on mp(3)",
6383		    queue_t *, wq, char *, opt_storage, mblk_t *, mp);
6384	} else {
6385		udp->udp_v6lastdst = *dst;
6386	}
6387	return (err);
6388}
6389
6390void
6391udp_output_connected(void *arg, mblk_t *mp)
6392{
6393	conn_t	*connp = (conn_t *)arg;
6394	udp_t	*udp = connp->conn_udp;
6395	udp_stack_t	*us = udp->udp_us;
6396	ipaddr_t	v4dst;
6397	in_port_t	dstport;
6398	boolean_t	mapped_addr;
6399	struct sockaddr_storage ss;
6400	sin_t		*sin;
6401	sin6_t		*sin6;
6402	struct sockaddr	*addr;
6403	socklen_t	addrlen;
6404	int		error;
6405	boolean_t	insert_spi = udp->udp_nat_t_endpoint;
6406
6407	/* M_DATA for connected socket */
6408
6409	ASSERT(udp->udp_issocket);
6410	UDP_DBGSTAT(us, udp_data_conn);
6411
6412	mutex_enter(&connp->conn_lock);
6413	if (udp->udp_state != TS_DATA_XFER) {
6414		mutex_exit(&connp->conn_lock);
6415		BUMP_MIB(&us->us_udp_mib, udpOutErrors);
6416		UDP_STAT(us, udp_out_err_notconn);
6417		freemsg(mp);
6418		TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
6419		    "udp_wput_end: connp %p (%S)", connp,
6420		    "not-connected; address required");
6421		return;
6422	}
6423
6424	mapped_addr = IN6_IS_ADDR_V4MAPPED(&udp->udp_v6dst);
6425	if (mapped_addr)
6426		IN6_V4MAPPED_TO_IPADDR(&udp->udp_v6dst, v4dst);
6427
6428	/* Initialize addr and addrlen as if they're passed in */
6429	if (udp->udp_family == AF_INET) {
6430		sin = (sin_t *)&ss;
6431		sin->sin_family = AF_INET;
6432		dstport = sin->sin_port = udp->udp_dstport;
6433		ASSERT(mapped_addr);
6434		sin->sin_addr.s_addr = v4dst;
6435		addr = (struct sockaddr *)sin;
6436		addrlen = sizeof (*sin);
6437	} else {
6438		sin6 = (sin6_t *)&ss;
6439		sin6->sin6_family = AF_INET6;
6440		dstport = sin6->sin6_port = udp->udp_dstport;
6441		sin6->sin6_flowinfo = udp->udp_flowinfo;
6442		sin6->sin6_addr = udp->udp_v6dst;
6443		sin6->sin6_scope_id = 0;
6444		sin6->__sin6_src_id = 0;
6445		addr = (struct sockaddr *)sin6;
6446		addrlen = sizeof (*sin6);
6447	}
6448	mutex_exit(&connp->conn_lock);
6449
6450	if (mapped_addr) {
6451		/*
6452		 * Handle both AF_INET and AF_INET6; the latter
6453		 * for IPV4 mapped destination addresses.  Note
6454		 * here that both addr and addrlen point to the
6455		 * corresponding struct depending on the address
6456		 * family of the socket.
6457		 */
6458		mp = udp_output_v4(connp, mp, v4dst, dstport, 0, &error,
6459		    insert_spi);
6460	} else {
6461		mp = udp_output_v6(connp, mp, sin6, &error);
6462	}
6463	if (error == 0) {
6464		ASSERT(mp == NULL);
6465		return;
6466	}
6467
6468	UDP_STAT(us, udp_out_err_output);
6469	ASSERT(mp != NULL);
6470	/* mp is freed by the following routine */
6471	udp_ud_err(connp->conn_wq, mp, (uchar_t *)addr, (t_scalar_t)addrlen,
6472	    (t_scalar_t)error);
6473}
6474
6475/*
6476 * This routine handles all messages passed downstream.  It either
6477 * consumes the message or passes it downstream; it never queues a
6478 * a message.
6479 *
6480 * Also entry point for sockfs when udp is in "direct sockfs" mode.  This mode
6481 * is valid when we are directly beneath the stream head, and thus sockfs
6482 * is able to bypass STREAMS and directly call us, passing along the sockaddr
6483 * structure without the cumbersome T_UNITDATA_REQ interface for the case of
6484 * connected endpoints.
6485 */
6486void
6487udp_wput(queue_t *q, mblk_t *mp)
6488{
6489	sin6_t		*sin6;
6490	sin_t		*sin;
6491	ipaddr_t	v4dst;
6492	uint16_t	port;
6493	uint_t		srcid;
6494	conn_t		*connp = Q_TO_CONN(q);
6495	udp_t		*udp = connp->conn_udp;
6496	int		error = 0;
6497	struct sockaddr	*addr;
6498	socklen_t	addrlen;
6499	udp_stack_t *us = udp->udp_us;
6500	boolean_t	insert_spi = udp->udp_nat_t_endpoint;
6501
6502	TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_START,
6503	    "udp_wput_start: queue %p mp %p", q, mp);
6504
6505	/*
6506	 * We directly handle several cases here: T_UNITDATA_REQ message
6507	 * coming down as M_PROTO/M_PCPROTO and M_DATA messages for connected
6508	 * socket.
6509	 */
6510	switch (DB_TYPE(mp)) {
6511	case M_DATA:
6512		/*
6513		 * Quick check for error cases. Checks will be done again
6514		 * under the lock later on
6515		 */
6516		if (!udp->udp_direct_sockfs || udp->udp_state != TS_DATA_XFER) {
6517			/* Not connected; address is required */
6518			BUMP_MIB(&us->us_udp_mib, udpOutErrors);
6519			UDP_STAT(us, udp_out_err_notconn);
6520			freemsg(mp);
6521			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
6522			    "udp_wput_end: connp %p (%S)", connp,
6523			    "not-connected; address required");
6524			return;
6525		}
6526		udp_output_connected(connp, mp);
6527		return;
6528
6529	case M_PROTO:
6530	case M_PCPROTO: {
6531		struct T_unitdata_req *tudr;
6532
6533		ASSERT((uintptr_t)MBLKL(mp) <= (uintptr_t)INT_MAX);
6534		tudr = (struct T_unitdata_req *)mp->b_rptr;
6535
6536		/* Handle valid T_UNITDATA_REQ here */
6537		if (MBLKL(mp) >= sizeof (*tudr) &&
6538		    ((t_primp_t)mp->b_rptr)->type == T_UNITDATA_REQ) {
6539			if (mp->b_cont == NULL) {
6540				TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
6541				    "udp_wput_end: q %p (%S)", q, "badaddr");
6542				error = EPROTO;
6543				goto ud_error;
6544			}
6545
6546			if (!MBLKIN(mp, 0, tudr->DEST_offset +
6547			    tudr->DEST_length)) {
6548				TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
6549				    "udp_wput_end: q %p (%S)", q, "badaddr");
6550				error = EADDRNOTAVAIL;
6551				goto ud_error;
6552			}
6553			/*
6554			 * If a port has not been bound to the stream, fail.
6555			 * This is not a problem when sockfs is directly
6556			 * above us, because it will ensure that the socket
6557			 * is first bound before allowing data to be sent.
6558			 */
6559			if (udp->udp_state == TS_UNBND) {
6560				TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
6561				    "udp_wput_end: q %p (%S)", q, "outstate");
6562				error = EPROTO;
6563				goto ud_error;
6564			}
6565			addr = (struct sockaddr *)
6566			    &mp->b_rptr[tudr->DEST_offset];
6567			addrlen = tudr->DEST_length;
6568			if (tudr->OPT_length != 0)
6569				UDP_STAT(us, udp_out_opt);
6570			break;
6571		}
6572		/* FALLTHRU */
6573	}
6574	default:
6575		udp_wput_other(q, mp);
6576		return;
6577	}
6578	ASSERT(addr != NULL);
6579
6580	switch (udp->udp_family) {
6581	case AF_INET6:
6582		sin6 = (sin6_t *)addr;
6583		if (!OK_32PTR((char *)sin6) || (addrlen != sizeof (sin6_t)) ||
6584		    (sin6->sin6_family != AF_INET6)) {
6585			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
6586			    "udp_wput_end: q %p (%S)", q, "badaddr");
6587			error = EADDRNOTAVAIL;
6588			goto ud_error;
6589		}
6590
6591		if (!IN6_IS_ADDR_V4MAPPED(&sin6->sin6_addr)) {
6592			/*
6593			 * Destination is a non-IPv4-compatible IPv6 address.
6594			 * Send out an IPv6 format packet.
6595			 */
6596			mp = udp_output_v6(connp, mp, sin6, &error);
6597			if (error != 0)
6598				goto ud_error;
6599
6600			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
6601			    "udp_wput_end: q %p (%S)", q, "udp_output_v6");
6602			return;
6603		}
6604		/*
6605		 * If the local address is not zero or a mapped address
6606		 * return an error.  It would be possible to send an IPv4
6607		 * packet but the response would never make it back to the
6608		 * application since it is bound to a non-mapped address.
6609		 */
6610		if (!IN6_IS_ADDR_V4MAPPED(&udp->udp_v6src) &&
6611		    !IN6_IS_ADDR_UNSPECIFIED(&udp->udp_v6src)) {
6612			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
6613			    "udp_wput_end: q %p (%S)", q, "badaddr");
6614			error = EADDRNOTAVAIL;
6615			goto ud_error;
6616		}
6617		/* Send IPv4 packet without modifying udp_ipversion */
6618		/* Extract port and ipaddr */
6619		port = sin6->sin6_port;
6620		IN6_V4MAPPED_TO_IPADDR(&sin6->sin6_addr, v4dst);
6621		srcid = sin6->__sin6_src_id;
6622		break;
6623
6624	case AF_INET:
6625		sin = (sin_t *)addr;
6626		if ((!OK_32PTR((char *)sin) || addrlen != sizeof (sin_t)) ||
6627		    (sin->sin_family != AF_INET)) {
6628			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_END,
6629			    "udp_wput_end: q %p (%S)", q, "badaddr");
6630			error = EADDRNOTAVAIL;
6631			goto ud_error;
6632		}
6633		/* Extract port and ipaddr */
6634		port = sin->sin_port;
6635		v4dst = sin->sin_addr.s_addr;
6636		srcid = 0;
6637		break;
6638	}
6639
6640	mp = udp_output_v4(connp, mp, v4dst, port, srcid, &error, insert_spi);
6641	if (error != 0) {
6642ud_error:
6643		UDP_STAT(us, udp_out_err_output);
6644		ASSERT(mp != NULL);
6645		/* mp is freed by the following routine */
6646		udp_ud_err(q, mp, (uchar_t *)addr, (t_scalar_t)addrlen,
6647		    (t_scalar_t)error);
6648	}
6649}
6650
6651/*
6652 * udp_output_v6():
6653 * Assumes that udp_wput did some sanity checking on the destination
6654 * address.
6655 */
6656static mblk_t *
6657udp_output_v6(conn_t *connp, mblk_t *mp, sin6_t *sin6, int *error)
6658{
6659	ip6_t		*ip6h;
6660	ip6i_t		*ip6i;	/* mp1->b_rptr even if no ip6i_t */
6661	mblk_t		*mp1 = mp;
6662	mblk_t		*mp2;
6663	int		udp_ip_hdr_len = IPV6_HDR_LEN + UDPH_SIZE;
6664	size_t		ip_len;
6665	udpha_t		*udph;
6666	udp_t		*udp = connp->conn_udp;
6667	queue_t		*q = connp->conn_wq;
6668	ip6_pkt_t	ipp_s;	/* For ancillary data options */
6669	ip6_pkt_t	*ipp = &ipp_s;
6670	ip6_pkt_t	*tipp;	/* temporary ipp */
6671	uint32_t	csum = 0;
6672	uint_t		ignore = 0;
6673	uint_t		option_exists = 0, is_sticky = 0;
6674	uint8_t		*cp;
6675	uint8_t		*nxthdr_ptr;
6676	in6_addr_t	ip6_dst;
6677	udpattrs_t	attrs;
6678	boolean_t	opt_present;
6679	ip6_hbh_t	*hopoptsptr = NULL;
6680	uint_t		hopoptslen = 0;
6681	boolean_t	is_ancillary = B_FALSE;
6682	udp_stack_t	*us = udp->udp_us;
6683	size_t		sth_wroff = 0;
6684
6685	*error = 0;
6686
6687	/*
6688	 * If the local address is a mapped address return
6689	 * an error.
6690	 * It would be possible to send an IPv6 packet but the
6691	 * response would never make it back to the application
6692	 * since it is bound to a mapped address.
6693	 */
6694	if (IN6_IS_ADDR_V4MAPPED(&udp->udp_v6src)) {
6695		*error = EADDRNOTAVAIL;
6696		goto done;
6697	}
6698
6699	ipp->ipp_fields = 0;
6700	ipp->ipp_sticky_ignored = 0;
6701
6702	/*
6703	 * If TPI options passed in, feed it for verification and handling
6704	 */
6705	attrs.udpattr_credset = B_FALSE;
6706	opt_present = B_FALSE;
6707	if (DB_TYPE(mp) != M_DATA) {
6708		mp1 = mp->b_cont;
6709		if (((struct T_unitdata_req *)mp->b_rptr)->OPT_length != 0) {
6710			attrs.udpattr_ipp6 = ipp;
6711			attrs.udpattr_mb = mp;
6712			if (udp_unitdata_opt_process(q, mp, error,
6713			    &attrs) < 0) {
6714				goto done;
6715			}
6716			ASSERT(*error == 0);
6717			opt_present = B_TRUE;
6718		}
6719	}
6720	rw_enter(&udp->udp_rwlock, RW_READER);
6721	ignore = ipp->ipp_sticky_ignored;
6722
6723	/* mp1 points to the M_DATA mblk carrying the packet */
6724	ASSERT(mp1 != NULL && DB_TYPE(mp1) == M_DATA);
6725
6726	if (sin6->sin6_scope_id != 0 &&
6727	    IN6_IS_ADDR_LINKLOCAL(&sin6->sin6_addr)) {
6728		/*
6729		 * IPPF_SCOPE_ID is special.  It's neither a sticky
6730		 * option nor ancillary data.  It needs to be
6731		 * explicitly set in options_exists.
6732		 */
6733		option_exists |= IPPF_SCOPE_ID;
6734	}
6735
6736	/*
6737	 * Compute the destination address
6738	 */
6739	ip6_dst = sin6->sin6_addr;
6740	if (IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr))
6741		ip6_dst = ipv6_loopback;
6742
6743	/*
6744	 * If we're not going to the same destination as last time, then
6745	 * recompute the label required.  This is done in a separate routine to
6746	 * avoid blowing up our stack here.
6747	 *
6748	 * TSOL Note: Since we are not in WRITER mode, UDP packets
6749	 * to different destination may require different labels,
6750	 * or worse, UDP packets to same IP address may require
6751	 * different labels due to use of shared all-zones address.
6752	 * We use conn_lock to ensure that lastdst, sticky ipp_hopopts,
6753	 * and sticky ipp_hopoptslen are consistent for the current
6754	 * destination and are updated atomically.
6755	 */
6756	mutex_enter(&connp->conn_lock);
6757	if (is_system_labeled()) {
6758		/* Using UDP MLP requires SCM_UCRED from user */
6759		if (connp->conn_mlp_type != mlptSingle &&
6760		    !attrs.udpattr_credset) {
6761			DTRACE_PROBE4(
6762			    tx__ip__log__info__output__udp6,
6763			    char *, "MLP mp(1) lacks SCM_UCRED attr(2) on q(3)",
6764			    mblk_t *, mp1, udpattrs_t *, &attrs, queue_t *, q);
6765			*error = ECONNREFUSED;
6766			rw_exit(&udp->udp_rwlock);
6767			mutex_exit(&connp->conn_lock);
6768			goto done;
6769		}
6770		/*
6771		 * update label option for this UDP socket if
6772		 * - the destination has changed, or
6773		 * - the UDP socket is MLP
6774		 */
6775		if ((opt_present ||
6776		    !IN6_ARE_ADDR_EQUAL(&udp->udp_v6lastdst, &ip6_dst) ||
6777		    connp->conn_mlp_type != mlptSingle) &&
6778		    (*error = udp_update_label_v6(q, mp, &ip6_dst)) != 0) {
6779			rw_exit(&udp->udp_rwlock);
6780			mutex_exit(&connp->conn_lock);
6781			goto done;
6782		}
6783	}
6784
6785	/*
6786	 * If there's a security label here, then we ignore any options the
6787	 * user may try to set.  We keep the peer's label as a hidden sticky
6788	 * option. We make a private copy of this label before releasing the
6789	 * lock so that label is kept consistent with the destination addr.
6790	 */
6791	if (udp->udp_label_len_v6 > 0) {
6792		ignore &= ~IPPF_HOPOPTS;
6793		ipp->ipp_fields &= ~IPPF_HOPOPTS;
6794	}
6795
6796	if ((udp->udp_sticky_ipp.ipp_fields == 0) && (ipp->ipp_fields == 0)) {
6797		/* No sticky options nor ancillary data. */
6798		mutex_exit(&connp->conn_lock);
6799		goto no_options;
6800	}
6801
6802	/*
6803	 * Go through the options figuring out where each is going to
6804	 * come from and build two masks.  The first mask indicates if
6805	 * the option exists at all.  The second mask indicates if the
6806	 * option is sticky or ancillary.
6807	 */
6808	if (!(ignore & IPPF_HOPOPTS)) {
6809		if (ipp->ipp_fields & IPPF_HOPOPTS) {
6810			option_exists |= IPPF_HOPOPTS;
6811			udp_ip_hdr_len += ipp->ipp_hopoptslen;
6812		} else if (udp->udp_sticky_ipp.ipp_fields & IPPF_HOPOPTS) {
6813			option_exists |= IPPF_HOPOPTS;
6814			is_sticky |= IPPF_HOPOPTS;
6815			ASSERT(udp->udp_sticky_ipp.ipp_hopoptslen != 0);
6816			hopoptsptr = kmem_alloc(
6817			    udp->udp_sticky_ipp.ipp_hopoptslen, KM_NOSLEEP);
6818			if (hopoptsptr == NULL) {
6819				*error = ENOMEM;
6820				mutex_exit(&connp->conn_lock);
6821				goto done;
6822			}
6823			hopoptslen = udp->udp_sticky_ipp.ipp_hopoptslen;
6824			bcopy(udp->udp_sticky_ipp.ipp_hopopts, hopoptsptr,
6825			    hopoptslen);
6826			udp_ip_hdr_len += hopoptslen;
6827		}
6828	}
6829	mutex_exit(&connp->conn_lock);
6830
6831	if (!(ignore & IPPF_RTHDR)) {
6832		if (ipp->ipp_fields & IPPF_RTHDR) {
6833			option_exists |= IPPF_RTHDR;
6834			udp_ip_hdr_len += ipp->ipp_rthdrlen;
6835		} else if (udp->udp_sticky_ipp.ipp_fields & IPPF_RTHDR) {
6836			option_exists |= IPPF_RTHDR;
6837			is_sticky |= IPPF_RTHDR;
6838			udp_ip_hdr_len += udp->udp_sticky_ipp.ipp_rthdrlen;
6839		}
6840	}
6841
6842	if (!(ignore & IPPF_RTDSTOPTS) && (option_exists & IPPF_RTHDR)) {
6843		if (ipp->ipp_fields & IPPF_RTDSTOPTS) {
6844			option_exists |= IPPF_RTDSTOPTS;
6845			udp_ip_hdr_len += ipp->ipp_rtdstoptslen;
6846		} else if (udp->udp_sticky_ipp.ipp_fields & IPPF_RTDSTOPTS) {
6847			option_exists |= IPPF_RTDSTOPTS;
6848			is_sticky |= IPPF_RTDSTOPTS;
6849			udp_ip_hdr_len += udp->udp_sticky_ipp.ipp_rtdstoptslen;
6850		}
6851	}
6852
6853	if (!(ignore & IPPF_DSTOPTS)) {
6854		if (ipp->ipp_fields & IPPF_DSTOPTS) {
6855			option_exists |= IPPF_DSTOPTS;
6856			udp_ip_hdr_len += ipp->ipp_dstoptslen;
6857		} else if (udp->udp_sticky_ipp.ipp_fields & IPPF_DSTOPTS) {
6858			option_exists |= IPPF_DSTOPTS;
6859			is_sticky |= IPPF_DSTOPTS;
6860			udp_ip_hdr_len += udp->udp_sticky_ipp.ipp_dstoptslen;
6861		}
6862	}
6863
6864	if (!(ignore & IPPF_IFINDEX)) {
6865		if (ipp->ipp_fields & IPPF_IFINDEX) {
6866			option_exists |= IPPF_IFINDEX;
6867		} else if (udp->udp_sticky_ipp.ipp_fields & IPPF_IFINDEX) {
6868			option_exists |= IPPF_IFINDEX;
6869			is_sticky |= IPPF_IFINDEX;
6870		}
6871	}
6872
6873	if (!(ignore & IPPF_ADDR)) {
6874		if (ipp->ipp_fields & IPPF_ADDR) {
6875			option_exists |= IPPF_ADDR;
6876		} else if (udp->udp_sticky_ipp.ipp_fields & IPPF_ADDR) {
6877			option_exists |= IPPF_ADDR;
6878			is_sticky |= IPPF_ADDR;
6879		}
6880	}
6881
6882	if (!(ignore & IPPF_DONTFRAG)) {
6883		if (ipp->ipp_fields & IPPF_DONTFRAG) {
6884			option_exists |= IPPF_DONTFRAG;
6885		} else if (udp->udp_sticky_ipp.ipp_fields & IPPF_DONTFRAG) {
6886			option_exists |= IPPF_DONTFRAG;
6887			is_sticky |= IPPF_DONTFRAG;
6888		}
6889	}
6890
6891	if (!(ignore & IPPF_USE_MIN_MTU)) {
6892		if (ipp->ipp_fields & IPPF_USE_MIN_MTU) {
6893			option_exists |= IPPF_USE_MIN_MTU;
6894		} else if (udp->udp_sticky_ipp.ipp_fields &
6895		    IPPF_USE_MIN_MTU) {
6896			option_exists |= IPPF_USE_MIN_MTU;
6897			is_sticky |= IPPF_USE_MIN_MTU;
6898		}
6899	}
6900
6901	if (!(ignore & IPPF_HOPLIMIT) && (ipp->ipp_fields & IPPF_HOPLIMIT))
6902		option_exists |= IPPF_HOPLIMIT;
6903	/* IPV6_HOPLIMIT can never be sticky */
6904	ASSERT(!(udp->udp_sticky_ipp.ipp_fields & IPPF_HOPLIMIT));
6905
6906	if (!(ignore & IPPF_UNICAST_HOPS) &&
6907	    (udp->udp_sticky_ipp.ipp_fields & IPPF_UNICAST_HOPS)) {
6908		option_exists |= IPPF_UNICAST_HOPS;
6909		is_sticky |= IPPF_UNICAST_HOPS;
6910	}
6911
6912	if (!(ignore & IPPF_MULTICAST_HOPS) &&
6913	    (udp->udp_sticky_ipp.ipp_fields & IPPF_MULTICAST_HOPS)) {
6914		option_exists |= IPPF_MULTICAST_HOPS;
6915		is_sticky |= IPPF_MULTICAST_HOPS;
6916	}
6917
6918	if (!(ignore & IPPF_TCLASS)) {
6919		if (ipp->ipp_fields & IPPF_TCLASS) {
6920			option_exists |= IPPF_TCLASS;
6921		} else if (udp->udp_sticky_ipp.ipp_fields & IPPF_TCLASS) {
6922			option_exists |= IPPF_TCLASS;
6923			is_sticky |= IPPF_TCLASS;
6924		}
6925	}
6926
6927	if (!(ignore & IPPF_NEXTHOP) &&
6928	    (udp->udp_sticky_ipp.ipp_fields & IPPF_NEXTHOP)) {
6929		option_exists |= IPPF_NEXTHOP;
6930		is_sticky |= IPPF_NEXTHOP;
6931	}
6932
6933no_options:
6934
6935	/*
6936	 * If any options carried in the ip6i_t were specified, we
6937	 * need to account for the ip6i_t in the data we'll be sending
6938	 * down.
6939	 */
6940	if (option_exists & IPPF_HAS_IP6I)
6941		udp_ip_hdr_len += sizeof (ip6i_t);
6942
6943	/* check/fix buffer config, setup pointers into it */
6944	ip6h = (ip6_t *)&mp1->b_rptr[-udp_ip_hdr_len];
6945	if (DB_REF(mp1) != 1 || ((unsigned char *)ip6h < DB_BASE(mp1)) ||
6946	    !OK_32PTR(ip6h)) {
6947
6948		/* Try to get everything in a single mblk next time */
6949		if (udp_ip_hdr_len > udp->udp_max_hdr_len) {
6950			udp->udp_max_hdr_len = udp_ip_hdr_len;
6951			sth_wroff = udp->udp_max_hdr_len + us->us_wroff_extra;
6952		}
6953
6954		mp2 = allocb(udp_ip_hdr_len + us->us_wroff_extra, BPRI_LO);
6955		if (mp2 == NULL) {
6956			*error = ENOMEM;
6957			rw_exit(&udp->udp_rwlock);
6958			goto done;
6959		}
6960		mp2->b_wptr = DB_LIM(mp2);
6961		mp2->b_cont = mp1;
6962		mp1 = mp2;
6963		if (DB_TYPE(mp) != M_DATA)
6964			mp->b_cont = mp1;
6965		else
6966			mp = mp1;
6967
6968		ip6h = (ip6_t *)(mp1->b_wptr - udp_ip_hdr_len);
6969	}
6970	mp1->b_rptr = (unsigned char *)ip6h;
6971	ip6i = (ip6i_t *)ip6h;
6972
6973#define	ANCIL_OR_STICKY_PTR(f) ((is_sticky & f) ? &udp->udp_sticky_ipp : ipp)
6974	if (option_exists & IPPF_HAS_IP6I) {
6975		ip6h = (ip6_t *)&ip6i[1];
6976		ip6i->ip6i_flags = 0;
6977		ip6i->ip6i_vcf = IPV6_DEFAULT_VERS_AND_FLOW;
6978
6979		/* sin6_scope_id takes precendence over IPPF_IFINDEX */
6980		if (option_exists & IPPF_SCOPE_ID) {
6981			ip6i->ip6i_flags |= IP6I_IFINDEX;
6982			ip6i->ip6i_ifindex = sin6->sin6_scope_id;
6983		} else if (option_exists & IPPF_IFINDEX) {
6984			tipp = ANCIL_OR_STICKY_PTR(IPPF_IFINDEX);
6985			ASSERT(tipp->ipp_ifindex != 0);
6986			ip6i->ip6i_flags |= IP6I_IFINDEX;
6987			ip6i->ip6i_ifindex = tipp->ipp_ifindex;
6988		}
6989
6990		if (option_exists & IPPF_ADDR) {
6991			/*
6992			 * Enable per-packet source address verification if
6993			 * IPV6_PKTINFO specified the source address.
6994			 * ip6_src is set in the transport's _wput function.
6995			 */
6996			ip6i->ip6i_flags |= IP6I_VERIFY_SRC;
6997		}
6998
6999		if (option_exists & IPPF_DONTFRAG) {
7000			ip6i->ip6i_flags |= IP6I_DONTFRAG;
7001		}
7002
7003		if (option_exists & IPPF_USE_MIN_MTU) {
7004			ip6i->ip6i_flags = IP6I_API_USE_MIN_MTU(
7005			    ip6i->ip6i_flags, ipp->ipp_use_min_mtu);
7006		}
7007
7008		if (option_exists & IPPF_NEXTHOP) {
7009			tipp = ANCIL_OR_STICKY_PTR(IPPF_NEXTHOP);
7010			ASSERT(!IN6_IS_ADDR_UNSPECIFIED(&tipp->ipp_nexthop));
7011			ip6i->ip6i_flags |= IP6I_NEXTHOP;
7012			ip6i->ip6i_nexthop = tipp->ipp_nexthop;
7013		}
7014
7015		/*
7016		 * tell IP this is an ip6i_t private header
7017		 */
7018		ip6i->ip6i_nxt = IPPROTO_RAW;
7019	}
7020
7021	/* Initialize IPv6 header */
7022	ip6h->ip6_vcf = IPV6_DEFAULT_VERS_AND_FLOW;
7023	bzero(&ip6h->ip6_src, sizeof (ip6h->ip6_src));
7024
7025	/* Set the hoplimit of the outgoing packet. */
7026	if (option_exists & IPPF_HOPLIMIT) {
7027		/* IPV6_HOPLIMIT ancillary data overrides all other settings. */
7028		ip6h->ip6_hops = ipp->ipp_hoplimit;
7029		ip6i->ip6i_flags |= IP6I_HOPLIMIT;
7030	} else if (IN6_IS_ADDR_MULTICAST(&sin6->sin6_addr)) {
7031		ip6h->ip6_hops = udp->udp_multicast_ttl;
7032		if (option_exists & IPPF_MULTICAST_HOPS)
7033			ip6i->ip6i_flags |= IP6I_HOPLIMIT;
7034	} else {
7035		ip6h->ip6_hops = udp->udp_ttl;
7036		if (option_exists & IPPF_UNICAST_HOPS)
7037			ip6i->ip6i_flags |= IP6I_HOPLIMIT;
7038	}
7039
7040	if (option_exists & IPPF_ADDR) {
7041		tipp = ANCIL_OR_STICKY_PTR(IPPF_ADDR);
7042		ASSERT(!IN6_IS_ADDR_UNSPECIFIED(&tipp->ipp_addr));
7043		ip6h->ip6_src = tipp->ipp_addr;
7044	} else {
7045		/*
7046		 * The source address was not set using IPV6_PKTINFO.
7047		 * First look at the bound source.
7048		 * If unspecified fallback to __sin6_src_id.
7049		 */
7050		ip6h->ip6_src = udp->udp_v6src;
7051		if (sin6->__sin6_src_id != 0 &&
7052		    IN6_IS_ADDR_UNSPECIFIED(&ip6h->ip6_src)) {
7053			ip_srcid_find_id(sin6->__sin6_src_id,
7054			    &ip6h->ip6_src, connp->conn_zoneid,
7055			    us->us_netstack);
7056		}
7057	}
7058
7059	nxthdr_ptr = (uint8_t *)&ip6h->ip6_nxt;
7060	cp = (uint8_t *)&ip6h[1];
7061
7062	/*
7063	 * Here's where we have to start stringing together
7064	 * any extension headers in the right order:
7065	 * Hop-by-hop, destination, routing, and final destination opts.
7066	 */
7067	if (option_exists & IPPF_HOPOPTS) {
7068		/* Hop-by-hop options */
7069		ip6_hbh_t *hbh = (ip6_hbh_t *)cp;
7070		tipp = ANCIL_OR_STICKY_PTR(IPPF_HOPOPTS);
7071		if (hopoptslen == 0) {
7072			hopoptsptr = tipp->ipp_hopopts;
7073			hopoptslen = tipp->ipp_hopoptslen;
7074			is_ancillary = B_TRUE;
7075		}
7076
7077		*nxthdr_ptr = IPPROTO_HOPOPTS;
7078		nxthdr_ptr = &hbh->ip6h_nxt;
7079
7080		bcopy(hopoptsptr, cp, hopoptslen);
7081		cp += hopoptslen;
7082
7083		if (hopoptsptr != NULL && !is_ancillary) {
7084			kmem_free(hopoptsptr, hopoptslen);
7085			hopoptsptr = NULL;
7086			hopoptslen = 0;
7087		}
7088	}
7089	/*
7090	 * En-route destination options
7091	 * Only do them if there's a routing header as well
7092	 */
7093	if (option_exists & IPPF_RTDSTOPTS) {
7094		ip6_dest_t *dst = (ip6_dest_t *)cp;
7095		tipp = ANCIL_OR_STICKY_PTR(IPPF_RTDSTOPTS);
7096
7097		*nxthdr_ptr = IPPROTO_DSTOPTS;
7098		nxthdr_ptr = &dst->ip6d_nxt;
7099
7100		bcopy(tipp->ipp_rtdstopts, cp, tipp->ipp_rtdstoptslen);
7101		cp += tipp->ipp_rtdstoptslen;
7102	}
7103	/*
7104	 * Routing header next
7105	 */
7106	if (option_exists & IPPF_RTHDR) {
7107		ip6_rthdr_t *rt = (ip6_rthdr_t *)cp;
7108		tipp = ANCIL_OR_STICKY_PTR(IPPF_RTHDR);
7109
7110		*nxthdr_ptr = IPPROTO_ROUTING;
7111		nxthdr_ptr = &rt->ip6r_nxt;
7112
7113		bcopy(tipp->ipp_rthdr, cp, tipp->ipp_rthdrlen);
7114		cp += tipp->ipp_rthdrlen;
7115	}
7116	/*
7117	 * Do ultimate destination options
7118	 */
7119	if (option_exists & IPPF_DSTOPTS) {
7120		ip6_dest_t *dest = (ip6_dest_t *)cp;
7121		tipp = ANCIL_OR_STICKY_PTR(IPPF_DSTOPTS);
7122
7123		*nxthdr_ptr = IPPROTO_DSTOPTS;
7124		nxthdr_ptr = &dest->ip6d_nxt;
7125
7126		bcopy(tipp->ipp_dstopts, cp, tipp->ipp_dstoptslen);
7127		cp += tipp->ipp_dstoptslen;
7128	}
7129	/*
7130	 * Now set the last header pointer to the proto passed in
7131	 */
7132	ASSERT((int)(cp - (uint8_t *)ip6i) == (udp_ip_hdr_len - UDPH_SIZE));
7133	*nxthdr_ptr = IPPROTO_UDP;
7134
7135	/* Update UDP header */
7136	udph = (udpha_t *)((uchar_t *)ip6i + udp_ip_hdr_len - UDPH_SIZE);
7137	udph->uha_dst_port = sin6->sin6_port;
7138	udph->uha_src_port = udp->udp_port;
7139
7140	/*
7141	 * Copy in the destination address
7142	 */
7143	ip6h->ip6_dst = ip6_dst;
7144
7145	ip6h->ip6_vcf =
7146	    (IPV6_DEFAULT_VERS_AND_FLOW & IPV6_VERS_AND_FLOW_MASK) |
7147	    (sin6->sin6_flowinfo & ~IPV6_VERS_AND_FLOW_MASK);
7148
7149	if (option_exists & IPPF_TCLASS) {
7150		tipp = ANCIL_OR_STICKY_PTR(IPPF_TCLASS);
7151		ip6h->ip6_vcf = IPV6_TCLASS_FLOW(ip6h->ip6_vcf,
7152		    tipp->ipp_tclass);
7153	}
7154	rw_exit(&udp->udp_rwlock);
7155
7156	if (option_exists & IPPF_RTHDR) {
7157		ip6_rthdr_t	*rth;
7158
7159		/*
7160		 * Perform any processing needed for source routing.
7161		 * We know that all extension headers will be in the same mblk
7162		 * as the IPv6 header.
7163		 */
7164		rth = ip_find_rthdr_v6(ip6h, mp1->b_wptr);
7165		if (rth != NULL && rth->ip6r_segleft != 0) {
7166			if (rth->ip6r_type != IPV6_RTHDR_TYPE_0) {
7167				/*
7168				 * Drop packet - only support Type 0 routing.
7169				 * Notify the application as well.
7170				 */
7171				*error = EPROTO;
7172				goto done;
7173			}
7174
7175			/*
7176			 * rth->ip6r_len is twice the number of
7177			 * addresses in the header. Thus it must be even.
7178			 */
7179			if (rth->ip6r_len & 0x1) {
7180				*error = EPROTO;
7181				goto done;
7182			}
7183			/*
7184			 * Shuffle the routing header and ip6_dst
7185			 * addresses, and get the checksum difference
7186			 * between the first hop (in ip6_dst) and
7187			 * the destination (in the last routing hdr entry).
7188			 */
7189			csum = ip_massage_options_v6(ip6h, rth,
7190			    us->us_netstack);
7191			/*
7192			 * Verify that the first hop isn't a mapped address.
7193			 * Routers along the path need to do this verification
7194			 * for subsequent hops.
7195			 */
7196			if (IN6_IS_ADDR_V4MAPPED(&ip6h->ip6_dst)) {
7197				*error = EADDRNOTAVAIL;
7198				goto done;
7199			}
7200
7201			cp += (rth->ip6r_len + 1)*8;
7202		}
7203	}
7204
7205	/* count up length of UDP packet */
7206	ip_len = (mp1->b_wptr - (unsigned char *)ip6h) - IPV6_HDR_LEN;
7207	if ((mp2 = mp1->b_cont) != NULL) {
7208		do {
7209			ASSERT((uintptr_t)MBLKL(mp2) <= (uintptr_t)UINT_MAX);
7210			ip_len += (uint32_t)MBLKL(mp2);
7211		} while ((mp2 = mp2->b_cont) != NULL);
7212	}
7213
7214	/*
7215	 * If the size of the packet is greater than the maximum allowed by
7216	 * ip, return an error. Passing this down could cause panics because
7217	 * the size will have wrapped and be inconsistent with the msg size.
7218	 */
7219	if (ip_len > IP_MAXPACKET) {
7220		*error = EMSGSIZE;
7221		goto done;
7222	}
7223
7224	/* Store the UDP length. Subtract length of extension hdrs */
7225	udph->uha_length = htons(ip_len + IPV6_HDR_LEN -
7226	    (int)((uchar_t *)udph - (uchar_t *)ip6h));
7227
7228	/*
7229	 * We make it easy for IP to include our pseudo header
7230	 * by putting our length in uh_checksum, modified (if
7231	 * we have a routing header) by the checksum difference
7232	 * between the ultimate destination and first hop addresses.
7233	 * Note: UDP over IPv6 must always checksum the packet.
7234	 */
7235	csum += udph->uha_length;
7236	csum = (csum & 0xFFFF) + (csum >> 16);
7237	udph->uha_checksum = (uint16_t)csum;
7238
7239#ifdef _LITTLE_ENDIAN
7240	ip_len = htons(ip_len);
7241#endif
7242	ip6h->ip6_plen = ip_len;
7243	if (DB_CRED(mp) != NULL)
7244		mblk_setcred(mp1, DB_CRED(mp));
7245
7246	if (DB_TYPE(mp) != M_DATA) {
7247		ASSERT(mp != mp1);
7248		freeb(mp);
7249	}
7250
7251	/* mp has been consumed and we'll return success */
7252	ASSERT(*error == 0);
7253	mp = NULL;
7254
7255	/* We're done. Pass the packet to IP */
7256	BUMP_MIB(&us->us_udp_mib, udpHCOutDatagrams);
7257	ip_output_v6(connp, mp1, q, IP_WPUT);
7258
7259done:
7260	if (sth_wroff != 0) {
7261		(void) mi_set_sth_wroff(RD(q),
7262		    udp->udp_max_hdr_len + us->us_wroff_extra);
7263	}
7264	if (hopoptsptr != NULL && !is_ancillary) {
7265		kmem_free(hopoptsptr, hopoptslen);
7266		hopoptsptr = NULL;
7267	}
7268	if (*error != 0) {
7269		ASSERT(mp != NULL);
7270		BUMP_MIB(&us->us_udp_mib, udpOutErrors);
7271	}
7272	return (mp);
7273}
7274
7275
7276static int
7277udp_getpeername(udp_t *udp, struct sockaddr *sa, uint_t *salenp)
7278{
7279	sin_t *sin = (sin_t *)sa;
7280	sin6_t *sin6 = (sin6_t *)sa;
7281
7282	ASSERT(RW_LOCK_HELD(&udp->udp_rwlock));
7283
7284	if (udp->udp_state != TS_DATA_XFER)
7285		return (ENOTCONN);
7286
7287	switch (udp->udp_family) {
7288	case AF_INET:
7289		ASSERT(udp->udp_ipversion == IPV4_VERSION);
7290
7291		if (*salenp < sizeof (sin_t))
7292			return (EINVAL);
7293
7294		*salenp = sizeof (sin_t);
7295		*sin = sin_null;
7296		sin->sin_family = AF_INET;
7297		sin->sin_port = udp->udp_dstport;
7298		sin->sin_addr.s_addr = V4_PART_OF_V6(udp->udp_v6dst);
7299		break;
7300
7301	case AF_INET6:
7302		if (*salenp < sizeof (sin6_t))
7303			return (EINVAL);
7304
7305		*salenp = sizeof (sin6_t);
7306		*sin6 = sin6_null;
7307		sin6->sin6_family = AF_INET6;
7308		sin6->sin6_port = udp->udp_dstport;
7309		sin6->sin6_addr = udp->udp_v6dst;
7310		sin6->sin6_flowinfo = udp->udp_flowinfo;
7311		break;
7312	}
7313
7314	return (0);
7315}
7316
7317static int
7318udp_getmyname(udp_t *udp, struct sockaddr *sa, uint_t *salenp)
7319{
7320	sin_t *sin = (sin_t *)sa;
7321	sin6_t *sin6 = (sin6_t *)sa;
7322
7323	ASSERT(RW_LOCK_HELD(&udp->udp_rwlock));
7324
7325	switch (udp->udp_family) {
7326	case AF_INET:
7327		ASSERT(udp->udp_ipversion == IPV4_VERSION);
7328
7329		if (*salenp < sizeof (sin_t))
7330			return (EINVAL);
7331
7332		*salenp = sizeof (sin_t);
7333		*sin = sin_null;
7334		sin->sin_family = AF_INET;
7335		sin->sin_port = udp->udp_port;
7336
7337		/*
7338		 * If udp_v6src is unspecified, we might be bound to broadcast
7339		 * / multicast.  Use udp_bound_v6src as local address instead
7340		 * (that could also still be unspecified).
7341		 */
7342		if (!IN6_IS_ADDR_V4MAPPED_ANY(&udp->udp_v6src) &&
7343		    !IN6_IS_ADDR_UNSPECIFIED(&udp->udp_v6src)) {
7344			sin->sin_addr.s_addr = V4_PART_OF_V6(udp->udp_v6src);
7345		} else {
7346			sin->sin_addr.s_addr =
7347			    V4_PART_OF_V6(udp->udp_bound_v6src);
7348		}
7349		break;
7350
7351	case AF_INET6:
7352		if (*salenp < sizeof (sin6_t))
7353			return (EINVAL);
7354
7355		*salenp = sizeof (sin6_t);
7356		*sin6 = sin6_null;
7357		sin6->sin6_family = AF_INET6;
7358		sin6->sin6_port = udp->udp_port;
7359		sin6->sin6_flowinfo = udp->udp_flowinfo;
7360
7361		/*
7362		 * If udp_v6src is unspecified, we might be bound to broadcast
7363		 * / multicast.  Use udp_bound_v6src as local address instead
7364		 * (that could also still be unspecified).
7365		 */
7366		if (!IN6_IS_ADDR_UNSPECIFIED(&udp->udp_v6src))
7367			sin6->sin6_addr = udp->udp_v6src;
7368		else
7369			sin6->sin6_addr = udp->udp_bound_v6src;
7370		break;
7371	}
7372
7373	return (0);
7374}
7375
7376/*
7377 * Handle special out-of-band ioctl requests (see PSARC/2008/265).
7378 */
7379static void
7380udp_wput_cmdblk(queue_t *q, mblk_t *mp)
7381{
7382	void	*data;
7383	mblk_t	*datamp = mp->b_cont;
7384	udp_t	*udp = Q_TO_UDP(q);
7385	cmdblk_t *cmdp = (cmdblk_t *)mp->b_rptr;
7386
7387	if (datamp == NULL || MBLKL(datamp) < cmdp->cb_len) {
7388		cmdp->cb_error = EPROTO;
7389		qreply(q, mp);
7390		return;
7391	}
7392	data = datamp->b_rptr;
7393
7394	rw_enter(&udp->udp_rwlock, RW_READER);
7395	switch (cmdp->cb_cmd) {
7396	case TI_GETPEERNAME:
7397		cmdp->cb_error = udp_getpeername(udp, data, &cmdp->cb_len);
7398		break;
7399	case TI_GETMYNAME:
7400		cmdp->cb_error = udp_getmyname(udp, data, &cmdp->cb_len);
7401		break;
7402	default:
7403		cmdp->cb_error = EINVAL;
7404		break;
7405	}
7406	rw_exit(&udp->udp_rwlock);
7407
7408	qreply(q, mp);
7409}
7410
7411static void
7412udp_wput_other(queue_t *q, mblk_t *mp)
7413{
7414	uchar_t	*rptr = mp->b_rptr;
7415	struct datab *db;
7416	struct iocblk *iocp;
7417	cred_t	*cr;
7418	conn_t	*connp = Q_TO_CONN(q);
7419	udp_t	*udp = connp->conn_udp;
7420	udp_stack_t *us;
7421
7422	TRACE_1(TR_FAC_UDP, TR_UDP_WPUT_OTHER_START,
7423	    "udp_wput_other_start: q %p", q);
7424
7425	us = udp->udp_us;
7426	db = mp->b_datap;
7427
7428	cr = DB_CREDDEF(mp, connp->conn_cred);
7429
7430	switch (db->db_type) {
7431	case M_CMD:
7432		udp_wput_cmdblk(q, mp);
7433		return;
7434
7435	case M_PROTO:
7436	case M_PCPROTO:
7437		if (mp->b_wptr - rptr < sizeof (t_scalar_t)) {
7438			freemsg(mp);
7439			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7440			    "udp_wput_other_end: q %p (%S)", q, "protoshort");
7441			return;
7442		}
7443		switch (((t_primp_t)rptr)->type) {
7444		case T_ADDR_REQ:
7445			udp_addr_req(q, mp);
7446			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7447			    "udp_wput_other_end: q %p (%S)", q, "addrreq");
7448			return;
7449		case O_T_BIND_REQ:
7450		case T_BIND_REQ:
7451			udp_bind(q, mp);
7452			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7453			    "udp_wput_other_end: q %p (%S)", q, "bindreq");
7454			return;
7455		case T_CONN_REQ:
7456			udp_connect(q, mp);
7457			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7458			    "udp_wput_other_end: q %p (%S)", q, "connreq");
7459			return;
7460		case T_CAPABILITY_REQ:
7461			udp_capability_req(q, mp);
7462			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7463			    "udp_wput_other_end: q %p (%S)", q, "capabreq");
7464			return;
7465		case T_INFO_REQ:
7466			udp_info_req(q, mp);
7467			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7468			    "udp_wput_other_end: q %p (%S)", q, "inforeq");
7469			return;
7470		case T_UNITDATA_REQ:
7471			/*
7472			 * If a T_UNITDATA_REQ gets here, the address must
7473			 * be bad.  Valid T_UNITDATA_REQs are handled
7474			 * in udp_wput.
7475			 */
7476			udp_ud_err(q, mp, NULL, 0, EADDRNOTAVAIL);
7477			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7478			    "udp_wput_other_end: q %p (%S)", q, "unitdatareq");
7479			return;
7480		case T_UNBIND_REQ:
7481			udp_unbind(q, mp);
7482			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7483			    "udp_wput_other_end: q %p (%S)", q, "unbindreq");
7484			return;
7485		case T_SVR4_OPTMGMT_REQ:
7486			if (!snmpcom_req(q, mp, udp_snmp_set, ip_snmp_get,
7487			    cr)) {
7488				(void) svr4_optcom_req(q,
7489				    mp, cr, &udp_opt_obj, B_TRUE);
7490			}
7491			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7492			    "udp_wput_other_end: q %p (%S)", q, "optmgmtreq");
7493			return;
7494
7495		case T_OPTMGMT_REQ:
7496			(void) tpi_optcom_req(q, mp, cr, &udp_opt_obj, B_TRUE);
7497			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7498			    "udp_wput_other_end: q %p (%S)", q, "optmgmtreq");
7499			return;
7500
7501		case T_DISCON_REQ:
7502			udp_disconnect(q, mp);
7503			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7504			    "udp_wput_other_end: q %p (%S)", q, "disconreq");
7505			return;
7506
7507		/* The following TPI message is not supported by udp. */
7508		case O_T_CONN_RES:
7509		case T_CONN_RES:
7510			udp_err_ack(q, mp, TNOTSUPPORT, 0);
7511			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7512			    "udp_wput_other_end: q %p (%S)", q,
7513			    "connres/disconreq");
7514			return;
7515
7516		/* The following 3 TPI messages are illegal for udp. */
7517		case T_DATA_REQ:
7518		case T_EXDATA_REQ:
7519		case T_ORDREL_REQ:
7520			udp_err_ack(q, mp, TNOTSUPPORT, 0);
7521			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7522			    "udp_wput_other_end: q %p (%S)", q,
7523			    "data/exdata/ordrel");
7524			return;
7525		default:
7526			break;
7527		}
7528		break;
7529	case M_FLUSH:
7530		if (*rptr & FLUSHW)
7531			flushq(q, FLUSHDATA);
7532		break;
7533	case M_IOCTL:
7534		iocp = (struct iocblk *)mp->b_rptr;
7535		switch (iocp->ioc_cmd) {
7536		case TI_GETPEERNAME:
7537			if (udp->udp_state != TS_DATA_XFER) {
7538				/*
7539				 * If a default destination address has not
7540				 * been associated with the stream, then we
7541				 * don't know the peer's name.
7542				 */
7543				iocp->ioc_error = ENOTCONN;
7544				iocp->ioc_count = 0;
7545				mp->b_datap->db_type = M_IOCACK;
7546				qreply(q, mp);
7547				TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7548				    "udp_wput_other_end: q %p (%S)", q,
7549				    "getpeername");
7550				return;
7551			}
7552			/* FALLTHRU */
7553		case TI_GETMYNAME: {
7554			/*
7555			 * For TI_GETPEERNAME and TI_GETMYNAME, we first
7556			 * need to copyin the user's strbuf structure.
7557			 * Processing will continue in the M_IOCDATA case
7558			 * below.
7559			 */
7560			mi_copyin(q, mp, NULL,
7561			    SIZEOF_STRUCT(strbuf, iocp->ioc_flag));
7562			TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7563			    "udp_wput_other_end: q %p (%S)", q, "getmyname");
7564			return;
7565			}
7566		case ND_SET:
7567			/* nd_getset performs the necessary checking */
7568		case ND_GET:
7569			if (nd_getset(q, us->us_nd, mp)) {
7570				qreply(q, mp);
7571				TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7572				    "udp_wput_other_end: q %p (%S)", q, "get");
7573				return;
7574			}
7575			break;
7576		case _SIOCSOCKFALLBACK:
7577			/*
7578			 * Either sockmod is about to be popped and the
7579			 * socket would now be treated as a plain stream,
7580			 * or a module is about to be pushed so we could
7581			 * no longer use read-side synchronous stream.
7582			 * Drain any queued data and disable direct sockfs
7583			 * interface from now on.
7584			 */
7585			if (!udp->udp_issocket) {
7586				DB_TYPE(mp) = M_IOCNAK;
7587				iocp->ioc_error = EINVAL;
7588			} else {
7589				udp->udp_issocket = B_FALSE;
7590				if (udp->udp_direct_sockfs) {
7591					/*
7592					 * Disable read-side synchronous
7593					 * stream interface and drain any
7594					 * queued data.
7595					 */
7596					udp_rcv_drain(RD(q), udp,
7597					    B_FALSE);
7598					ASSERT(!udp->udp_direct_sockfs);
7599					UDP_STAT(us, udp_sock_fallback);
7600				}
7601				DB_TYPE(mp) = M_IOCACK;
7602				iocp->ioc_error = 0;
7603			}
7604			iocp->ioc_count = 0;
7605			iocp->ioc_rval = 0;
7606			qreply(q, mp);
7607			return;
7608		default:
7609			break;
7610		}
7611		break;
7612	case M_IOCDATA:
7613		udp_wput_iocdata(q, mp);
7614		TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7615		    "udp_wput_other_end: q %p (%S)", q, "iocdata");
7616		return;
7617	default:
7618		/* Unrecognized messages are passed through without change. */
7619		break;
7620	}
7621	TRACE_2(TR_FAC_UDP, TR_UDP_WPUT_OTHER_END,
7622	    "udp_wput_other_end: q %p (%S)", q, "end");
7623	ip_output(connp, mp, q, IP_WPUT);
7624}
7625
7626/*
7627 * udp_wput_iocdata is called by udp_wput_other to handle all M_IOCDATA
7628 * messages.
7629 */
7630static void
7631udp_wput_iocdata(queue_t *q, mblk_t *mp)
7632{
7633	mblk_t	*mp1;
7634	struct iocblk *iocp = (struct iocblk *)mp->b_rptr;
7635	STRUCT_HANDLE(strbuf, sb);
7636	udp_t	*udp = Q_TO_UDP(q);
7637	int	error;
7638	uint_t	addrlen;
7639
7640	/* Make sure it is one of ours. */
7641	switch (iocp->ioc_cmd) {
7642	case TI_GETMYNAME:
7643	case TI_GETPEERNAME:
7644		break;
7645	default:
7646		ip_output(udp->udp_connp, mp, q, IP_WPUT);
7647		return;
7648	}
7649
7650	switch (mi_copy_state(q, mp, &mp1)) {
7651	case -1:
7652		return;
7653	case MI_COPY_CASE(MI_COPY_IN, 1):
7654		break;
7655	case MI_COPY_CASE(MI_COPY_OUT, 1):
7656		/*
7657		 * The address has been copied out, so now
7658		 * copyout the strbuf.
7659		 */
7660		mi_copyout(q, mp);
7661		return;
7662	case MI_COPY_CASE(MI_COPY_OUT, 2):
7663		/*
7664		 * The address and strbuf have been copied out.
7665		 * We're done, so just acknowledge the original
7666		 * M_IOCTL.
7667		 */
7668		mi_copy_done(q, mp, 0);
7669		return;
7670	default:
7671		/*
7672		 * Something strange has happened, so acknowledge
7673		 * the original M_IOCTL with an EPROTO error.
7674		 */
7675		mi_copy_done(q, mp, EPROTO);
7676		return;
7677	}
7678
7679	/*
7680	 * Now we have the strbuf structure for TI_GETMYNAME
7681	 * and TI_GETPEERNAME.  Next we copyout the requested
7682	 * address and then we'll copyout the strbuf.
7683	 */
7684	STRUCT_SET_HANDLE(sb, iocp->ioc_flag, (void *)mp1->b_rptr);
7685	addrlen = udp->udp_family == AF_INET ? sizeof (sin_t) : sizeof (sin6_t);
7686	if (STRUCT_FGET(sb, maxlen) < addrlen) {
7687		mi_copy_done(q, mp, EINVAL);
7688		return;
7689	}
7690
7691	mp1 = mi_copyout_alloc(q, mp, STRUCT_FGETP(sb, buf), addrlen, B_TRUE);
7692	if (mp1 == NULL)
7693		return;
7694
7695	rw_enter(&udp->udp_rwlock, RW_READER);
7696	switch (iocp->ioc_cmd) {
7697	case TI_GETMYNAME:
7698		error = udp_getmyname(udp, (void *)mp1->b_rptr, &addrlen);
7699		break;
7700	case TI_GETPEERNAME:
7701		error = udp_getpeername(udp, (void *)mp1->b_rptr, &addrlen);
7702		break;
7703	}
7704	rw_exit(&udp->udp_rwlock);
7705
7706	if (error != 0) {
7707		mi_copy_done(q, mp, error);
7708	} else {
7709		mp1->b_wptr += addrlen;
7710		STRUCT_FSET(sb, len, addrlen);
7711
7712		/* Copy out the address */
7713		mi_copyout(q, mp);
7714	}
7715}
7716
7717static int
7718udp_unitdata_opt_process(queue_t *q, mblk_t *mp, int *errorp,
7719    udpattrs_t *udpattrs)
7720{
7721	struct T_unitdata_req *udreqp;
7722	int is_absreq_failure;
7723	cred_t *cr;
7724	conn_t	*connp = Q_TO_CONN(q);
7725
7726	ASSERT(((t_primp_t)mp->b_rptr)->type);
7727
7728	cr = DB_CREDDEF(mp, connp->conn_cred);
7729
7730	udreqp = (struct T_unitdata_req *)mp->b_rptr;
7731
7732	*errorp = tpi_optcom_buf(q, mp, &udreqp->OPT_length,
7733	    udreqp->OPT_offset, cr, &udp_opt_obj,
7734	    udpattrs, &is_absreq_failure);
7735
7736	if (*errorp != 0) {
7737		/*
7738		 * Note: No special action needed in this
7739		 * module for "is_absreq_failure"
7740		 */
7741		return (-1);		/* failure */
7742	}
7743	ASSERT(is_absreq_failure == 0);
7744	return (0);	/* success */
7745}
7746
7747void
7748udp_ddi_init(void)
7749{
7750	udp_max_optsize = optcom_max_optsize(udp_opt_obj.odb_opt_des_arr,
7751	    udp_opt_obj.odb_opt_arr_cnt);
7752
7753	/*
7754	 * We want to be informed each time a stack is created or
7755	 * destroyed in the kernel, so we can maintain the
7756	 * set of udp_stack_t's.
7757	 */
7758	netstack_register(NS_UDP, udp_stack_init, NULL, udp_stack_fini);
7759}
7760
7761void
7762udp_ddi_destroy(void)
7763{
7764	netstack_unregister(NS_UDP);
7765}
7766
7767/*
7768 * Initialize the UDP stack instance.
7769 */
7770static void *
7771udp_stack_init(netstackid_t stackid, netstack_t *ns)
7772{
7773	udp_stack_t	*us;
7774	udpparam_t	*pa;
7775	int		i;
7776
7777	us = (udp_stack_t *)kmem_zalloc(sizeof (*us), KM_SLEEP);
7778	us->us_netstack = ns;
7779
7780	us->us_num_epriv_ports = UDP_NUM_EPRIV_PORTS;
7781	us->us_epriv_ports[0] = 2049;
7782	us->us_epriv_ports[1] = 4045;
7783
7784	/*
7785	 * The smallest anonymous port in the priviledged port range which UDP
7786	 * looks for free port.  Use in the option UDP_ANONPRIVBIND.
7787	 */
7788	us->us_min_anonpriv_port = 512;
7789
7790	us->us_bind_fanout_size = udp_bind_fanout_size;
7791
7792	/* Roundup variable that might have been modified in /etc/system */
7793	if (us->us_bind_fanout_size & (us->us_bind_fanout_size - 1)) {
7794		/* Not a power of two. Round up to nearest power of two */
7795		for (i = 0; i < 31; i++) {
7796			if (us->us_bind_fanout_size < (1 << i))
7797				break;
7798		}
7799		us->us_bind_fanout_size = 1 << i;
7800	}
7801	us->us_bind_fanout = kmem_zalloc(us->us_bind_fanout_size *
7802	    sizeof (udp_fanout_t), KM_SLEEP);
7803	for (i = 0; i < us->us_bind_fanout_size; i++) {
7804		mutex_init(&us->us_bind_fanout[i].uf_lock, NULL, MUTEX_DEFAULT,
7805		    NULL);
7806	}
7807
7808	pa = (udpparam_t *)kmem_alloc(sizeof (udp_param_arr), KM_SLEEP);
7809
7810	us->us_param_arr = pa;
7811	bcopy(udp_param_arr, us->us_param_arr, sizeof (udp_param_arr));
7812
7813	(void) udp_param_register(&us->us_nd,
7814	    us->us_param_arr, A_CNT(udp_param_arr));
7815
7816	us->us_kstat = udp_kstat2_init(stackid, &us->us_statistics);
7817	us->us_mibkp = udp_kstat_init(stackid);
7818	return (us);
7819}
7820
7821/*
7822 * Free the UDP stack instance.
7823 */
7824static void
7825udp_stack_fini(netstackid_t stackid, void *arg)
7826{
7827	udp_stack_t *us = (udp_stack_t *)arg;
7828	int i;
7829
7830	for (i = 0; i < us->us_bind_fanout_size; i++) {
7831		mutex_destroy(&us->us_bind_fanout[i].uf_lock);
7832	}
7833
7834	kmem_free(us->us_bind_fanout, us->us_bind_fanout_size *
7835	    sizeof (udp_fanout_t));
7836
7837	us->us_bind_fanout = NULL;
7838
7839	nd_free(&us->us_nd);
7840	kmem_free(us->us_param_arr, sizeof (udp_param_arr));
7841	us->us_param_arr = NULL;
7842
7843	udp_kstat_fini(stackid, us->us_mibkp);
7844	us->us_mibkp = NULL;
7845
7846	udp_kstat2_fini(stackid, us->us_kstat);
7847	us->us_kstat = NULL;
7848	bzero(&us->us_statistics, sizeof (us->us_statistics));
7849	kmem_free(us, sizeof (*us));
7850}
7851
7852static void *
7853udp_kstat2_init(netstackid_t stackid, udp_stat_t *us_statisticsp)
7854{
7855	kstat_t *ksp;
7856
7857	udp_stat_t template = {
7858		{ "udp_ip_send",		KSTAT_DATA_UINT64 },
7859		{ "udp_ip_ire_send",		KSTAT_DATA_UINT64 },
7860		{ "udp_ire_null",		KSTAT_DATA_UINT64 },
7861		{ "udp_drain",			KSTAT_DATA_UINT64 },
7862		{ "udp_sock_fallback",		KSTAT_DATA_UINT64 },
7863		{ "udp_rrw_busy",		KSTAT_DATA_UINT64 },
7864		{ "udp_rrw_msgcnt",		KSTAT_DATA_UINT64 },
7865		{ "udp_out_sw_cksum",		KSTAT_DATA_UINT64 },
7866		{ "udp_out_sw_cksum_bytes",	KSTAT_DATA_UINT64 },
7867		{ "udp_out_opt",		KSTAT_DATA_UINT64 },
7868		{ "udp_out_err_notconn",	KSTAT_DATA_UINT64 },
7869		{ "udp_out_err_output",		KSTAT_DATA_UINT64 },
7870		{ "udp_out_err_tudr",		KSTAT_DATA_UINT64 },
7871		{ "udp_in_pktinfo",		KSTAT_DATA_UINT64 },
7872		{ "udp_in_recvdstaddr",		KSTAT_DATA_UINT64 },
7873		{ "udp_in_recvopts",		KSTAT_DATA_UINT64 },
7874		{ "udp_in_recvif",		KSTAT_DATA_UINT64 },
7875		{ "udp_in_recvslla",		KSTAT_DATA_UINT64 },
7876		{ "udp_in_recvucred",		KSTAT_DATA_UINT64 },
7877		{ "udp_in_recvttl",		KSTAT_DATA_UINT64 },
7878		{ "udp_in_recvhopopts",		KSTAT_DATA_UINT64 },
7879		{ "udp_in_recvhoplimit",	KSTAT_DATA_UINT64 },
7880		{ "udp_in_recvdstopts",		KSTAT_DATA_UINT64 },
7881		{ "udp_in_recvrtdstopts",	KSTAT_DATA_UINT64 },
7882		{ "udp_in_recvrthdr",		KSTAT_DATA_UINT64 },
7883		{ "udp_in_recvpktinfo",		KSTAT_DATA_UINT64 },
7884		{ "udp_in_recvtclass",		KSTAT_DATA_UINT64 },
7885		{ "udp_in_timestamp",		KSTAT_DATA_UINT64 },
7886#ifdef DEBUG
7887		{ "udp_data_conn",		KSTAT_DATA_UINT64 },
7888		{ "udp_data_notconn",		KSTAT_DATA_UINT64 },
7889#endif
7890	};
7891
7892	ksp = kstat_create_netstack(UDP_MOD_NAME, 0, "udpstat", "net",
7893	    KSTAT_TYPE_NAMED, sizeof (template) / sizeof (kstat_named_t),
7894	    KSTAT_FLAG_VIRTUAL, stackid);
7895
7896	if (ksp == NULL)
7897		return (NULL);
7898
7899	bcopy(&template, us_statisticsp, sizeof (template));
7900	ksp->ks_data = (void *)us_statisticsp;
7901	ksp->ks_private = (void *)(uintptr_t)stackid;
7902
7903	kstat_install(ksp);
7904	return (ksp);
7905}
7906
7907static void
7908udp_kstat2_fini(netstackid_t stackid, kstat_t *ksp)
7909{
7910	if (ksp != NULL) {
7911		ASSERT(stackid == (netstackid_t)(uintptr_t)ksp->ks_private);
7912		kstat_delete_netstack(ksp, stackid);
7913	}
7914}
7915
7916static void *
7917udp_kstat_init(netstackid_t stackid)
7918{
7919	kstat_t	*ksp;
7920
7921	udp_named_kstat_t template = {
7922		{ "inDatagrams",	KSTAT_DATA_UINT64, 0 },
7923		{ "inErrors",		KSTAT_DATA_UINT32, 0 },
7924		{ "outDatagrams",	KSTAT_DATA_UINT64, 0 },
7925		{ "entrySize",		KSTAT_DATA_INT32, 0 },
7926		{ "entry6Size",		KSTAT_DATA_INT32, 0 },
7927		{ "outErrors",		KSTAT_DATA_UINT32, 0 },
7928	};
7929
7930	ksp = kstat_create_netstack(UDP_MOD_NAME, 0, UDP_MOD_NAME, "mib2",
7931	    KSTAT_TYPE_NAMED,
7932	    NUM_OF_FIELDS(udp_named_kstat_t), 0, stackid);
7933
7934	if (ksp == NULL || ksp->ks_data == NULL)
7935		return (NULL);
7936
7937	template.entrySize.value.ui32 = sizeof (mib2_udpEntry_t);
7938	template.entry6Size.value.ui32 = sizeof (mib2_udp6Entry_t);
7939
7940	bcopy(&template, ksp->ks_data, sizeof (template));
7941	ksp->ks_update = udp_kstat_update;
7942	ksp->ks_private = (void *)(uintptr_t)stackid;
7943
7944	kstat_install(ksp);
7945	return (ksp);
7946}
7947
7948static void
7949udp_kstat_fini(netstackid_t stackid, kstat_t *ksp)
7950{
7951	if (ksp != NULL) {
7952		ASSERT(stackid == (netstackid_t)(uintptr_t)ksp->ks_private);
7953		kstat_delete_netstack(ksp, stackid);
7954	}
7955}
7956
7957static int
7958udp_kstat_update(kstat_t *kp, int rw)
7959{
7960	udp_named_kstat_t *udpkp;
7961	netstackid_t	stackid = (netstackid_t)(uintptr_t)kp->ks_private;
7962	netstack_t	*ns;
7963	udp_stack_t	*us;
7964
7965	if ((kp == NULL) || (kp->ks_data == NULL))
7966		return (EIO);
7967
7968	if (rw == KSTAT_WRITE)
7969		return (EACCES);
7970
7971	ns = netstack_find_by_stackid(stackid);
7972	if (ns == NULL)
7973		return (-1);
7974	us = ns->netstack_udp;
7975	if (us == NULL) {
7976		netstack_rele(ns);
7977		return (-1);
7978	}
7979	udpkp = (udp_named_kstat_t *)kp->ks_data;
7980
7981	udpkp->inDatagrams.value.ui64 =	us->us_udp_mib.udpHCInDatagrams;
7982	udpkp->inErrors.value.ui32 =	us->us_udp_mib.udpInErrors;
7983	udpkp->outDatagrams.value.ui64 = us->us_udp_mib.udpHCOutDatagrams;
7984	udpkp->outErrors.value.ui32 =	us->us_udp_mib.udpOutErrors;
7985	netstack_rele(ns);
7986	return (0);
7987}
7988
7989/*
7990 * Read-side synchronous stream info entry point, called as a
7991 * result of handling certain STREAMS ioctl operations.
7992 */
7993static int
7994udp_rinfop(queue_t *q, infod_t *dp)
7995{
7996	mblk_t	*mp;
7997	uint_t	cmd = dp->d_cmd;
7998	int	res = 0;
7999	int	error = 0;
8000	udp_t	*udp = Q_TO_UDP(q);
8001	struct stdata *stp = STREAM(q);
8002
8003	mutex_enter(&udp->udp_drain_lock);
8004	/* If shutdown on read has happened, return nothing */
8005	mutex_enter(&stp->sd_lock);
8006	if (stp->sd_flag & STREOF) {
8007		mutex_exit(&stp->sd_lock);
8008		goto done;
8009	}
8010	mutex_exit(&stp->sd_lock);
8011
8012	if ((mp = udp->udp_rcv_list_head) == NULL)
8013		goto done;
8014
8015	ASSERT(DB_TYPE(mp) != M_DATA && mp->b_cont != NULL);
8016
8017	if (cmd & INFOD_COUNT) {
8018		/*
8019		 * Return the number of messages.
8020		 */
8021		dp->d_count += udp->udp_rcv_msgcnt;
8022		res |= INFOD_COUNT;
8023	}
8024	if (cmd & INFOD_BYTES) {
8025		/*
8026		 * Return size of all data messages.
8027		 */
8028		dp->d_bytes += udp->udp_rcv_cnt;
8029		res |= INFOD_BYTES;
8030	}
8031	if (cmd & INFOD_FIRSTBYTES) {
8032		/*
8033		 * Return size of first data message.
8034		 */
8035		dp->d_bytes = msgdsize(mp);
8036		res |= INFOD_FIRSTBYTES;
8037		dp->d_cmd &= ~INFOD_FIRSTBYTES;
8038	}
8039	if (cmd & INFOD_COPYOUT) {
8040		mblk_t *mp1 = mp->b_cont;
8041		int n;
8042		/*
8043		 * Return data contents of first message.
8044		 */
8045		ASSERT(DB_TYPE(mp1) == M_DATA);
8046		while (mp1 != NULL && dp->d_uiop->uio_resid > 0) {
8047			n = MIN(dp->d_uiop->uio_resid, MBLKL(mp1));
8048			if (n != 0 && (error = uiomove((char *)mp1->b_rptr, n,
8049			    UIO_READ, dp->d_uiop)) != 0) {
8050				goto done;
8051			}
8052			mp1 = mp1->b_cont;
8053		}
8054		res |= INFOD_COPYOUT;
8055		dp->d_cmd &= ~INFOD_COPYOUT;
8056	}
8057done:
8058	mutex_exit(&udp->udp_drain_lock);
8059
8060	dp->d_res |= res;
8061
8062	return (error);
8063}
8064
8065/*
8066 * Read-side synchronous stream entry point.  This is called as a result
8067 * of recv/read operation done at sockfs, and is guaranteed to execute
8068 * outside of the interrupt thread context.  It returns a single datagram
8069 * (b_cont chain of T_UNITDATA_IND plus data) to the upper layer.
8070 */
8071static int
8072udp_rrw(queue_t *q, struiod_t *dp)
8073{
8074	mblk_t	*mp;
8075	udp_t	*udp = Q_TO_UDP(q);
8076	udp_stack_t *us = udp->udp_us;
8077
8078	/*
8079	 * Dequeue datagram from the head of the list and return
8080	 * it to caller; also ensure that RSLEEP sd_wakeq flag is
8081	 * set/cleared depending on whether or not there's data
8082	 * remaining in the list.
8083	 */
8084	mutex_enter(&udp->udp_drain_lock);
8085	if (!udp->udp_direct_sockfs) {
8086		mutex_exit(&udp->udp_drain_lock);
8087		UDP_STAT(us, udp_rrw_busy);
8088		return (EBUSY);
8089	}
8090	if ((mp = udp->udp_rcv_list_head) != NULL) {
8091		uint_t size = msgdsize(mp);
8092
8093		/* Last datagram in the list? */
8094		if ((udp->udp_rcv_list_head = mp->b_next) == NULL)
8095			udp->udp_rcv_list_tail = NULL;
8096		mp->b_next = NULL;
8097
8098		udp->udp_rcv_cnt -= size;
8099		udp->udp_rcv_msgcnt--;
8100		UDP_STAT(us, udp_rrw_msgcnt);
8101
8102		/* No longer flow-controlling? */
8103		if (udp->udp_rcv_cnt < udp->udp_rcv_hiwat &&
8104		    udp->udp_rcv_msgcnt < udp->udp_rcv_hiwat)
8105			udp->udp_drain_qfull = B_FALSE;
8106	}
8107	if (udp->udp_rcv_list_head == NULL) {
8108		/*
8109		 * Either we just dequeued the last datagram or
8110		 * we get here from sockfs and have nothing to
8111		 * return; in this case clear RSLEEP.
8112		 */
8113		ASSERT(udp->udp_rcv_cnt == 0);
8114		ASSERT(udp->udp_rcv_msgcnt == 0);
8115		ASSERT(udp->udp_rcv_list_tail == NULL);
8116		STR_WAKEUP_CLEAR(STREAM(q));
8117	} else {
8118		/*
8119		 * More data follows; we need udp_rrw() to be
8120		 * called in future to pick up the rest.
8121		 */
8122		STR_WAKEUP_SET(STREAM(q));
8123	}
8124	mutex_exit(&udp->udp_drain_lock);
8125	dp->d_mp = mp;
8126	return (0);
8127}
8128
8129/*
8130 * Enqueue a completely-built T_UNITDATA_IND message into the receive
8131 * list; this is typically executed within the interrupt thread context
8132 * and so we do things as quickly as possible.
8133 */
8134static void
8135udp_rcv_enqueue(queue_t *q, udp_t *udp, mblk_t *mp, uint_t pkt_len)
8136{
8137	ASSERT(q == RD(q));
8138	ASSERT(pkt_len == msgdsize(mp));
8139	ASSERT(mp->b_next == NULL && mp->b_cont != NULL);
8140	ASSERT(DB_TYPE(mp) == M_PROTO && DB_TYPE(mp->b_cont) == M_DATA);
8141	ASSERT(MBLKL(mp) >= sizeof (struct T_unitdata_ind));
8142
8143	mutex_enter(&udp->udp_drain_lock);
8144	/*
8145	 * Wake up and signal the receiving app; it is okay to do this
8146	 * before enqueueing the mp because we are holding the drain lock.
8147	 * One of the advantages of synchronous stream is the ability for
8148	 * us to find out when the application performs a read on the
8149	 * socket by way of udp_rrw() entry point being called.  We need
8150	 * to generate SIGPOLL/SIGIO for each received data in the case
8151	 * of asynchronous socket just as in the strrput() case.  However,
8152	 * we only wake the application up when necessary, i.e. during the
8153	 * first enqueue.  When udp_rrw() is called, we send up a single
8154	 * datagram upstream and call STR_WAKEUP_SET() again when there
8155	 * are still data remaining in our receive queue.
8156	 */
8157	STR_WAKEUP_SENDSIG(STREAM(q), udp->udp_rcv_list_head);
8158	if (udp->udp_rcv_list_head == NULL)
8159		udp->udp_rcv_list_head = mp;
8160	else
8161		udp->udp_rcv_list_tail->b_next = mp;
8162	udp->udp_rcv_list_tail = mp;
8163	udp->udp_rcv_cnt += pkt_len;
8164	udp->udp_rcv_msgcnt++;
8165
8166	/* Need to flow-control? */
8167	if (udp->udp_rcv_cnt >= udp->udp_rcv_hiwat ||
8168	    udp->udp_rcv_msgcnt >= udp->udp_rcv_hiwat)
8169		udp->udp_drain_qfull = B_TRUE;
8170
8171	mutex_exit(&udp->udp_drain_lock);
8172}
8173
8174/*
8175 * Drain the contents of receive list to the module upstream; we do
8176 * this during close or when we fallback to the slow mode due to
8177 * sockmod being popped or a module being pushed on top of us.
8178 */
8179static void
8180udp_rcv_drain(queue_t *q, udp_t *udp, boolean_t closing)
8181{
8182	mblk_t *mp;
8183	udp_stack_t *us = udp->udp_us;
8184
8185	ASSERT(q == RD(q));
8186
8187	mutex_enter(&udp->udp_drain_lock);
8188	/*
8189	 * There is no race with a concurrent udp_input() sending
8190	 * up packets using putnext() after we have cleared the
8191	 * udp_direct_sockfs flag but before we have completed
8192	 * sending up the packets in udp_rcv_list, since we are
8193	 * either a writer or we have quiesced the conn.
8194	 */
8195	udp->udp_direct_sockfs = B_FALSE;
8196	mutex_exit(&udp->udp_drain_lock);
8197
8198	if (udp->udp_rcv_list_head != NULL)
8199		UDP_STAT(us, udp_drain);
8200
8201	/*
8202	 * Send up everything via putnext(); note here that we
8203	 * don't need the udp_drain_lock to protect us since
8204	 * nothing can enter udp_rrw() and that we currently
8205	 * have exclusive access to this udp.
8206	 */
8207	while ((mp = udp->udp_rcv_list_head) != NULL) {
8208		udp->udp_rcv_list_head = mp->b_next;
8209		mp->b_next = NULL;
8210		udp->udp_rcv_cnt -= msgdsize(mp);
8211		udp->udp_rcv_msgcnt--;
8212		if (closing) {
8213			freemsg(mp);
8214		} else {
8215			putnext(q, mp);
8216		}
8217	}
8218	ASSERT(udp->udp_rcv_cnt == 0);
8219	ASSERT(udp->udp_rcv_msgcnt == 0);
8220	ASSERT(udp->udp_rcv_list_head == NULL);
8221	udp->udp_rcv_list_tail = NULL;
8222	udp->udp_drain_qfull = B_FALSE;
8223}
8224
8225static size_t
8226udp_set_rcv_hiwat(udp_t *udp, size_t size)
8227{
8228	udp_stack_t *us = udp->udp_us;
8229
8230	/* We add a bit of extra buffering */
8231	size += size >> 1;
8232	if (size > us->us_max_buf)
8233		size = us->us_max_buf;
8234
8235	udp->udp_rcv_hiwat = size;
8236	return (size);
8237}
8238
8239/*
8240 * For the lower queue so that UDP can be a dummy mux.
8241 * Nobody should be sending
8242 * packets up this stream
8243 */
8244static void
8245udp_lrput(queue_t *q, mblk_t *mp)
8246{
8247	mblk_t *mp1;
8248
8249	switch (mp->b_datap->db_type) {
8250	case M_FLUSH:
8251		/* Turn around */
8252		if (*mp->b_rptr & FLUSHW) {
8253			*mp->b_rptr &= ~FLUSHR;
8254			qreply(q, mp);
8255			return;
8256		}
8257		break;
8258	}
8259	/* Could receive messages that passed through ar_rput */
8260	for (mp1 = mp; mp1; mp1 = mp1->b_cont)
8261		mp1->b_prev = mp1->b_next = NULL;
8262	freemsg(mp);
8263}
8264
8265/*
8266 * For the lower queue so that UDP can be a dummy mux.
8267 * Nobody should be sending packets down this stream.
8268 */
8269/* ARGSUSED */
8270void
8271udp_lwput(queue_t *q, mblk_t *mp)
8272{
8273	freemsg(mp);
8274}
8275