svc.c revision 267740
1177633Sdfr/*	$NetBSD: svc.c,v 1.21 2000/07/06 03:10:35 christos Exp $	*/
2177633Sdfr
3261046Smav/*-
4261046Smav * Copyright (c) 2009, Sun Microsystems, Inc.
5261046Smav * All rights reserved.
6177633Sdfr *
7261046Smav * Redistribution and use in source and binary forms, with or without
8261046Smav * modification, are permitted provided that the following conditions are met:
9261046Smav * - Redistributions of source code must retain the above copyright notice,
10261046Smav *   this list of conditions and the following disclaimer.
11261046Smav * - Redistributions in binary form must reproduce the above copyright notice,
12261046Smav *   this list of conditions and the following disclaimer in the documentation
13261046Smav *   and/or other materials provided with the distribution.
14261046Smav * - Neither the name of Sun Microsystems, Inc. nor the names of its
15261046Smav *   contributors may be used to endorse or promote products derived
16261046Smav *   from this software without specific prior written permission.
17261046Smav *
18261046Smav * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
19261046Smav * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
20261046Smav * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
21261046Smav * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
22261046Smav * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
23261046Smav * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
24261046Smav * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
25261046Smav * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
26261046Smav * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
27261046Smav * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
28261046Smav * POSSIBILITY OF SUCH DAMAGE.
29177633Sdfr */
30177633Sdfr
31177633Sdfr#if defined(LIBC_SCCS) && !defined(lint)
32177633Sdfrstatic char *sccsid2 = "@(#)svc.c 1.44 88/02/08 Copyr 1984 Sun Micro";
33177633Sdfrstatic char *sccsid = "@(#)svc.c	2.4 88/08/11 4.0 RPCSRC";
34177633Sdfr#endif
35177633Sdfr#include <sys/cdefs.h>
36177633Sdfr__FBSDID("$FreeBSD: stable/10/sys/rpc/svc.c 267740 2014-06-22 18:01:40Z mav $");
37177633Sdfr
38177633Sdfr/*
39177633Sdfr * svc.c, Server-side remote procedure call interface.
40177633Sdfr *
41177633Sdfr * There are two sets of procedures here.  The xprt routines are
42177633Sdfr * for handling transport handles.  The svc routines handle the
43177633Sdfr * list of service routines.
44177633Sdfr *
45177633Sdfr * Copyright (C) 1984, Sun Microsystems, Inc.
46177633Sdfr */
47177633Sdfr
48177633Sdfr#include <sys/param.h>
49177633Sdfr#include <sys/lock.h>
50177633Sdfr#include <sys/kernel.h>
51184588Sdfr#include <sys/kthread.h>
52177633Sdfr#include <sys/malloc.h>
53184588Sdfr#include <sys/mbuf.h>
54177633Sdfr#include <sys/mutex.h>
55184588Sdfr#include <sys/proc.h>
56177633Sdfr#include <sys/queue.h>
57184588Sdfr#include <sys/socketvar.h>
58177633Sdfr#include <sys/systm.h>
59261055Smav#include <sys/sx.h>
60177633Sdfr#include <sys/ucred.h>
61177633Sdfr
62177633Sdfr#include <rpc/rpc.h>
63177633Sdfr#include <rpc/rpcb_clnt.h>
64184588Sdfr#include <rpc/replay.h>
65177633Sdfr
66177685Sdfr#include <rpc/rpc_com.h>
67177633Sdfr
68177633Sdfr#define SVC_VERSQUIET 0x0001		/* keep quiet about vers mismatch */
69184588Sdfr#define version_keepquiet(xp) (SVC_EXT(xp)->xp_flags & SVC_VERSQUIET)
70177633Sdfr
71177633Sdfrstatic struct svc_callout *svc_find(SVCPOOL *pool, rpcprog_t, rpcvers_t,
72177633Sdfr    char *);
73184588Sdfrstatic void svc_new_thread(SVCPOOL *pool);
74184588Sdfrstatic void xprt_unregister_locked(SVCXPRT *xprt);
75261054Smavstatic void svc_change_space_used(SVCPOOL *pool, int delta);
76261054Smavstatic bool_t svc_request_space_available(SVCPOOL *pool);
77177633Sdfr
78177633Sdfr/* ***************  SVCXPRT related stuff **************** */
79177633Sdfr
80184588Sdfrstatic int svcpool_minthread_sysctl(SYSCTL_HANDLER_ARGS);
81184588Sdfrstatic int svcpool_maxthread_sysctl(SYSCTL_HANDLER_ARGS);
82184588Sdfr
83177633SdfrSVCPOOL*
84184588Sdfrsvcpool_create(const char *name, struct sysctl_oid_list *sysctl_base)
85177633Sdfr{
86177633Sdfr	SVCPOOL *pool;
87177633Sdfr
88177633Sdfr	pool = malloc(sizeof(SVCPOOL), M_RPC, M_WAITOK|M_ZERO);
89177633Sdfr
90177633Sdfr	mtx_init(&pool->sp_lock, "sp_lock", NULL, MTX_DEF);
91184588Sdfr	pool->sp_name = name;
92184588Sdfr	pool->sp_state = SVCPOOL_INIT;
93184588Sdfr	pool->sp_proc = NULL;
94177633Sdfr	TAILQ_INIT(&pool->sp_xlist);
95177633Sdfr	TAILQ_INIT(&pool->sp_active);
96177633Sdfr	TAILQ_INIT(&pool->sp_callouts);
97261055Smav	TAILQ_INIT(&pool->sp_lcallouts);
98184588Sdfr	LIST_INIT(&pool->sp_threads);
99184588Sdfr	LIST_INIT(&pool->sp_idlethreads);
100184588Sdfr	pool->sp_minthreads = 1;
101184588Sdfr	pool->sp_maxthreads = 1;
102184588Sdfr	pool->sp_threadcount = 0;
103177633Sdfr
104184588Sdfr	/*
105184588Sdfr	 * Don't use more than a quarter of mbuf clusters or more than
106184588Sdfr	 * 45Mb buffering requests.
107184588Sdfr	 */
108184588Sdfr	pool->sp_space_high = nmbclusters * MCLBYTES / 4;
109184588Sdfr	if (pool->sp_space_high > 45 << 20)
110184588Sdfr		pool->sp_space_high = 45 << 20;
111184588Sdfr	pool->sp_space_low = 2 * pool->sp_space_high / 3;
112184588Sdfr
113184588Sdfr	sysctl_ctx_init(&pool->sp_sysctl);
114184588Sdfr	if (sysctl_base) {
115184588Sdfr		SYSCTL_ADD_PROC(&pool->sp_sysctl, sysctl_base, OID_AUTO,
116184588Sdfr		    "minthreads", CTLTYPE_INT | CTLFLAG_RW,
117184588Sdfr		    pool, 0, svcpool_minthread_sysctl, "I", "");
118184588Sdfr		SYSCTL_ADD_PROC(&pool->sp_sysctl, sysctl_base, OID_AUTO,
119184588Sdfr		    "maxthreads", CTLTYPE_INT | CTLFLAG_RW,
120184588Sdfr		    pool, 0, svcpool_maxthread_sysctl, "I", "");
121184588Sdfr		SYSCTL_ADD_INT(&pool->sp_sysctl, sysctl_base, OID_AUTO,
122184588Sdfr		    "threads", CTLFLAG_RD, &pool->sp_threadcount, 0, "");
123184588Sdfr
124184588Sdfr		SYSCTL_ADD_UINT(&pool->sp_sysctl, sysctl_base, OID_AUTO,
125184588Sdfr		    "request_space_used", CTLFLAG_RD,
126184588Sdfr		    &pool->sp_space_used, 0,
127184588Sdfr		    "Space in parsed but not handled requests.");
128184588Sdfr
129184588Sdfr		SYSCTL_ADD_UINT(&pool->sp_sysctl, sysctl_base, OID_AUTO,
130184588Sdfr		    "request_space_used_highest", CTLFLAG_RD,
131184588Sdfr		    &pool->sp_space_used_highest, 0,
132184588Sdfr		    "Highest space used since reboot.");
133184588Sdfr
134184588Sdfr		SYSCTL_ADD_UINT(&pool->sp_sysctl, sysctl_base, OID_AUTO,
135184588Sdfr		    "request_space_high", CTLFLAG_RW,
136184588Sdfr		    &pool->sp_space_high, 0,
137184588Sdfr		    "Maximum space in parsed but not handled requests.");
138184588Sdfr
139184588Sdfr		SYSCTL_ADD_UINT(&pool->sp_sysctl, sysctl_base, OID_AUTO,
140184588Sdfr		    "request_space_low", CTLFLAG_RW,
141184588Sdfr		    &pool->sp_space_low, 0,
142184588Sdfr		    "Low water mark for request space.");
143184588Sdfr
144217326Smdf		SYSCTL_ADD_INT(&pool->sp_sysctl, sysctl_base, OID_AUTO,
145184588Sdfr		    "request_space_throttled", CTLFLAG_RD,
146184588Sdfr		    &pool->sp_space_throttled, 0,
147184588Sdfr		    "Whether nfs requests are currently throttled");
148184588Sdfr
149217326Smdf		SYSCTL_ADD_INT(&pool->sp_sysctl, sysctl_base, OID_AUTO,
150184588Sdfr		    "request_space_throttle_count", CTLFLAG_RD,
151184588Sdfr		    &pool->sp_space_throttle_count, 0,
152184588Sdfr		    "Count of times throttling based on request space has occurred");
153184588Sdfr	}
154184588Sdfr
155177633Sdfr	return pool;
156177633Sdfr}
157177633Sdfr
158177633Sdfrvoid
159177633Sdfrsvcpool_destroy(SVCPOOL *pool)
160177633Sdfr{
161184588Sdfr	SVCXPRT *xprt, *nxprt;
162177633Sdfr	struct svc_callout *s;
163261055Smav	struct svc_loss_callout *sl;
164184588Sdfr	struct svcxprt_list cleanup;
165177633Sdfr
166184588Sdfr	TAILQ_INIT(&cleanup);
167177633Sdfr	mtx_lock(&pool->sp_lock);
168177633Sdfr
169177633Sdfr	while (TAILQ_FIRST(&pool->sp_xlist)) {
170177633Sdfr		xprt = TAILQ_FIRST(&pool->sp_xlist);
171184588Sdfr		xprt_unregister_locked(xprt);
172184588Sdfr		TAILQ_INSERT_TAIL(&cleanup, xprt, xp_link);
173177633Sdfr	}
174177633Sdfr
175261055Smav	while ((s = TAILQ_FIRST(&pool->sp_callouts)) != NULL) {
176177633Sdfr		mtx_unlock(&pool->sp_lock);
177177633Sdfr		svc_unreg(pool, s->sc_prog, s->sc_vers);
178177633Sdfr		mtx_lock(&pool->sp_lock);
179177633Sdfr	}
180261055Smav	while ((sl = TAILQ_FIRST(&pool->sp_lcallouts)) != NULL) {
181261055Smav		mtx_unlock(&pool->sp_lock);
182261055Smav		svc_loss_unreg(pool, sl->slc_dispatch);
183261055Smav		mtx_lock(&pool->sp_lock);
184261055Smav	}
185193603Srmacklem	mtx_unlock(&pool->sp_lock);
186177633Sdfr
187184588Sdfr	TAILQ_FOREACH_SAFE(xprt, &cleanup, xp_link, nxprt) {
188184588Sdfr		SVC_RELEASE(xprt);
189184588Sdfr	}
190184588Sdfr
191193436Srmacklem	mtx_destroy(&pool->sp_lock);
192193436Srmacklem
193184588Sdfr	if (pool->sp_rcache)
194184588Sdfr		replay_freecache(pool->sp_rcache);
195184588Sdfr
196184588Sdfr	sysctl_ctx_free(&pool->sp_sysctl);
197177633Sdfr	free(pool, M_RPC);
198177633Sdfr}
199177633Sdfr
200184588Sdfrstatic bool_t
201184588Sdfrsvcpool_active(SVCPOOL *pool)
202184588Sdfr{
203184588Sdfr	enum svcpool_state state = pool->sp_state;
204184588Sdfr
205184588Sdfr	if (state == SVCPOOL_INIT || state == SVCPOOL_CLOSING)
206184588Sdfr		return (FALSE);
207184588Sdfr	return (TRUE);
208184588Sdfr}
209184588Sdfr
210177633Sdfr/*
211184588Sdfr * Sysctl handler to set the minimum thread count on a pool
212184588Sdfr */
213184588Sdfrstatic int
214184588Sdfrsvcpool_minthread_sysctl(SYSCTL_HANDLER_ARGS)
215184588Sdfr{
216184588Sdfr	SVCPOOL *pool;
217184588Sdfr	int newminthreads, error, n;
218184588Sdfr
219184588Sdfr	pool = oidp->oid_arg1;
220184588Sdfr	newminthreads = pool->sp_minthreads;
221184588Sdfr	error = sysctl_handle_int(oidp, &newminthreads, 0, req);
222184588Sdfr	if (error == 0 && newminthreads != pool->sp_minthreads) {
223184588Sdfr		if (newminthreads > pool->sp_maxthreads)
224184588Sdfr			return (EINVAL);
225184588Sdfr		mtx_lock(&pool->sp_lock);
226184588Sdfr		if (newminthreads > pool->sp_minthreads
227184588Sdfr		    && svcpool_active(pool)) {
228184588Sdfr			/*
229184588Sdfr			 * If the pool is running and we are
230184588Sdfr			 * increasing, create some more threads now.
231184588Sdfr			 */
232184588Sdfr			n = newminthreads - pool->sp_threadcount;
233184588Sdfr			if (n > 0) {
234184588Sdfr				mtx_unlock(&pool->sp_lock);
235184588Sdfr				while (n--)
236184588Sdfr					svc_new_thread(pool);
237184588Sdfr				mtx_lock(&pool->sp_lock);
238184588Sdfr			}
239184588Sdfr		}
240184588Sdfr		pool->sp_minthreads = newminthreads;
241184588Sdfr		mtx_unlock(&pool->sp_lock);
242184588Sdfr	}
243184588Sdfr	return (error);
244184588Sdfr}
245184588Sdfr
246184588Sdfr/*
247184588Sdfr * Sysctl handler to set the maximum thread count on a pool
248184588Sdfr */
249184588Sdfrstatic int
250184588Sdfrsvcpool_maxthread_sysctl(SYSCTL_HANDLER_ARGS)
251184588Sdfr{
252184588Sdfr	SVCPOOL *pool;
253184588Sdfr	SVCTHREAD *st;
254184588Sdfr	int newmaxthreads, error;
255184588Sdfr
256184588Sdfr	pool = oidp->oid_arg1;
257184588Sdfr	newmaxthreads = pool->sp_maxthreads;
258184588Sdfr	error = sysctl_handle_int(oidp, &newmaxthreads, 0, req);
259184588Sdfr	if (error == 0 && newmaxthreads != pool->sp_maxthreads) {
260184588Sdfr		if (newmaxthreads < pool->sp_minthreads)
261184588Sdfr			return (EINVAL);
262184588Sdfr		mtx_lock(&pool->sp_lock);
263184588Sdfr		if (newmaxthreads < pool->sp_maxthreads
264184588Sdfr		    && svcpool_active(pool)) {
265184588Sdfr			/*
266184588Sdfr			 * If the pool is running and we are
267184588Sdfr			 * decreasing, wake up some idle threads to
268184588Sdfr			 * encourage them to exit.
269184588Sdfr			 */
270184588Sdfr			LIST_FOREACH(st, &pool->sp_idlethreads, st_ilink)
271184588Sdfr				cv_signal(&st->st_cond);
272184588Sdfr		}
273184588Sdfr		pool->sp_maxthreads = newmaxthreads;
274184588Sdfr		mtx_unlock(&pool->sp_lock);
275184588Sdfr	}
276184588Sdfr	return (error);
277184588Sdfr}
278184588Sdfr
279184588Sdfr/*
280177633Sdfr * Activate a transport handle.
281177633Sdfr */
282177633Sdfrvoid
283177633Sdfrxprt_register(SVCXPRT *xprt)
284177633Sdfr{
285177633Sdfr	SVCPOOL *pool = xprt->xp_pool;
286177633Sdfr
287194407Srmacklem	SVC_ACQUIRE(xprt);
288177633Sdfr	mtx_lock(&pool->sp_lock);
289177633Sdfr	xprt->xp_registered = TRUE;
290177633Sdfr	xprt->xp_active = FALSE;
291177633Sdfr	TAILQ_INSERT_TAIL(&pool->sp_xlist, xprt, xp_link);
292177633Sdfr	mtx_unlock(&pool->sp_lock);
293177633Sdfr}
294177633Sdfr
295177633Sdfr/*
296184588Sdfr * De-activate a transport handle. Note: the locked version doesn't
297184588Sdfr * release the transport - caller must do that after dropping the pool
298184588Sdfr * lock.
299177633Sdfr */
300177633Sdfrstatic void
301184588Sdfrxprt_unregister_locked(SVCXPRT *xprt)
302177633Sdfr{
303177633Sdfr	SVCPOOL *pool = xprt->xp_pool;
304177633Sdfr
305261048Smav	mtx_assert(&pool->sp_lock, MA_OWNED);
306193649Srmacklem	KASSERT(xprt->xp_registered == TRUE,
307193649Srmacklem	    ("xprt_unregister_locked: not registered"));
308261048Smav	xprt_inactive_locked(xprt);
309177633Sdfr	TAILQ_REMOVE(&pool->sp_xlist, xprt, xp_link);
310177633Sdfr	xprt->xp_registered = FALSE;
311184588Sdfr}
312177633Sdfr
313184588Sdfrvoid
314184588Sdfrxprt_unregister(SVCXPRT *xprt)
315184588Sdfr{
316184588Sdfr	SVCPOOL *pool = xprt->xp_pool;
317184588Sdfr
318184588Sdfr	mtx_lock(&pool->sp_lock);
319193649Srmacklem	if (xprt->xp_registered == FALSE) {
320193649Srmacklem		/* Already unregistered by another thread */
321193649Srmacklem		mtx_unlock(&pool->sp_lock);
322193649Srmacklem		return;
323193649Srmacklem	}
324184588Sdfr	xprt_unregister_locked(xprt);
325184588Sdfr	mtx_unlock(&pool->sp_lock);
326184588Sdfr
327184588Sdfr	SVC_RELEASE(xprt);
328177633Sdfr}
329177633Sdfr
330261048Smav/*
331261048Smav * Attempt to assign a service thread to this transport.
332261048Smav */
333261048Smavstatic int
334184588Sdfrxprt_assignthread(SVCXPRT *xprt)
335184588Sdfr{
336184588Sdfr	SVCPOOL *pool = xprt->xp_pool;
337184588Sdfr	SVCTHREAD *st;
338184588Sdfr
339261048Smav	mtx_assert(&pool->sp_lock, MA_OWNED);
340261048Smav	st = LIST_FIRST(&pool->sp_idlethreads);
341184588Sdfr	if (st) {
342261048Smav		LIST_REMOVE(st, st_ilink);
343261048Smav		st->st_idle = FALSE;
344184588Sdfr		SVC_ACQUIRE(xprt);
345184588Sdfr		xprt->xp_thread = st;
346184588Sdfr		st->st_xprt = xprt;
347184588Sdfr		cv_signal(&st->st_cond);
348261048Smav		return (TRUE);
349184588Sdfr	} else {
350184588Sdfr		/*
351184588Sdfr		 * See if we can create a new thread. The
352184588Sdfr		 * actual thread creation happens in
353184588Sdfr		 * svc_run_internal because our locking state
354184588Sdfr		 * is poorly defined (we are typically called
355184588Sdfr		 * from a socket upcall). Don't create more
356184588Sdfr		 * than one thread per second.
357184588Sdfr		 */
358184588Sdfr		if (pool->sp_state == SVCPOOL_ACTIVE
359184588Sdfr		    && pool->sp_lastcreatetime < time_uptime
360184588Sdfr		    && pool->sp_threadcount < pool->sp_maxthreads) {
361184588Sdfr			pool->sp_state = SVCPOOL_THREADWANTED;
362184588Sdfr		}
363184588Sdfr	}
364261048Smav	return (FALSE);
365184588Sdfr}
366184588Sdfr
367177633Sdfrvoid
368177633Sdfrxprt_active(SVCXPRT *xprt)
369177633Sdfr{
370177633Sdfr	SVCPOOL *pool = xprt->xp_pool;
371177633Sdfr
372193436Srmacklem	mtx_lock(&pool->sp_lock);
373193436Srmacklem
374184588Sdfr	if (!xprt->xp_registered) {
375184588Sdfr		/*
376184588Sdfr		 * Race with xprt_unregister - we lose.
377184588Sdfr		 */
378193436Srmacklem		mtx_unlock(&pool->sp_lock);
379184588Sdfr		return;
380184588Sdfr	}
381184588Sdfr
382177633Sdfr	if (!xprt->xp_active) {
383177633Sdfr		xprt->xp_active = TRUE;
384261048Smav		if (xprt->xp_thread == NULL) {
385261054Smav			if (!svc_request_space_available(pool) ||
386261054Smav			    !xprt_assignthread(xprt))
387261048Smav				TAILQ_INSERT_TAIL(&pool->sp_active, xprt,
388261048Smav				    xp_alink);
389261048Smav		}
390177633Sdfr	}
391177633Sdfr
392177633Sdfr	mtx_unlock(&pool->sp_lock);
393177633Sdfr}
394177633Sdfr
395177633Sdfrvoid
396184588Sdfrxprt_inactive_locked(SVCXPRT *xprt)
397177633Sdfr{
398177633Sdfr	SVCPOOL *pool = xprt->xp_pool;
399177633Sdfr
400261048Smav	mtx_assert(&pool->sp_lock, MA_OWNED);
401177633Sdfr	if (xprt->xp_active) {
402261048Smav		if (xprt->xp_thread == NULL)
403261048Smav			TAILQ_REMOVE(&pool->sp_active, xprt, xp_alink);
404177633Sdfr		xprt->xp_active = FALSE;
405177633Sdfr	}
406184588Sdfr}
407177633Sdfr
408184588Sdfrvoid
409184588Sdfrxprt_inactive(SVCXPRT *xprt)
410184588Sdfr{
411184588Sdfr	SVCPOOL *pool = xprt->xp_pool;
412184588Sdfr
413184588Sdfr	mtx_lock(&pool->sp_lock);
414184588Sdfr	xprt_inactive_locked(xprt);
415177633Sdfr	mtx_unlock(&pool->sp_lock);
416177633Sdfr}
417177633Sdfr
418177633Sdfr/*
419261053Smav * Variant of xprt_inactive() for use only when sure that port is
420261053Smav * assigned to thread. For example, withing receive handlers.
421261053Smav */
422261053Smavvoid
423261053Smavxprt_inactive_self(SVCXPRT *xprt)
424261053Smav{
425261053Smav
426261053Smav	KASSERT(xprt->xp_thread != NULL,
427261053Smav	    ("xprt_inactive_self(%p) with NULL xp_thread", xprt));
428261053Smav	xprt->xp_active = FALSE;
429261053Smav}
430261053Smav
431261053Smav/*
432177633Sdfr * Add a service program to the callout list.
433177633Sdfr * The dispatch routine will be called when a rpc request for this
434177633Sdfr * program number comes in.
435177633Sdfr */
436177633Sdfrbool_t
437177633Sdfrsvc_reg(SVCXPRT *xprt, const rpcprog_t prog, const rpcvers_t vers,
438177633Sdfr    void (*dispatch)(struct svc_req *, SVCXPRT *),
439177633Sdfr    const struct netconfig *nconf)
440177633Sdfr{
441177633Sdfr	SVCPOOL *pool = xprt->xp_pool;
442177633Sdfr	struct svc_callout *s;
443177633Sdfr	char *netid = NULL;
444177633Sdfr	int flag = 0;
445177633Sdfr
446177633Sdfr/* VARIABLES PROTECTED BY svc_lock: s, svc_head */
447177633Sdfr
448177633Sdfr	if (xprt->xp_netid) {
449177633Sdfr		netid = strdup(xprt->xp_netid, M_RPC);
450177633Sdfr		flag = 1;
451177633Sdfr	} else if (nconf && nconf->nc_netid) {
452177633Sdfr		netid = strdup(nconf->nc_netid, M_RPC);
453177633Sdfr		flag = 1;
454177633Sdfr	} /* must have been created with svc_raw_create */
455177633Sdfr	if ((netid == NULL) && (flag == 1)) {
456177633Sdfr		return (FALSE);
457177633Sdfr	}
458177633Sdfr
459177633Sdfr	mtx_lock(&pool->sp_lock);
460177633Sdfr	if ((s = svc_find(pool, prog, vers, netid)) != NULL) {
461177633Sdfr		if (netid)
462177633Sdfr			free(netid, M_RPC);
463177633Sdfr		if (s->sc_dispatch == dispatch)
464177633Sdfr			goto rpcb_it; /* he is registering another xptr */
465177633Sdfr		mtx_unlock(&pool->sp_lock);
466177633Sdfr		return (FALSE);
467177633Sdfr	}
468177633Sdfr	s = malloc(sizeof (struct svc_callout), M_RPC, M_NOWAIT);
469177633Sdfr	if (s == NULL) {
470177633Sdfr		if (netid)
471177633Sdfr			free(netid, M_RPC);
472177633Sdfr		mtx_unlock(&pool->sp_lock);
473177633Sdfr		return (FALSE);
474177633Sdfr	}
475177633Sdfr
476177633Sdfr	s->sc_prog = prog;
477177633Sdfr	s->sc_vers = vers;
478177633Sdfr	s->sc_dispatch = dispatch;
479177633Sdfr	s->sc_netid = netid;
480177633Sdfr	TAILQ_INSERT_TAIL(&pool->sp_callouts, s, sc_link);
481177633Sdfr
482177633Sdfr	if ((xprt->xp_netid == NULL) && (flag == 1) && netid)
483177633Sdfr		((SVCXPRT *) xprt)->xp_netid = strdup(netid, M_RPC);
484177633Sdfr
485177633Sdfrrpcb_it:
486177633Sdfr	mtx_unlock(&pool->sp_lock);
487177633Sdfr	/* now register the information with the local binder service */
488177633Sdfr	if (nconf) {
489177633Sdfr		bool_t dummy;
490177633Sdfr		struct netconfig tnc;
491184588Sdfr		struct netbuf nb;
492177633Sdfr		tnc = *nconf;
493184588Sdfr		nb.buf = &xprt->xp_ltaddr;
494184588Sdfr		nb.len = xprt->xp_ltaddr.ss_len;
495184588Sdfr		dummy = rpcb_set(prog, vers, &tnc, &nb);
496177633Sdfr		return (dummy);
497177633Sdfr	}
498177633Sdfr	return (TRUE);
499177633Sdfr}
500177633Sdfr
501177633Sdfr/*
502177633Sdfr * Remove a service program from the callout list.
503177633Sdfr */
504177633Sdfrvoid
505177633Sdfrsvc_unreg(SVCPOOL *pool, const rpcprog_t prog, const rpcvers_t vers)
506177633Sdfr{
507177633Sdfr	struct svc_callout *s;
508177633Sdfr
509177633Sdfr	/* unregister the information anyway */
510177633Sdfr	(void) rpcb_unset(prog, vers, NULL);
511177633Sdfr	mtx_lock(&pool->sp_lock);
512177633Sdfr	while ((s = svc_find(pool, prog, vers, NULL)) != NULL) {
513177633Sdfr		TAILQ_REMOVE(&pool->sp_callouts, s, sc_link);
514177633Sdfr		if (s->sc_netid)
515177633Sdfr			mem_free(s->sc_netid, sizeof (s->sc_netid) + 1);
516177633Sdfr		mem_free(s, sizeof (struct svc_callout));
517177633Sdfr	}
518177633Sdfr	mtx_unlock(&pool->sp_lock);
519177633Sdfr}
520177633Sdfr
521261055Smav/*
522261055Smav * Add a service connection loss program to the callout list.
523261055Smav * The dispatch routine will be called when some port in ths pool die.
524261055Smav */
525261055Smavbool_t
526261055Smavsvc_loss_reg(SVCXPRT *xprt, void (*dispatch)(SVCXPRT *))
527261055Smav{
528261055Smav	SVCPOOL *pool = xprt->xp_pool;
529261055Smav	struct svc_loss_callout *s;
530261055Smav
531261055Smav	mtx_lock(&pool->sp_lock);
532261055Smav	TAILQ_FOREACH(s, &pool->sp_lcallouts, slc_link) {
533261055Smav		if (s->slc_dispatch == dispatch)
534261055Smav			break;
535261055Smav	}
536261055Smav	if (s != NULL) {
537261055Smav		mtx_unlock(&pool->sp_lock);
538261055Smav		return (TRUE);
539261055Smav	}
540261055Smav	s = malloc(sizeof (struct svc_callout), M_RPC, M_NOWAIT);
541261055Smav	if (s == NULL) {
542261055Smav		mtx_unlock(&pool->sp_lock);
543261055Smav		return (FALSE);
544261055Smav	}
545261055Smav	s->slc_dispatch = dispatch;
546261055Smav	TAILQ_INSERT_TAIL(&pool->sp_lcallouts, s, slc_link);
547261055Smav	mtx_unlock(&pool->sp_lock);
548261055Smav	return (TRUE);
549261055Smav}
550261055Smav
551261055Smav/*
552261055Smav * Remove a service connection loss program from the callout list.
553261055Smav */
554261055Smavvoid
555261055Smavsvc_loss_unreg(SVCPOOL *pool, void (*dispatch)(SVCXPRT *))
556261055Smav{
557261055Smav	struct svc_loss_callout *s;
558261055Smav
559261055Smav	mtx_lock(&pool->sp_lock);
560261055Smav	TAILQ_FOREACH(s, &pool->sp_lcallouts, slc_link) {
561261055Smav		if (s->slc_dispatch == dispatch) {
562261055Smav			TAILQ_REMOVE(&pool->sp_lcallouts, s, slc_link);
563261055Smav			free(s, M_RPC);
564261055Smav			break;
565261055Smav		}
566261055Smav	}
567261055Smav	mtx_unlock(&pool->sp_lock);
568261055Smav}
569261055Smav
570177633Sdfr/* ********************** CALLOUT list related stuff ************* */
571177633Sdfr
572177633Sdfr/*
573177633Sdfr * Search the callout list for a program number, return the callout
574177633Sdfr * struct.
575177633Sdfr */
576177633Sdfrstatic struct svc_callout *
577177633Sdfrsvc_find(SVCPOOL *pool, rpcprog_t prog, rpcvers_t vers, char *netid)
578177633Sdfr{
579177633Sdfr	struct svc_callout *s;
580177633Sdfr
581177633Sdfr	mtx_assert(&pool->sp_lock, MA_OWNED);
582177633Sdfr	TAILQ_FOREACH(s, &pool->sp_callouts, sc_link) {
583177633Sdfr		if (s->sc_prog == prog && s->sc_vers == vers
584177633Sdfr		    && (netid == NULL || s->sc_netid == NULL ||
585177633Sdfr			strcmp(netid, s->sc_netid) == 0))
586177633Sdfr			break;
587177633Sdfr	}
588177633Sdfr
589177633Sdfr	return (s);
590177633Sdfr}
591177633Sdfr
592177633Sdfr/* ******************* REPLY GENERATION ROUTINES  ************ */
593177633Sdfr
594184588Sdfrstatic bool_t
595184588Sdfrsvc_sendreply_common(struct svc_req *rqstp, struct rpc_msg *rply,
596184588Sdfr    struct mbuf *body)
597184588Sdfr{
598184588Sdfr	SVCXPRT *xprt = rqstp->rq_xprt;
599184588Sdfr	bool_t ok;
600184588Sdfr
601184588Sdfr	if (rqstp->rq_args) {
602184588Sdfr		m_freem(rqstp->rq_args);
603184588Sdfr		rqstp->rq_args = NULL;
604184588Sdfr	}
605184588Sdfr
606184588Sdfr	if (xprt->xp_pool->sp_rcache)
607184588Sdfr		replay_setreply(xprt->xp_pool->sp_rcache,
608184588Sdfr		    rply, svc_getrpccaller(rqstp), body);
609184588Sdfr
610184588Sdfr	if (!SVCAUTH_WRAP(&rqstp->rq_auth, &body))
611184588Sdfr		return (FALSE);
612184588Sdfr
613261055Smav	ok = SVC_REPLY(xprt, rply, rqstp->rq_addr, body, &rqstp->rq_reply_seq);
614184588Sdfr	if (rqstp->rq_addr) {
615184588Sdfr		free(rqstp->rq_addr, M_SONAME);
616184588Sdfr		rqstp->rq_addr = NULL;
617184588Sdfr	}
618184588Sdfr
619184588Sdfr	return (ok);
620184588Sdfr}
621184588Sdfr
622177633Sdfr/*
623177633Sdfr * Send a reply to an rpc request
624177633Sdfr */
625177633Sdfrbool_t
626184588Sdfrsvc_sendreply(struct svc_req *rqstp, xdrproc_t xdr_results, void * xdr_location)
627177633Sdfr{
628177633Sdfr	struct rpc_msg rply;
629184588Sdfr	struct mbuf *m;
630184588Sdfr	XDR xdrs;
631184588Sdfr	bool_t ok;
632177633Sdfr
633184588Sdfr	rply.rm_xid = rqstp->rq_xid;
634177633Sdfr	rply.rm_direction = REPLY;
635177633Sdfr	rply.rm_reply.rp_stat = MSG_ACCEPTED;
636184588Sdfr	rply.acpted_rply.ar_verf = rqstp->rq_verf;
637177633Sdfr	rply.acpted_rply.ar_stat = SUCCESS;
638184588Sdfr	rply.acpted_rply.ar_results.where = NULL;
639184588Sdfr	rply.acpted_rply.ar_results.proc = (xdrproc_t) xdr_void;
640177633Sdfr
641248195Sglebius	m = m_getcl(M_WAITOK, MT_DATA, 0);
642184588Sdfr	xdrmbuf_create(&xdrs, m, XDR_ENCODE);
643184588Sdfr	ok = xdr_results(&xdrs, xdr_location);
644184588Sdfr	XDR_DESTROY(&xdrs);
645184588Sdfr
646184588Sdfr	if (ok) {
647184588Sdfr		return (svc_sendreply_common(rqstp, &rply, m));
648184588Sdfr	} else {
649184588Sdfr		m_freem(m);
650184588Sdfr		return (FALSE);
651184588Sdfr	}
652177633Sdfr}
653177633Sdfr
654184588Sdfrbool_t
655184588Sdfrsvc_sendreply_mbuf(struct svc_req *rqstp, struct mbuf *m)
656184588Sdfr{
657184588Sdfr	struct rpc_msg rply;
658184588Sdfr
659184588Sdfr	rply.rm_xid = rqstp->rq_xid;
660184588Sdfr	rply.rm_direction = REPLY;
661184588Sdfr	rply.rm_reply.rp_stat = MSG_ACCEPTED;
662184588Sdfr	rply.acpted_rply.ar_verf = rqstp->rq_verf;
663184588Sdfr	rply.acpted_rply.ar_stat = SUCCESS;
664184588Sdfr	rply.acpted_rply.ar_results.where = NULL;
665184588Sdfr	rply.acpted_rply.ar_results.proc = (xdrproc_t) xdr_void;
666184588Sdfr
667184588Sdfr	return (svc_sendreply_common(rqstp, &rply, m));
668184588Sdfr}
669184588Sdfr
670177633Sdfr/*
671177633Sdfr * No procedure error reply
672177633Sdfr */
673177633Sdfrvoid
674184588Sdfrsvcerr_noproc(struct svc_req *rqstp)
675177633Sdfr{
676184588Sdfr	SVCXPRT *xprt = rqstp->rq_xprt;
677177633Sdfr	struct rpc_msg rply;
678177633Sdfr
679184588Sdfr	rply.rm_xid = rqstp->rq_xid;
680177633Sdfr	rply.rm_direction = REPLY;
681177633Sdfr	rply.rm_reply.rp_stat = MSG_ACCEPTED;
682184588Sdfr	rply.acpted_rply.ar_verf = rqstp->rq_verf;
683177633Sdfr	rply.acpted_rply.ar_stat = PROC_UNAVAIL;
684177633Sdfr
685184588Sdfr	if (xprt->xp_pool->sp_rcache)
686184588Sdfr		replay_setreply(xprt->xp_pool->sp_rcache,
687184588Sdfr		    &rply, svc_getrpccaller(rqstp), NULL);
688184588Sdfr
689184588Sdfr	svc_sendreply_common(rqstp, &rply, NULL);
690177633Sdfr}
691177633Sdfr
692177633Sdfr/*
693177633Sdfr * Can't decode args error reply
694177633Sdfr */
695177633Sdfrvoid
696184588Sdfrsvcerr_decode(struct svc_req *rqstp)
697177633Sdfr{
698184588Sdfr	SVCXPRT *xprt = rqstp->rq_xprt;
699177633Sdfr	struct rpc_msg rply;
700177633Sdfr
701184588Sdfr	rply.rm_xid = rqstp->rq_xid;
702177633Sdfr	rply.rm_direction = REPLY;
703177633Sdfr	rply.rm_reply.rp_stat = MSG_ACCEPTED;
704184588Sdfr	rply.acpted_rply.ar_verf = rqstp->rq_verf;
705177633Sdfr	rply.acpted_rply.ar_stat = GARBAGE_ARGS;
706177633Sdfr
707184588Sdfr	if (xprt->xp_pool->sp_rcache)
708184588Sdfr		replay_setreply(xprt->xp_pool->sp_rcache,
709184588Sdfr		    &rply, (struct sockaddr *) &xprt->xp_rtaddr, NULL);
710184588Sdfr
711184588Sdfr	svc_sendreply_common(rqstp, &rply, NULL);
712177633Sdfr}
713177633Sdfr
714177633Sdfr/*
715177633Sdfr * Some system error
716177633Sdfr */
717177633Sdfrvoid
718184588Sdfrsvcerr_systemerr(struct svc_req *rqstp)
719177633Sdfr{
720184588Sdfr	SVCXPRT *xprt = rqstp->rq_xprt;
721177633Sdfr	struct rpc_msg rply;
722177633Sdfr
723184588Sdfr	rply.rm_xid = rqstp->rq_xid;
724177633Sdfr	rply.rm_direction = REPLY;
725177633Sdfr	rply.rm_reply.rp_stat = MSG_ACCEPTED;
726184588Sdfr	rply.acpted_rply.ar_verf = rqstp->rq_verf;
727177633Sdfr	rply.acpted_rply.ar_stat = SYSTEM_ERR;
728177633Sdfr
729184588Sdfr	if (xprt->xp_pool->sp_rcache)
730184588Sdfr		replay_setreply(xprt->xp_pool->sp_rcache,
731184588Sdfr		    &rply, svc_getrpccaller(rqstp), NULL);
732184588Sdfr
733184588Sdfr	svc_sendreply_common(rqstp, &rply, NULL);
734177633Sdfr}
735177633Sdfr
736177633Sdfr/*
737177633Sdfr * Authentication error reply
738177633Sdfr */
739177633Sdfrvoid
740184588Sdfrsvcerr_auth(struct svc_req *rqstp, enum auth_stat why)
741177633Sdfr{
742184588Sdfr	SVCXPRT *xprt = rqstp->rq_xprt;
743177633Sdfr	struct rpc_msg rply;
744177633Sdfr
745184588Sdfr	rply.rm_xid = rqstp->rq_xid;
746177633Sdfr	rply.rm_direction = REPLY;
747177633Sdfr	rply.rm_reply.rp_stat = MSG_DENIED;
748177633Sdfr	rply.rjcted_rply.rj_stat = AUTH_ERROR;
749177633Sdfr	rply.rjcted_rply.rj_why = why;
750177633Sdfr
751184588Sdfr	if (xprt->xp_pool->sp_rcache)
752184588Sdfr		replay_setreply(xprt->xp_pool->sp_rcache,
753184588Sdfr		    &rply, svc_getrpccaller(rqstp), NULL);
754184588Sdfr
755184588Sdfr	svc_sendreply_common(rqstp, &rply, NULL);
756177633Sdfr}
757177633Sdfr
758177633Sdfr/*
759177633Sdfr * Auth too weak error reply
760177633Sdfr */
761177633Sdfrvoid
762184588Sdfrsvcerr_weakauth(struct svc_req *rqstp)
763177633Sdfr{
764177633Sdfr
765184588Sdfr	svcerr_auth(rqstp, AUTH_TOOWEAK);
766177633Sdfr}
767177633Sdfr
768177633Sdfr/*
769177633Sdfr * Program unavailable error reply
770177633Sdfr */
771177633Sdfrvoid
772184588Sdfrsvcerr_noprog(struct svc_req *rqstp)
773177633Sdfr{
774184588Sdfr	SVCXPRT *xprt = rqstp->rq_xprt;
775177633Sdfr	struct rpc_msg rply;
776177633Sdfr
777184588Sdfr	rply.rm_xid = rqstp->rq_xid;
778177633Sdfr	rply.rm_direction = REPLY;
779177633Sdfr	rply.rm_reply.rp_stat = MSG_ACCEPTED;
780184588Sdfr	rply.acpted_rply.ar_verf = rqstp->rq_verf;
781177633Sdfr	rply.acpted_rply.ar_stat = PROG_UNAVAIL;
782177633Sdfr
783184588Sdfr	if (xprt->xp_pool->sp_rcache)
784184588Sdfr		replay_setreply(xprt->xp_pool->sp_rcache,
785184588Sdfr		    &rply, svc_getrpccaller(rqstp), NULL);
786184588Sdfr
787184588Sdfr	svc_sendreply_common(rqstp, &rply, NULL);
788177633Sdfr}
789177633Sdfr
790177633Sdfr/*
791177633Sdfr * Program version mismatch error reply
792177633Sdfr */
793177633Sdfrvoid
794184588Sdfrsvcerr_progvers(struct svc_req *rqstp, rpcvers_t low_vers, rpcvers_t high_vers)
795177633Sdfr{
796184588Sdfr	SVCXPRT *xprt = rqstp->rq_xprt;
797177633Sdfr	struct rpc_msg rply;
798177633Sdfr
799184588Sdfr	rply.rm_xid = rqstp->rq_xid;
800177633Sdfr	rply.rm_direction = REPLY;
801177633Sdfr	rply.rm_reply.rp_stat = MSG_ACCEPTED;
802184588Sdfr	rply.acpted_rply.ar_verf = rqstp->rq_verf;
803177633Sdfr	rply.acpted_rply.ar_stat = PROG_MISMATCH;
804177633Sdfr	rply.acpted_rply.ar_vers.low = (uint32_t)low_vers;
805177633Sdfr	rply.acpted_rply.ar_vers.high = (uint32_t)high_vers;
806177633Sdfr
807184588Sdfr	if (xprt->xp_pool->sp_rcache)
808184588Sdfr		replay_setreply(xprt->xp_pool->sp_rcache,
809184588Sdfr		    &rply, svc_getrpccaller(rqstp), NULL);
810184588Sdfr
811184588Sdfr	svc_sendreply_common(rqstp, &rply, NULL);
812177633Sdfr}
813177633Sdfr
814184588Sdfr/*
815184588Sdfr * Allocate a new server transport structure. All fields are
816184588Sdfr * initialized to zero and xp_p3 is initialized to point at an
817184588Sdfr * extension structure to hold various flags and authentication
818184588Sdfr * parameters.
819184588Sdfr */
820184588SdfrSVCXPRT *
821184588Sdfrsvc_xprt_alloc()
822184588Sdfr{
823184588Sdfr	SVCXPRT *xprt;
824184588Sdfr	SVCXPRT_EXT *ext;
825184588Sdfr
826184588Sdfr	xprt = mem_alloc(sizeof(SVCXPRT));
827184588Sdfr	memset(xprt, 0, sizeof(SVCXPRT));
828184588Sdfr	ext = mem_alloc(sizeof(SVCXPRT_EXT));
829184588Sdfr	memset(ext, 0, sizeof(SVCXPRT_EXT));
830184588Sdfr	xprt->xp_p3 = ext;
831184588Sdfr	refcount_init(&xprt->xp_refs, 1);
832184588Sdfr
833184588Sdfr	return (xprt);
834184588Sdfr}
835184588Sdfr
836184588Sdfr/*
837184588Sdfr * Free a server transport structure.
838184588Sdfr */
839184588Sdfrvoid
840184588Sdfrsvc_xprt_free(xprt)
841184588Sdfr	SVCXPRT *xprt;
842184588Sdfr{
843184588Sdfr
844184588Sdfr	mem_free(xprt->xp_p3, sizeof(SVCXPRT_EXT));
845184588Sdfr	mem_free(xprt, sizeof(SVCXPRT));
846184588Sdfr}
847184588Sdfr
848177633Sdfr/* ******************* SERVER INPUT STUFF ******************* */
849177633Sdfr
850177633Sdfr/*
851184588Sdfr * Read RPC requests from a transport and queue them to be
852184588Sdfr * executed. We handle authentication and replay cache replies here.
853184588Sdfr * Actually dispatching the RPC is deferred till svc_executereq.
854177633Sdfr */
855184588Sdfrstatic enum xprt_stat
856184588Sdfrsvc_getreq(SVCXPRT *xprt, struct svc_req **rqstp_ret)
857177633Sdfr{
858177633Sdfr	SVCPOOL *pool = xprt->xp_pool;
859184588Sdfr	struct svc_req *r;
860177633Sdfr	struct rpc_msg msg;
861184588Sdfr	struct mbuf *args;
862261055Smav	struct svc_loss_callout *s;
863177633Sdfr	enum xprt_stat stat;
864177633Sdfr
865177633Sdfr	/* now receive msgs from xprtprt (support batch calls) */
866184588Sdfr	r = malloc(sizeof(*r), M_RPC, M_WAITOK|M_ZERO);
867177633Sdfr
868184588Sdfr	msg.rm_call.cb_cred.oa_base = r->rq_credarea;
869184588Sdfr	msg.rm_call.cb_verf.oa_base = &r->rq_credarea[MAX_AUTH_BYTES];
870184588Sdfr	r->rq_clntcred = &r->rq_credarea[2*MAX_AUTH_BYTES];
871184588Sdfr	if (SVC_RECV(xprt, &msg, &r->rq_addr, &args)) {
872184588Sdfr		enum auth_stat why;
873177633Sdfr
874184588Sdfr		/*
875184588Sdfr		 * Handle replays and authenticate before queuing the
876184588Sdfr		 * request to be executed.
877184588Sdfr		 */
878184588Sdfr		SVC_ACQUIRE(xprt);
879184588Sdfr		r->rq_xprt = xprt;
880184588Sdfr		if (pool->sp_rcache) {
881184588Sdfr			struct rpc_msg repmsg;
882184588Sdfr			struct mbuf *repbody;
883184588Sdfr			enum replay_state rs;
884184588Sdfr			rs = replay_find(pool->sp_rcache, &msg,
885184588Sdfr			    svc_getrpccaller(r), &repmsg, &repbody);
886184588Sdfr			switch (rs) {
887184588Sdfr			case RS_NEW:
888184588Sdfr				break;
889184588Sdfr			case RS_DONE:
890184588Sdfr				SVC_REPLY(xprt, &repmsg, r->rq_addr,
891261055Smav				    repbody, &r->rq_reply_seq);
892184588Sdfr				if (r->rq_addr) {
893184588Sdfr					free(r->rq_addr, M_SONAME);
894184588Sdfr					r->rq_addr = NULL;
895184588Sdfr				}
896205562Srmacklem				m_freem(args);
897177633Sdfr				goto call_done;
898184588Sdfr
899184588Sdfr			default:
900205562Srmacklem				m_freem(args);
901184588Sdfr				goto call_done;
902177633Sdfr			}
903184588Sdfr		}
904184588Sdfr
905184588Sdfr		r->rq_xid = msg.rm_xid;
906184588Sdfr		r->rq_prog = msg.rm_call.cb_prog;
907184588Sdfr		r->rq_vers = msg.rm_call.cb_vers;
908184588Sdfr		r->rq_proc = msg.rm_call.cb_proc;
909184588Sdfr		r->rq_size = sizeof(*r) + m_length(args, NULL);
910184588Sdfr		r->rq_args = args;
911184588Sdfr		if ((why = _authenticate(r, &msg)) != AUTH_OK) {
912177633Sdfr			/*
913184588Sdfr			 * RPCSEC_GSS uses this return code
914184588Sdfr			 * for requests that form part of its
915184588Sdfr			 * context establishment protocol and
916184588Sdfr			 * should not be dispatched to the
917184588Sdfr			 * application.
918177633Sdfr			 */
919184588Sdfr			if (why != RPCSEC_GSS_NODISPATCH)
920184588Sdfr				svcerr_auth(r, why);
921184588Sdfr			goto call_done;
922177633Sdfr		}
923184588Sdfr
924184588Sdfr		if (!SVCAUTH_UNWRAP(&r->rq_auth, &r->rq_args)) {
925184588Sdfr			svcerr_decode(r);
926184588Sdfr			goto call_done;
927184588Sdfr		}
928184588Sdfr
929177633Sdfr		/*
930184588Sdfr		 * Everything checks out, return request to caller.
931177633Sdfr		 */
932184588Sdfr		*rqstp_ret = r;
933184588Sdfr		r = NULL;
934184588Sdfr	}
935177633Sdfrcall_done:
936184588Sdfr	if (r) {
937184588Sdfr		svc_freereq(r);
938184588Sdfr		r = NULL;
939184588Sdfr	}
940184588Sdfr	if ((stat = SVC_STAT(xprt)) == XPRT_DIED) {
941261055Smav		TAILQ_FOREACH(s, &pool->sp_lcallouts, slc_link)
942261055Smav			(*s->slc_dispatch)(xprt);
943184588Sdfr		xprt_unregister(xprt);
944184588Sdfr	}
945184588Sdfr
946184588Sdfr	return (stat);
947184588Sdfr}
948184588Sdfr
949184588Sdfrstatic void
950184588Sdfrsvc_executereq(struct svc_req *rqstp)
951184588Sdfr{
952184588Sdfr	SVCXPRT *xprt = rqstp->rq_xprt;
953184588Sdfr	SVCPOOL *pool = xprt->xp_pool;
954184588Sdfr	int prog_found;
955184588Sdfr	rpcvers_t low_vers;
956184588Sdfr	rpcvers_t high_vers;
957184588Sdfr	struct svc_callout *s;
958184588Sdfr
959184588Sdfr	/* now match message with a registered service*/
960184588Sdfr	prog_found = FALSE;
961184588Sdfr	low_vers = (rpcvers_t) -1L;
962184588Sdfr	high_vers = (rpcvers_t) 0L;
963184588Sdfr	TAILQ_FOREACH(s, &pool->sp_callouts, sc_link) {
964184588Sdfr		if (s->sc_prog == rqstp->rq_prog) {
965184588Sdfr			if (s->sc_vers == rqstp->rq_vers) {
966184588Sdfr				/*
967184588Sdfr				 * We hand ownership of r to the
968184588Sdfr				 * dispatch method - they must call
969184588Sdfr				 * svc_freereq.
970184588Sdfr				 */
971184588Sdfr				(*s->sc_dispatch)(rqstp, xprt);
972184588Sdfr				return;
973184588Sdfr			}  /* found correct version */
974184588Sdfr			prog_found = TRUE;
975184588Sdfr			if (s->sc_vers < low_vers)
976184588Sdfr				low_vers = s->sc_vers;
977184588Sdfr			if (s->sc_vers > high_vers)
978184588Sdfr				high_vers = s->sc_vers;
979184588Sdfr		}   /* found correct program */
980184588Sdfr	}
981184588Sdfr
982184588Sdfr	/*
983184588Sdfr	 * if we got here, the program or version
984184588Sdfr	 * is not served ...
985184588Sdfr	 */
986184588Sdfr	if (prog_found)
987184588Sdfr		svcerr_progvers(rqstp, low_vers, high_vers);
988184588Sdfr	else
989184588Sdfr		svcerr_noprog(rqstp);
990184588Sdfr
991184588Sdfr	svc_freereq(rqstp);
992184588Sdfr}
993184588Sdfr
994184588Sdfrstatic void
995184588Sdfrsvc_checkidle(SVCPOOL *pool)
996184588Sdfr{
997184588Sdfr	SVCXPRT *xprt, *nxprt;
998184588Sdfr	time_t timo;
999184588Sdfr	struct svcxprt_list cleanup;
1000184588Sdfr
1001184588Sdfr	TAILQ_INIT(&cleanup);
1002184588Sdfr	TAILQ_FOREACH_SAFE(xprt, &pool->sp_xlist, xp_link, nxprt) {
1003184588Sdfr		/*
1004184588Sdfr		 * Only some transports have idle timers. Don't time
1005184588Sdfr		 * something out which is just waking up.
1006184588Sdfr		 */
1007184588Sdfr		if (!xprt->xp_idletimeout || xprt->xp_thread)
1008184588Sdfr			continue;
1009184588Sdfr
1010184588Sdfr		timo = xprt->xp_lastactive + xprt->xp_idletimeout;
1011184588Sdfr		if (time_uptime > timo) {
1012184588Sdfr			xprt_unregister_locked(xprt);
1013184588Sdfr			TAILQ_INSERT_TAIL(&cleanup, xprt, xp_link);
1014177633Sdfr		}
1015184588Sdfr	}
1016184588Sdfr
1017184588Sdfr	mtx_unlock(&pool->sp_lock);
1018184588Sdfr	TAILQ_FOREACH_SAFE(xprt, &cleanup, xp_link, nxprt) {
1019184588Sdfr		SVC_RELEASE(xprt);
1020184588Sdfr	}
1021184588Sdfr	mtx_lock(&pool->sp_lock);
1022184588Sdfr
1023177633Sdfr}
1024177633Sdfr
1025184588Sdfrstatic void
1026184588Sdfrsvc_assign_waiting_sockets(SVCPOOL *pool)
1027177633Sdfr{
1028177633Sdfr	SVCXPRT *xprt;
1029184588Sdfr
1030261054Smav	mtx_lock(&pool->sp_lock);
1031261048Smav	while ((xprt = TAILQ_FIRST(&pool->sp_active)) != NULL) {
1032261048Smav		if (xprt_assignthread(xprt))
1033261048Smav			TAILQ_REMOVE(&pool->sp_active, xprt, xp_alink);
1034261048Smav		else
1035261048Smav			break;
1036184588Sdfr	}
1037261054Smav	mtx_unlock(&pool->sp_lock);
1038184588Sdfr}
1039184588Sdfr
1040261054Smavstatic void
1041261054Smavsvc_change_space_used(SVCPOOL *pool, int delta)
1042184588Sdfr{
1043261054Smav	unsigned int value;
1044184588Sdfr
1045261054Smav	value = atomic_fetchadd_int(&pool->sp_space_used, delta) + delta;
1046261054Smav	if (delta > 0) {
1047261054Smav		if (value >= pool->sp_space_high && !pool->sp_space_throttled) {
1048261054Smav			pool->sp_space_throttled = TRUE;
1049261054Smav			pool->sp_space_throttle_count++;
1050261054Smav		}
1051261054Smav		if (value > pool->sp_space_used_highest)
1052261054Smav			pool->sp_space_used_highest = value;
1053261054Smav	} else {
1054261054Smav		if (value < pool->sp_space_low && pool->sp_space_throttled) {
1055184588Sdfr			pool->sp_space_throttled = FALSE;
1056184588Sdfr			svc_assign_waiting_sockets(pool);
1057184588Sdfr		}
1058184588Sdfr	}
1059184588Sdfr}
1060184588Sdfr
1061261054Smavstatic bool_t
1062261054Smavsvc_request_space_available(SVCPOOL *pool)
1063261054Smav{
1064261054Smav
1065261054Smav	if (pool->sp_space_throttled)
1066261054Smav		return (FALSE);
1067261054Smav	return (TRUE);
1068261054Smav}
1069261054Smav
1070184588Sdfrstatic void
1071184588Sdfrsvc_run_internal(SVCPOOL *pool, bool_t ismaster)
1072184588Sdfr{
1073184588Sdfr	SVCTHREAD *st, *stpref;
1074184588Sdfr	SVCXPRT *xprt;
1075184588Sdfr	enum xprt_stat stat;
1076184588Sdfr	struct svc_req *rqstp;
1077261054Smav	size_t sz;
1078177633Sdfr	int error;
1079177633Sdfr
1080184588Sdfr	st = mem_alloc(sizeof(*st));
1081267740Smav	mtx_init(&st->st_lock, "st_lock", NULL, MTX_DEF);
1082261054Smav	st->st_pool = pool;
1083184588Sdfr	st->st_xprt = NULL;
1084184588Sdfr	STAILQ_INIT(&st->st_reqs);
1085184588Sdfr	cv_init(&st->st_cond, "rpcsvc");
1086184588Sdfr
1087177633Sdfr	mtx_lock(&pool->sp_lock);
1088184588Sdfr	LIST_INSERT_HEAD(&pool->sp_threads, st, st_link);
1089177633Sdfr
1090184588Sdfr	/*
1091184588Sdfr	 * If we are a new thread which was spawned to cope with
1092184588Sdfr	 * increased load, set the state back to SVCPOOL_ACTIVE.
1093184588Sdfr	 */
1094184588Sdfr	if (pool->sp_state == SVCPOOL_THREADSTARTING)
1095184588Sdfr		pool->sp_state = SVCPOOL_ACTIVE;
1096177633Sdfr
1097184588Sdfr	while (pool->sp_state != SVCPOOL_CLOSING) {
1098184588Sdfr		/*
1099261045Smav		 * Create new thread if requested.
1100261045Smav		 */
1101261045Smav		if (pool->sp_state == SVCPOOL_THREADWANTED) {
1102261045Smav			pool->sp_state = SVCPOOL_THREADSTARTING;
1103261045Smav			pool->sp_lastcreatetime = time_uptime;
1104261045Smav			mtx_unlock(&pool->sp_lock);
1105261045Smav			svc_new_thread(pool);
1106261045Smav			mtx_lock(&pool->sp_lock);
1107261045Smav			continue;
1108261045Smav		}
1109261045Smav
1110261045Smav		/*
1111184588Sdfr		 * Check for idle transports once per second.
1112184588Sdfr		 */
1113184588Sdfr		if (time_uptime > pool->sp_lastidlecheck) {
1114184588Sdfr			pool->sp_lastidlecheck = time_uptime;
1115184588Sdfr			svc_checkidle(pool);
1116184588Sdfr		}
1117184588Sdfr
1118184588Sdfr		xprt = st->st_xprt;
1119267740Smav		if (!xprt) {
1120184588Sdfr			/*
1121184588Sdfr			 * Enforce maxthreads count.
1122184588Sdfr			 */
1123184588Sdfr			if (pool->sp_threadcount > pool->sp_maxthreads)
1124177633Sdfr				break;
1125184588Sdfr
1126184588Sdfr			/*
1127184588Sdfr			 * Before sleeping, see if we can find an
1128184588Sdfr			 * active transport which isn't being serviced
1129184588Sdfr			 * by a thread.
1130184588Sdfr			 */
1131261048Smav			if (svc_request_space_available(pool) &&
1132261048Smav			    (xprt = TAILQ_FIRST(&pool->sp_active)) != NULL) {
1133261048Smav				TAILQ_REMOVE(&pool->sp_active, xprt, xp_alink);
1134261048Smav				SVC_ACQUIRE(xprt);
1135261048Smav				xprt->xp_thread = st;
1136261048Smav				st->st_xprt = xprt;
1137261048Smav				continue;
1138184588Sdfr			}
1139184588Sdfr
1140184588Sdfr			LIST_INSERT_HEAD(&pool->sp_idlethreads, st, st_ilink);
1141261048Smav			st->st_idle = TRUE;
1142261045Smav			if (ismaster || (!ismaster &&
1143261045Smav			    pool->sp_threadcount > pool->sp_minthreads))
1144261045Smav				error = cv_timedwait_sig(&st->st_cond,
1145261045Smav				    &pool->sp_lock, 5 * hz);
1146261045Smav			else
1147261045Smav				error = cv_wait_sig(&st->st_cond,
1148261045Smav				    &pool->sp_lock);
1149261048Smav			if (st->st_idle) {
1150261048Smav				LIST_REMOVE(st, st_ilink);
1151261048Smav				st->st_idle = FALSE;
1152261048Smav			}
1153184588Sdfr
1154184588Sdfr			/*
1155184588Sdfr			 * Reduce worker thread count when idle.
1156184588Sdfr			 */
1157184588Sdfr			if (error == EWOULDBLOCK) {
1158184588Sdfr				if (!ismaster
1159184588Sdfr				    && (pool->sp_threadcount
1160184588Sdfr					> pool->sp_minthreads)
1161267740Smav					&& !st->st_xprt)
1162184588Sdfr					break;
1163261045Smav			} else if (error) {
1164184588Sdfr				mtx_unlock(&pool->sp_lock);
1165261045Smav				svc_exit(pool);
1166184588Sdfr				mtx_lock(&pool->sp_lock);
1167261045Smav				break;
1168184588Sdfr			}
1169177633Sdfr			continue;
1170177633Sdfr		}
1171267740Smav		mtx_unlock(&pool->sp_lock);
1172177633Sdfr
1173267740Smav		/*
1174267740Smav		 * Drain the transport socket and queue up any RPCs.
1175267740Smav		 */
1176267740Smav		xprt->xp_lastactive = time_uptime;
1177267740Smav		do {
1178267740Smav			if (!svc_request_space_available(pool))
1179267740Smav				break;
1180267740Smav			rqstp = NULL;
1181267740Smav			stat = svc_getreq(xprt, &rqstp);
1182267740Smav			if (rqstp) {
1183267740Smav				svc_change_space_used(pool, rqstp->rq_size);
1184267740Smav				/*
1185267740Smav				 * See if the application has a preference
1186267740Smav				 * for some other thread.
1187267740Smav				 */
1188267740Smav				if (pool->sp_assign) {
1189267740Smav					stpref = pool->sp_assign(st, rqstp);
1190184588Sdfr					rqstp->rq_thread = stpref;
1191184588Sdfr					STAILQ_INSERT_TAIL(&stpref->st_reqs,
1192184588Sdfr					    rqstp, rq_link);
1193267740Smav					mtx_unlock(&stpref->st_lock);
1194267740Smav					if (stpref != st)
1195267740Smav						rqstp = NULL;
1196267740Smav				} else {
1197267740Smav					rqstp->rq_thread = st;
1198267740Smav					STAILQ_INSERT_TAIL(&st->st_reqs,
1199267740Smav					    rqstp, rq_link);
1200267740Smav				}
1201267740Smav			}
1202267740Smav		} while (rqstp == NULL && stat == XPRT_MOREREQS
1203267740Smav		    && pool->sp_state != SVCPOOL_CLOSING);
1204184588Sdfr
1205267740Smav		/*
1206267740Smav		 * Move this transport to the end of the active list to
1207267740Smav		 * ensure fairness when multiple transports are active.
1208267740Smav		 * If this was the last queued request, svc_getreq will end
1209267740Smav		 * up calling xprt_inactive to remove from the active list.
1210267740Smav		 */
1211267740Smav		mtx_lock(&pool->sp_lock);
1212267740Smav		xprt->xp_thread = NULL;
1213267740Smav		st->st_xprt = NULL;
1214267740Smav		if (xprt->xp_active) {
1215267740Smav			if (!svc_request_space_available(pool) ||
1216267740Smav			    !xprt_assignthread(xprt))
1217267740Smav				TAILQ_INSERT_TAIL(&pool->sp_active,
1218267740Smav				    xprt, xp_alink);
1219184588Sdfr		}
1220267740Smav		mtx_unlock(&pool->sp_lock);
1221267740Smav		SVC_RELEASE(xprt);
1222184588Sdfr
1223177633Sdfr		/*
1224184588Sdfr		 * Execute what we have queued.
1225177633Sdfr		 */
1226261054Smav		sz = 0;
1227267740Smav		mtx_lock(&st->st_lock);
1228267740Smav		while ((rqstp = STAILQ_FIRST(&st->st_reqs)) != NULL) {
1229267740Smav			STAILQ_REMOVE_HEAD(&st->st_reqs, rq_link);
1230267740Smav			mtx_unlock(&st->st_lock);
1231261054Smav			sz += rqstp->rq_size;
1232184588Sdfr			svc_executereq(rqstp);
1233267740Smav			mtx_lock(&st->st_lock);
1234184588Sdfr		}
1235267740Smav		mtx_unlock(&st->st_lock);
1236261054Smav		svc_change_space_used(pool, -sz);
1237261054Smav		mtx_lock(&pool->sp_lock);
1238184588Sdfr	}
1239177633Sdfr
1240184588Sdfr	if (st->st_xprt) {
1241184588Sdfr		xprt = st->st_xprt;
1242184588Sdfr		st->st_xprt = NULL;
1243184588Sdfr		SVC_RELEASE(xprt);
1244177633Sdfr	}
1245177633Sdfr
1246184588Sdfr	KASSERT(STAILQ_EMPTY(&st->st_reqs), ("stray reqs on exit"));
1247184588Sdfr	LIST_REMOVE(st, st_link);
1248184588Sdfr	pool->sp_threadcount--;
1249184588Sdfr
1250177633Sdfr	mtx_unlock(&pool->sp_lock);
1251184588Sdfr
1252267740Smav	mtx_destroy(&st->st_lock);
1253184588Sdfr	cv_destroy(&st->st_cond);
1254184588Sdfr	mem_free(st, sizeof(*st));
1255184588Sdfr
1256184588Sdfr	if (!ismaster)
1257184588Sdfr		wakeup(pool);
1258177633Sdfr}
1259177633Sdfr
1260184588Sdfrstatic void
1261184588Sdfrsvc_thread_start(void *arg)
1262184588Sdfr{
1263184588Sdfr
1264184588Sdfr	svc_run_internal((SVCPOOL *) arg, FALSE);
1265184588Sdfr	kthread_exit();
1266184588Sdfr}
1267184588Sdfr
1268184588Sdfrstatic void
1269184588Sdfrsvc_new_thread(SVCPOOL *pool)
1270184588Sdfr{
1271184588Sdfr	struct thread *td;
1272184588Sdfr
1273184588Sdfr	pool->sp_threadcount++;
1274184588Sdfr	kthread_add(svc_thread_start, pool,
1275184588Sdfr	    pool->sp_proc, &td, 0, 0,
1276184588Sdfr	    "%s: service", pool->sp_name);
1277184588Sdfr}
1278184588Sdfr
1279177633Sdfrvoid
1280184588Sdfrsvc_run(SVCPOOL *pool)
1281184588Sdfr{
1282184588Sdfr	int i;
1283184588Sdfr	struct proc *p;
1284184588Sdfr	struct thread *td;
1285184588Sdfr
1286184588Sdfr	p = curproc;
1287184588Sdfr	td = curthread;
1288184588Sdfr	snprintf(td->td_name, sizeof(td->td_name),
1289184588Sdfr	    "%s: master", pool->sp_name);
1290184588Sdfr	pool->sp_state = SVCPOOL_ACTIVE;
1291184588Sdfr	pool->sp_proc = p;
1292184588Sdfr	pool->sp_lastcreatetime = time_uptime;
1293184588Sdfr	pool->sp_threadcount = 1;
1294184588Sdfr
1295184588Sdfr	for (i = 1; i < pool->sp_minthreads; i++) {
1296184588Sdfr		svc_new_thread(pool);
1297184588Sdfr	}
1298184588Sdfr
1299184588Sdfr	svc_run_internal(pool, TRUE);
1300184588Sdfr
1301184588Sdfr	mtx_lock(&pool->sp_lock);
1302184588Sdfr	while (pool->sp_threadcount > 0)
1303184588Sdfr		msleep(pool, &pool->sp_lock, 0, "svcexit", 0);
1304184588Sdfr	mtx_unlock(&pool->sp_lock);
1305184588Sdfr}
1306184588Sdfr
1307184588Sdfrvoid
1308177633Sdfrsvc_exit(SVCPOOL *pool)
1309177633Sdfr{
1310184588Sdfr	SVCTHREAD *st;
1311184588Sdfr
1312177633Sdfr	mtx_lock(&pool->sp_lock);
1313184588Sdfr
1314261045Smav	if (pool->sp_state != SVCPOOL_CLOSING) {
1315261045Smav		pool->sp_state = SVCPOOL_CLOSING;
1316261045Smav		LIST_FOREACH(st, &pool->sp_idlethreads, st_ilink)
1317261045Smav			cv_signal(&st->st_cond);
1318261045Smav	}
1319184588Sdfr
1320177633Sdfr	mtx_unlock(&pool->sp_lock);
1321177633Sdfr}
1322184588Sdfr
1323184588Sdfrbool_t
1324184588Sdfrsvc_getargs(struct svc_req *rqstp, xdrproc_t xargs, void *args)
1325184588Sdfr{
1326184588Sdfr	struct mbuf *m;
1327184588Sdfr	XDR xdrs;
1328184588Sdfr	bool_t stat;
1329184588Sdfr
1330184588Sdfr	m = rqstp->rq_args;
1331184588Sdfr	rqstp->rq_args = NULL;
1332184588Sdfr
1333184588Sdfr	xdrmbuf_create(&xdrs, m, XDR_DECODE);
1334184588Sdfr	stat = xargs(&xdrs, args);
1335184588Sdfr	XDR_DESTROY(&xdrs);
1336184588Sdfr
1337184588Sdfr	return (stat);
1338184588Sdfr}
1339184588Sdfr
1340184588Sdfrbool_t
1341184588Sdfrsvc_freeargs(struct svc_req *rqstp, xdrproc_t xargs, void *args)
1342184588Sdfr{
1343184588Sdfr	XDR xdrs;
1344184588Sdfr
1345184588Sdfr	if (rqstp->rq_addr) {
1346184588Sdfr		free(rqstp->rq_addr, M_SONAME);
1347184588Sdfr		rqstp->rq_addr = NULL;
1348184588Sdfr	}
1349184588Sdfr
1350184588Sdfr	xdrs.x_op = XDR_FREE;
1351184588Sdfr	return (xargs(&xdrs, args));
1352184588Sdfr}
1353184588Sdfr
1354184588Sdfrvoid
1355184588Sdfrsvc_freereq(struct svc_req *rqstp)
1356184588Sdfr{
1357184588Sdfr	SVCTHREAD *st;
1358184588Sdfr	SVCPOOL *pool;
1359184588Sdfr
1360184588Sdfr	st = rqstp->rq_thread;
1361184588Sdfr	if (st) {
1362261054Smav		pool = st->st_pool;
1363184588Sdfr		if (pool->sp_done)
1364184588Sdfr			pool->sp_done(st, rqstp);
1365184588Sdfr	}
1366184588Sdfr
1367184588Sdfr	if (rqstp->rq_auth.svc_ah_ops)
1368184588Sdfr		SVCAUTH_RELEASE(&rqstp->rq_auth);
1369184588Sdfr
1370184588Sdfr	if (rqstp->rq_xprt) {
1371184588Sdfr		SVC_RELEASE(rqstp->rq_xprt);
1372184588Sdfr	}
1373184588Sdfr
1374184588Sdfr	if (rqstp->rq_addr)
1375184588Sdfr		free(rqstp->rq_addr, M_SONAME);
1376184588Sdfr
1377184588Sdfr	if (rqstp->rq_args)
1378184588Sdfr		m_freem(rqstp->rq_args);
1379184588Sdfr
1380184588Sdfr	free(rqstp, M_RPC);
1381184588Sdfr}
1382