1/*
2 * Neil Brown <neilb@cse.unsw.edu.au>
3 * J. Bruce Fields <bfields@umich.edu>
4 * Andy Adamson <andros@umich.edu>
5 * Dug Song <dugsong@monkey.org>
6 *
7 * RPCSEC_GSS server authentication.
8 * This implements RPCSEC_GSS as defined in rfc2203 (rpcsec_gss) and rfc2078
9 * (gssapi)
10 *
11 * The RPCSEC_GSS involves three stages:
12 *  1/ context creation
13 *  2/ data exchange
14 *  3/ context destruction
15 *
16 * Context creation is handled largely by upcalls to user-space.
17 *  In particular, GSS_Accept_sec_context is handled by an upcall
18 * Data exchange is handled entirely within the kernel
19 *  In particular, GSS_GetMIC, GSS_VerifyMIC, GSS_Seal, GSS_Unseal are in-kernel.
20 * Context destruction is handled in-kernel
21 *  GSS_Delete_sec_context is in-kernel
22 *
23 * Context creation is initiated by a RPCSEC_GSS_INIT request arriving.
24 * The context handle and gss_token are used as a key into the rpcsec_init cache.
25 * The content of this cache includes some of the outputs of GSS_Accept_sec_context,
26 * being major_status, minor_status, context_handle, reply_token.
27 * These are sent back to the client.
28 * Sequence window management is handled by the kernel.  The window size if currently
29 * a compile time constant.
30 *
31 * When user-space is happy that a context is established, it places an entry
32 * in the rpcsec_context cache. The key for this cache is the context_handle.
33 * The content includes:
34 *   uid/gidlist - for determining access rights
35 *   mechanism type
36 *   mechanism specific information, such as a key
37 *
38 */
39
40#include <linux/types.h>
41#include <linux/module.h>
42#include <linux/pagemap.h>
43
44#include <linux/sunrpc/auth_gss.h>
45#include <linux/sunrpc/svcauth.h>
46#include <linux/sunrpc/gss_err.h>
47#include <linux/sunrpc/svcauth.h>
48#include <linux/sunrpc/svcauth_gss.h>
49#include <linux/sunrpc/cache.h>
50
51#ifdef RPC_DEBUG
52# define RPCDBG_FACILITY	RPCDBG_AUTH
53#endif
54
55/* The rpcsec_init cache is used for mapping RPCSEC_GSS_{,CONT_}INIT requests
56 * into replies.
57 *
58 * Key is context handle (\x if empty) and gss_token.
59 * Content is major_status minor_status (integers) context_handle, reply_token.
60 *
61 */
62
63static int netobj_equal(struct xdr_netobj *a, struct xdr_netobj *b)
64{
65	return a->len == b->len && 0 == memcmp(a->data, b->data, a->len);
66}
67
68#define	RSI_HASHBITS	6
69#define	RSI_HASHMAX	(1<<RSI_HASHBITS)
70#define	RSI_HASHMASK	(RSI_HASHMAX-1)
71
72struct rsi {
73	struct cache_head	h;
74	struct xdr_netobj	in_handle, in_token;
75	struct xdr_netobj	out_handle, out_token;
76	int			major_status, minor_status;
77};
78
79static struct cache_head *rsi_table[RSI_HASHMAX];
80static struct cache_detail rsi_cache;
81static struct rsi *rsi_update(struct rsi *new, struct rsi *old);
82static struct rsi *rsi_lookup(struct rsi *item);
83
84static void rsi_free(struct rsi *rsii)
85{
86	kfree(rsii->in_handle.data);
87	kfree(rsii->in_token.data);
88	kfree(rsii->out_handle.data);
89	kfree(rsii->out_token.data);
90}
91
92static void rsi_put(struct kref *ref)
93{
94	struct rsi *rsii = container_of(ref, struct rsi, h.ref);
95	rsi_free(rsii);
96	kfree(rsii);
97}
98
99static inline int rsi_hash(struct rsi *item)
100{
101	return hash_mem(item->in_handle.data, item->in_handle.len, RSI_HASHBITS)
102	     ^ hash_mem(item->in_token.data, item->in_token.len, RSI_HASHBITS);
103}
104
105static int rsi_match(struct cache_head *a, struct cache_head *b)
106{
107	struct rsi *item = container_of(a, struct rsi, h);
108	struct rsi *tmp = container_of(b, struct rsi, h);
109	return netobj_equal(&item->in_handle, &tmp->in_handle)
110		&& netobj_equal(&item->in_token, &tmp->in_token);
111}
112
113static int dup_to_netobj(struct xdr_netobj *dst, char *src, int len)
114{
115	dst->len = len;
116	dst->data = (len ? kmemdup(src, len, GFP_KERNEL) : NULL);
117	if (len && !dst->data)
118		return -ENOMEM;
119	return 0;
120}
121
122static inline int dup_netobj(struct xdr_netobj *dst, struct xdr_netobj *src)
123{
124	return dup_to_netobj(dst, src->data, src->len);
125}
126
127static void rsi_init(struct cache_head *cnew, struct cache_head *citem)
128{
129	struct rsi *new = container_of(cnew, struct rsi, h);
130	struct rsi *item = container_of(citem, struct rsi, h);
131
132	new->out_handle.data = NULL;
133	new->out_handle.len = 0;
134	new->out_token.data = NULL;
135	new->out_token.len = 0;
136	new->in_handle.len = item->in_handle.len;
137	item->in_handle.len = 0;
138	new->in_token.len = item->in_token.len;
139	item->in_token.len = 0;
140	new->in_handle.data = item->in_handle.data;
141	item->in_handle.data = NULL;
142	new->in_token.data = item->in_token.data;
143	item->in_token.data = NULL;
144}
145
146static void update_rsi(struct cache_head *cnew, struct cache_head *citem)
147{
148	struct rsi *new = container_of(cnew, struct rsi, h);
149	struct rsi *item = container_of(citem, struct rsi, h);
150
151	BUG_ON(new->out_handle.data || new->out_token.data);
152	new->out_handle.len = item->out_handle.len;
153	item->out_handle.len = 0;
154	new->out_token.len = item->out_token.len;
155	item->out_token.len = 0;
156	new->out_handle.data = item->out_handle.data;
157	item->out_handle.data = NULL;
158	new->out_token.data = item->out_token.data;
159	item->out_token.data = NULL;
160
161	new->major_status = item->major_status;
162	new->minor_status = item->minor_status;
163}
164
165static struct cache_head *rsi_alloc(void)
166{
167	struct rsi *rsii = kmalloc(sizeof(*rsii), GFP_KERNEL);
168	if (rsii)
169		return &rsii->h;
170	else
171		return NULL;
172}
173
174static void rsi_request(struct cache_detail *cd,
175		       struct cache_head *h,
176		       char **bpp, int *blen)
177{
178	struct rsi *rsii = container_of(h, struct rsi, h);
179
180	qword_addhex(bpp, blen, rsii->in_handle.data, rsii->in_handle.len);
181	qword_addhex(bpp, blen, rsii->in_token.data, rsii->in_token.len);
182	(*bpp)[-1] = '\n';
183}
184
185
186static int rsi_parse(struct cache_detail *cd,
187		    char *mesg, int mlen)
188{
189	/* context token expiry major minor context token */
190	char *buf = mesg;
191	char *ep;
192	int len;
193	struct rsi rsii, *rsip = NULL;
194	time_t expiry;
195	int status = -EINVAL;
196
197	memset(&rsii, 0, sizeof(rsii));
198	/* handle */
199	len = qword_get(&mesg, buf, mlen);
200	if (len < 0)
201		goto out;
202	status = -ENOMEM;
203	if (dup_to_netobj(&rsii.in_handle, buf, len))
204		goto out;
205
206	/* token */
207	len = qword_get(&mesg, buf, mlen);
208	status = -EINVAL;
209	if (len < 0)
210		goto out;
211	status = -ENOMEM;
212	if (dup_to_netobj(&rsii.in_token, buf, len))
213		goto out;
214
215	rsip = rsi_lookup(&rsii);
216	if (!rsip)
217		goto out;
218
219	rsii.h.flags = 0;
220	/* expiry */
221	expiry = get_expiry(&mesg);
222	status = -EINVAL;
223	if (expiry == 0)
224		goto out;
225
226	/* major/minor */
227	len = qword_get(&mesg, buf, mlen);
228	if (len < 0)
229		goto out;
230	if (len == 0) {
231		goto out;
232	} else {
233		rsii.major_status = simple_strtoul(buf, &ep, 10);
234		if (*ep)
235			goto out;
236		len = qword_get(&mesg, buf, mlen);
237		if (len <= 0)
238			goto out;
239		rsii.minor_status = simple_strtoul(buf, &ep, 10);
240		if (*ep)
241			goto out;
242
243		/* out_handle */
244		len = qword_get(&mesg, buf, mlen);
245		if (len < 0)
246			goto out;
247		status = -ENOMEM;
248		if (dup_to_netobj(&rsii.out_handle, buf, len))
249			goto out;
250
251		/* out_token */
252		len = qword_get(&mesg, buf, mlen);
253		status = -EINVAL;
254		if (len < 0)
255			goto out;
256		status = -ENOMEM;
257		if (dup_to_netobj(&rsii.out_token, buf, len))
258			goto out;
259	}
260	rsii.h.expiry_time = expiry;
261	rsip = rsi_update(&rsii, rsip);
262	status = 0;
263out:
264	rsi_free(&rsii);
265	if (rsip)
266		cache_put(&rsip->h, &rsi_cache);
267	else
268		status = -ENOMEM;
269	return status;
270}
271
272static struct cache_detail rsi_cache = {
273	.owner		= THIS_MODULE,
274	.hash_size	= RSI_HASHMAX,
275	.hash_table     = rsi_table,
276	.name           = "auth.rpcsec.init",
277	.cache_put      = rsi_put,
278	.cache_request  = rsi_request,
279	.cache_parse    = rsi_parse,
280	.match		= rsi_match,
281	.init		= rsi_init,
282	.update		= update_rsi,
283	.alloc		= rsi_alloc,
284};
285
286static struct rsi *rsi_lookup(struct rsi *item)
287{
288	struct cache_head *ch;
289	int hash = rsi_hash(item);
290
291	ch = sunrpc_cache_lookup(&rsi_cache, &item->h, hash);
292	if (ch)
293		return container_of(ch, struct rsi, h);
294	else
295		return NULL;
296}
297
298static struct rsi *rsi_update(struct rsi *new, struct rsi *old)
299{
300	struct cache_head *ch;
301	int hash = rsi_hash(new);
302
303	ch = sunrpc_cache_update(&rsi_cache, &new->h,
304				 &old->h, hash);
305	if (ch)
306		return container_of(ch, struct rsi, h);
307	else
308		return NULL;
309}
310
311
312/*
313 * The rpcsec_context cache is used to store a context that is
314 * used in data exchange.
315 * The key is a context handle. The content is:
316 *  uid, gidlist, mechanism, service-set, mech-specific-data
317 */
318
319#define	RSC_HASHBITS	10
320#define	RSC_HASHMAX	(1<<RSC_HASHBITS)
321#define	RSC_HASHMASK	(RSC_HASHMAX-1)
322
323#define GSS_SEQ_WIN	128
324
325struct gss_svc_seq_data {
326	/* highest seq number seen so far: */
327	int			sd_max;
328	/* for i such that sd_max-GSS_SEQ_WIN < i <= sd_max, the i-th bit of
329	 * sd_win is nonzero iff sequence number i has been seen already: */
330	unsigned long		sd_win[GSS_SEQ_WIN/BITS_PER_LONG];
331	spinlock_t		sd_lock;
332};
333
334struct rsc {
335	struct cache_head	h;
336	struct xdr_netobj	handle;
337	struct svc_cred		cred;
338	struct gss_svc_seq_data	seqdata;
339	struct gss_ctx		*mechctx;
340};
341
342static struct cache_head *rsc_table[RSC_HASHMAX];
343static struct cache_detail rsc_cache;
344static struct rsc *rsc_update(struct rsc *new, struct rsc *old);
345static struct rsc *rsc_lookup(struct rsc *item);
346
347static void rsc_free(struct rsc *rsci)
348{
349	kfree(rsci->handle.data);
350	if (rsci->mechctx)
351		gss_delete_sec_context(&rsci->mechctx);
352	if (rsci->cred.cr_group_info)
353		put_group_info(rsci->cred.cr_group_info);
354}
355
356static void rsc_put(struct kref *ref)
357{
358	struct rsc *rsci = container_of(ref, struct rsc, h.ref);
359
360	rsc_free(rsci);
361	kfree(rsci);
362}
363
364static inline int
365rsc_hash(struct rsc *rsci)
366{
367	return hash_mem(rsci->handle.data, rsci->handle.len, RSC_HASHBITS);
368}
369
370static int
371rsc_match(struct cache_head *a, struct cache_head *b)
372{
373	struct rsc *new = container_of(a, struct rsc, h);
374	struct rsc *tmp = container_of(b, struct rsc, h);
375
376	return netobj_equal(&new->handle, &tmp->handle);
377}
378
379static void
380rsc_init(struct cache_head *cnew, struct cache_head *ctmp)
381{
382	struct rsc *new = container_of(cnew, struct rsc, h);
383	struct rsc *tmp = container_of(ctmp, struct rsc, h);
384
385	new->handle.len = tmp->handle.len;
386	tmp->handle.len = 0;
387	new->handle.data = tmp->handle.data;
388	tmp->handle.data = NULL;
389	new->mechctx = NULL;
390	new->cred.cr_group_info = NULL;
391}
392
393static void
394update_rsc(struct cache_head *cnew, struct cache_head *ctmp)
395{
396	struct rsc *new = container_of(cnew, struct rsc, h);
397	struct rsc *tmp = container_of(ctmp, struct rsc, h);
398
399	new->mechctx = tmp->mechctx;
400	tmp->mechctx = NULL;
401	memset(&new->seqdata, 0, sizeof(new->seqdata));
402	spin_lock_init(&new->seqdata.sd_lock);
403	new->cred = tmp->cred;
404	tmp->cred.cr_group_info = NULL;
405}
406
407static struct cache_head *
408rsc_alloc(void)
409{
410	struct rsc *rsci = kmalloc(sizeof(*rsci), GFP_KERNEL);
411	if (rsci)
412		return &rsci->h;
413	else
414		return NULL;
415}
416
417static int rsc_parse(struct cache_detail *cd,
418		     char *mesg, int mlen)
419{
420	/* contexthandle expiry [ uid gid N <n gids> mechname ...mechdata... ] */
421	char *buf = mesg;
422	int len, rv;
423	struct rsc rsci, *rscp = NULL;
424	time_t expiry;
425	int status = -EINVAL;
426	struct gss_api_mech *gm = NULL;
427
428	memset(&rsci, 0, sizeof(rsci));
429	/* context handle */
430	len = qword_get(&mesg, buf, mlen);
431	if (len < 0) goto out;
432	status = -ENOMEM;
433	if (dup_to_netobj(&rsci.handle, buf, len))
434		goto out;
435
436	rsci.h.flags = 0;
437	/* expiry */
438	expiry = get_expiry(&mesg);
439	status = -EINVAL;
440	if (expiry == 0)
441		goto out;
442
443	rscp = rsc_lookup(&rsci);
444	if (!rscp)
445		goto out;
446
447	/* uid, or NEGATIVE */
448	rv = get_int(&mesg, &rsci.cred.cr_uid);
449	if (rv == -EINVAL)
450		goto out;
451	if (rv == -ENOENT)
452		set_bit(CACHE_NEGATIVE, &rsci.h.flags);
453	else {
454		int N, i;
455
456		/* gid */
457		if (get_int(&mesg, &rsci.cred.cr_gid))
458			goto out;
459
460		/* number of additional gid's */
461		if (get_int(&mesg, &N))
462			goto out;
463		status = -ENOMEM;
464		rsci.cred.cr_group_info = groups_alloc(N);
465		if (rsci.cred.cr_group_info == NULL)
466			goto out;
467
468		/* gid's */
469		status = -EINVAL;
470		for (i=0; i<N; i++) {
471			gid_t gid;
472			if (get_int(&mesg, &gid))
473				goto out;
474			GROUP_AT(rsci.cred.cr_group_info, i) = gid;
475		}
476
477		/* mech name */
478		len = qword_get(&mesg, buf, mlen);
479		if (len < 0)
480			goto out;
481		gm = gss_mech_get_by_name(buf);
482		status = -EOPNOTSUPP;
483		if (!gm)
484			goto out;
485
486		status = -EINVAL;
487		/* mech-specific data: */
488		len = qword_get(&mesg, buf, mlen);
489		if (len < 0)
490			goto out;
491		status = gss_import_sec_context(buf, len, gm, &rsci.mechctx);
492		if (status)
493			goto out;
494	}
495	rsci.h.expiry_time = expiry;
496	rscp = rsc_update(&rsci, rscp);
497	status = 0;
498out:
499	gss_mech_put(gm);
500	rsc_free(&rsci);
501	if (rscp)
502		cache_put(&rscp->h, &rsc_cache);
503	else
504		status = -ENOMEM;
505	return status;
506}
507
508static struct cache_detail rsc_cache = {
509	.owner		= THIS_MODULE,
510	.hash_size	= RSC_HASHMAX,
511	.hash_table	= rsc_table,
512	.name		= "auth.rpcsec.context",
513	.cache_put	= rsc_put,
514	.cache_parse	= rsc_parse,
515	.match		= rsc_match,
516	.init		= rsc_init,
517	.update		= update_rsc,
518	.alloc		= rsc_alloc,
519};
520
521static struct rsc *rsc_lookup(struct rsc *item)
522{
523	struct cache_head *ch;
524	int hash = rsc_hash(item);
525
526	ch = sunrpc_cache_lookup(&rsc_cache, &item->h, hash);
527	if (ch)
528		return container_of(ch, struct rsc, h);
529	else
530		return NULL;
531}
532
533static struct rsc *rsc_update(struct rsc *new, struct rsc *old)
534{
535	struct cache_head *ch;
536	int hash = rsc_hash(new);
537
538	ch = sunrpc_cache_update(&rsc_cache, &new->h,
539				 &old->h, hash);
540	if (ch)
541		return container_of(ch, struct rsc, h);
542	else
543		return NULL;
544}
545
546
547static struct rsc *
548gss_svc_searchbyctx(struct xdr_netobj *handle)
549{
550	struct rsc rsci;
551	struct rsc *found;
552
553	memset(&rsci, 0, sizeof(rsci));
554	if (dup_to_netobj(&rsci.handle, handle->data, handle->len))
555		return NULL;
556	found = rsc_lookup(&rsci);
557	rsc_free(&rsci);
558	if (!found)
559		return NULL;
560	if (cache_check(&rsc_cache, &found->h, NULL))
561		return NULL;
562	return found;
563}
564
565/* Implements sequence number algorithm as specified in RFC 2203. */
566static int
567gss_check_seq_num(struct rsc *rsci, int seq_num)
568{
569	struct gss_svc_seq_data *sd = &rsci->seqdata;
570
571	spin_lock(&sd->sd_lock);
572	if (seq_num > sd->sd_max) {
573		if (seq_num >= sd->sd_max + GSS_SEQ_WIN) {
574			memset(sd->sd_win,0,sizeof(sd->sd_win));
575			sd->sd_max = seq_num;
576		} else while (sd->sd_max < seq_num) {
577			sd->sd_max++;
578			__clear_bit(sd->sd_max % GSS_SEQ_WIN, sd->sd_win);
579		}
580		__set_bit(seq_num % GSS_SEQ_WIN, sd->sd_win);
581		goto ok;
582	} else if (seq_num <= sd->sd_max - GSS_SEQ_WIN) {
583		goto drop;
584	}
585	/* sd_max - GSS_SEQ_WIN < seq_num <= sd_max */
586	if (__test_and_set_bit(seq_num % GSS_SEQ_WIN, sd->sd_win))
587		goto drop;
588ok:
589	spin_unlock(&sd->sd_lock);
590	return 1;
591drop:
592	spin_unlock(&sd->sd_lock);
593	return 0;
594}
595
596static inline u32 round_up_to_quad(u32 i)
597{
598	return (i + 3 ) & ~3;
599}
600
601static inline int
602svc_safe_getnetobj(struct kvec *argv, struct xdr_netobj *o)
603{
604	int l;
605
606	if (argv->iov_len < 4)
607		return -1;
608	o->len = svc_getnl(argv);
609	l = round_up_to_quad(o->len);
610	if (argv->iov_len < l)
611		return -1;
612	o->data = argv->iov_base;
613	argv->iov_base += l;
614	argv->iov_len -= l;
615	return 0;
616}
617
618static inline int
619svc_safe_putnetobj(struct kvec *resv, struct xdr_netobj *o)
620{
621	u8 *p;
622
623	if (resv->iov_len + 4 > PAGE_SIZE)
624		return -1;
625	svc_putnl(resv, o->len);
626	p = resv->iov_base + resv->iov_len;
627	resv->iov_len += round_up_to_quad(o->len);
628	if (resv->iov_len > PAGE_SIZE)
629		return -1;
630	memcpy(p, o->data, o->len);
631	memset(p + o->len, 0, round_up_to_quad(o->len) - o->len);
632	return 0;
633}
634
635/* Verify the checksum on the header and return SVC_OK on success.
636 * Otherwise, return SVC_DROP (in the case of a bad sequence number)
637 * or return SVC_DENIED and indicate error in authp.
638 */
639static int
640gss_verify_header(struct svc_rqst *rqstp, struct rsc *rsci,
641		  __be32 *rpcstart, struct rpc_gss_wire_cred *gc, __be32 *authp)
642{
643	struct gss_ctx		*ctx_id = rsci->mechctx;
644	struct xdr_buf		rpchdr;
645	struct xdr_netobj	checksum;
646	u32			flavor = 0;
647	struct kvec		*argv = &rqstp->rq_arg.head[0];
648	struct kvec		iov;
649
650	/* data to compute the checksum over: */
651	iov.iov_base = rpcstart;
652	iov.iov_len = (u8 *)argv->iov_base - (u8 *)rpcstart;
653	xdr_buf_from_iov(&iov, &rpchdr);
654
655	*authp = rpc_autherr_badverf;
656	if (argv->iov_len < 4)
657		return SVC_DENIED;
658	flavor = svc_getnl(argv);
659	if (flavor != RPC_AUTH_GSS)
660		return SVC_DENIED;
661	if (svc_safe_getnetobj(argv, &checksum))
662		return SVC_DENIED;
663
664	if (rqstp->rq_deferred) /* skip verification of revisited request */
665		return SVC_OK;
666	if (gss_verify_mic(ctx_id, &rpchdr, &checksum) != GSS_S_COMPLETE) {
667		*authp = rpcsec_gsserr_credproblem;
668		return SVC_DENIED;
669	}
670
671	if (gc->gc_seq > MAXSEQ) {
672		dprintk("RPC:       svcauth_gss: discarding request with "
673				"large sequence number %d\n", gc->gc_seq);
674		*authp = rpcsec_gsserr_ctxproblem;
675		return SVC_DENIED;
676	}
677	if (!gss_check_seq_num(rsci, gc->gc_seq)) {
678		dprintk("RPC:       svcauth_gss: discarding request with "
679				"old sequence number %d\n", gc->gc_seq);
680		return SVC_DROP;
681	}
682	return SVC_OK;
683}
684
685static int
686gss_write_null_verf(struct svc_rqst *rqstp)
687{
688	__be32     *p;
689
690	svc_putnl(rqstp->rq_res.head, RPC_AUTH_NULL);
691	p = rqstp->rq_res.head->iov_base + rqstp->rq_res.head->iov_len;
692	/* don't really need to check if head->iov_len > PAGE_SIZE ... */
693	*p++ = 0;
694	if (!xdr_ressize_check(rqstp, p))
695		return -1;
696	return 0;
697}
698
699static int
700gss_write_verf(struct svc_rqst *rqstp, struct gss_ctx *ctx_id, u32 seq)
701{
702	__be32			xdr_seq;
703	u32			maj_stat;
704	struct xdr_buf		verf_data;
705	struct xdr_netobj	mic;
706	__be32			*p;
707	struct kvec		iov;
708
709	svc_putnl(rqstp->rq_res.head, RPC_AUTH_GSS);
710	xdr_seq = htonl(seq);
711
712	iov.iov_base = &xdr_seq;
713	iov.iov_len = sizeof(xdr_seq);
714	xdr_buf_from_iov(&iov, &verf_data);
715	p = rqstp->rq_res.head->iov_base + rqstp->rq_res.head->iov_len;
716	mic.data = (u8 *)(p + 1);
717	maj_stat = gss_get_mic(ctx_id, &verf_data, &mic);
718	if (maj_stat != GSS_S_COMPLETE)
719		return -1;
720	*p++ = htonl(mic.len);
721	memset((u8 *)p + mic.len, 0, round_up_to_quad(mic.len) - mic.len);
722	p += XDR_QUADLEN(mic.len);
723	if (!xdr_ressize_check(rqstp, p))
724		return -1;
725	return 0;
726}
727
728struct gss_domain {
729	struct auth_domain	h;
730	u32			pseudoflavor;
731};
732
733static struct auth_domain *
734find_gss_auth_domain(struct gss_ctx *ctx, u32 svc)
735{
736	char *name;
737
738	name = gss_service_to_auth_domain_name(ctx->mech_type, svc);
739	if (!name)
740		return NULL;
741	return auth_domain_find(name);
742}
743
744static struct auth_ops svcauthops_gss;
745
746int
747svcauth_gss_register_pseudoflavor(u32 pseudoflavor, char * name)
748{
749	struct gss_domain	*new;
750	struct auth_domain	*test;
751	int			stat = -ENOMEM;
752
753	new = kmalloc(sizeof(*new), GFP_KERNEL);
754	if (!new)
755		goto out;
756	kref_init(&new->h.ref);
757	new->h.name = kstrdup(name, GFP_KERNEL);
758	if (!new->h.name)
759		goto out_free_dom;
760	new->h.flavour = &svcauthops_gss;
761	new->pseudoflavor = pseudoflavor;
762
763	test = auth_domain_lookup(name, &new->h);
764	if (test != &new->h) {
765		auth_domain_put(&new->h);
766		/* dangling ref-count... */
767		goto out;
768	}
769	return 0;
770
771out_free_dom:
772	kfree(new);
773out:
774	return stat;
775}
776
777EXPORT_SYMBOL(svcauth_gss_register_pseudoflavor);
778
779static inline int
780read_u32_from_xdr_buf(struct xdr_buf *buf, int base, u32 *obj)
781{
782	__be32  raw;
783	int     status;
784
785	status = read_bytes_from_xdr_buf(buf, base, &raw, sizeof(*obj));
786	if (status)
787		return status;
788	*obj = ntohl(raw);
789	return 0;
790}
791
792/* It would be nice if this bit of code could be shared with the client.
793 * Obstacles:
794 *	The client shouldn't malloc(), would have to pass in own memory.
795 *	The server uses base of head iovec as read pointer, while the
796 *	client uses separate pointer. */
797static int
798unwrap_integ_data(struct xdr_buf *buf, u32 seq, struct gss_ctx *ctx)
799{
800	int stat = -EINVAL;
801	u32 integ_len, maj_stat;
802	struct xdr_netobj mic;
803	struct xdr_buf integ_buf;
804
805	integ_len = svc_getnl(&buf->head[0]);
806	if (integ_len & 3)
807		return stat;
808	if (integ_len > buf->len)
809		return stat;
810	if (xdr_buf_subsegment(buf, &integ_buf, 0, integ_len))
811		BUG();
812	/* copy out mic... */
813	if (read_u32_from_xdr_buf(buf, integ_len, &mic.len))
814		BUG();
815	if (mic.len > RPC_MAX_AUTH_SIZE)
816		return stat;
817	mic.data = kmalloc(mic.len, GFP_KERNEL);
818	if (!mic.data)
819		return stat;
820	if (read_bytes_from_xdr_buf(buf, integ_len + 4, mic.data, mic.len))
821		goto out;
822	maj_stat = gss_verify_mic(ctx, &integ_buf, &mic);
823	if (maj_stat != GSS_S_COMPLETE)
824		goto out;
825	if (svc_getnl(&buf->head[0]) != seq)
826		goto out;
827	stat = 0;
828out:
829	kfree(mic.data);
830	return stat;
831}
832
833static inline int
834total_buf_len(struct xdr_buf *buf)
835{
836	return buf->head[0].iov_len + buf->page_len + buf->tail[0].iov_len;
837}
838
839static void
840fix_priv_head(struct xdr_buf *buf, int pad)
841{
842	if (buf->page_len == 0) {
843		/* We need to adjust head and buf->len in tandem in this
844		 * case to make svc_defer() work--it finds the original
845		 * buffer start using buf->len - buf->head[0].iov_len. */
846		buf->head[0].iov_len -= pad;
847	}
848}
849
850static int
851unwrap_priv_data(struct svc_rqst *rqstp, struct xdr_buf *buf, u32 seq, struct gss_ctx *ctx)
852{
853	u32 priv_len, maj_stat;
854	int pad, saved_len, remaining_len, offset;
855
856	rqstp->rq_sendfile_ok = 0;
857
858	priv_len = svc_getnl(&buf->head[0]);
859	if (rqstp->rq_deferred) {
860		/* Already decrypted last time through! The sequence number
861		 * check at out_seq is unnecessary but harmless: */
862		goto out_seq;
863	}
864	/* buf->len is the number of bytes from the original start of the
865	 * request to the end, where head[0].iov_len is just the bytes
866	 * not yet read from the head, so these two values are different: */
867	remaining_len = total_buf_len(buf);
868	if (priv_len > remaining_len)
869		return -EINVAL;
870	pad = remaining_len - priv_len;
871	buf->len -= pad;
872	fix_priv_head(buf, pad);
873
874	/* Maybe it would be better to give gss_unwrap a length parameter: */
875	saved_len = buf->len;
876	buf->len = priv_len;
877	maj_stat = gss_unwrap(ctx, 0, buf);
878	pad = priv_len - buf->len;
879	buf->len = saved_len;
880	buf->len -= pad;
881	/* The upper layers assume the buffer is aligned on 4-byte boundaries.
882	 * In the krb5p case, at least, the data ends up offset, so we need to
883	 * move it around. */
884	offset = buf->head[0].iov_len % 4;
885	if (offset) {
886		buf->buflen = RPCSVC_MAXPAYLOAD;
887		xdr_shift_buf(buf, offset);
888		fix_priv_head(buf, pad);
889	}
890	if (maj_stat != GSS_S_COMPLETE)
891		return -EINVAL;
892out_seq:
893	if (svc_getnl(&buf->head[0]) != seq)
894		return -EINVAL;
895	return 0;
896}
897
898struct gss_svc_data {
899	/* decoded gss client cred: */
900	struct rpc_gss_wire_cred	clcred;
901	/* save a pointer to the beginning of the encoded verifier,
902	 * for use in encryption/checksumming in svcauth_gss_release: */
903	__be32				*verf_start;
904	struct rsc			*rsci;
905};
906
907static int
908svcauth_gss_set_client(struct svc_rqst *rqstp)
909{
910	struct gss_svc_data *svcdata = rqstp->rq_auth_data;
911	struct rsc *rsci = svcdata->rsci;
912	struct rpc_gss_wire_cred *gc = &svcdata->clcred;
913
914	rqstp->rq_client = find_gss_auth_domain(rsci->mechctx, gc->gc_svc);
915	if (rqstp->rq_client == NULL)
916		return SVC_DENIED;
917	return SVC_OK;
918}
919
920static inline int
921gss_write_init_verf(struct svc_rqst *rqstp, struct rsi *rsip)
922{
923	struct rsc *rsci;
924	int        rc;
925
926	if (rsip->major_status != GSS_S_COMPLETE)
927		return gss_write_null_verf(rqstp);
928	rsci = gss_svc_searchbyctx(&rsip->out_handle);
929	if (rsci == NULL) {
930		rsip->major_status = GSS_S_NO_CONTEXT;
931		return gss_write_null_verf(rqstp);
932	}
933	rc = gss_write_verf(rqstp, rsci->mechctx, GSS_SEQ_WIN);
934	cache_put(&rsci->h, &rsc_cache);
935	return rc;
936}
937
938/*
939 * Accept an rpcsec packet.
940 * If context establishment, punt to user space
941 * If data exchange, verify/decrypt
942 * If context destruction, handle here
943 * In the context establishment and destruction case we encode
944 * response here and return SVC_COMPLETE.
945 */
946static int
947svcauth_gss_accept(struct svc_rqst *rqstp, __be32 *authp)
948{
949	struct kvec	*argv = &rqstp->rq_arg.head[0];
950	struct kvec	*resv = &rqstp->rq_res.head[0];
951	u32		crlen;
952	struct xdr_netobj tmpobj;
953	struct gss_svc_data *svcdata = rqstp->rq_auth_data;
954	struct rpc_gss_wire_cred *gc;
955	struct rsc	*rsci = NULL;
956	struct rsi	*rsip, rsikey;
957	__be32		*rpcstart;
958	__be32		*reject_stat = resv->iov_base + resv->iov_len;
959	int		ret;
960
961	dprintk("RPC:       svcauth_gss: argv->iov_len = %zd\n",
962			argv->iov_len);
963
964	*authp = rpc_autherr_badcred;
965	if (!svcdata)
966		svcdata = kmalloc(sizeof(*svcdata), GFP_KERNEL);
967	if (!svcdata)
968		goto auth_err;
969	rqstp->rq_auth_data = svcdata;
970	svcdata->verf_start = NULL;
971	svcdata->rsci = NULL;
972	gc = &svcdata->clcred;
973
974	/* start of rpc packet is 7 u32's back from here:
975	 * xid direction rpcversion prog vers proc flavour
976	 */
977	rpcstart = argv->iov_base;
978	rpcstart -= 7;
979
980	/* credential is:
981	 *   version(==1), proc(0,1,2,3), seq, service (1,2,3), handle
982	 * at least 5 u32s, and is preceeded by length, so that makes 6.
983	 */
984
985	if (argv->iov_len < 5 * 4)
986		goto auth_err;
987	crlen = svc_getnl(argv);
988	if (svc_getnl(argv) != RPC_GSS_VERSION)
989		goto auth_err;
990	gc->gc_proc = svc_getnl(argv);
991	gc->gc_seq = svc_getnl(argv);
992	gc->gc_svc = svc_getnl(argv);
993	if (svc_safe_getnetobj(argv, &gc->gc_ctx))
994		goto auth_err;
995	if (crlen != round_up_to_quad(gc->gc_ctx.len) + 5 * 4)
996		goto auth_err;
997
998	if ((gc->gc_proc != RPC_GSS_PROC_DATA) && (rqstp->rq_proc != 0))
999		goto auth_err;
1000
1001	/*
1002	 * We've successfully parsed the credential. Let's check out the
1003	 * verifier.  An AUTH_NULL verifier is allowed (and required) for
1004	 * INIT and CONTINUE_INIT requests. AUTH_RPCSEC_GSS is required for
1005	 * PROC_DATA and PROC_DESTROY.
1006	 *
1007	 * AUTH_NULL verifier is 0 (AUTH_NULL), 0 (length).
1008	 * AUTH_RPCSEC_GSS verifier is:
1009	 *   6 (AUTH_RPCSEC_GSS), length, checksum.
1010	 * checksum is calculated over rpcheader from xid up to here.
1011	 */
1012	*authp = rpc_autherr_badverf;
1013	switch (gc->gc_proc) {
1014	case RPC_GSS_PROC_INIT:
1015	case RPC_GSS_PROC_CONTINUE_INIT:
1016		if (argv->iov_len < 2 * 4)
1017			goto auth_err;
1018		if (svc_getnl(argv) != RPC_AUTH_NULL)
1019			goto auth_err;
1020		if (svc_getnl(argv) != 0)
1021			goto auth_err;
1022		break;
1023	case RPC_GSS_PROC_DATA:
1024	case RPC_GSS_PROC_DESTROY:
1025		*authp = rpcsec_gsserr_credproblem;
1026		rsci = gss_svc_searchbyctx(&gc->gc_ctx);
1027		if (!rsci)
1028			goto auth_err;
1029		switch (gss_verify_header(rqstp, rsci, rpcstart, gc, authp)) {
1030		case SVC_OK:
1031			break;
1032		case SVC_DENIED:
1033			goto auth_err;
1034		case SVC_DROP:
1035			goto drop;
1036		}
1037		break;
1038	default:
1039		*authp = rpc_autherr_rejectedcred;
1040		goto auth_err;
1041	}
1042
1043	/* now act upon the command: */
1044	switch (gc->gc_proc) {
1045	case RPC_GSS_PROC_INIT:
1046	case RPC_GSS_PROC_CONTINUE_INIT:
1047		*authp = rpc_autherr_badcred;
1048		if (gc->gc_proc == RPC_GSS_PROC_INIT && gc->gc_ctx.len != 0)
1049			goto auth_err;
1050		memset(&rsikey, 0, sizeof(rsikey));
1051		if (dup_netobj(&rsikey.in_handle, &gc->gc_ctx))
1052			goto drop;
1053		*authp = rpc_autherr_badverf;
1054		if (svc_safe_getnetobj(argv, &tmpobj)) {
1055			kfree(rsikey.in_handle.data);
1056			goto auth_err;
1057		}
1058		if (dup_netobj(&rsikey.in_token, &tmpobj)) {
1059			kfree(rsikey.in_handle.data);
1060			goto drop;
1061		}
1062
1063		rsip = rsi_lookup(&rsikey);
1064		rsi_free(&rsikey);
1065		if (!rsip) {
1066			goto drop;
1067		}
1068		switch(cache_check(&rsi_cache, &rsip->h, &rqstp->rq_chandle)) {
1069		case -EAGAIN:
1070		case -ETIMEDOUT:
1071		case -ENOENT:
1072			goto drop;
1073		case 0:
1074			if (gss_write_init_verf(rqstp, rsip))
1075				goto drop;
1076			if (resv->iov_len + 4 > PAGE_SIZE)
1077				goto drop;
1078			svc_putnl(resv, RPC_SUCCESS);
1079			if (svc_safe_putnetobj(resv, &rsip->out_handle))
1080				goto drop;
1081			if (resv->iov_len + 3 * 4 > PAGE_SIZE)
1082				goto drop;
1083			svc_putnl(resv, rsip->major_status);
1084			svc_putnl(resv, rsip->minor_status);
1085			svc_putnl(resv, GSS_SEQ_WIN);
1086			if (svc_safe_putnetobj(resv, &rsip->out_token))
1087				goto drop;
1088			rqstp->rq_client = NULL;
1089		}
1090		goto complete;
1091	case RPC_GSS_PROC_DESTROY:
1092		if (gss_write_verf(rqstp, rsci->mechctx, gc->gc_seq))
1093			goto auth_err;
1094		set_bit(CACHE_NEGATIVE, &rsci->h.flags);
1095		if (resv->iov_len + 4 > PAGE_SIZE)
1096			goto drop;
1097		svc_putnl(resv, RPC_SUCCESS);
1098		goto complete;
1099	case RPC_GSS_PROC_DATA:
1100		*authp = rpcsec_gsserr_ctxproblem;
1101		svcdata->verf_start = resv->iov_base + resv->iov_len;
1102		if (gss_write_verf(rqstp, rsci->mechctx, gc->gc_seq))
1103			goto auth_err;
1104		rqstp->rq_cred = rsci->cred;
1105		get_group_info(rsci->cred.cr_group_info);
1106		*authp = rpc_autherr_badcred;
1107		switch (gc->gc_svc) {
1108		case RPC_GSS_SVC_NONE:
1109			break;
1110		case RPC_GSS_SVC_INTEGRITY:
1111			if (unwrap_integ_data(&rqstp->rq_arg,
1112					gc->gc_seq, rsci->mechctx))
1113				goto auth_err;
1114			/* placeholders for length and seq. number: */
1115			svc_putnl(resv, 0);
1116			svc_putnl(resv, 0);
1117			break;
1118		case RPC_GSS_SVC_PRIVACY:
1119			if (unwrap_priv_data(rqstp, &rqstp->rq_arg,
1120					gc->gc_seq, rsci->mechctx))
1121				goto auth_err;
1122			/* placeholders for length and seq. number: */
1123			svc_putnl(resv, 0);
1124			svc_putnl(resv, 0);
1125			break;
1126		default:
1127			goto auth_err;
1128		}
1129		svcdata->rsci = rsci;
1130		cache_get(&rsci->h);
1131		ret = SVC_OK;
1132		goto out;
1133	}
1134auth_err:
1135	/* Restore write pointer to original value: */
1136	xdr_ressize_check(rqstp, reject_stat);
1137	ret = SVC_DENIED;
1138	goto out;
1139complete:
1140	ret = SVC_COMPLETE;
1141	goto out;
1142drop:
1143	ret = SVC_DROP;
1144out:
1145	if (rsci)
1146		cache_put(&rsci->h, &rsc_cache);
1147	return ret;
1148}
1149
1150static __be32 *
1151svcauth_gss_prepare_to_wrap(struct xdr_buf *resbuf, struct gss_svc_data *gsd)
1152{
1153	__be32 *p;
1154	u32 verf_len;
1155
1156	p = gsd->verf_start;
1157	gsd->verf_start = NULL;
1158
1159	/* If the reply stat is nonzero, don't wrap: */
1160	if (*(p-1) != rpc_success)
1161		return NULL;
1162	/* Skip the verifier: */
1163	p += 1;
1164	verf_len = ntohl(*p++);
1165	p += XDR_QUADLEN(verf_len);
1166	/* move accept_stat to right place: */
1167	memcpy(p, p + 2, 4);
1168	/* Also don't wrap if the accept stat is nonzero: */
1169	if (*p != rpc_success) {
1170		resbuf->head[0].iov_len -= 2 * 4;
1171		return NULL;
1172	}
1173	p++;
1174	return p;
1175}
1176
1177static inline int
1178svcauth_gss_wrap_resp_integ(struct svc_rqst *rqstp)
1179{
1180	struct gss_svc_data *gsd = (struct gss_svc_data *)rqstp->rq_auth_data;
1181	struct rpc_gss_wire_cred *gc = &gsd->clcred;
1182	struct xdr_buf *resbuf = &rqstp->rq_res;
1183	struct xdr_buf integ_buf;
1184	struct xdr_netobj mic;
1185	struct kvec *resv;
1186	__be32 *p;
1187	int integ_offset, integ_len;
1188	int stat = -EINVAL;
1189
1190	p = svcauth_gss_prepare_to_wrap(resbuf, gsd);
1191	if (p == NULL)
1192		goto out;
1193	integ_offset = (u8 *)(p + 1) - (u8 *)resbuf->head[0].iov_base;
1194	integ_len = resbuf->len - integ_offset;
1195	BUG_ON(integ_len % 4);
1196	*p++ = htonl(integ_len);
1197	*p++ = htonl(gc->gc_seq);
1198	if (xdr_buf_subsegment(resbuf, &integ_buf, integ_offset,
1199				integ_len))
1200		BUG();
1201	if (resbuf->tail[0].iov_base == NULL) {
1202		if (resbuf->head[0].iov_len + RPC_MAX_AUTH_SIZE > PAGE_SIZE)
1203			goto out_err;
1204		resbuf->tail[0].iov_base = resbuf->head[0].iov_base
1205						+ resbuf->head[0].iov_len;
1206		resbuf->tail[0].iov_len = 0;
1207		resv = &resbuf->tail[0];
1208	} else {
1209		resv = &resbuf->tail[0];
1210	}
1211	mic.data = (u8 *)resv->iov_base + resv->iov_len + 4;
1212	if (gss_get_mic(gsd->rsci->mechctx, &integ_buf, &mic))
1213		goto out_err;
1214	svc_putnl(resv, mic.len);
1215	memset(mic.data + mic.len, 0,
1216			round_up_to_quad(mic.len) - mic.len);
1217	resv->iov_len += XDR_QUADLEN(mic.len) << 2;
1218	/* not strictly required: */
1219	resbuf->len += XDR_QUADLEN(mic.len) << 2;
1220	BUG_ON(resv->iov_len > PAGE_SIZE);
1221out:
1222	stat = 0;
1223out_err:
1224	return stat;
1225}
1226
1227static inline int
1228svcauth_gss_wrap_resp_priv(struct svc_rqst *rqstp)
1229{
1230	struct gss_svc_data *gsd = (struct gss_svc_data *)rqstp->rq_auth_data;
1231	struct rpc_gss_wire_cred *gc = &gsd->clcred;
1232	struct xdr_buf *resbuf = &rqstp->rq_res;
1233	struct page **inpages = NULL;
1234	__be32 *p, *len;
1235	int offset;
1236	int pad;
1237
1238	p = svcauth_gss_prepare_to_wrap(resbuf, gsd);
1239	if (p == NULL)
1240		return 0;
1241	len = p++;
1242	offset = (u8 *)p - (u8 *)resbuf->head[0].iov_base;
1243	*p++ = htonl(gc->gc_seq);
1244	inpages = resbuf->pages;
1245	if (resbuf->tail[0].iov_base) {
1246		BUG_ON(resbuf->tail[0].iov_base >= resbuf->head[0].iov_base
1247							+ PAGE_SIZE);
1248		BUG_ON(resbuf->tail[0].iov_base < resbuf->head[0].iov_base);
1249		if (resbuf->tail[0].iov_len + resbuf->head[0].iov_len
1250				+ 2 * RPC_MAX_AUTH_SIZE > PAGE_SIZE)
1251			return -ENOMEM;
1252		memmove(resbuf->tail[0].iov_base + RPC_MAX_AUTH_SIZE,
1253			resbuf->tail[0].iov_base,
1254			resbuf->tail[0].iov_len);
1255		resbuf->tail[0].iov_base += RPC_MAX_AUTH_SIZE;
1256	}
1257	if (resbuf->tail[0].iov_base == NULL) {
1258		if (resbuf->head[0].iov_len + 2*RPC_MAX_AUTH_SIZE > PAGE_SIZE)
1259			return -ENOMEM;
1260		resbuf->tail[0].iov_base = resbuf->head[0].iov_base
1261			+ resbuf->head[0].iov_len + RPC_MAX_AUTH_SIZE;
1262		resbuf->tail[0].iov_len = 0;
1263	}
1264	if (gss_wrap(gsd->rsci->mechctx, offset, resbuf, inpages))
1265		return -ENOMEM;
1266	*len = htonl(resbuf->len - offset);
1267	pad = 3 - ((resbuf->len - offset - 1)&3);
1268	p = (__be32 *)(resbuf->tail[0].iov_base + resbuf->tail[0].iov_len);
1269	memset(p, 0, pad);
1270	resbuf->tail[0].iov_len += pad;
1271	resbuf->len += pad;
1272	return 0;
1273}
1274
1275static int
1276svcauth_gss_release(struct svc_rqst *rqstp)
1277{
1278	struct gss_svc_data *gsd = (struct gss_svc_data *)rqstp->rq_auth_data;
1279	struct rpc_gss_wire_cred *gc = &gsd->clcred;
1280	struct xdr_buf *resbuf = &rqstp->rq_res;
1281	int stat = -EINVAL;
1282
1283	if (gc->gc_proc != RPC_GSS_PROC_DATA)
1284		goto out;
1285	/* Release can be called twice, but we only wrap once. */
1286	if (gsd->verf_start == NULL)
1287		goto out;
1288	/* normally not set till svc_send, but we need it here: */
1289	resbuf->len = total_buf_len(resbuf);
1290	switch (gc->gc_svc) {
1291	case RPC_GSS_SVC_NONE:
1292		break;
1293	case RPC_GSS_SVC_INTEGRITY:
1294		stat = svcauth_gss_wrap_resp_integ(rqstp);
1295		if (stat)
1296			goto out_err;
1297		break;
1298	case RPC_GSS_SVC_PRIVACY:
1299		stat = svcauth_gss_wrap_resp_priv(rqstp);
1300		if (stat)
1301			goto out_err;
1302		break;
1303	default:
1304		goto out_err;
1305	}
1306
1307out:
1308	stat = 0;
1309out_err:
1310	if (rqstp->rq_client)
1311		auth_domain_put(rqstp->rq_client);
1312	rqstp->rq_client = NULL;
1313	if (rqstp->rq_cred.cr_group_info)
1314		put_group_info(rqstp->rq_cred.cr_group_info);
1315	rqstp->rq_cred.cr_group_info = NULL;
1316	if (gsd->rsci)
1317		cache_put(&gsd->rsci->h, &rsc_cache);
1318	gsd->rsci = NULL;
1319
1320	return stat;
1321}
1322
1323static void
1324svcauth_gss_domain_release(struct auth_domain *dom)
1325{
1326	struct gss_domain *gd = container_of(dom, struct gss_domain, h);
1327
1328	kfree(dom->name);
1329	kfree(gd);
1330}
1331
1332static struct auth_ops svcauthops_gss = {
1333	.name		= "rpcsec_gss",
1334	.owner		= THIS_MODULE,
1335	.flavour	= RPC_AUTH_GSS,
1336	.accept		= svcauth_gss_accept,
1337	.release	= svcauth_gss_release,
1338	.domain_release = svcauth_gss_domain_release,
1339	.set_client	= svcauth_gss_set_client,
1340};
1341
1342int
1343gss_svc_init(void)
1344{
1345	int rv = svc_auth_register(RPC_AUTH_GSS, &svcauthops_gss);
1346	if (rv == 0) {
1347		cache_register(&rsc_cache);
1348		cache_register(&rsi_cache);
1349	}
1350	return rv;
1351}
1352
1353void
1354gss_svc_shutdown(void)
1355{
1356	if (cache_unregister(&rsc_cache))
1357		printk(KERN_ERR "auth_rpcgss: failed to unregister rsc cache\n");
1358	if (cache_unregister(&rsi_cache))
1359		printk(KERN_ERR "auth_rpcgss: failed to unregister rsi cache\n");
1360	svc_auth_unregister(RPC_AUTH_GSS);
1361}
1362