1 // SPDX-License-Identifier: BSD-3-Clause
3 * linux/net/sunrpc/auth_gss/auth_gss.c
5 * RPCSEC_GSS client authentication.
7 * Copyright (c) 2000 The Regents of the University of Michigan.
10 * Dug Song <dugsong@monkey.org>
11 * Andy Adamson <andros@umich.edu>
14 #include <linux/module.h>
15 #include <linux/init.h>
16 #include <linux/types.h>
17 #include <linux/slab.h>
18 #include <linux/sched.h>
19 #include <linux/pagemap.h>
20 #include <linux/sunrpc/clnt.h>
21 #include <linux/sunrpc/auth.h>
22 #include <linux/sunrpc/auth_gss.h>
23 #include <linux/sunrpc/svcauth_gss.h>
24 #include <linux/sunrpc/gss_err.h>
25 #include <linux/workqueue.h>
26 #include <linux/sunrpc/rpc_pipe_fs.h>
27 #include <linux/sunrpc/gss_api.h>
28 #include <linux/uaccess.h>
29 #include <linux/hashtable.h>
33 #include <trace/events/rpcgss.h>
35 static const struct rpc_authops authgss_ops
;
37 static const struct rpc_credops gss_credops
;
38 static const struct rpc_credops gss_nullops
;
40 #define GSS_RETRY_EXPIRED 5
41 static unsigned int gss_expired_cred_retry_delay
= GSS_RETRY_EXPIRED
;
43 #define GSS_KEY_EXPIRE_TIMEO 240
44 static unsigned int gss_key_expire_timeo
= GSS_KEY_EXPIRE_TIMEO
;
46 #if IS_ENABLED(CONFIG_SUNRPC_DEBUG)
47 # define RPCDBG_FACILITY RPCDBG_AUTH
50 #define GSS_CRED_SLACK (RPC_MAX_AUTH_SIZE * 2)
51 /* length of a krb5 verifier (48), plus data added before arguments when
52 * using integrity (two 4-byte integers): */
53 #define GSS_VERF_SLACK 100
55 static DEFINE_HASHTABLE(gss_auth_hash_table
, 4);
56 static DEFINE_SPINLOCK(gss_auth_hash_lock
);
59 struct rpc_pipe_dir_object pdo
;
60 struct rpc_pipe
*pipe
;
61 struct rpc_clnt
*clnt
;
68 struct hlist_node hash
;
69 struct rpc_auth rpc_auth
;
70 struct gss_api_mech
*mech
;
71 enum rpc_gss_svc service
;
72 struct rpc_clnt
*client
;
75 * There are two upcall pipes; dentry[1], named "gssd", is used
76 * for the new text-based upcall; dentry[0] is named after the
77 * mechanism (for example, "krb5") and exists for
78 * backwards-compatibility with older gssd's.
80 struct gss_pipe
*gss_pipe
[2];
81 const char *target_name
;
84 /* pipe_version >= 0 if and only if someone has a pipe open. */
85 static DEFINE_SPINLOCK(pipe_version_lock
);
86 static struct rpc_wait_queue pipe_version_rpc_waitqueue
;
87 static DECLARE_WAIT_QUEUE_HEAD(pipe_version_waitqueue
);
88 static void gss_put_auth(struct gss_auth
*gss_auth
);
90 static void gss_free_ctx(struct gss_cl_ctx
*);
91 static const struct rpc_pipe_ops gss_upcall_ops_v0
;
92 static const struct rpc_pipe_ops gss_upcall_ops_v1
;
94 static inline struct gss_cl_ctx
*
95 gss_get_ctx(struct gss_cl_ctx
*ctx
)
97 refcount_inc(&ctx
->count
);
102 gss_put_ctx(struct gss_cl_ctx
*ctx
)
104 if (refcount_dec_and_test(&ctx
->count
))
109 * called by gss_upcall_callback and gss_create_upcall in order
110 * to set the gss context. The actual exchange of an old context
111 * and a new one is protected by the pipe->lock.
114 gss_cred_set_ctx(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
)
116 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
118 if (!test_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
))
121 rcu_assign_pointer(gss_cred
->gc_ctx
, ctx
);
122 set_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
123 smp_mb__before_atomic();
124 clear_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
);
128 simple_get_bytes(const void *p
, const void *end
, void *res
, size_t len
)
130 const void *q
= (const void *)((const char *)p
+ len
);
131 if (unlikely(q
> end
|| q
< p
))
132 return ERR_PTR(-EFAULT
);
137 static inline const void *
138 simple_get_netobj(const void *p
, const void *end
, struct xdr_netobj
*dest
)
143 p
= simple_get_bytes(p
, end
, &len
, sizeof(len
));
146 q
= (const void *)((const char *)p
+ len
);
147 if (unlikely(q
> end
|| q
< p
))
148 return ERR_PTR(-EFAULT
);
149 dest
->data
= kmemdup(p
, len
, GFP_NOFS
);
150 if (unlikely(dest
->data
== NULL
))
151 return ERR_PTR(-ENOMEM
);
156 static struct gss_cl_ctx
*
157 gss_cred_get_ctx(struct rpc_cred
*cred
)
159 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
160 struct gss_cl_ctx
*ctx
= NULL
;
163 ctx
= rcu_dereference(gss_cred
->gc_ctx
);
170 static struct gss_cl_ctx
*
171 gss_alloc_context(void)
173 struct gss_cl_ctx
*ctx
;
175 ctx
= kzalloc(sizeof(*ctx
), GFP_NOFS
);
177 ctx
->gc_proc
= RPC_GSS_PROC_DATA
;
178 ctx
->gc_seq
= 1; /* NetApp 6.4R1 doesn't accept seq. no. 0 */
179 spin_lock_init(&ctx
->gc_seq_lock
);
180 refcount_set(&ctx
->count
,1);
185 #define GSSD_MIN_TIMEOUT (60 * 60)
187 gss_fill_context(const void *p
, const void *end
, struct gss_cl_ctx
*ctx
, struct gss_api_mech
*gm
)
191 unsigned int timeout
;
192 unsigned long now
= jiffies
;
196 /* First unsigned int gives the remaining lifetime in seconds of the
197 * credential - e.g. the remaining TGT lifetime for Kerberos or
198 * the -t value passed to GSSD.
200 p
= simple_get_bytes(p
, end
, &timeout
, sizeof(timeout
));
204 timeout
= GSSD_MIN_TIMEOUT
;
205 ctx
->gc_expiry
= now
+ ((unsigned long)timeout
* HZ
);
206 /* Sequence number window. Determines the maximum number of
207 * simultaneous requests
209 p
= simple_get_bytes(p
, end
, &window_size
, sizeof(window_size
));
212 ctx
->gc_win
= window_size
;
213 /* gssd signals an error by passing ctx->gc_win = 0: */
214 if (ctx
->gc_win
== 0) {
216 * in which case, p points to an error code. Anything other
217 * than -EKEYEXPIRED gets converted to -EACCES.
219 p
= simple_get_bytes(p
, end
, &ret
, sizeof(ret
));
221 p
= (ret
== -EKEYEXPIRED
) ? ERR_PTR(-EKEYEXPIRED
) :
225 /* copy the opaque wire context */
226 p
= simple_get_netobj(p
, end
, &ctx
->gc_wire_ctx
);
229 /* import the opaque security context */
230 p
= simple_get_bytes(p
, end
, &seclen
, sizeof(seclen
));
233 q
= (const void *)((const char *)p
+ seclen
);
234 if (unlikely(q
> end
|| q
< p
)) {
235 p
= ERR_PTR(-EFAULT
);
238 ret
= gss_import_sec_context(p
, seclen
, gm
, &ctx
->gc_gss_ctx
, NULL
, GFP_NOFS
);
240 trace_rpcgss_import_ctx(ret
);
245 /* is there any trailing data? */
251 /* pull in acceptor name (if there is one) */
252 p
= simple_get_netobj(q
, end
, &ctx
->gc_acceptor
);
256 trace_rpcgss_context(ctx
->gc_expiry
, now
, timeout
,
257 ctx
->gc_acceptor
.len
, ctx
->gc_acceptor
.data
);
262 /* XXX: Need some documentation about why UPCALL_BUF_LEN is so small.
263 * Is user space expecting no more than UPCALL_BUF_LEN bytes?
264 * Note that there are now _two_ NI_MAXHOST sized data items
265 * being passed in this string.
267 #define UPCALL_BUF_LEN 256
269 struct gss_upcall_msg
{
272 const char *service_name
;
273 struct rpc_pipe_msg msg
;
274 struct list_head list
;
275 struct gss_auth
*auth
;
276 struct rpc_pipe
*pipe
;
277 struct rpc_wait_queue rpc_waitqueue
;
278 wait_queue_head_t waitqueue
;
279 struct gss_cl_ctx
*ctx
;
280 char databuf
[UPCALL_BUF_LEN
];
283 static int get_pipe_version(struct net
*net
)
285 struct sunrpc_net
*sn
= net_generic(net
, sunrpc_net_id
);
288 spin_lock(&pipe_version_lock
);
289 if (sn
->pipe_version
>= 0) {
290 atomic_inc(&sn
->pipe_users
);
291 ret
= sn
->pipe_version
;
294 spin_unlock(&pipe_version_lock
);
298 static void put_pipe_version(struct net
*net
)
300 struct sunrpc_net
*sn
= net_generic(net
, sunrpc_net_id
);
302 if (atomic_dec_and_lock(&sn
->pipe_users
, &pipe_version_lock
)) {
303 sn
->pipe_version
= -1;
304 spin_unlock(&pipe_version_lock
);
309 gss_release_msg(struct gss_upcall_msg
*gss_msg
)
311 struct net
*net
= gss_msg
->auth
->net
;
312 if (!refcount_dec_and_test(&gss_msg
->count
))
314 put_pipe_version(net
);
315 BUG_ON(!list_empty(&gss_msg
->list
));
316 if (gss_msg
->ctx
!= NULL
)
317 gss_put_ctx(gss_msg
->ctx
);
318 rpc_destroy_wait_queue(&gss_msg
->rpc_waitqueue
);
319 gss_put_auth(gss_msg
->auth
);
320 kfree_const(gss_msg
->service_name
);
324 static struct gss_upcall_msg
*
325 __gss_find_upcall(struct rpc_pipe
*pipe
, kuid_t uid
, const struct gss_auth
*auth
)
327 struct gss_upcall_msg
*pos
;
328 list_for_each_entry(pos
, &pipe
->in_downcall
, list
) {
329 if (!uid_eq(pos
->uid
, uid
))
331 if (auth
&& pos
->auth
->service
!= auth
->service
)
333 refcount_inc(&pos
->count
);
339 /* Try to add an upcall to the pipefs queue.
340 * If an upcall owned by our uid already exists, then we return a reference
341 * to that upcall instead of adding the new upcall.
343 static inline struct gss_upcall_msg
*
344 gss_add_msg(struct gss_upcall_msg
*gss_msg
)
346 struct rpc_pipe
*pipe
= gss_msg
->pipe
;
347 struct gss_upcall_msg
*old
;
349 spin_lock(&pipe
->lock
);
350 old
= __gss_find_upcall(pipe
, gss_msg
->uid
, gss_msg
->auth
);
352 refcount_inc(&gss_msg
->count
);
353 list_add(&gss_msg
->list
, &pipe
->in_downcall
);
356 spin_unlock(&pipe
->lock
);
361 __gss_unhash_msg(struct gss_upcall_msg
*gss_msg
)
363 list_del_init(&gss_msg
->list
);
364 rpc_wake_up_status(&gss_msg
->rpc_waitqueue
, gss_msg
->msg
.errno
);
365 wake_up_all(&gss_msg
->waitqueue
);
366 refcount_dec(&gss_msg
->count
);
370 gss_unhash_msg(struct gss_upcall_msg
*gss_msg
)
372 struct rpc_pipe
*pipe
= gss_msg
->pipe
;
374 if (list_empty(&gss_msg
->list
))
376 spin_lock(&pipe
->lock
);
377 if (!list_empty(&gss_msg
->list
))
378 __gss_unhash_msg(gss_msg
);
379 spin_unlock(&pipe
->lock
);
383 gss_handle_downcall_result(struct gss_cred
*gss_cred
, struct gss_upcall_msg
*gss_msg
)
385 switch (gss_msg
->msg
.errno
) {
387 if (gss_msg
->ctx
== NULL
)
389 clear_bit(RPCAUTH_CRED_NEGATIVE
, &gss_cred
->gc_base
.cr_flags
);
390 gss_cred_set_ctx(&gss_cred
->gc_base
, gss_msg
->ctx
);
393 set_bit(RPCAUTH_CRED_NEGATIVE
, &gss_cred
->gc_base
.cr_flags
);
395 gss_cred
->gc_upcall_timestamp
= jiffies
;
396 gss_cred
->gc_upcall
= NULL
;
397 rpc_wake_up_status(&gss_msg
->rpc_waitqueue
, gss_msg
->msg
.errno
);
401 gss_upcall_callback(struct rpc_task
*task
)
403 struct gss_cred
*gss_cred
= container_of(task
->tk_rqstp
->rq_cred
,
404 struct gss_cred
, gc_base
);
405 struct gss_upcall_msg
*gss_msg
= gss_cred
->gc_upcall
;
406 struct rpc_pipe
*pipe
= gss_msg
->pipe
;
408 spin_lock(&pipe
->lock
);
409 gss_handle_downcall_result(gss_cred
, gss_msg
);
410 spin_unlock(&pipe
->lock
);
411 task
->tk_status
= gss_msg
->msg
.errno
;
412 gss_release_msg(gss_msg
);
415 static void gss_encode_v0_msg(struct gss_upcall_msg
*gss_msg
,
416 const struct cred
*cred
)
418 struct user_namespace
*userns
= cred
->user_ns
;
420 uid_t uid
= from_kuid_munged(userns
, gss_msg
->uid
);
421 memcpy(gss_msg
->databuf
, &uid
, sizeof(uid
));
422 gss_msg
->msg
.data
= gss_msg
->databuf
;
423 gss_msg
->msg
.len
= sizeof(uid
);
425 BUILD_BUG_ON(sizeof(uid
) > sizeof(gss_msg
->databuf
));
429 gss_v0_upcall(struct file
*file
, struct rpc_pipe_msg
*msg
,
430 char __user
*buf
, size_t buflen
)
432 struct gss_upcall_msg
*gss_msg
= container_of(msg
,
433 struct gss_upcall_msg
,
435 if (msg
->copied
== 0)
436 gss_encode_v0_msg(gss_msg
, file
->f_cred
);
437 return rpc_pipe_generic_upcall(file
, msg
, buf
, buflen
);
440 static int gss_encode_v1_msg(struct gss_upcall_msg
*gss_msg
,
441 const char *service_name
,
442 const char *target_name
,
443 const struct cred
*cred
)
445 struct user_namespace
*userns
= cred
->user_ns
;
446 struct gss_api_mech
*mech
= gss_msg
->auth
->mech
;
447 char *p
= gss_msg
->databuf
;
448 size_t buflen
= sizeof(gss_msg
->databuf
);
451 len
= scnprintf(p
, buflen
, "mech=%s uid=%d", mech
->gm_name
,
452 from_kuid_munged(userns
, gss_msg
->uid
));
455 gss_msg
->msg
.len
= len
;
458 * target= is a full service principal that names the remote
459 * identity that we are authenticating to.
462 len
= scnprintf(p
, buflen
, " target=%s", target_name
);
465 gss_msg
->msg
.len
+= len
;
469 * gssd uses service= and srchost= to select a matching key from
470 * the system's keytab to use as the source principal.
472 * service= is the service name part of the source principal,
473 * or "*" (meaning choose any).
475 * srchost= is the hostname part of the source principal. When
476 * not provided, gssd uses the local hostname.
479 char *c
= strchr(service_name
, '@');
482 len
= scnprintf(p
, buflen
, " service=%s",
485 len
= scnprintf(p
, buflen
,
486 " service=%.*s srchost=%s",
487 (int)(c
- service_name
),
488 service_name
, c
+ 1);
491 gss_msg
->msg
.len
+= len
;
494 if (mech
->gm_upcall_enctypes
) {
495 len
= scnprintf(p
, buflen
, " enctypes=%s",
496 mech
->gm_upcall_enctypes
);
499 gss_msg
->msg
.len
+= len
;
501 trace_rpcgss_upcall_msg(gss_msg
->databuf
);
502 len
= scnprintf(p
, buflen
, "\n");
505 gss_msg
->msg
.len
+= len
;
506 gss_msg
->msg
.data
= gss_msg
->databuf
;
514 gss_v1_upcall(struct file
*file
, struct rpc_pipe_msg
*msg
,
515 char __user
*buf
, size_t buflen
)
517 struct gss_upcall_msg
*gss_msg
= container_of(msg
,
518 struct gss_upcall_msg
,
521 if (msg
->copied
== 0) {
522 err
= gss_encode_v1_msg(gss_msg
,
523 gss_msg
->service_name
,
524 gss_msg
->auth
->target_name
,
529 return rpc_pipe_generic_upcall(file
, msg
, buf
, buflen
);
532 static struct gss_upcall_msg
*
533 gss_alloc_msg(struct gss_auth
*gss_auth
,
534 kuid_t uid
, const char *service_name
)
536 struct gss_upcall_msg
*gss_msg
;
540 gss_msg
= kzalloc(sizeof(*gss_msg
), GFP_NOFS
);
543 vers
= get_pipe_version(gss_auth
->net
);
547 gss_msg
->pipe
= gss_auth
->gss_pipe
[vers
]->pipe
;
548 INIT_LIST_HEAD(&gss_msg
->list
);
549 rpc_init_wait_queue(&gss_msg
->rpc_waitqueue
, "RPCSEC_GSS upcall waitq");
550 init_waitqueue_head(&gss_msg
->waitqueue
);
551 refcount_set(&gss_msg
->count
, 1);
553 gss_msg
->auth
= gss_auth
;
554 kref_get(&gss_auth
->kref
);
556 gss_msg
->service_name
= kstrdup_const(service_name
, GFP_NOFS
);
557 if (!gss_msg
->service_name
) {
559 goto err_put_pipe_version
;
563 err_put_pipe_version
:
564 put_pipe_version(gss_auth
->net
);
571 static struct gss_upcall_msg
*
572 gss_setup_upcall(struct gss_auth
*gss_auth
, struct rpc_cred
*cred
)
574 struct gss_cred
*gss_cred
= container_of(cred
,
575 struct gss_cred
, gc_base
);
576 struct gss_upcall_msg
*gss_new
, *gss_msg
;
577 kuid_t uid
= cred
->cr_cred
->fsuid
;
579 gss_new
= gss_alloc_msg(gss_auth
, uid
, gss_cred
->gc_principal
);
582 gss_msg
= gss_add_msg(gss_new
);
583 if (gss_msg
== gss_new
) {
585 refcount_inc(&gss_msg
->count
);
586 res
= rpc_queue_upcall(gss_new
->pipe
, &gss_new
->msg
);
588 gss_unhash_msg(gss_new
);
589 refcount_dec(&gss_msg
->count
);
590 gss_release_msg(gss_new
);
591 gss_msg
= ERR_PTR(res
);
594 gss_release_msg(gss_new
);
598 static void warn_gssd(void)
600 dprintk("AUTH_GSS upcall failed. Please check user daemon is running.\n");
604 gss_refresh_upcall(struct rpc_task
*task
)
606 struct rpc_cred
*cred
= task
->tk_rqstp
->rq_cred
;
607 struct gss_auth
*gss_auth
= container_of(cred
->cr_auth
,
608 struct gss_auth
, rpc_auth
);
609 struct gss_cred
*gss_cred
= container_of(cred
,
610 struct gss_cred
, gc_base
);
611 struct gss_upcall_msg
*gss_msg
;
612 struct rpc_pipe
*pipe
;
615 gss_msg
= gss_setup_upcall(gss_auth
, cred
);
616 if (PTR_ERR(gss_msg
) == -EAGAIN
) {
617 /* XXX: warning on the first, under the assumption we
618 * shouldn't normally hit this case on a refresh. */
620 rpc_sleep_on_timeout(&pipe_version_rpc_waitqueue
,
621 task
, NULL
, jiffies
+ (15 * HZ
));
625 if (IS_ERR(gss_msg
)) {
626 err
= PTR_ERR(gss_msg
);
629 pipe
= gss_msg
->pipe
;
630 spin_lock(&pipe
->lock
);
631 if (gss_cred
->gc_upcall
!= NULL
)
632 rpc_sleep_on(&gss_cred
->gc_upcall
->rpc_waitqueue
, task
, NULL
);
633 else if (gss_msg
->ctx
== NULL
&& gss_msg
->msg
.errno
>= 0) {
634 gss_cred
->gc_upcall
= gss_msg
;
635 /* gss_upcall_callback will release the reference to gss_upcall_msg */
636 refcount_inc(&gss_msg
->count
);
637 rpc_sleep_on(&gss_msg
->rpc_waitqueue
, task
, gss_upcall_callback
);
639 gss_handle_downcall_result(gss_cred
, gss_msg
);
640 err
= gss_msg
->msg
.errno
;
642 spin_unlock(&pipe
->lock
);
643 gss_release_msg(gss_msg
);
645 trace_rpcgss_upcall_result(from_kuid(&init_user_ns
,
646 cred
->cr_cred
->fsuid
), err
);
651 gss_create_upcall(struct gss_auth
*gss_auth
, struct gss_cred
*gss_cred
)
653 struct net
*net
= gss_auth
->net
;
654 struct sunrpc_net
*sn
= net_generic(net
, sunrpc_net_id
);
655 struct rpc_pipe
*pipe
;
656 struct rpc_cred
*cred
= &gss_cred
->gc_base
;
657 struct gss_upcall_msg
*gss_msg
;
663 /* if gssd is down, just skip upcalling altogether */
664 if (!gssd_running(net
)) {
669 gss_msg
= gss_setup_upcall(gss_auth
, cred
);
670 if (PTR_ERR(gss_msg
) == -EAGAIN
) {
671 err
= wait_event_interruptible_timeout(pipe_version_waitqueue
,
672 sn
->pipe_version
>= 0, 15 * HZ
);
673 if (sn
->pipe_version
< 0) {
681 if (IS_ERR(gss_msg
)) {
682 err
= PTR_ERR(gss_msg
);
685 pipe
= gss_msg
->pipe
;
687 prepare_to_wait(&gss_msg
->waitqueue
, &wait
, TASK_KILLABLE
);
688 spin_lock(&pipe
->lock
);
689 if (gss_msg
->ctx
!= NULL
|| gss_msg
->msg
.errno
< 0) {
692 spin_unlock(&pipe
->lock
);
693 if (fatal_signal_pending(current
)) {
700 gss_cred_set_ctx(cred
, gss_msg
->ctx
);
702 err
= gss_msg
->msg
.errno
;
703 spin_unlock(&pipe
->lock
);
705 finish_wait(&gss_msg
->waitqueue
, &wait
);
706 gss_release_msg(gss_msg
);
708 trace_rpcgss_upcall_result(from_kuid(&init_user_ns
,
709 cred
->cr_cred
->fsuid
), err
);
713 #define MSG_BUF_MAXSIZE 1024
716 gss_pipe_downcall(struct file
*filp
, const char __user
*src
, size_t mlen
)
720 struct gss_upcall_msg
*gss_msg
;
721 struct rpc_pipe
*pipe
= RPC_I(file_inode(filp
))->pipe
;
722 struct gss_cl_ctx
*ctx
;
725 ssize_t err
= -EFBIG
;
727 if (mlen
> MSG_BUF_MAXSIZE
)
730 buf
= kmalloc(mlen
, GFP_NOFS
);
735 if (copy_from_user(buf
, src
, mlen
))
738 end
= (const void *)((char *)buf
+ mlen
);
739 p
= simple_get_bytes(buf
, end
, &id
, sizeof(id
));
745 uid
= make_kuid(current_user_ns(), id
);
746 if (!uid_valid(uid
)) {
752 ctx
= gss_alloc_context();
757 /* Find a matching upcall */
758 spin_lock(&pipe
->lock
);
759 gss_msg
= __gss_find_upcall(pipe
, uid
, NULL
);
760 if (gss_msg
== NULL
) {
761 spin_unlock(&pipe
->lock
);
764 list_del_init(&gss_msg
->list
);
765 spin_unlock(&pipe
->lock
);
767 p
= gss_fill_context(p
, end
, ctx
, gss_msg
->auth
->mech
);
773 gss_msg
->msg
.errno
= err
;
780 gss_msg
->msg
.errno
= -EAGAIN
;
783 printk(KERN_CRIT
"%s: bad return from "
784 "gss_fill_context: %zd\n", __func__
, err
);
785 gss_msg
->msg
.errno
= -EIO
;
787 goto err_release_msg
;
789 gss_msg
->ctx
= gss_get_ctx(ctx
);
793 spin_lock(&pipe
->lock
);
794 __gss_unhash_msg(gss_msg
);
795 spin_unlock(&pipe
->lock
);
796 gss_release_msg(gss_msg
);
805 static int gss_pipe_open(struct inode
*inode
, int new_version
)
807 struct net
*net
= inode
->i_sb
->s_fs_info
;
808 struct sunrpc_net
*sn
= net_generic(net
, sunrpc_net_id
);
811 spin_lock(&pipe_version_lock
);
812 if (sn
->pipe_version
< 0) {
813 /* First open of any gss pipe determines the version: */
814 sn
->pipe_version
= new_version
;
815 rpc_wake_up(&pipe_version_rpc_waitqueue
);
816 wake_up(&pipe_version_waitqueue
);
817 } else if (sn
->pipe_version
!= new_version
) {
818 /* Trying to open a pipe of a different version */
822 atomic_inc(&sn
->pipe_users
);
824 spin_unlock(&pipe_version_lock
);
829 static int gss_pipe_open_v0(struct inode
*inode
)
831 return gss_pipe_open(inode
, 0);
834 static int gss_pipe_open_v1(struct inode
*inode
)
836 return gss_pipe_open(inode
, 1);
840 gss_pipe_release(struct inode
*inode
)
842 struct net
*net
= inode
->i_sb
->s_fs_info
;
843 struct rpc_pipe
*pipe
= RPC_I(inode
)->pipe
;
844 struct gss_upcall_msg
*gss_msg
;
847 spin_lock(&pipe
->lock
);
848 list_for_each_entry(gss_msg
, &pipe
->in_downcall
, list
) {
850 if (!list_empty(&gss_msg
->msg
.list
))
852 gss_msg
->msg
.errno
= -EPIPE
;
853 refcount_inc(&gss_msg
->count
);
854 __gss_unhash_msg(gss_msg
);
855 spin_unlock(&pipe
->lock
);
856 gss_release_msg(gss_msg
);
859 spin_unlock(&pipe
->lock
);
861 put_pipe_version(net
);
865 gss_pipe_destroy_msg(struct rpc_pipe_msg
*msg
)
867 struct gss_upcall_msg
*gss_msg
= container_of(msg
, struct gss_upcall_msg
, msg
);
869 if (msg
->errno
< 0) {
870 refcount_inc(&gss_msg
->count
);
871 gss_unhash_msg(gss_msg
);
872 if (msg
->errno
== -ETIMEDOUT
)
874 gss_release_msg(gss_msg
);
876 gss_release_msg(gss_msg
);
879 static void gss_pipe_dentry_destroy(struct dentry
*dir
,
880 struct rpc_pipe_dir_object
*pdo
)
882 struct gss_pipe
*gss_pipe
= pdo
->pdo_data
;
883 struct rpc_pipe
*pipe
= gss_pipe
->pipe
;
885 if (pipe
->dentry
!= NULL
) {
886 rpc_unlink(pipe
->dentry
);
891 static int gss_pipe_dentry_create(struct dentry
*dir
,
892 struct rpc_pipe_dir_object
*pdo
)
894 struct gss_pipe
*p
= pdo
->pdo_data
;
895 struct dentry
*dentry
;
897 dentry
= rpc_mkpipe_dentry(dir
, p
->name
, p
->clnt
, p
->pipe
);
899 return PTR_ERR(dentry
);
900 p
->pipe
->dentry
= dentry
;
904 static const struct rpc_pipe_dir_object_ops gss_pipe_dir_object_ops
= {
905 .create
= gss_pipe_dentry_create
,
906 .destroy
= gss_pipe_dentry_destroy
,
909 static struct gss_pipe
*gss_pipe_alloc(struct rpc_clnt
*clnt
,
911 const struct rpc_pipe_ops
*upcall_ops
)
916 p
= kmalloc(sizeof(*p
), GFP_KERNEL
);
919 p
->pipe
= rpc_mkpipe_data(upcall_ops
, RPC_PIPE_WAIT_FOR_OPEN
);
920 if (IS_ERR(p
->pipe
)) {
921 err
= PTR_ERR(p
->pipe
);
922 goto err_free_gss_pipe
;
927 rpc_init_pipe_dir_object(&p
->pdo
,
928 &gss_pipe_dir_object_ops
,
937 struct gss_alloc_pdo
{
938 struct rpc_clnt
*clnt
;
940 const struct rpc_pipe_ops
*upcall_ops
;
943 static int gss_pipe_match_pdo(struct rpc_pipe_dir_object
*pdo
, void *data
)
945 struct gss_pipe
*gss_pipe
;
946 struct gss_alloc_pdo
*args
= data
;
948 if (pdo
->pdo_ops
!= &gss_pipe_dir_object_ops
)
950 gss_pipe
= container_of(pdo
, struct gss_pipe
, pdo
);
951 if (strcmp(gss_pipe
->name
, args
->name
) != 0)
953 if (!kref_get_unless_zero(&gss_pipe
->kref
))
958 static struct rpc_pipe_dir_object
*gss_pipe_alloc_pdo(void *data
)
960 struct gss_pipe
*gss_pipe
;
961 struct gss_alloc_pdo
*args
= data
;
963 gss_pipe
= gss_pipe_alloc(args
->clnt
, args
->name
, args
->upcall_ops
);
964 if (!IS_ERR(gss_pipe
))
965 return &gss_pipe
->pdo
;
969 static struct gss_pipe
*gss_pipe_get(struct rpc_clnt
*clnt
,
971 const struct rpc_pipe_ops
*upcall_ops
)
973 struct net
*net
= rpc_net_ns(clnt
);
974 struct rpc_pipe_dir_object
*pdo
;
975 struct gss_alloc_pdo args
= {
978 .upcall_ops
= upcall_ops
,
981 pdo
= rpc_find_or_alloc_pipe_dir_object(net
,
982 &clnt
->cl_pipedir_objects
,
987 return container_of(pdo
, struct gss_pipe
, pdo
);
988 return ERR_PTR(-ENOMEM
);
991 static void __gss_pipe_free(struct gss_pipe
*p
)
993 struct rpc_clnt
*clnt
= p
->clnt
;
994 struct net
*net
= rpc_net_ns(clnt
);
996 rpc_remove_pipe_dir_object(net
,
997 &clnt
->cl_pipedir_objects
,
999 rpc_destroy_pipe_data(p
->pipe
);
1003 static void __gss_pipe_release(struct kref
*kref
)
1005 struct gss_pipe
*p
= container_of(kref
, struct gss_pipe
, kref
);
1010 static void gss_pipe_free(struct gss_pipe
*p
)
1013 kref_put(&p
->kref
, __gss_pipe_release
);
1017 * NOTE: we have the opportunity to use different
1018 * parameters based on the input flavor (which must be a pseudoflavor)
1020 static struct gss_auth
*
1021 gss_create_new(const struct rpc_auth_create_args
*args
, struct rpc_clnt
*clnt
)
1023 rpc_authflavor_t flavor
= args
->pseudoflavor
;
1024 struct gss_auth
*gss_auth
;
1025 struct gss_pipe
*gss_pipe
;
1026 struct rpc_auth
* auth
;
1027 int err
= -ENOMEM
; /* XXX? */
1029 if (!try_module_get(THIS_MODULE
))
1030 return ERR_PTR(err
);
1031 if (!(gss_auth
= kmalloc(sizeof(*gss_auth
), GFP_KERNEL
)))
1033 INIT_HLIST_NODE(&gss_auth
->hash
);
1034 gss_auth
->target_name
= NULL
;
1035 if (args
->target_name
) {
1036 gss_auth
->target_name
= kstrdup(args
->target_name
, GFP_KERNEL
);
1037 if (gss_auth
->target_name
== NULL
)
1040 gss_auth
->client
= clnt
;
1041 gss_auth
->net
= get_net(rpc_net_ns(clnt
));
1043 gss_auth
->mech
= gss_mech_get_by_pseudoflavor(flavor
);
1044 if (!gss_auth
->mech
)
1046 gss_auth
->service
= gss_pseudoflavor_to_service(gss_auth
->mech
, flavor
);
1047 if (gss_auth
->service
== 0)
1049 if (!gssd_running(gss_auth
->net
))
1051 auth
= &gss_auth
->rpc_auth
;
1052 auth
->au_cslack
= GSS_CRED_SLACK
>> 2;
1053 auth
->au_rslack
= GSS_VERF_SLACK
>> 2;
1054 auth
->au_verfsize
= GSS_VERF_SLACK
>> 2;
1055 auth
->au_ralign
= GSS_VERF_SLACK
>> 2;
1057 auth
->au_ops
= &authgss_ops
;
1058 auth
->au_flavor
= flavor
;
1059 if (gss_pseudoflavor_to_datatouch(gss_auth
->mech
, flavor
))
1060 auth
->au_flags
|= RPCAUTH_AUTH_DATATOUCH
;
1061 refcount_set(&auth
->au_count
, 1);
1062 kref_init(&gss_auth
->kref
);
1064 err
= rpcauth_init_credcache(auth
);
1068 * Note: if we created the old pipe first, then someone who
1069 * examined the directory at the right moment might conclude
1070 * that we supported only the old pipe. So we instead create
1071 * the new pipe first.
1073 gss_pipe
= gss_pipe_get(clnt
, "gssd", &gss_upcall_ops_v1
);
1074 if (IS_ERR(gss_pipe
)) {
1075 err
= PTR_ERR(gss_pipe
);
1076 goto err_destroy_credcache
;
1078 gss_auth
->gss_pipe
[1] = gss_pipe
;
1080 gss_pipe
= gss_pipe_get(clnt
, gss_auth
->mech
->gm_name
,
1081 &gss_upcall_ops_v0
);
1082 if (IS_ERR(gss_pipe
)) {
1083 err
= PTR_ERR(gss_pipe
);
1084 goto err_destroy_pipe_1
;
1086 gss_auth
->gss_pipe
[0] = gss_pipe
;
1090 gss_pipe_free(gss_auth
->gss_pipe
[1]);
1091 err_destroy_credcache
:
1092 rpcauth_destroy_credcache(auth
);
1094 gss_mech_put(gss_auth
->mech
);
1096 put_net(gss_auth
->net
);
1098 kfree(gss_auth
->target_name
);
1101 module_put(THIS_MODULE
);
1102 trace_rpcgss_createauth(flavor
, err
);
1103 return ERR_PTR(err
);
1107 gss_free(struct gss_auth
*gss_auth
)
1109 gss_pipe_free(gss_auth
->gss_pipe
[0]);
1110 gss_pipe_free(gss_auth
->gss_pipe
[1]);
1111 gss_mech_put(gss_auth
->mech
);
1112 put_net(gss_auth
->net
);
1113 kfree(gss_auth
->target_name
);
1116 module_put(THIS_MODULE
);
1120 gss_free_callback(struct kref
*kref
)
1122 struct gss_auth
*gss_auth
= container_of(kref
, struct gss_auth
, kref
);
1128 gss_put_auth(struct gss_auth
*gss_auth
)
1130 kref_put(&gss_auth
->kref
, gss_free_callback
);
1134 gss_destroy(struct rpc_auth
*auth
)
1136 struct gss_auth
*gss_auth
= container_of(auth
,
1137 struct gss_auth
, rpc_auth
);
1139 if (hash_hashed(&gss_auth
->hash
)) {
1140 spin_lock(&gss_auth_hash_lock
);
1141 hash_del(&gss_auth
->hash
);
1142 spin_unlock(&gss_auth_hash_lock
);
1145 gss_pipe_free(gss_auth
->gss_pipe
[0]);
1146 gss_auth
->gss_pipe
[0] = NULL
;
1147 gss_pipe_free(gss_auth
->gss_pipe
[1]);
1148 gss_auth
->gss_pipe
[1] = NULL
;
1149 rpcauth_destroy_credcache(auth
);
1151 gss_put_auth(gss_auth
);
1155 * Auths may be shared between rpc clients that were cloned from a
1156 * common client with the same xprt, if they also share the flavor and
1159 * The auth is looked up from the oldest parent sharing the same
1160 * cl_xprt, and the auth itself references only that common parent
1161 * (which is guaranteed to last as long as any of its descendants).
1163 static struct gss_auth
*
1164 gss_auth_find_or_add_hashed(const struct rpc_auth_create_args
*args
,
1165 struct rpc_clnt
*clnt
,
1166 struct gss_auth
*new)
1168 struct gss_auth
*gss_auth
;
1169 unsigned long hashval
= (unsigned long)clnt
;
1171 spin_lock(&gss_auth_hash_lock
);
1172 hash_for_each_possible(gss_auth_hash_table
,
1176 if (gss_auth
->client
!= clnt
)
1178 if (gss_auth
->rpc_auth
.au_flavor
!= args
->pseudoflavor
)
1180 if (gss_auth
->target_name
!= args
->target_name
) {
1181 if (gss_auth
->target_name
== NULL
)
1183 if (args
->target_name
== NULL
)
1185 if (strcmp(gss_auth
->target_name
, args
->target_name
))
1188 if (!refcount_inc_not_zero(&gss_auth
->rpc_auth
.au_count
))
1193 hash_add(gss_auth_hash_table
, &new->hash
, hashval
);
1196 spin_unlock(&gss_auth_hash_lock
);
1200 static struct gss_auth
*
1201 gss_create_hashed(const struct rpc_auth_create_args
*args
,
1202 struct rpc_clnt
*clnt
)
1204 struct gss_auth
*gss_auth
;
1205 struct gss_auth
*new;
1207 gss_auth
= gss_auth_find_or_add_hashed(args
, clnt
, NULL
);
1208 if (gss_auth
!= NULL
)
1210 new = gss_create_new(args
, clnt
);
1213 gss_auth
= gss_auth_find_or_add_hashed(args
, clnt
, new);
1214 if (gss_auth
!= new)
1215 gss_destroy(&new->rpc_auth
);
1220 static struct rpc_auth
*
1221 gss_create(const struct rpc_auth_create_args
*args
, struct rpc_clnt
*clnt
)
1223 struct gss_auth
*gss_auth
;
1224 struct rpc_xprt_switch
*xps
= rcu_access_pointer(clnt
->cl_xpi
.xpi_xpswitch
);
1226 while (clnt
!= clnt
->cl_parent
) {
1227 struct rpc_clnt
*parent
= clnt
->cl_parent
;
1228 /* Find the original parent for this transport */
1229 if (rcu_access_pointer(parent
->cl_xpi
.xpi_xpswitch
) != xps
)
1234 gss_auth
= gss_create_hashed(args
, clnt
);
1235 if (IS_ERR(gss_auth
))
1236 return ERR_CAST(gss_auth
);
1237 return &gss_auth
->rpc_auth
;
1240 static struct gss_cred
*
1241 gss_dup_cred(struct gss_auth
*gss_auth
, struct gss_cred
*gss_cred
)
1243 struct gss_cred
*new;
1245 /* Make a copy of the cred so that we can reference count it */
1246 new = kzalloc(sizeof(*gss_cred
), GFP_NOFS
);
1248 struct auth_cred acred
= {
1249 .cred
= gss_cred
->gc_base
.cr_cred
,
1251 struct gss_cl_ctx
*ctx
=
1252 rcu_dereference_protected(gss_cred
->gc_ctx
, 1);
1254 rpcauth_init_cred(&new->gc_base
, &acred
,
1255 &gss_auth
->rpc_auth
,
1257 new->gc_base
.cr_flags
= 1UL << RPCAUTH_CRED_UPTODATE
;
1258 new->gc_service
= gss_cred
->gc_service
;
1259 new->gc_principal
= gss_cred
->gc_principal
;
1260 kref_get(&gss_auth
->kref
);
1261 rcu_assign_pointer(new->gc_ctx
, ctx
);
1268 * gss_send_destroy_context will cause the RPCSEC_GSS to send a NULL RPC call
1269 * to the server with the GSS control procedure field set to
1270 * RPC_GSS_PROC_DESTROY. This should normally cause the server to release
1271 * all RPCSEC_GSS state associated with that context.
1274 gss_send_destroy_context(struct rpc_cred
*cred
)
1276 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
1277 struct gss_auth
*gss_auth
= container_of(cred
->cr_auth
, struct gss_auth
, rpc_auth
);
1278 struct gss_cl_ctx
*ctx
= rcu_dereference_protected(gss_cred
->gc_ctx
, 1);
1279 struct gss_cred
*new;
1280 struct rpc_task
*task
;
1282 new = gss_dup_cred(gss_auth
, gss_cred
);
1284 ctx
->gc_proc
= RPC_GSS_PROC_DESTROY
;
1286 task
= rpc_call_null(gss_auth
->client
, &new->gc_base
,
1287 RPC_TASK_ASYNC
|RPC_TASK_SOFT
);
1291 put_rpccred(&new->gc_base
);
1295 /* gss_destroy_cred (and gss_free_ctx) are used to clean up after failure
1296 * to create a new cred or context, so they check that things have been
1297 * allocated before freeing them. */
1299 gss_do_free_ctx(struct gss_cl_ctx
*ctx
)
1301 gss_delete_sec_context(&ctx
->gc_gss_ctx
);
1302 kfree(ctx
->gc_wire_ctx
.data
);
1303 kfree(ctx
->gc_acceptor
.data
);
1308 gss_free_ctx_callback(struct rcu_head
*head
)
1310 struct gss_cl_ctx
*ctx
= container_of(head
, struct gss_cl_ctx
, gc_rcu
);
1311 gss_do_free_ctx(ctx
);
1315 gss_free_ctx(struct gss_cl_ctx
*ctx
)
1317 call_rcu(&ctx
->gc_rcu
, gss_free_ctx_callback
);
1321 gss_free_cred(struct gss_cred
*gss_cred
)
1327 gss_free_cred_callback(struct rcu_head
*head
)
1329 struct gss_cred
*gss_cred
= container_of(head
, struct gss_cred
, gc_base
.cr_rcu
);
1330 gss_free_cred(gss_cred
);
1334 gss_destroy_nullcred(struct rpc_cred
*cred
)
1336 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
1337 struct gss_auth
*gss_auth
= container_of(cred
->cr_auth
, struct gss_auth
, rpc_auth
);
1338 struct gss_cl_ctx
*ctx
= rcu_dereference_protected(gss_cred
->gc_ctx
, 1);
1340 RCU_INIT_POINTER(gss_cred
->gc_ctx
, NULL
);
1341 put_cred(cred
->cr_cred
);
1342 call_rcu(&cred
->cr_rcu
, gss_free_cred_callback
);
1345 gss_put_auth(gss_auth
);
1349 gss_destroy_cred(struct rpc_cred
*cred
)
1352 if (test_and_clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
) != 0)
1353 gss_send_destroy_context(cred
);
1354 gss_destroy_nullcred(cred
);
1358 gss_hash_cred(struct auth_cred
*acred
, unsigned int hashbits
)
1360 return hash_64(from_kuid(&init_user_ns
, acred
->cred
->fsuid
), hashbits
);
1364 * Lookup RPCSEC_GSS cred for the current process
1366 static struct rpc_cred
*
1367 gss_lookup_cred(struct rpc_auth
*auth
, struct auth_cred
*acred
, int flags
)
1369 return rpcauth_lookup_credcache(auth
, acred
, flags
, GFP_NOFS
);
1372 static struct rpc_cred
*
1373 gss_create_cred(struct rpc_auth
*auth
, struct auth_cred
*acred
, int flags
, gfp_t gfp
)
1375 struct gss_auth
*gss_auth
= container_of(auth
, struct gss_auth
, rpc_auth
);
1376 struct gss_cred
*cred
= NULL
;
1379 if (!(cred
= kzalloc(sizeof(*cred
), gfp
)))
1382 rpcauth_init_cred(&cred
->gc_base
, acred
, auth
, &gss_credops
);
1384 * Note: in order to force a call to call_refresh(), we deliberately
1385 * fail to flag the credential as RPCAUTH_CRED_UPTODATE.
1387 cred
->gc_base
.cr_flags
= 1UL << RPCAUTH_CRED_NEW
;
1388 cred
->gc_service
= gss_auth
->service
;
1389 cred
->gc_principal
= acred
->principal
;
1390 kref_get(&gss_auth
->kref
);
1391 return &cred
->gc_base
;
1394 return ERR_PTR(err
);
1398 gss_cred_init(struct rpc_auth
*auth
, struct rpc_cred
*cred
)
1400 struct gss_auth
*gss_auth
= container_of(auth
, struct gss_auth
, rpc_auth
);
1401 struct gss_cred
*gss_cred
= container_of(cred
,struct gss_cred
, gc_base
);
1405 err
= gss_create_upcall(gss_auth
, gss_cred
);
1406 } while (err
== -EAGAIN
);
1411 gss_stringify_acceptor(struct rpc_cred
*cred
)
1413 char *string
= NULL
;
1414 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
1415 struct gss_cl_ctx
*ctx
;
1417 struct xdr_netobj
*acceptor
;
1420 ctx
= rcu_dereference(gss_cred
->gc_ctx
);
1424 len
= ctx
->gc_acceptor
.len
;
1427 /* no point if there's no string */
1431 string
= kmalloc(len
+ 1, GFP_KERNEL
);
1436 ctx
= rcu_dereference(gss_cred
->gc_ctx
);
1438 /* did the ctx disappear or was it replaced by one with no acceptor? */
1439 if (!ctx
|| !ctx
->gc_acceptor
.len
) {
1445 acceptor
= &ctx
->gc_acceptor
;
1448 * Did we find a new acceptor that's longer than the original? Allocate
1449 * a longer buffer and try again.
1451 if (len
< acceptor
->len
) {
1452 len
= acceptor
->len
;
1458 memcpy(string
, acceptor
->data
, acceptor
->len
);
1459 string
[acceptor
->len
] = '\0';
1466 * Returns -EACCES if GSS context is NULL or will expire within the
1467 * timeout (miliseconds)
1470 gss_key_timeout(struct rpc_cred
*rc
)
1472 struct gss_cred
*gss_cred
= container_of(rc
, struct gss_cred
, gc_base
);
1473 struct gss_cl_ctx
*ctx
;
1474 unsigned long timeout
= jiffies
+ (gss_key_expire_timeo
* HZ
);
1478 ctx
= rcu_dereference(gss_cred
->gc_ctx
);
1479 if (!ctx
|| time_after(timeout
, ctx
->gc_expiry
))
1487 gss_match(struct auth_cred
*acred
, struct rpc_cred
*rc
, int flags
)
1489 struct gss_cred
*gss_cred
= container_of(rc
, struct gss_cred
, gc_base
);
1490 struct gss_cl_ctx
*ctx
;
1493 if (test_bit(RPCAUTH_CRED_NEW
, &rc
->cr_flags
))
1495 /* Don't match with creds that have expired. */
1497 ctx
= rcu_dereference(gss_cred
->gc_ctx
);
1498 if (!ctx
|| time_after(jiffies
, ctx
->gc_expiry
)) {
1503 if (!test_bit(RPCAUTH_CRED_UPTODATE
, &rc
->cr_flags
))
1506 if (acred
->principal
!= NULL
) {
1507 if (gss_cred
->gc_principal
== NULL
)
1509 ret
= strcmp(acred
->principal
, gss_cred
->gc_principal
) == 0;
1511 if (gss_cred
->gc_principal
!= NULL
)
1513 ret
= uid_eq(rc
->cr_cred
->fsuid
, acred
->cred
->fsuid
);
1519 * Marshal credentials.
1521 * The expensive part is computing the verifier. We can't cache a
1522 * pre-computed version of the verifier because the seqno, which
1523 * is different every time, is included in the MIC.
1525 static int gss_marshal(struct rpc_task
*task
, struct xdr_stream
*xdr
)
1527 struct rpc_rqst
*req
= task
->tk_rqstp
;
1528 struct rpc_cred
*cred
= req
->rq_cred
;
1529 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
,
1531 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
1532 __be32
*p
, *cred_len
;
1534 struct xdr_netobj mic
;
1536 struct xdr_buf verf_buf
;
1541 p
= xdr_reserve_space(xdr
, 7 * sizeof(*p
) +
1542 ctx
->gc_wire_ctx
.len
);
1544 goto marshal_failed
;
1545 *p
++ = rpc_auth_gss
;
1548 spin_lock(&ctx
->gc_seq_lock
);
1549 req
->rq_seqno
= (ctx
->gc_seq
< MAXSEQ
) ? ctx
->gc_seq
++ : MAXSEQ
;
1550 spin_unlock(&ctx
->gc_seq_lock
);
1551 if (req
->rq_seqno
== MAXSEQ
)
1553 trace_rpcgss_seqno(task
);
1555 *p
++ = cpu_to_be32(RPC_GSS_VERSION
);
1556 *p
++ = cpu_to_be32(ctx
->gc_proc
);
1557 *p
++ = cpu_to_be32(req
->rq_seqno
);
1558 *p
++ = cpu_to_be32(gss_cred
->gc_service
);
1559 p
= xdr_encode_netobj(p
, &ctx
->gc_wire_ctx
);
1560 *cred_len
= cpu_to_be32((p
- (cred_len
+ 1)) << 2);
1564 /* We compute the checksum for the verifier over the xdr-encoded bytes
1565 * starting with the xid and ending at the end of the credential: */
1566 iov
.iov_base
= req
->rq_snd_buf
.head
[0].iov_base
;
1567 iov
.iov_len
= (u8
*)p
- (u8
*)iov
.iov_base
;
1568 xdr_buf_from_iov(&iov
, &verf_buf
);
1570 p
= xdr_reserve_space(xdr
, sizeof(*p
));
1572 goto marshal_failed
;
1573 *p
++ = rpc_auth_gss
;
1574 mic
.data
= (u8
*)(p
+ 1);
1575 maj_stat
= gss_get_mic(ctx
->gc_gss_ctx
, &verf_buf
, &mic
);
1576 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1578 else if (maj_stat
!= 0)
1580 if (xdr_stream_encode_opaque_inline(xdr
, (void **)&p
, mic
.len
) < 0)
1581 goto marshal_failed
;
1587 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1588 status
= -EKEYEXPIRED
;
1594 trace_rpcgss_get_mic(task
, maj_stat
);
1599 static int gss_renew_cred(struct rpc_task
*task
)
1601 struct rpc_cred
*oldcred
= task
->tk_rqstp
->rq_cred
;
1602 struct gss_cred
*gss_cred
= container_of(oldcred
,
1605 struct rpc_auth
*auth
= oldcred
->cr_auth
;
1606 struct auth_cred acred
= {
1607 .cred
= oldcred
->cr_cred
,
1608 .principal
= gss_cred
->gc_principal
,
1610 struct rpc_cred
*new;
1612 new = gss_lookup_cred(auth
, &acred
, RPCAUTH_LOOKUP_NEW
);
1614 return PTR_ERR(new);
1615 task
->tk_rqstp
->rq_cred
= new;
1616 put_rpccred(oldcred
);
1620 static int gss_cred_is_negative_entry(struct rpc_cred
*cred
)
1622 if (test_bit(RPCAUTH_CRED_NEGATIVE
, &cred
->cr_flags
)) {
1623 unsigned long now
= jiffies
;
1624 unsigned long begin
, expire
;
1625 struct gss_cred
*gss_cred
;
1627 gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
1628 begin
= gss_cred
->gc_upcall_timestamp
;
1629 expire
= begin
+ gss_expired_cred_retry_delay
* HZ
;
1631 if (time_in_range_open(now
, begin
, expire
))
1638 * Refresh credentials. XXX - finish
1641 gss_refresh(struct rpc_task
*task
)
1643 struct rpc_cred
*cred
= task
->tk_rqstp
->rq_cred
;
1646 if (gss_cred_is_negative_entry(cred
))
1647 return -EKEYEXPIRED
;
1649 if (!test_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
) &&
1650 !test_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
)) {
1651 ret
= gss_renew_cred(task
);
1654 cred
= task
->tk_rqstp
->rq_cred
;
1657 if (test_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
))
1658 ret
= gss_refresh_upcall(task
);
1663 /* Dummy refresh routine: used only when destroying the context */
1665 gss_refresh_null(struct rpc_task
*task
)
1671 gss_validate(struct rpc_task
*task
, struct xdr_stream
*xdr
)
1673 struct rpc_cred
*cred
= task
->tk_rqstp
->rq_cred
;
1674 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
1675 __be32
*p
, *seq
= NULL
;
1677 struct xdr_buf verf_buf
;
1678 struct xdr_netobj mic
;
1682 p
= xdr_inline_decode(xdr
, 2 * sizeof(*p
));
1684 goto validate_failed
;
1685 if (*p
++ != rpc_auth_gss
)
1686 goto validate_failed
;
1687 len
= be32_to_cpup(p
);
1688 if (len
> RPC_MAX_AUTH_SIZE
)
1689 goto validate_failed
;
1690 p
= xdr_inline_decode(xdr
, len
);
1692 goto validate_failed
;
1694 seq
= kmalloc(4, GFP_NOFS
);
1696 goto validate_failed
;
1697 *seq
= cpu_to_be32(task
->tk_rqstp
->rq_seqno
);
1700 xdr_buf_from_iov(&iov
, &verf_buf
);
1703 maj_stat
= gss_verify_mic(ctx
->gc_gss_ctx
, &verf_buf
, &mic
);
1704 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1705 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1709 /* We leave it to unwrap to calculate au_rslack. For now we just
1710 * calculate the length of the verifier: */
1711 cred
->cr_auth
->au_verfsize
= XDR_QUADLEN(len
) + 2;
1722 trace_rpcgss_verify_mic(task
, maj_stat
);
1727 static int gss_wrap_req_integ(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
,
1728 struct rpc_task
*task
, struct xdr_stream
*xdr
)
1730 struct rpc_rqst
*rqstp
= task
->tk_rqstp
;
1731 struct xdr_buf integ_buf
, *snd_buf
= &rqstp
->rq_snd_buf
;
1732 struct xdr_netobj mic
;
1733 __be32
*p
, *integ_len
;
1734 u32 offset
, maj_stat
;
1736 p
= xdr_reserve_space(xdr
, 2 * sizeof(*p
));
1740 *p
= cpu_to_be32(rqstp
->rq_seqno
);
1742 if (rpcauth_wrap_req_encode(task
, xdr
))
1745 offset
= (u8
*)p
- (u8
*)snd_buf
->head
[0].iov_base
;
1746 if (xdr_buf_subsegment(snd_buf
, &integ_buf
,
1747 offset
, snd_buf
->len
- offset
))
1749 *integ_len
= cpu_to_be32(integ_buf
.len
);
1751 p
= xdr_reserve_space(xdr
, 0);
1754 mic
.data
= (u8
*)(p
+ 1);
1755 maj_stat
= gss_get_mic(ctx
->gc_gss_ctx
, &integ_buf
, &mic
);
1756 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1757 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1760 /* Check that the trailing MIC fit in the buffer, after the fact */
1761 if (xdr_stream_encode_opaque_inline(xdr
, (void **)&p
, mic
.len
) < 0)
1767 trace_rpcgss_get_mic(task
, maj_stat
);
1772 priv_release_snd_buf(struct rpc_rqst
*rqstp
)
1776 for (i
=0; i
< rqstp
->rq_enc_pages_num
; i
++)
1777 __free_page(rqstp
->rq_enc_pages
[i
]);
1778 kfree(rqstp
->rq_enc_pages
);
1779 rqstp
->rq_release_snd_buf
= NULL
;
1783 alloc_enc_pages(struct rpc_rqst
*rqstp
)
1785 struct xdr_buf
*snd_buf
= &rqstp
->rq_snd_buf
;
1788 if (rqstp
->rq_release_snd_buf
)
1789 rqstp
->rq_release_snd_buf(rqstp
);
1791 if (snd_buf
->page_len
== 0) {
1792 rqstp
->rq_enc_pages_num
= 0;
1796 first
= snd_buf
->page_base
>> PAGE_SHIFT
;
1797 last
= (snd_buf
->page_base
+ snd_buf
->page_len
- 1) >> PAGE_SHIFT
;
1798 rqstp
->rq_enc_pages_num
= last
- first
+ 1 + 1;
1800 = kmalloc_array(rqstp
->rq_enc_pages_num
,
1801 sizeof(struct page
*),
1803 if (!rqstp
->rq_enc_pages
)
1805 for (i
=0; i
< rqstp
->rq_enc_pages_num
; i
++) {
1806 rqstp
->rq_enc_pages
[i
] = alloc_page(GFP_NOFS
);
1807 if (rqstp
->rq_enc_pages
[i
] == NULL
)
1810 rqstp
->rq_release_snd_buf
= priv_release_snd_buf
;
1813 rqstp
->rq_enc_pages_num
= i
;
1814 priv_release_snd_buf(rqstp
);
1819 static int gss_wrap_req_priv(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
,
1820 struct rpc_task
*task
, struct xdr_stream
*xdr
)
1822 struct rpc_rqst
*rqstp
= task
->tk_rqstp
;
1823 struct xdr_buf
*snd_buf
= &rqstp
->rq_snd_buf
;
1824 u32 pad
, offset
, maj_stat
;
1826 __be32
*p
, *opaque_len
;
1827 struct page
**inpages
;
1832 p
= xdr_reserve_space(xdr
, 2 * sizeof(*p
));
1836 *p
= cpu_to_be32(rqstp
->rq_seqno
);
1838 if (rpcauth_wrap_req_encode(task
, xdr
))
1841 status
= alloc_enc_pages(rqstp
);
1842 if (unlikely(status
))
1844 first
= snd_buf
->page_base
>> PAGE_SHIFT
;
1845 inpages
= snd_buf
->pages
+ first
;
1846 snd_buf
->pages
= rqstp
->rq_enc_pages
;
1847 snd_buf
->page_base
-= first
<< PAGE_SHIFT
;
1849 * Move the tail into its own page, in case gss_wrap needs
1850 * more space in the head when wrapping.
1852 * Still... Why can't gss_wrap just slide the tail down?
1854 if (snd_buf
->page_len
|| snd_buf
->tail
[0].iov_len
) {
1857 tmp
= page_address(rqstp
->rq_enc_pages
[rqstp
->rq_enc_pages_num
- 1]);
1858 memcpy(tmp
, snd_buf
->tail
[0].iov_base
, snd_buf
->tail
[0].iov_len
);
1859 snd_buf
->tail
[0].iov_base
= tmp
;
1861 offset
= (u8
*)p
- (u8
*)snd_buf
->head
[0].iov_base
;
1862 maj_stat
= gss_wrap(ctx
->gc_gss_ctx
, offset
, snd_buf
, inpages
);
1863 /* slack space should prevent this ever happening: */
1864 if (unlikely(snd_buf
->len
> snd_buf
->buflen
))
1866 /* We're assuming that when GSS_S_CONTEXT_EXPIRED, the encryption was
1867 * done anyway, so it's safe to put the request on the wire: */
1868 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1869 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1873 *opaque_len
= cpu_to_be32(snd_buf
->len
- offset
);
1874 /* guess whether the pad goes into the head or the tail: */
1875 if (snd_buf
->page_len
|| snd_buf
->tail
[0].iov_len
)
1876 iov
= snd_buf
->tail
;
1878 iov
= snd_buf
->head
;
1879 p
= iov
->iov_base
+ iov
->iov_len
;
1880 pad
= 3 - ((snd_buf
->len
- offset
- 1) & 3);
1882 iov
->iov_len
+= pad
;
1883 snd_buf
->len
+= pad
;
1889 trace_rpcgss_wrap(task
, maj_stat
);
1893 static int gss_wrap_req(struct rpc_task
*task
, struct xdr_stream
*xdr
)
1895 struct rpc_cred
*cred
= task
->tk_rqstp
->rq_cred
;
1896 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
,
1898 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
1902 if (ctx
->gc_proc
!= RPC_GSS_PROC_DATA
) {
1903 /* The spec seems a little ambiguous here, but I think that not
1904 * wrapping context destruction requests makes the most sense.
1906 status
= rpcauth_wrap_req_encode(task
, xdr
);
1909 switch (gss_cred
->gc_service
) {
1910 case RPC_GSS_SVC_NONE
:
1911 status
= rpcauth_wrap_req_encode(task
, xdr
);
1913 case RPC_GSS_SVC_INTEGRITY
:
1914 status
= gss_wrap_req_integ(cred
, ctx
, task
, xdr
);
1916 case RPC_GSS_SVC_PRIVACY
:
1917 status
= gss_wrap_req_priv(cred
, ctx
, task
, xdr
);
1928 gss_unwrap_resp_auth(struct rpc_cred
*cred
)
1930 struct rpc_auth
*auth
= cred
->cr_auth
;
1932 auth
->au_rslack
= auth
->au_verfsize
;
1933 auth
->au_ralign
= auth
->au_verfsize
;
1938 gss_unwrap_resp_integ(struct rpc_task
*task
, struct rpc_cred
*cred
,
1939 struct gss_cl_ctx
*ctx
, struct rpc_rqst
*rqstp
,
1940 struct xdr_stream
*xdr
)
1942 struct xdr_buf integ_buf
, *rcv_buf
= &rqstp
->rq_rcv_buf
;
1943 u32 data_offset
, mic_offset
, integ_len
, maj_stat
;
1944 struct rpc_auth
*auth
= cred
->cr_auth
;
1945 struct xdr_netobj mic
;
1948 p
= xdr_inline_decode(xdr
, 2 * sizeof(*p
));
1951 integ_len
= be32_to_cpup(p
++);
1954 data_offset
= (u8
*)(p
) - (u8
*)rcv_buf
->head
[0].iov_base
;
1955 mic_offset
= integ_len
+ data_offset
;
1956 if (mic_offset
> rcv_buf
->len
)
1958 if (be32_to_cpup(p
) != rqstp
->rq_seqno
)
1961 if (xdr_buf_subsegment(rcv_buf
, &integ_buf
, data_offset
, integ_len
))
1963 if (xdr_buf_read_mic(rcv_buf
, &mic
, mic_offset
))
1965 maj_stat
= gss_verify_mic(ctx
->gc_gss_ctx
, &integ_buf
, &mic
);
1966 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1967 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1968 if (maj_stat
!= GSS_S_COMPLETE
)
1971 auth
->au_rslack
= auth
->au_verfsize
+ 2 + 1 + XDR_QUADLEN(mic
.len
);
1972 auth
->au_ralign
= auth
->au_verfsize
+ 2;
1975 trace_rpcgss_unwrap_failed(task
);
1978 trace_rpcgss_bad_seqno(task
, rqstp
->rq_seqno
, be32_to_cpup(p
));
1981 trace_rpcgss_verify_mic(task
, maj_stat
);
1986 gss_unwrap_resp_priv(struct rpc_task
*task
, struct rpc_cred
*cred
,
1987 struct gss_cl_ctx
*ctx
, struct rpc_rqst
*rqstp
,
1988 struct xdr_stream
*xdr
)
1990 struct xdr_buf
*rcv_buf
= &rqstp
->rq_rcv_buf
;
1991 struct kvec
*head
= rqstp
->rq_rcv_buf
.head
;
1992 struct rpc_auth
*auth
= cred
->cr_auth
;
1993 unsigned int savedlen
= rcv_buf
->len
;
1994 u32 offset
, opaque_len
, maj_stat
;
1997 p
= xdr_inline_decode(xdr
, 2 * sizeof(*p
));
2000 opaque_len
= be32_to_cpup(p
++);
2001 offset
= (u8
*)(p
) - (u8
*)head
->iov_base
;
2002 if (offset
+ opaque_len
> rcv_buf
->len
)
2004 rcv_buf
->len
= offset
+ opaque_len
;
2006 maj_stat
= gss_unwrap(ctx
->gc_gss_ctx
, offset
, rcv_buf
);
2007 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
2008 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
2009 if (maj_stat
!= GSS_S_COMPLETE
)
2011 /* gss_unwrap decrypted the sequence number */
2012 if (be32_to_cpup(p
++) != rqstp
->rq_seqno
)
2015 /* gss_unwrap redacts the opaque blob from the head iovec.
2016 * rcv_buf has changed, thus the stream needs to be reset.
2018 xdr_init_decode(xdr
, rcv_buf
, p
, rqstp
);
2020 auth
->au_rslack
= auth
->au_verfsize
+ 2 +
2021 XDR_QUADLEN(savedlen
- rcv_buf
->len
);
2022 auth
->au_ralign
= auth
->au_verfsize
+ 2 +
2023 XDR_QUADLEN(savedlen
- rcv_buf
->len
);
2026 trace_rpcgss_unwrap_failed(task
);
2029 trace_rpcgss_bad_seqno(task
, rqstp
->rq_seqno
, be32_to_cpup(--p
));
2032 trace_rpcgss_unwrap(task
, maj_stat
);
2037 gss_seq_is_newer(u32
new, u32 old
)
2039 return (s32
)(new - old
) > 0;
2043 gss_xmit_need_reencode(struct rpc_task
*task
)
2045 struct rpc_rqst
*req
= task
->tk_rqstp
;
2046 struct rpc_cred
*cred
= req
->rq_cred
;
2047 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
2048 u32 win
, seq_xmit
= 0;
2054 if (gss_seq_is_newer(req
->rq_seqno
, READ_ONCE(ctx
->gc_seq
)))
2057 seq_xmit
= READ_ONCE(ctx
->gc_seq_xmit
);
2058 while (gss_seq_is_newer(req
->rq_seqno
, seq_xmit
)) {
2061 seq_xmit
= cmpxchg(&ctx
->gc_seq_xmit
, tmp
, req
->rq_seqno
);
2062 if (seq_xmit
== tmp
) {
2070 ret
= !gss_seq_is_newer(req
->rq_seqno
, seq_xmit
- win
);
2075 trace_rpcgss_need_reencode(task
, seq_xmit
, ret
);
2080 gss_unwrap_resp(struct rpc_task
*task
, struct xdr_stream
*xdr
)
2082 struct rpc_rqst
*rqstp
= task
->tk_rqstp
;
2083 struct rpc_cred
*cred
= rqstp
->rq_cred
;
2084 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
,
2086 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
2089 if (ctx
->gc_proc
!= RPC_GSS_PROC_DATA
)
2091 switch (gss_cred
->gc_service
) {
2092 case RPC_GSS_SVC_NONE
:
2093 status
= gss_unwrap_resp_auth(cred
);
2095 case RPC_GSS_SVC_INTEGRITY
:
2096 status
= gss_unwrap_resp_integ(task
, cred
, ctx
, rqstp
, xdr
);
2098 case RPC_GSS_SVC_PRIVACY
:
2099 status
= gss_unwrap_resp_priv(task
, cred
, ctx
, rqstp
, xdr
);
2106 status
= rpcauth_unwrap_resp_decode(task
, xdr
);
2112 static const struct rpc_authops authgss_ops
= {
2113 .owner
= THIS_MODULE
,
2114 .au_flavor
= RPC_AUTH_GSS
,
2115 .au_name
= "RPCSEC_GSS",
2116 .create
= gss_create
,
2117 .destroy
= gss_destroy
,
2118 .hash_cred
= gss_hash_cred
,
2119 .lookup_cred
= gss_lookup_cred
,
2120 .crcreate
= gss_create_cred
,
2121 .list_pseudoflavors
= gss_mech_list_pseudoflavors
,
2122 .info2flavor
= gss_mech_info2flavor
,
2123 .flavor2info
= gss_mech_flavor2info
,
2126 static const struct rpc_credops gss_credops
= {
2127 .cr_name
= "AUTH_GSS",
2128 .crdestroy
= gss_destroy_cred
,
2129 .cr_init
= gss_cred_init
,
2130 .crmatch
= gss_match
,
2131 .crmarshal
= gss_marshal
,
2132 .crrefresh
= gss_refresh
,
2133 .crvalidate
= gss_validate
,
2134 .crwrap_req
= gss_wrap_req
,
2135 .crunwrap_resp
= gss_unwrap_resp
,
2136 .crkey_timeout
= gss_key_timeout
,
2137 .crstringify_acceptor
= gss_stringify_acceptor
,
2138 .crneed_reencode
= gss_xmit_need_reencode
,
2141 static const struct rpc_credops gss_nullops
= {
2142 .cr_name
= "AUTH_GSS",
2143 .crdestroy
= gss_destroy_nullcred
,
2144 .crmatch
= gss_match
,
2145 .crmarshal
= gss_marshal
,
2146 .crrefresh
= gss_refresh_null
,
2147 .crvalidate
= gss_validate
,
2148 .crwrap_req
= gss_wrap_req
,
2149 .crunwrap_resp
= gss_unwrap_resp
,
2150 .crstringify_acceptor
= gss_stringify_acceptor
,
2153 static const struct rpc_pipe_ops gss_upcall_ops_v0
= {
2154 .upcall
= gss_v0_upcall
,
2155 .downcall
= gss_pipe_downcall
,
2156 .destroy_msg
= gss_pipe_destroy_msg
,
2157 .open_pipe
= gss_pipe_open_v0
,
2158 .release_pipe
= gss_pipe_release
,
2161 static const struct rpc_pipe_ops gss_upcall_ops_v1
= {
2162 .upcall
= gss_v1_upcall
,
2163 .downcall
= gss_pipe_downcall
,
2164 .destroy_msg
= gss_pipe_destroy_msg
,
2165 .open_pipe
= gss_pipe_open_v1
,
2166 .release_pipe
= gss_pipe_release
,
2169 static __net_init
int rpcsec_gss_init_net(struct net
*net
)
2171 return gss_svc_init_net(net
);
2174 static __net_exit
void rpcsec_gss_exit_net(struct net
*net
)
2176 gss_svc_shutdown_net(net
);
2179 static struct pernet_operations rpcsec_gss_net_ops
= {
2180 .init
= rpcsec_gss_init_net
,
2181 .exit
= rpcsec_gss_exit_net
,
2185 * Initialize RPCSEC_GSS module
2187 static int __init
init_rpcsec_gss(void)
2191 err
= rpcauth_register(&authgss_ops
);
2194 err
= gss_svc_init();
2196 goto out_unregister
;
2197 err
= register_pernet_subsys(&rpcsec_gss_net_ops
);
2200 rpc_init_wait_queue(&pipe_version_rpc_waitqueue
, "gss pipe version");
2205 rpcauth_unregister(&authgss_ops
);
2210 static void __exit
exit_rpcsec_gss(void)
2212 unregister_pernet_subsys(&rpcsec_gss_net_ops
);
2214 rpcauth_unregister(&authgss_ops
);
2215 rcu_barrier(); /* Wait for completion of call_rcu()'s */
2218 MODULE_ALIAS("rpc-auth-6");
2219 MODULE_LICENSE("GPL");
2220 module_param_named(expired_cred_retry_delay
,
2221 gss_expired_cred_retry_delay
,
2223 MODULE_PARM_DESC(expired_cred_retry_delay
, "Timeout (in seconds) until "
2224 "the RPC engine retries an expired credential");
2226 module_param_named(key_expire_timeo
,
2227 gss_key_expire_timeo
,
2229 MODULE_PARM_DESC(key_expire_timeo
, "Time (in seconds) at the end of a "
2230 "credential keys lifetime where the NFS layer cleans up "
2231 "prior to key expiration");
2233 module_init(init_rpcsec_gss
)
2234 module_exit(exit_rpcsec_gss
)