2 * linux/net/sunrpc/auth_gss/auth_gss.c
4 * RPCSEC_GSS client authentication.
6 * Copyright (c) 2000 The Regents of the University of Michigan.
9 * Dug Song <dugsong@monkey.org>
10 * Andy Adamson <andros@umich.edu>
12 * Redistribution and use in source and binary forms, with or without
13 * modification, are permitted provided that the following conditions
16 * 1. Redistributions of source code must retain the above copyright
17 * notice, this list of conditions and the following disclaimer.
18 * 2. Redistributions in binary form must reproduce the above copyright
19 * notice, this list of conditions and the following disclaimer in the
20 * documentation and/or other materials provided with the distribution.
21 * 3. Neither the name of the University nor the names of its
22 * contributors may be used to endorse or promote products derived
23 * from this software without specific prior written permission.
25 * THIS SOFTWARE IS PROVIDED ``AS IS'' AND ANY EXPRESS OR IMPLIED
26 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
27 * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
28 * DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
29 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
30 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
31 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
32 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
33 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
34 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
35 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
39 #include <linux/module.h>
40 #include <linux/init.h>
41 #include <linux/types.h>
42 #include <linux/slab.h>
43 #include <linux/sched.h>
44 #include <linux/pagemap.h>
45 #include <linux/sunrpc/clnt.h>
46 #include <linux/sunrpc/auth.h>
47 #include <linux/sunrpc/auth_gss.h>
48 #include <linux/sunrpc/svcauth_gss.h>
49 #include <linux/sunrpc/gss_err.h>
50 #include <linux/workqueue.h>
51 #include <linux/sunrpc/rpc_pipe_fs.h>
52 #include <linux/sunrpc/gss_api.h>
53 #include <asm/uaccess.h>
54 #include <linux/hashtable.h>
58 static const struct rpc_authops authgss_ops
;
60 static const struct rpc_credops gss_credops
;
61 static const struct rpc_credops gss_nullops
;
63 #define GSS_RETRY_EXPIRED 5
64 static unsigned int gss_expired_cred_retry_delay
= GSS_RETRY_EXPIRED
;
66 #define GSS_KEY_EXPIRE_TIMEO 240
67 static unsigned int gss_key_expire_timeo
= GSS_KEY_EXPIRE_TIMEO
;
69 #if IS_ENABLED(CONFIG_SUNRPC_DEBUG)
70 # define RPCDBG_FACILITY RPCDBG_AUTH
73 #define GSS_CRED_SLACK (RPC_MAX_AUTH_SIZE * 2)
74 /* length of a krb5 verifier (48), plus data added before arguments when
75 * using integrity (two 4-byte integers): */
76 #define GSS_VERF_SLACK 100
78 static DEFINE_HASHTABLE(gss_auth_hash_table
, 4);
79 static DEFINE_SPINLOCK(gss_auth_hash_lock
);
82 struct rpc_pipe_dir_object pdo
;
83 struct rpc_pipe
*pipe
;
84 struct rpc_clnt
*clnt
;
91 struct hlist_node hash
;
92 struct rpc_auth rpc_auth
;
93 struct gss_api_mech
*mech
;
94 enum rpc_gss_svc service
;
95 struct rpc_clnt
*client
;
98 * There are two upcall pipes; dentry[1], named "gssd", is used
99 * for the new text-based upcall; dentry[0] is named after the
100 * mechanism (for example, "krb5") and exists for
101 * backwards-compatibility with older gssd's.
103 struct gss_pipe
*gss_pipe
[2];
104 const char *target_name
;
107 /* pipe_version >= 0 if and only if someone has a pipe open. */
108 static DEFINE_SPINLOCK(pipe_version_lock
);
109 static struct rpc_wait_queue pipe_version_rpc_waitqueue
;
110 static DECLARE_WAIT_QUEUE_HEAD(pipe_version_waitqueue
);
111 static void gss_put_auth(struct gss_auth
*gss_auth
);
113 static void gss_free_ctx(struct gss_cl_ctx
*);
114 static const struct rpc_pipe_ops gss_upcall_ops_v0
;
115 static const struct rpc_pipe_ops gss_upcall_ops_v1
;
117 static inline struct gss_cl_ctx
*
118 gss_get_ctx(struct gss_cl_ctx
*ctx
)
120 atomic_inc(&ctx
->count
);
125 gss_put_ctx(struct gss_cl_ctx
*ctx
)
127 if (atomic_dec_and_test(&ctx
->count
))
132 * called by gss_upcall_callback and gss_create_upcall in order
133 * to set the gss context. The actual exchange of an old context
134 * and a new one is protected by the pipe->lock.
137 gss_cred_set_ctx(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
)
139 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
141 if (!test_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
))
144 rcu_assign_pointer(gss_cred
->gc_ctx
, ctx
);
145 set_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
146 smp_mb__before_atomic();
147 clear_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
);
151 simple_get_bytes(const void *p
, const void *end
, void *res
, size_t len
)
153 const void *q
= (const void *)((const char *)p
+ len
);
154 if (unlikely(q
> end
|| q
< p
))
155 return ERR_PTR(-EFAULT
);
160 static inline const void *
161 simple_get_netobj(const void *p
, const void *end
, struct xdr_netobj
*dest
)
166 p
= simple_get_bytes(p
, end
, &len
, sizeof(len
));
169 q
= (const void *)((const char *)p
+ len
);
170 if (unlikely(q
> end
|| q
< p
))
171 return ERR_PTR(-EFAULT
);
172 dest
->data
= kmemdup(p
, len
, GFP_NOFS
);
173 if (unlikely(dest
->data
== NULL
))
174 return ERR_PTR(-ENOMEM
);
179 static struct gss_cl_ctx
*
180 gss_cred_get_ctx(struct rpc_cred
*cred
)
182 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
183 struct gss_cl_ctx
*ctx
= NULL
;
186 ctx
= rcu_dereference(gss_cred
->gc_ctx
);
193 static struct gss_cl_ctx
*
194 gss_alloc_context(void)
196 struct gss_cl_ctx
*ctx
;
198 ctx
= kzalloc(sizeof(*ctx
), GFP_NOFS
);
200 ctx
->gc_proc
= RPC_GSS_PROC_DATA
;
201 ctx
->gc_seq
= 1; /* NetApp 6.4R1 doesn't accept seq. no. 0 */
202 spin_lock_init(&ctx
->gc_seq_lock
);
203 atomic_set(&ctx
->count
,1);
208 #define GSSD_MIN_TIMEOUT (60 * 60)
210 gss_fill_context(const void *p
, const void *end
, struct gss_cl_ctx
*ctx
, struct gss_api_mech
*gm
)
214 unsigned int timeout
;
215 unsigned long now
= jiffies
;
219 /* First unsigned int gives the remaining lifetime in seconds of the
220 * credential - e.g. the remaining TGT lifetime for Kerberos or
221 * the -t value passed to GSSD.
223 p
= simple_get_bytes(p
, end
, &timeout
, sizeof(timeout
));
227 timeout
= GSSD_MIN_TIMEOUT
;
228 ctx
->gc_expiry
= now
+ ((unsigned long)timeout
* HZ
);
229 /* Sequence number window. Determines the maximum number of
230 * simultaneous requests
232 p
= simple_get_bytes(p
, end
, &window_size
, sizeof(window_size
));
235 ctx
->gc_win
= window_size
;
236 /* gssd signals an error by passing ctx->gc_win = 0: */
237 if (ctx
->gc_win
== 0) {
239 * in which case, p points to an error code. Anything other
240 * than -EKEYEXPIRED gets converted to -EACCES.
242 p
= simple_get_bytes(p
, end
, &ret
, sizeof(ret
));
244 p
= (ret
== -EKEYEXPIRED
) ? ERR_PTR(-EKEYEXPIRED
) :
248 /* copy the opaque wire context */
249 p
= simple_get_netobj(p
, end
, &ctx
->gc_wire_ctx
);
252 /* import the opaque security context */
253 p
= simple_get_bytes(p
, end
, &seclen
, sizeof(seclen
));
256 q
= (const void *)((const char *)p
+ seclen
);
257 if (unlikely(q
> end
|| q
< p
)) {
258 p
= ERR_PTR(-EFAULT
);
261 ret
= gss_import_sec_context(p
, seclen
, gm
, &ctx
->gc_gss_ctx
, NULL
, GFP_NOFS
);
267 /* is there any trailing data? */
273 /* pull in acceptor name (if there is one) */
274 p
= simple_get_netobj(q
, end
, &ctx
->gc_acceptor
);
278 dprintk("RPC: %s Success. gc_expiry %lu now %lu timeout %u acceptor %.*s\n",
279 __func__
, ctx
->gc_expiry
, now
, timeout
, ctx
->gc_acceptor
.len
,
280 ctx
->gc_acceptor
.data
);
283 dprintk("RPC: %s returns error %ld\n", __func__
, -PTR_ERR(p
));
287 #define UPCALL_BUF_LEN 128
289 struct gss_upcall_msg
{
292 struct rpc_pipe_msg msg
;
293 struct list_head list
;
294 struct gss_auth
*auth
;
295 struct rpc_pipe
*pipe
;
296 struct rpc_wait_queue rpc_waitqueue
;
297 wait_queue_head_t waitqueue
;
298 struct gss_cl_ctx
*ctx
;
299 char databuf
[UPCALL_BUF_LEN
];
302 static int get_pipe_version(struct net
*net
)
304 struct sunrpc_net
*sn
= net_generic(net
, sunrpc_net_id
);
307 spin_lock(&pipe_version_lock
);
308 if (sn
->pipe_version
>= 0) {
309 atomic_inc(&sn
->pipe_users
);
310 ret
= sn
->pipe_version
;
313 spin_unlock(&pipe_version_lock
);
317 static void put_pipe_version(struct net
*net
)
319 struct sunrpc_net
*sn
= net_generic(net
, sunrpc_net_id
);
321 if (atomic_dec_and_lock(&sn
->pipe_users
, &pipe_version_lock
)) {
322 sn
->pipe_version
= -1;
323 spin_unlock(&pipe_version_lock
);
328 gss_release_msg(struct gss_upcall_msg
*gss_msg
)
330 struct net
*net
= gss_msg
->auth
->net
;
331 if (!atomic_dec_and_test(&gss_msg
->count
))
333 put_pipe_version(net
);
334 BUG_ON(!list_empty(&gss_msg
->list
));
335 if (gss_msg
->ctx
!= NULL
)
336 gss_put_ctx(gss_msg
->ctx
);
337 rpc_destroy_wait_queue(&gss_msg
->rpc_waitqueue
);
338 gss_put_auth(gss_msg
->auth
);
342 static struct gss_upcall_msg
*
343 __gss_find_upcall(struct rpc_pipe
*pipe
, kuid_t uid
, const struct gss_auth
*auth
)
345 struct gss_upcall_msg
*pos
;
346 list_for_each_entry(pos
, &pipe
->in_downcall
, list
) {
347 if (!uid_eq(pos
->uid
, uid
))
349 if (auth
&& pos
->auth
->service
!= auth
->service
)
351 atomic_inc(&pos
->count
);
352 dprintk("RPC: %s found msg %p\n", __func__
, pos
);
355 dprintk("RPC: %s found nothing\n", __func__
);
359 /* Try to add an upcall to the pipefs queue.
360 * If an upcall owned by our uid already exists, then we return a reference
361 * to that upcall instead of adding the new upcall.
363 static inline struct gss_upcall_msg
*
364 gss_add_msg(struct gss_upcall_msg
*gss_msg
)
366 struct rpc_pipe
*pipe
= gss_msg
->pipe
;
367 struct gss_upcall_msg
*old
;
369 spin_lock(&pipe
->lock
);
370 old
= __gss_find_upcall(pipe
, gss_msg
->uid
, gss_msg
->auth
);
372 atomic_inc(&gss_msg
->count
);
373 list_add(&gss_msg
->list
, &pipe
->in_downcall
);
376 spin_unlock(&pipe
->lock
);
381 __gss_unhash_msg(struct gss_upcall_msg
*gss_msg
)
383 list_del_init(&gss_msg
->list
);
384 rpc_wake_up_status(&gss_msg
->rpc_waitqueue
, gss_msg
->msg
.errno
);
385 wake_up_all(&gss_msg
->waitqueue
);
386 atomic_dec(&gss_msg
->count
);
390 gss_unhash_msg(struct gss_upcall_msg
*gss_msg
)
392 struct rpc_pipe
*pipe
= gss_msg
->pipe
;
394 if (list_empty(&gss_msg
->list
))
396 spin_lock(&pipe
->lock
);
397 if (!list_empty(&gss_msg
->list
))
398 __gss_unhash_msg(gss_msg
);
399 spin_unlock(&pipe
->lock
);
403 gss_handle_downcall_result(struct gss_cred
*gss_cred
, struct gss_upcall_msg
*gss_msg
)
405 switch (gss_msg
->msg
.errno
) {
407 if (gss_msg
->ctx
== NULL
)
409 clear_bit(RPCAUTH_CRED_NEGATIVE
, &gss_cred
->gc_base
.cr_flags
);
410 gss_cred_set_ctx(&gss_cred
->gc_base
, gss_msg
->ctx
);
413 set_bit(RPCAUTH_CRED_NEGATIVE
, &gss_cred
->gc_base
.cr_flags
);
415 gss_cred
->gc_upcall_timestamp
= jiffies
;
416 gss_cred
->gc_upcall
= NULL
;
417 rpc_wake_up_status(&gss_msg
->rpc_waitqueue
, gss_msg
->msg
.errno
);
421 gss_upcall_callback(struct rpc_task
*task
)
423 struct gss_cred
*gss_cred
= container_of(task
->tk_rqstp
->rq_cred
,
424 struct gss_cred
, gc_base
);
425 struct gss_upcall_msg
*gss_msg
= gss_cred
->gc_upcall
;
426 struct rpc_pipe
*pipe
= gss_msg
->pipe
;
428 spin_lock(&pipe
->lock
);
429 gss_handle_downcall_result(gss_cred
, gss_msg
);
430 spin_unlock(&pipe
->lock
);
431 task
->tk_status
= gss_msg
->msg
.errno
;
432 gss_release_msg(gss_msg
);
435 static void gss_encode_v0_msg(struct gss_upcall_msg
*gss_msg
)
437 uid_t uid
= from_kuid(&init_user_ns
, gss_msg
->uid
);
438 memcpy(gss_msg
->databuf
, &uid
, sizeof(uid
));
439 gss_msg
->msg
.data
= gss_msg
->databuf
;
440 gss_msg
->msg
.len
= sizeof(uid
);
442 BUILD_BUG_ON(sizeof(uid
) > sizeof(gss_msg
->databuf
));
445 static int gss_encode_v1_msg(struct gss_upcall_msg
*gss_msg
,
446 const char *service_name
,
447 const char *target_name
)
449 struct gss_api_mech
*mech
= gss_msg
->auth
->mech
;
450 char *p
= gss_msg
->databuf
;
451 size_t buflen
= sizeof(gss_msg
->databuf
);
454 len
= scnprintf(p
, buflen
, "mech=%s uid=%d ", mech
->gm_name
,
455 from_kuid(&init_user_ns
, gss_msg
->uid
));
458 gss_msg
->msg
.len
= len
;
460 len
= scnprintf(p
, buflen
, "target=%s ", target_name
);
463 gss_msg
->msg
.len
+= len
;
465 if (service_name
!= NULL
) {
466 len
= scnprintf(p
, buflen
, "service=%s ", service_name
);
469 gss_msg
->msg
.len
+= len
;
471 if (mech
->gm_upcall_enctypes
) {
472 len
= scnprintf(p
, buflen
, "enctypes=%s ",
473 mech
->gm_upcall_enctypes
);
476 gss_msg
->msg
.len
+= len
;
478 len
= scnprintf(p
, buflen
, "\n");
481 gss_msg
->msg
.len
+= len
;
483 gss_msg
->msg
.data
= gss_msg
->databuf
;
490 static struct gss_upcall_msg
*
491 gss_alloc_msg(struct gss_auth
*gss_auth
,
492 kuid_t uid
, const char *service_name
)
494 struct gss_upcall_msg
*gss_msg
;
498 gss_msg
= kzalloc(sizeof(*gss_msg
), GFP_NOFS
);
501 vers
= get_pipe_version(gss_auth
->net
);
505 gss_msg
->pipe
= gss_auth
->gss_pipe
[vers
]->pipe
;
506 INIT_LIST_HEAD(&gss_msg
->list
);
507 rpc_init_wait_queue(&gss_msg
->rpc_waitqueue
, "RPCSEC_GSS upcall waitq");
508 init_waitqueue_head(&gss_msg
->waitqueue
);
509 atomic_set(&gss_msg
->count
, 1);
511 gss_msg
->auth
= gss_auth
;
514 gss_encode_v0_msg(gss_msg
);
517 err
= gss_encode_v1_msg(gss_msg
, service_name
, gss_auth
->target_name
);
519 goto err_put_pipe_version
;
521 kref_get(&gss_auth
->kref
);
523 err_put_pipe_version
:
524 put_pipe_version(gss_auth
->net
);
531 static struct gss_upcall_msg
*
532 gss_setup_upcall(struct gss_auth
*gss_auth
, struct rpc_cred
*cred
)
534 struct gss_cred
*gss_cred
= container_of(cred
,
535 struct gss_cred
, gc_base
);
536 struct gss_upcall_msg
*gss_new
, *gss_msg
;
537 kuid_t uid
= cred
->cr_uid
;
539 gss_new
= gss_alloc_msg(gss_auth
, uid
, gss_cred
->gc_principal
);
542 gss_msg
= gss_add_msg(gss_new
);
543 if (gss_msg
== gss_new
) {
544 int res
= rpc_queue_upcall(gss_new
->pipe
, &gss_new
->msg
);
546 gss_unhash_msg(gss_new
);
547 gss_msg
= ERR_PTR(res
);
550 gss_release_msg(gss_new
);
554 static void warn_gssd(void)
556 dprintk("AUTH_GSS upcall failed. Please check user daemon is running.\n");
560 gss_refresh_upcall(struct rpc_task
*task
)
562 struct rpc_cred
*cred
= task
->tk_rqstp
->rq_cred
;
563 struct gss_auth
*gss_auth
= container_of(cred
->cr_auth
,
564 struct gss_auth
, rpc_auth
);
565 struct gss_cred
*gss_cred
= container_of(cred
,
566 struct gss_cred
, gc_base
);
567 struct gss_upcall_msg
*gss_msg
;
568 struct rpc_pipe
*pipe
;
571 dprintk("RPC: %5u %s for uid %u\n",
572 task
->tk_pid
, __func__
, from_kuid(&init_user_ns
, cred
->cr_uid
));
573 gss_msg
= gss_setup_upcall(gss_auth
, cred
);
574 if (PTR_ERR(gss_msg
) == -EAGAIN
) {
575 /* XXX: warning on the first, under the assumption we
576 * shouldn't normally hit this case on a refresh. */
578 task
->tk_timeout
= 15*HZ
;
579 rpc_sleep_on(&pipe_version_rpc_waitqueue
, task
, NULL
);
582 if (IS_ERR(gss_msg
)) {
583 err
= PTR_ERR(gss_msg
);
586 pipe
= gss_msg
->pipe
;
587 spin_lock(&pipe
->lock
);
588 if (gss_cred
->gc_upcall
!= NULL
)
589 rpc_sleep_on(&gss_cred
->gc_upcall
->rpc_waitqueue
, task
, NULL
);
590 else if (gss_msg
->ctx
== NULL
&& gss_msg
->msg
.errno
>= 0) {
591 task
->tk_timeout
= 0;
592 gss_cred
->gc_upcall
= gss_msg
;
593 /* gss_upcall_callback will release the reference to gss_upcall_msg */
594 atomic_inc(&gss_msg
->count
);
595 rpc_sleep_on(&gss_msg
->rpc_waitqueue
, task
, gss_upcall_callback
);
597 gss_handle_downcall_result(gss_cred
, gss_msg
);
598 err
= gss_msg
->msg
.errno
;
600 spin_unlock(&pipe
->lock
);
601 gss_release_msg(gss_msg
);
603 dprintk("RPC: %5u %s for uid %u result %d\n",
604 task
->tk_pid
, __func__
,
605 from_kuid(&init_user_ns
, cred
->cr_uid
), err
);
610 gss_create_upcall(struct gss_auth
*gss_auth
, struct gss_cred
*gss_cred
)
612 struct net
*net
= gss_auth
->net
;
613 struct sunrpc_net
*sn
= net_generic(net
, sunrpc_net_id
);
614 struct rpc_pipe
*pipe
;
615 struct rpc_cred
*cred
= &gss_cred
->gc_base
;
616 struct gss_upcall_msg
*gss_msg
;
620 dprintk("RPC: %s for uid %u\n",
621 __func__
, from_kuid(&init_user_ns
, cred
->cr_uid
));
624 /* if gssd is down, just skip upcalling altogether */
625 if (!gssd_running(net
)) {
629 gss_msg
= gss_setup_upcall(gss_auth
, cred
);
630 if (PTR_ERR(gss_msg
) == -EAGAIN
) {
631 err
= wait_event_interruptible_timeout(pipe_version_waitqueue
,
632 sn
->pipe_version
>= 0, 15 * HZ
);
633 if (sn
->pipe_version
< 0) {
641 if (IS_ERR(gss_msg
)) {
642 err
= PTR_ERR(gss_msg
);
645 pipe
= gss_msg
->pipe
;
647 prepare_to_wait(&gss_msg
->waitqueue
, &wait
, TASK_KILLABLE
);
648 spin_lock(&pipe
->lock
);
649 if (gss_msg
->ctx
!= NULL
|| gss_msg
->msg
.errno
< 0) {
652 spin_unlock(&pipe
->lock
);
653 if (fatal_signal_pending(current
)) {
660 gss_cred_set_ctx(cred
, gss_msg
->ctx
);
662 err
= gss_msg
->msg
.errno
;
663 spin_unlock(&pipe
->lock
);
665 finish_wait(&gss_msg
->waitqueue
, &wait
);
666 gss_release_msg(gss_msg
);
668 dprintk("RPC: %s for uid %u result %d\n",
669 __func__
, from_kuid(&init_user_ns
, cred
->cr_uid
), err
);
673 #define MSG_BUF_MAXSIZE 1024
676 gss_pipe_downcall(struct file
*filp
, const char __user
*src
, size_t mlen
)
680 struct gss_upcall_msg
*gss_msg
;
681 struct rpc_pipe
*pipe
= RPC_I(file_inode(filp
))->pipe
;
682 struct gss_cl_ctx
*ctx
;
685 ssize_t err
= -EFBIG
;
687 if (mlen
> MSG_BUF_MAXSIZE
)
690 buf
= kmalloc(mlen
, GFP_NOFS
);
695 if (copy_from_user(buf
, src
, mlen
))
698 end
= (const void *)((char *)buf
+ mlen
);
699 p
= simple_get_bytes(buf
, end
, &id
, sizeof(id
));
705 uid
= make_kuid(&init_user_ns
, id
);
706 if (!uid_valid(uid
)) {
712 ctx
= gss_alloc_context();
717 /* Find a matching upcall */
718 spin_lock(&pipe
->lock
);
719 gss_msg
= __gss_find_upcall(pipe
, uid
, NULL
);
720 if (gss_msg
== NULL
) {
721 spin_unlock(&pipe
->lock
);
724 list_del_init(&gss_msg
->list
);
725 spin_unlock(&pipe
->lock
);
727 p
= gss_fill_context(p
, end
, ctx
, gss_msg
->auth
->mech
);
733 gss_msg
->msg
.errno
= err
;
740 gss_msg
->msg
.errno
= -EAGAIN
;
743 printk(KERN_CRIT
"%s: bad return from "
744 "gss_fill_context: %zd\n", __func__
, err
);
745 gss_msg
->msg
.errno
= -EIO
;
747 goto err_release_msg
;
749 gss_msg
->ctx
= gss_get_ctx(ctx
);
753 spin_lock(&pipe
->lock
);
754 __gss_unhash_msg(gss_msg
);
755 spin_unlock(&pipe
->lock
);
756 gss_release_msg(gss_msg
);
762 dprintk("RPC: %s returning %Zd\n", __func__
, err
);
766 static int gss_pipe_open(struct inode
*inode
, int new_version
)
768 struct net
*net
= inode
->i_sb
->s_fs_info
;
769 struct sunrpc_net
*sn
= net_generic(net
, sunrpc_net_id
);
772 spin_lock(&pipe_version_lock
);
773 if (sn
->pipe_version
< 0) {
774 /* First open of any gss pipe determines the version: */
775 sn
->pipe_version
= new_version
;
776 rpc_wake_up(&pipe_version_rpc_waitqueue
);
777 wake_up(&pipe_version_waitqueue
);
778 } else if (sn
->pipe_version
!= new_version
) {
779 /* Trying to open a pipe of a different version */
783 atomic_inc(&sn
->pipe_users
);
785 spin_unlock(&pipe_version_lock
);
790 static int gss_pipe_open_v0(struct inode
*inode
)
792 return gss_pipe_open(inode
, 0);
795 static int gss_pipe_open_v1(struct inode
*inode
)
797 return gss_pipe_open(inode
, 1);
801 gss_pipe_release(struct inode
*inode
)
803 struct net
*net
= inode
->i_sb
->s_fs_info
;
804 struct rpc_pipe
*pipe
= RPC_I(inode
)->pipe
;
805 struct gss_upcall_msg
*gss_msg
;
808 spin_lock(&pipe
->lock
);
809 list_for_each_entry(gss_msg
, &pipe
->in_downcall
, list
) {
811 if (!list_empty(&gss_msg
->msg
.list
))
813 gss_msg
->msg
.errno
= -EPIPE
;
814 atomic_inc(&gss_msg
->count
);
815 __gss_unhash_msg(gss_msg
);
816 spin_unlock(&pipe
->lock
);
817 gss_release_msg(gss_msg
);
820 spin_unlock(&pipe
->lock
);
822 put_pipe_version(net
);
826 gss_pipe_destroy_msg(struct rpc_pipe_msg
*msg
)
828 struct gss_upcall_msg
*gss_msg
= container_of(msg
, struct gss_upcall_msg
, msg
);
830 if (msg
->errno
< 0) {
831 dprintk("RPC: %s releasing msg %p\n",
833 atomic_inc(&gss_msg
->count
);
834 gss_unhash_msg(gss_msg
);
835 if (msg
->errno
== -ETIMEDOUT
)
837 gss_release_msg(gss_msg
);
841 static void gss_pipe_dentry_destroy(struct dentry
*dir
,
842 struct rpc_pipe_dir_object
*pdo
)
844 struct gss_pipe
*gss_pipe
= pdo
->pdo_data
;
845 struct rpc_pipe
*pipe
= gss_pipe
->pipe
;
847 if (pipe
->dentry
!= NULL
) {
848 rpc_unlink(pipe
->dentry
);
853 static int gss_pipe_dentry_create(struct dentry
*dir
,
854 struct rpc_pipe_dir_object
*pdo
)
856 struct gss_pipe
*p
= pdo
->pdo_data
;
857 struct dentry
*dentry
;
859 dentry
= rpc_mkpipe_dentry(dir
, p
->name
, p
->clnt
, p
->pipe
);
861 return PTR_ERR(dentry
);
862 p
->pipe
->dentry
= dentry
;
866 static const struct rpc_pipe_dir_object_ops gss_pipe_dir_object_ops
= {
867 .create
= gss_pipe_dentry_create
,
868 .destroy
= gss_pipe_dentry_destroy
,
871 static struct gss_pipe
*gss_pipe_alloc(struct rpc_clnt
*clnt
,
873 const struct rpc_pipe_ops
*upcall_ops
)
878 p
= kmalloc(sizeof(*p
), GFP_KERNEL
);
881 p
->pipe
= rpc_mkpipe_data(upcall_ops
, RPC_PIPE_WAIT_FOR_OPEN
);
882 if (IS_ERR(p
->pipe
)) {
883 err
= PTR_ERR(p
->pipe
);
884 goto err_free_gss_pipe
;
889 rpc_init_pipe_dir_object(&p
->pdo
,
890 &gss_pipe_dir_object_ops
,
899 struct gss_alloc_pdo
{
900 struct rpc_clnt
*clnt
;
902 const struct rpc_pipe_ops
*upcall_ops
;
905 static int gss_pipe_match_pdo(struct rpc_pipe_dir_object
*pdo
, void *data
)
907 struct gss_pipe
*gss_pipe
;
908 struct gss_alloc_pdo
*args
= data
;
910 if (pdo
->pdo_ops
!= &gss_pipe_dir_object_ops
)
912 gss_pipe
= container_of(pdo
, struct gss_pipe
, pdo
);
913 if (strcmp(gss_pipe
->name
, args
->name
) != 0)
915 if (!kref_get_unless_zero(&gss_pipe
->kref
))
920 static struct rpc_pipe_dir_object
*gss_pipe_alloc_pdo(void *data
)
922 struct gss_pipe
*gss_pipe
;
923 struct gss_alloc_pdo
*args
= data
;
925 gss_pipe
= gss_pipe_alloc(args
->clnt
, args
->name
, args
->upcall_ops
);
926 if (!IS_ERR(gss_pipe
))
927 return &gss_pipe
->pdo
;
931 static struct gss_pipe
*gss_pipe_get(struct rpc_clnt
*clnt
,
933 const struct rpc_pipe_ops
*upcall_ops
)
935 struct net
*net
= rpc_net_ns(clnt
);
936 struct rpc_pipe_dir_object
*pdo
;
937 struct gss_alloc_pdo args
= {
940 .upcall_ops
= upcall_ops
,
943 pdo
= rpc_find_or_alloc_pipe_dir_object(net
,
944 &clnt
->cl_pipedir_objects
,
949 return container_of(pdo
, struct gss_pipe
, pdo
);
950 return ERR_PTR(-ENOMEM
);
953 static void __gss_pipe_free(struct gss_pipe
*p
)
955 struct rpc_clnt
*clnt
= p
->clnt
;
956 struct net
*net
= rpc_net_ns(clnt
);
958 rpc_remove_pipe_dir_object(net
,
959 &clnt
->cl_pipedir_objects
,
961 rpc_destroy_pipe_data(p
->pipe
);
965 static void __gss_pipe_release(struct kref
*kref
)
967 struct gss_pipe
*p
= container_of(kref
, struct gss_pipe
, kref
);
972 static void gss_pipe_free(struct gss_pipe
*p
)
975 kref_put(&p
->kref
, __gss_pipe_release
);
979 * NOTE: we have the opportunity to use different
980 * parameters based on the input flavor (which must be a pseudoflavor)
982 static struct gss_auth
*
983 gss_create_new(struct rpc_auth_create_args
*args
, struct rpc_clnt
*clnt
)
985 rpc_authflavor_t flavor
= args
->pseudoflavor
;
986 struct gss_auth
*gss_auth
;
987 struct gss_pipe
*gss_pipe
;
988 struct rpc_auth
* auth
;
989 int err
= -ENOMEM
; /* XXX? */
991 dprintk("RPC: creating GSS authenticator for client %p\n", clnt
);
993 if (!try_module_get(THIS_MODULE
))
995 if (!(gss_auth
= kmalloc(sizeof(*gss_auth
), GFP_KERNEL
)))
997 INIT_HLIST_NODE(&gss_auth
->hash
);
998 gss_auth
->target_name
= NULL
;
999 if (args
->target_name
) {
1000 gss_auth
->target_name
= kstrdup(args
->target_name
, GFP_KERNEL
);
1001 if (gss_auth
->target_name
== NULL
)
1004 gss_auth
->client
= clnt
;
1005 gss_auth
->net
= get_net(rpc_net_ns(clnt
));
1007 gss_auth
->mech
= gss_mech_get_by_pseudoflavor(flavor
);
1008 if (!gss_auth
->mech
) {
1009 dprintk("RPC: Pseudoflavor %d not found!\n", flavor
);
1012 gss_auth
->service
= gss_pseudoflavor_to_service(gss_auth
->mech
, flavor
);
1013 if (gss_auth
->service
== 0)
1015 if (!gssd_running(gss_auth
->net
))
1017 auth
= &gss_auth
->rpc_auth
;
1018 auth
->au_cslack
= GSS_CRED_SLACK
>> 2;
1019 auth
->au_rslack
= GSS_VERF_SLACK
>> 2;
1021 auth
->au_ops
= &authgss_ops
;
1022 auth
->au_flavor
= flavor
;
1023 if (gss_pseudoflavor_to_datatouch(gss_auth
->mech
, flavor
))
1024 auth
->au_flags
|= RPCAUTH_AUTH_DATATOUCH
;
1025 atomic_set(&auth
->au_count
, 1);
1026 kref_init(&gss_auth
->kref
);
1028 err
= rpcauth_init_credcache(auth
);
1032 * Note: if we created the old pipe first, then someone who
1033 * examined the directory at the right moment might conclude
1034 * that we supported only the old pipe. So we instead create
1035 * the new pipe first.
1037 gss_pipe
= gss_pipe_get(clnt
, "gssd", &gss_upcall_ops_v1
);
1038 if (IS_ERR(gss_pipe
)) {
1039 err
= PTR_ERR(gss_pipe
);
1040 goto err_destroy_credcache
;
1042 gss_auth
->gss_pipe
[1] = gss_pipe
;
1044 gss_pipe
= gss_pipe_get(clnt
, gss_auth
->mech
->gm_name
,
1045 &gss_upcall_ops_v0
);
1046 if (IS_ERR(gss_pipe
)) {
1047 err
= PTR_ERR(gss_pipe
);
1048 goto err_destroy_pipe_1
;
1050 gss_auth
->gss_pipe
[0] = gss_pipe
;
1054 gss_pipe_free(gss_auth
->gss_pipe
[1]);
1055 err_destroy_credcache
:
1056 rpcauth_destroy_credcache(auth
);
1058 gss_mech_put(gss_auth
->mech
);
1060 put_net(gss_auth
->net
);
1062 kfree(gss_auth
->target_name
);
1065 module_put(THIS_MODULE
);
1066 return ERR_PTR(err
);
1070 gss_free(struct gss_auth
*gss_auth
)
1072 gss_pipe_free(gss_auth
->gss_pipe
[0]);
1073 gss_pipe_free(gss_auth
->gss_pipe
[1]);
1074 gss_mech_put(gss_auth
->mech
);
1075 put_net(gss_auth
->net
);
1076 kfree(gss_auth
->target_name
);
1079 module_put(THIS_MODULE
);
1083 gss_free_callback(struct kref
*kref
)
1085 struct gss_auth
*gss_auth
= container_of(kref
, struct gss_auth
, kref
);
1091 gss_put_auth(struct gss_auth
*gss_auth
)
1093 kref_put(&gss_auth
->kref
, gss_free_callback
);
1097 gss_destroy(struct rpc_auth
*auth
)
1099 struct gss_auth
*gss_auth
= container_of(auth
,
1100 struct gss_auth
, rpc_auth
);
1102 dprintk("RPC: destroying GSS authenticator %p flavor %d\n",
1103 auth
, auth
->au_flavor
);
1105 if (hash_hashed(&gss_auth
->hash
)) {
1106 spin_lock(&gss_auth_hash_lock
);
1107 hash_del(&gss_auth
->hash
);
1108 spin_unlock(&gss_auth_hash_lock
);
1111 gss_pipe_free(gss_auth
->gss_pipe
[0]);
1112 gss_auth
->gss_pipe
[0] = NULL
;
1113 gss_pipe_free(gss_auth
->gss_pipe
[1]);
1114 gss_auth
->gss_pipe
[1] = NULL
;
1115 rpcauth_destroy_credcache(auth
);
1117 gss_put_auth(gss_auth
);
1121 * Auths may be shared between rpc clients that were cloned from a
1122 * common client with the same xprt, if they also share the flavor and
1125 * The auth is looked up from the oldest parent sharing the same
1126 * cl_xprt, and the auth itself references only that common parent
1127 * (which is guaranteed to last as long as any of its descendants).
1129 static struct gss_auth
*
1130 gss_auth_find_or_add_hashed(struct rpc_auth_create_args
*args
,
1131 struct rpc_clnt
*clnt
,
1132 struct gss_auth
*new)
1134 struct gss_auth
*gss_auth
;
1135 unsigned long hashval
= (unsigned long)clnt
;
1137 spin_lock(&gss_auth_hash_lock
);
1138 hash_for_each_possible(gss_auth_hash_table
,
1142 if (gss_auth
->client
!= clnt
)
1144 if (gss_auth
->rpc_auth
.au_flavor
!= args
->pseudoflavor
)
1146 if (gss_auth
->target_name
!= args
->target_name
) {
1147 if (gss_auth
->target_name
== NULL
)
1149 if (args
->target_name
== NULL
)
1151 if (strcmp(gss_auth
->target_name
, args
->target_name
))
1154 if (!atomic_inc_not_zero(&gss_auth
->rpc_auth
.au_count
))
1159 hash_add(gss_auth_hash_table
, &new->hash
, hashval
);
1162 spin_unlock(&gss_auth_hash_lock
);
1166 static struct gss_auth
*
1167 gss_create_hashed(struct rpc_auth_create_args
*args
, struct rpc_clnt
*clnt
)
1169 struct gss_auth
*gss_auth
;
1170 struct gss_auth
*new;
1172 gss_auth
= gss_auth_find_or_add_hashed(args
, clnt
, NULL
);
1173 if (gss_auth
!= NULL
)
1175 new = gss_create_new(args
, clnt
);
1178 gss_auth
= gss_auth_find_or_add_hashed(args
, clnt
, new);
1179 if (gss_auth
!= new)
1180 gss_destroy(&new->rpc_auth
);
1185 static struct rpc_auth
*
1186 gss_create(struct rpc_auth_create_args
*args
, struct rpc_clnt
*clnt
)
1188 struct gss_auth
*gss_auth
;
1189 struct rpc_xprt_switch
*xps
= rcu_access_pointer(clnt
->cl_xpi
.xpi_xpswitch
);
1191 while (clnt
!= clnt
->cl_parent
) {
1192 struct rpc_clnt
*parent
= clnt
->cl_parent
;
1193 /* Find the original parent for this transport */
1194 if (rcu_access_pointer(parent
->cl_xpi
.xpi_xpswitch
) != xps
)
1199 gss_auth
= gss_create_hashed(args
, clnt
);
1200 if (IS_ERR(gss_auth
))
1201 return ERR_CAST(gss_auth
);
1202 return &gss_auth
->rpc_auth
;
1206 * gss_destroying_context will cause the RPCSEC_GSS to send a NULL RPC call
1207 * to the server with the GSS control procedure field set to
1208 * RPC_GSS_PROC_DESTROY. This should normally cause the server to release
1209 * all RPCSEC_GSS state associated with that context.
1212 gss_destroying_context(struct rpc_cred
*cred
)
1214 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
1215 struct gss_auth
*gss_auth
= container_of(cred
->cr_auth
, struct gss_auth
, rpc_auth
);
1216 struct gss_cl_ctx
*ctx
= rcu_dereference_protected(gss_cred
->gc_ctx
, 1);
1217 struct rpc_task
*task
;
1219 if (test_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
) == 0)
1222 ctx
->gc_proc
= RPC_GSS_PROC_DESTROY
;
1223 cred
->cr_ops
= &gss_nullops
;
1225 /* Take a reference to ensure the cred will be destroyed either
1226 * by the RPC call or by the put_rpccred() below */
1229 task
= rpc_call_null(gss_auth
->client
, cred
, RPC_TASK_ASYNC
|RPC_TASK_SOFT
);
1237 /* gss_destroy_cred (and gss_free_ctx) are used to clean up after failure
1238 * to create a new cred or context, so they check that things have been
1239 * allocated before freeing them. */
1241 gss_do_free_ctx(struct gss_cl_ctx
*ctx
)
1243 dprintk("RPC: %s\n", __func__
);
1245 gss_delete_sec_context(&ctx
->gc_gss_ctx
);
1246 kfree(ctx
->gc_wire_ctx
.data
);
1247 kfree(ctx
->gc_acceptor
.data
);
1252 gss_free_ctx_callback(struct rcu_head
*head
)
1254 struct gss_cl_ctx
*ctx
= container_of(head
, struct gss_cl_ctx
, gc_rcu
);
1255 gss_do_free_ctx(ctx
);
1259 gss_free_ctx(struct gss_cl_ctx
*ctx
)
1261 call_rcu(&ctx
->gc_rcu
, gss_free_ctx_callback
);
1265 gss_free_cred(struct gss_cred
*gss_cred
)
1267 dprintk("RPC: %s cred=%p\n", __func__
, gss_cred
);
1272 gss_free_cred_callback(struct rcu_head
*head
)
1274 struct gss_cred
*gss_cred
= container_of(head
, struct gss_cred
, gc_base
.cr_rcu
);
1275 gss_free_cred(gss_cred
);
1279 gss_destroy_nullcred(struct rpc_cred
*cred
)
1281 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
1282 struct gss_auth
*gss_auth
= container_of(cred
->cr_auth
, struct gss_auth
, rpc_auth
);
1283 struct gss_cl_ctx
*ctx
= rcu_dereference_protected(gss_cred
->gc_ctx
, 1);
1285 RCU_INIT_POINTER(gss_cred
->gc_ctx
, NULL
);
1286 call_rcu(&cred
->cr_rcu
, gss_free_cred_callback
);
1289 gss_put_auth(gss_auth
);
1293 gss_destroy_cred(struct rpc_cred
*cred
)
1296 if (gss_destroying_context(cred
))
1298 gss_destroy_nullcred(cred
);
1302 * Lookup RPCSEC_GSS cred for the current process
1304 static struct rpc_cred
*
1305 gss_lookup_cred(struct rpc_auth
*auth
, struct auth_cred
*acred
, int flags
)
1307 return rpcauth_lookup_credcache(auth
, acred
, flags
, GFP_NOFS
);
1310 static struct rpc_cred
*
1311 gss_create_cred(struct rpc_auth
*auth
, struct auth_cred
*acred
, int flags
, gfp_t gfp
)
1313 struct gss_auth
*gss_auth
= container_of(auth
, struct gss_auth
, rpc_auth
);
1314 struct gss_cred
*cred
= NULL
;
1317 dprintk("RPC: %s for uid %d, flavor %d\n",
1318 __func__
, from_kuid(&init_user_ns
, acred
->uid
),
1321 if (!(cred
= kzalloc(sizeof(*cred
), gfp
)))
1324 rpcauth_init_cred(&cred
->gc_base
, acred
, auth
, &gss_credops
);
1326 * Note: in order to force a call to call_refresh(), we deliberately
1327 * fail to flag the credential as RPCAUTH_CRED_UPTODATE.
1329 cred
->gc_base
.cr_flags
= 1UL << RPCAUTH_CRED_NEW
;
1330 cred
->gc_service
= gss_auth
->service
;
1331 cred
->gc_principal
= NULL
;
1332 if (acred
->machine_cred
)
1333 cred
->gc_principal
= acred
->principal
;
1334 kref_get(&gss_auth
->kref
);
1335 return &cred
->gc_base
;
1338 dprintk("RPC: %s failed with error %d\n", __func__
, err
);
1339 return ERR_PTR(err
);
1343 gss_cred_init(struct rpc_auth
*auth
, struct rpc_cred
*cred
)
1345 struct gss_auth
*gss_auth
= container_of(auth
, struct gss_auth
, rpc_auth
);
1346 struct gss_cred
*gss_cred
= container_of(cred
,struct gss_cred
, gc_base
);
1350 err
= gss_create_upcall(gss_auth
, gss_cred
);
1351 } while (err
== -EAGAIN
);
1356 gss_stringify_acceptor(struct rpc_cred
*cred
)
1358 char *string
= NULL
;
1359 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
1360 struct gss_cl_ctx
*ctx
;
1362 struct xdr_netobj
*acceptor
;
1365 ctx
= rcu_dereference(gss_cred
->gc_ctx
);
1369 len
= ctx
->gc_acceptor
.len
;
1372 /* no point if there's no string */
1376 string
= kmalloc(len
+ 1, GFP_KERNEL
);
1381 ctx
= rcu_dereference(gss_cred
->gc_ctx
);
1383 /* did the ctx disappear or was it replaced by one with no acceptor? */
1384 if (!ctx
|| !ctx
->gc_acceptor
.len
) {
1390 acceptor
= &ctx
->gc_acceptor
;
1393 * Did we find a new acceptor that's longer than the original? Allocate
1394 * a longer buffer and try again.
1396 if (len
< acceptor
->len
) {
1397 len
= acceptor
->len
;
1403 memcpy(string
, acceptor
->data
, acceptor
->len
);
1404 string
[acceptor
->len
] = '\0';
1411 * Returns -EACCES if GSS context is NULL or will expire within the
1412 * timeout (miliseconds)
1415 gss_key_timeout(struct rpc_cred
*rc
)
1417 struct gss_cred
*gss_cred
= container_of(rc
, struct gss_cred
, gc_base
);
1418 struct gss_cl_ctx
*ctx
;
1419 unsigned long timeout
= jiffies
+ (gss_key_expire_timeo
* HZ
);
1423 ctx
= rcu_dereference(gss_cred
->gc_ctx
);
1424 if (!ctx
|| time_after(timeout
, ctx
->gc_expiry
))
1432 gss_match(struct auth_cred
*acred
, struct rpc_cred
*rc
, int flags
)
1434 struct gss_cred
*gss_cred
= container_of(rc
, struct gss_cred
, gc_base
);
1435 struct gss_cl_ctx
*ctx
;
1438 if (test_bit(RPCAUTH_CRED_NEW
, &rc
->cr_flags
))
1440 /* Don't match with creds that have expired. */
1442 ctx
= rcu_dereference(gss_cred
->gc_ctx
);
1443 if (!ctx
|| time_after(jiffies
, ctx
->gc_expiry
)) {
1448 if (!test_bit(RPCAUTH_CRED_UPTODATE
, &rc
->cr_flags
))
1451 if (acred
->principal
!= NULL
) {
1452 if (gss_cred
->gc_principal
== NULL
)
1454 ret
= strcmp(acred
->principal
, gss_cred
->gc_principal
) == 0;
1457 if (gss_cred
->gc_principal
!= NULL
)
1459 ret
= uid_eq(rc
->cr_uid
, acred
->uid
);
1465 /* Notify acred users of GSS context expiration timeout */
1466 if (test_bit(RPC_CRED_NOTIFY_TIMEOUT
, &acred
->ac_flags
) &&
1467 (gss_key_timeout(rc
) != 0)) {
1468 /* test will now be done from generic cred */
1469 test_and_clear_bit(RPC_CRED_NOTIFY_TIMEOUT
, &acred
->ac_flags
);
1470 /* tell NFS layer that key will expire soon */
1471 set_bit(RPC_CRED_KEY_EXPIRE_SOON
, &acred
->ac_flags
);
1477 * Marshal credentials.
1478 * Maybe we should keep a cached credential for performance reasons.
1481 gss_marshal(struct rpc_task
*task
, __be32
*p
)
1483 struct rpc_rqst
*req
= task
->tk_rqstp
;
1484 struct rpc_cred
*cred
= req
->rq_cred
;
1485 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
,
1487 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
1490 struct xdr_netobj mic
;
1492 struct xdr_buf verf_buf
;
1494 dprintk("RPC: %5u %s\n", task
->tk_pid
, __func__
);
1496 *p
++ = htonl(RPC_AUTH_GSS
);
1499 spin_lock(&ctx
->gc_seq_lock
);
1500 req
->rq_seqno
= ctx
->gc_seq
++;
1501 spin_unlock(&ctx
->gc_seq_lock
);
1503 *p
++ = htonl((u32
) RPC_GSS_VERSION
);
1504 *p
++ = htonl((u32
) ctx
->gc_proc
);
1505 *p
++ = htonl((u32
) req
->rq_seqno
);
1506 *p
++ = htonl((u32
) gss_cred
->gc_service
);
1507 p
= xdr_encode_netobj(p
, &ctx
->gc_wire_ctx
);
1508 *cred_len
= htonl((p
- (cred_len
+ 1)) << 2);
1510 /* We compute the checksum for the verifier over the xdr-encoded bytes
1511 * starting with the xid and ending at the end of the credential: */
1512 iov
.iov_base
= xprt_skip_transport_header(req
->rq_xprt
,
1513 req
->rq_snd_buf
.head
[0].iov_base
);
1514 iov
.iov_len
= (u8
*)p
- (u8
*)iov
.iov_base
;
1515 xdr_buf_from_iov(&iov
, &verf_buf
);
1517 /* set verifier flavor*/
1518 *p
++ = htonl(RPC_AUTH_GSS
);
1520 mic
.data
= (u8
*)(p
+ 1);
1521 maj_stat
= gss_get_mic(ctx
->gc_gss_ctx
, &verf_buf
, &mic
);
1522 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
) {
1523 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1524 } else if (maj_stat
!= 0) {
1525 printk("gss_marshal: gss_get_mic FAILED (%d)\n", maj_stat
);
1528 p
= xdr_encode_opaque(p
, NULL
, mic
.len
);
1536 static int gss_renew_cred(struct rpc_task
*task
)
1538 struct rpc_cred
*oldcred
= task
->tk_rqstp
->rq_cred
;
1539 struct gss_cred
*gss_cred
= container_of(oldcred
,
1542 struct rpc_auth
*auth
= oldcred
->cr_auth
;
1543 struct auth_cred acred
= {
1544 .uid
= oldcred
->cr_uid
,
1545 .principal
= gss_cred
->gc_principal
,
1546 .machine_cred
= (gss_cred
->gc_principal
!= NULL
? 1 : 0),
1548 struct rpc_cred
*new;
1550 new = gss_lookup_cred(auth
, &acred
, RPCAUTH_LOOKUP_NEW
);
1552 return PTR_ERR(new);
1553 task
->tk_rqstp
->rq_cred
= new;
1554 put_rpccred(oldcred
);
1558 static int gss_cred_is_negative_entry(struct rpc_cred
*cred
)
1560 if (test_bit(RPCAUTH_CRED_NEGATIVE
, &cred
->cr_flags
)) {
1561 unsigned long now
= jiffies
;
1562 unsigned long begin
, expire
;
1563 struct gss_cred
*gss_cred
;
1565 gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
1566 begin
= gss_cred
->gc_upcall_timestamp
;
1567 expire
= begin
+ gss_expired_cred_retry_delay
* HZ
;
1569 if (time_in_range_open(now
, begin
, expire
))
1576 * Refresh credentials. XXX - finish
1579 gss_refresh(struct rpc_task
*task
)
1581 struct rpc_cred
*cred
= task
->tk_rqstp
->rq_cred
;
1584 if (gss_cred_is_negative_entry(cred
))
1585 return -EKEYEXPIRED
;
1587 if (!test_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
) &&
1588 !test_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
)) {
1589 ret
= gss_renew_cred(task
);
1592 cred
= task
->tk_rqstp
->rq_cred
;
1595 if (test_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
))
1596 ret
= gss_refresh_upcall(task
);
1601 /* Dummy refresh routine: used only when destroying the context */
1603 gss_refresh_null(struct rpc_task
*task
)
1609 gss_validate(struct rpc_task
*task
, __be32
*p
)
1611 struct rpc_cred
*cred
= task
->tk_rqstp
->rq_cred
;
1612 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
1615 struct xdr_buf verf_buf
;
1616 struct xdr_netobj mic
;
1619 __be32
*ret
= ERR_PTR(-EIO
);
1621 dprintk("RPC: %5u %s\n", task
->tk_pid
, __func__
);
1624 if ((len
= ntohl(*p
++)) > RPC_MAX_AUTH_SIZE
)
1626 if (flav
!= RPC_AUTH_GSS
)
1628 seq
= htonl(task
->tk_rqstp
->rq_seqno
);
1629 iov
.iov_base
= &seq
;
1630 iov
.iov_len
= sizeof(seq
);
1631 xdr_buf_from_iov(&iov
, &verf_buf
);
1635 ret
= ERR_PTR(-EACCES
);
1636 maj_stat
= gss_verify_mic(ctx
->gc_gss_ctx
, &verf_buf
, &mic
);
1637 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1638 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1640 dprintk("RPC: %5u %s: gss_verify_mic returned error 0x%08x\n",
1641 task
->tk_pid
, __func__
, maj_stat
);
1644 /* We leave it to unwrap to calculate au_rslack. For now we just
1645 * calculate the length of the verifier: */
1646 cred
->cr_auth
->au_verfsize
= XDR_QUADLEN(len
) + 2;
1648 dprintk("RPC: %5u %s: gss_verify_mic succeeded.\n",
1649 task
->tk_pid
, __func__
);
1650 return p
+ XDR_QUADLEN(len
);
1653 dprintk("RPC: %5u %s failed ret %ld.\n", task
->tk_pid
, __func__
,
1658 static void gss_wrap_req_encode(kxdreproc_t encode
, struct rpc_rqst
*rqstp
,
1659 __be32
*p
, void *obj
)
1661 struct xdr_stream xdr
;
1663 xdr_init_encode(&xdr
, &rqstp
->rq_snd_buf
, p
);
1664 encode(rqstp
, &xdr
, obj
);
1668 gss_wrap_req_integ(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
,
1669 kxdreproc_t encode
, struct rpc_rqst
*rqstp
,
1670 __be32
*p
, void *obj
)
1672 struct xdr_buf
*snd_buf
= &rqstp
->rq_snd_buf
;
1673 struct xdr_buf integ_buf
;
1674 __be32
*integ_len
= NULL
;
1675 struct xdr_netobj mic
;
1683 offset
= (u8
*)p
- (u8
*)snd_buf
->head
[0].iov_base
;
1684 *p
++ = htonl(rqstp
->rq_seqno
);
1686 gss_wrap_req_encode(encode
, rqstp
, p
, obj
);
1688 if (xdr_buf_subsegment(snd_buf
, &integ_buf
,
1689 offset
, snd_buf
->len
- offset
))
1691 *integ_len
= htonl(integ_buf
.len
);
1693 /* guess whether we're in the head or the tail: */
1694 if (snd_buf
->page_len
|| snd_buf
->tail
[0].iov_len
)
1695 iov
= snd_buf
->tail
;
1697 iov
= snd_buf
->head
;
1698 p
= iov
->iov_base
+ iov
->iov_len
;
1699 mic
.data
= (u8
*)(p
+ 1);
1701 maj_stat
= gss_get_mic(ctx
->gc_gss_ctx
, &integ_buf
, &mic
);
1702 status
= -EIO
; /* XXX? */
1703 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1704 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1707 q
= xdr_encode_opaque(p
, NULL
, mic
.len
);
1709 offset
= (u8
*)q
- (u8
*)p
;
1710 iov
->iov_len
+= offset
;
1711 snd_buf
->len
+= offset
;
1716 priv_release_snd_buf(struct rpc_rqst
*rqstp
)
1720 for (i
=0; i
< rqstp
->rq_enc_pages_num
; i
++)
1721 __free_page(rqstp
->rq_enc_pages
[i
]);
1722 kfree(rqstp
->rq_enc_pages
);
1726 alloc_enc_pages(struct rpc_rqst
*rqstp
)
1728 struct xdr_buf
*snd_buf
= &rqstp
->rq_snd_buf
;
1731 if (snd_buf
->page_len
== 0) {
1732 rqstp
->rq_enc_pages_num
= 0;
1736 first
= snd_buf
->page_base
>> PAGE_SHIFT
;
1737 last
= (snd_buf
->page_base
+ snd_buf
->page_len
- 1) >> PAGE_SHIFT
;
1738 rqstp
->rq_enc_pages_num
= last
- first
+ 1 + 1;
1740 = kmalloc(rqstp
->rq_enc_pages_num
* sizeof(struct page
*),
1742 if (!rqstp
->rq_enc_pages
)
1744 for (i
=0; i
< rqstp
->rq_enc_pages_num
; i
++) {
1745 rqstp
->rq_enc_pages
[i
] = alloc_page(GFP_NOFS
);
1746 if (rqstp
->rq_enc_pages
[i
] == NULL
)
1749 rqstp
->rq_release_snd_buf
= priv_release_snd_buf
;
1752 rqstp
->rq_enc_pages_num
= i
;
1753 priv_release_snd_buf(rqstp
);
1759 gss_wrap_req_priv(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
,
1760 kxdreproc_t encode
, struct rpc_rqst
*rqstp
,
1761 __be32
*p
, void *obj
)
1763 struct xdr_buf
*snd_buf
= &rqstp
->rq_snd_buf
;
1768 struct page
**inpages
;
1775 offset
= (u8
*)p
- (u8
*)snd_buf
->head
[0].iov_base
;
1776 *p
++ = htonl(rqstp
->rq_seqno
);
1778 gss_wrap_req_encode(encode
, rqstp
, p
, obj
);
1780 status
= alloc_enc_pages(rqstp
);
1783 first
= snd_buf
->page_base
>> PAGE_SHIFT
;
1784 inpages
= snd_buf
->pages
+ first
;
1785 snd_buf
->pages
= rqstp
->rq_enc_pages
;
1786 snd_buf
->page_base
-= first
<< PAGE_SHIFT
;
1788 * Give the tail its own page, in case we need extra space in the
1789 * head when wrapping:
1791 * call_allocate() allocates twice the slack space required
1792 * by the authentication flavor to rq_callsize.
1793 * For GSS, slack is GSS_CRED_SLACK.
1795 if (snd_buf
->page_len
|| snd_buf
->tail
[0].iov_len
) {
1796 tmp
= page_address(rqstp
->rq_enc_pages
[rqstp
->rq_enc_pages_num
- 1]);
1797 memcpy(tmp
, snd_buf
->tail
[0].iov_base
, snd_buf
->tail
[0].iov_len
);
1798 snd_buf
->tail
[0].iov_base
= tmp
;
1800 maj_stat
= gss_wrap(ctx
->gc_gss_ctx
, offset
, snd_buf
, inpages
);
1801 /* slack space should prevent this ever happening: */
1802 BUG_ON(snd_buf
->len
> snd_buf
->buflen
);
1804 /* We're assuming that when GSS_S_CONTEXT_EXPIRED, the encryption was
1805 * done anyway, so it's safe to put the request on the wire: */
1806 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1807 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1811 *opaque_len
= htonl(snd_buf
->len
- offset
);
1812 /* guess whether we're in the head or the tail: */
1813 if (snd_buf
->page_len
|| snd_buf
->tail
[0].iov_len
)
1814 iov
= snd_buf
->tail
;
1816 iov
= snd_buf
->head
;
1817 p
= iov
->iov_base
+ iov
->iov_len
;
1818 pad
= 3 - ((snd_buf
->len
- offset
- 1) & 3);
1820 iov
->iov_len
+= pad
;
1821 snd_buf
->len
+= pad
;
1827 gss_wrap_req(struct rpc_task
*task
,
1828 kxdreproc_t encode
, void *rqstp
, __be32
*p
, void *obj
)
1830 struct rpc_cred
*cred
= task
->tk_rqstp
->rq_cred
;
1831 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
,
1833 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
1836 dprintk("RPC: %5u %s\n", task
->tk_pid
, __func__
);
1837 if (ctx
->gc_proc
!= RPC_GSS_PROC_DATA
) {
1838 /* The spec seems a little ambiguous here, but I think that not
1839 * wrapping context destruction requests makes the most sense.
1841 gss_wrap_req_encode(encode
, rqstp
, p
, obj
);
1845 switch (gss_cred
->gc_service
) {
1846 case RPC_GSS_SVC_NONE
:
1847 gss_wrap_req_encode(encode
, rqstp
, p
, obj
);
1850 case RPC_GSS_SVC_INTEGRITY
:
1851 status
= gss_wrap_req_integ(cred
, ctx
, encode
, rqstp
, p
, obj
);
1853 case RPC_GSS_SVC_PRIVACY
:
1854 status
= gss_wrap_req_priv(cred
, ctx
, encode
, rqstp
, p
, obj
);
1859 dprintk("RPC: %5u %s returning %d\n", task
->tk_pid
, __func__
, status
);
1864 gss_unwrap_resp_integ(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
,
1865 struct rpc_rqst
*rqstp
, __be32
**p
)
1867 struct xdr_buf
*rcv_buf
= &rqstp
->rq_rcv_buf
;
1868 struct xdr_buf integ_buf
;
1869 struct xdr_netobj mic
;
1870 u32 data_offset
, mic_offset
;
1875 integ_len
= ntohl(*(*p
)++);
1878 data_offset
= (u8
*)(*p
) - (u8
*)rcv_buf
->head
[0].iov_base
;
1879 mic_offset
= integ_len
+ data_offset
;
1880 if (mic_offset
> rcv_buf
->len
)
1882 if (ntohl(*(*p
)++) != rqstp
->rq_seqno
)
1885 if (xdr_buf_subsegment(rcv_buf
, &integ_buf
, data_offset
,
1886 mic_offset
- data_offset
))
1889 if (xdr_buf_read_netobj(rcv_buf
, &mic
, mic_offset
))
1892 maj_stat
= gss_verify_mic(ctx
->gc_gss_ctx
, &integ_buf
, &mic
);
1893 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1894 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1895 if (maj_stat
!= GSS_S_COMPLETE
)
1901 gss_unwrap_resp_priv(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
,
1902 struct rpc_rqst
*rqstp
, __be32
**p
)
1904 struct xdr_buf
*rcv_buf
= &rqstp
->rq_rcv_buf
;
1910 opaque_len
= ntohl(*(*p
)++);
1911 offset
= (u8
*)(*p
) - (u8
*)rcv_buf
->head
[0].iov_base
;
1912 if (offset
+ opaque_len
> rcv_buf
->len
)
1914 /* remove padding: */
1915 rcv_buf
->len
= offset
+ opaque_len
;
1917 maj_stat
= gss_unwrap(ctx
->gc_gss_ctx
, offset
, rcv_buf
);
1918 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1919 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1920 if (maj_stat
!= GSS_S_COMPLETE
)
1922 if (ntohl(*(*p
)++) != rqstp
->rq_seqno
)
1929 gss_unwrap_req_decode(kxdrdproc_t decode
, struct rpc_rqst
*rqstp
,
1930 __be32
*p
, void *obj
)
1932 struct xdr_stream xdr
;
1934 xdr_init_decode(&xdr
, &rqstp
->rq_rcv_buf
, p
);
1935 return decode(rqstp
, &xdr
, obj
);
1939 gss_unwrap_resp(struct rpc_task
*task
,
1940 kxdrdproc_t decode
, void *rqstp
, __be32
*p
, void *obj
)
1942 struct rpc_cred
*cred
= task
->tk_rqstp
->rq_cred
;
1943 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
,
1945 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
1947 struct kvec
*head
= ((struct rpc_rqst
*)rqstp
)->rq_rcv_buf
.head
;
1948 int savedlen
= head
->iov_len
;
1951 if (ctx
->gc_proc
!= RPC_GSS_PROC_DATA
)
1953 switch (gss_cred
->gc_service
) {
1954 case RPC_GSS_SVC_NONE
:
1956 case RPC_GSS_SVC_INTEGRITY
:
1957 status
= gss_unwrap_resp_integ(cred
, ctx
, rqstp
, &p
);
1961 case RPC_GSS_SVC_PRIVACY
:
1962 status
= gss_unwrap_resp_priv(cred
, ctx
, rqstp
, &p
);
1967 /* take into account extra slack for integrity and privacy cases: */
1968 cred
->cr_auth
->au_rslack
= cred
->cr_auth
->au_verfsize
+ (p
- savedp
)
1969 + (savedlen
- head
->iov_len
);
1971 status
= gss_unwrap_req_decode(decode
, rqstp
, p
, obj
);
1974 dprintk("RPC: %5u %s returning %d\n",
1975 task
->tk_pid
, __func__
, status
);
1979 static const struct rpc_authops authgss_ops
= {
1980 .owner
= THIS_MODULE
,
1981 .au_flavor
= RPC_AUTH_GSS
,
1982 .au_name
= "RPCSEC_GSS",
1983 .create
= gss_create
,
1984 .destroy
= gss_destroy
,
1985 .lookup_cred
= gss_lookup_cred
,
1986 .crcreate
= gss_create_cred
,
1987 .list_pseudoflavors
= gss_mech_list_pseudoflavors
,
1988 .info2flavor
= gss_mech_info2flavor
,
1989 .flavor2info
= gss_mech_flavor2info
,
1992 static const struct rpc_credops gss_credops
= {
1993 .cr_name
= "AUTH_GSS",
1994 .crdestroy
= gss_destroy_cred
,
1995 .cr_init
= gss_cred_init
,
1996 .crbind
= rpcauth_generic_bind_cred
,
1997 .crmatch
= gss_match
,
1998 .crmarshal
= gss_marshal
,
1999 .crrefresh
= gss_refresh
,
2000 .crvalidate
= gss_validate
,
2001 .crwrap_req
= gss_wrap_req
,
2002 .crunwrap_resp
= gss_unwrap_resp
,
2003 .crkey_timeout
= gss_key_timeout
,
2004 .crstringify_acceptor
= gss_stringify_acceptor
,
2007 static const struct rpc_credops gss_nullops
= {
2008 .cr_name
= "AUTH_GSS",
2009 .crdestroy
= gss_destroy_nullcred
,
2010 .crbind
= rpcauth_generic_bind_cred
,
2011 .crmatch
= gss_match
,
2012 .crmarshal
= gss_marshal
,
2013 .crrefresh
= gss_refresh_null
,
2014 .crvalidate
= gss_validate
,
2015 .crwrap_req
= gss_wrap_req
,
2016 .crunwrap_resp
= gss_unwrap_resp
,
2017 .crstringify_acceptor
= gss_stringify_acceptor
,
2020 static const struct rpc_pipe_ops gss_upcall_ops_v0
= {
2021 .upcall
= rpc_pipe_generic_upcall
,
2022 .downcall
= gss_pipe_downcall
,
2023 .destroy_msg
= gss_pipe_destroy_msg
,
2024 .open_pipe
= gss_pipe_open_v0
,
2025 .release_pipe
= gss_pipe_release
,
2028 static const struct rpc_pipe_ops gss_upcall_ops_v1
= {
2029 .upcall
= rpc_pipe_generic_upcall
,
2030 .downcall
= gss_pipe_downcall
,
2031 .destroy_msg
= gss_pipe_destroy_msg
,
2032 .open_pipe
= gss_pipe_open_v1
,
2033 .release_pipe
= gss_pipe_release
,
2036 static __net_init
int rpcsec_gss_init_net(struct net
*net
)
2038 return gss_svc_init_net(net
);
2041 static __net_exit
void rpcsec_gss_exit_net(struct net
*net
)
2043 gss_svc_shutdown_net(net
);
2046 static struct pernet_operations rpcsec_gss_net_ops
= {
2047 .init
= rpcsec_gss_init_net
,
2048 .exit
= rpcsec_gss_exit_net
,
2052 * Initialize RPCSEC_GSS module
2054 static int __init
init_rpcsec_gss(void)
2058 err
= rpcauth_register(&authgss_ops
);
2061 err
= gss_svc_init();
2063 goto out_unregister
;
2064 err
= register_pernet_subsys(&rpcsec_gss_net_ops
);
2067 rpc_init_wait_queue(&pipe_version_rpc_waitqueue
, "gss pipe version");
2072 rpcauth_unregister(&authgss_ops
);
2077 static void __exit
exit_rpcsec_gss(void)
2079 unregister_pernet_subsys(&rpcsec_gss_net_ops
);
2081 rpcauth_unregister(&authgss_ops
);
2082 rcu_barrier(); /* Wait for completion of call_rcu()'s */
2085 MODULE_ALIAS("rpc-auth-6");
2086 MODULE_LICENSE("GPL");
2087 module_param_named(expired_cred_retry_delay
,
2088 gss_expired_cred_retry_delay
,
2090 MODULE_PARM_DESC(expired_cred_retry_delay
, "Timeout (in seconds) until "
2091 "the RPC engine retries an expired credential");
2093 module_param_named(key_expire_timeo
,
2094 gss_key_expire_timeo
,
2096 MODULE_PARM_DESC(key_expire_timeo
, "Time (in seconds) at the end of a "
2097 "credential keys lifetime where the NFS layer cleans up "
2098 "prior to key expiration");
2100 module_init(init_rpcsec_gss
)
2101 module_exit(exit_rpcsec_gss
)