2 * linux/net/sunrpc/auth_gss/auth_gss.c
4 * RPCSEC_GSS client authentication.
6 * Copyright (c) 2000 The Regents of the University of Michigan.
9 * Dug Song <dugsong@monkey.org>
10 * Andy Adamson <andros@umich.edu>
12 * Redistribution and use in source and binary forms, with or without
13 * modification, are permitted provided that the following conditions
16 * 1. Redistributions of source code must retain the above copyright
17 * notice, this list of conditions and the following disclaimer.
18 * 2. Redistributions in binary form must reproduce the above copyright
19 * notice, this list of conditions and the following disclaimer in the
20 * documentation and/or other materials provided with the distribution.
21 * 3. Neither the name of the University nor the names of its
22 * contributors may be used to endorse or promote products derived
23 * from this software without specific prior written permission.
25 * THIS SOFTWARE IS PROVIDED ``AS IS'' AND ANY EXPRESS OR IMPLIED
26 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
27 * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
28 * DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
29 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
30 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
31 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
32 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
33 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
34 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
35 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
39 #include <linux/module.h>
40 #include <linux/init.h>
41 #include <linux/types.h>
42 #include <linux/slab.h>
43 #include <linux/sched.h>
44 #include <linux/pagemap.h>
45 #include <linux/sunrpc/clnt.h>
46 #include <linux/sunrpc/auth.h>
47 #include <linux/sunrpc/auth_gss.h>
48 #include <linux/sunrpc/svcauth_gss.h>
49 #include <linux/sunrpc/gss_err.h>
50 #include <linux/workqueue.h>
51 #include <linux/sunrpc/rpc_pipe_fs.h>
52 #include <linux/sunrpc/gss_api.h>
53 #include <asm/uaccess.h>
55 static const struct rpc_authops authgss_ops
;
57 static const struct rpc_credops gss_credops
;
58 static const struct rpc_credops gss_nullops
;
61 # define RPCDBG_FACILITY RPCDBG_AUTH
64 #define NFS_NGROUPS 16
66 #define GSS_CRED_SLACK 1024 /* XXX: unused */
67 /* length of a krb5 verifier (48), plus data added before arguments when
68 * using integrity (two 4-byte integers): */
69 #define GSS_VERF_SLACK 100
71 /* XXX this define must match the gssd define
72 * as it is passed to gssd to signal the use of
73 * machine creds should be part of the shared rpc interface */
75 #define CA_RUN_AS_MACHINE 0x00000200
77 /* dump the buffer in `emacs-hexl' style */
78 #define isprint(c) ((c > 0x1f) && (c < 0x7f))
82 struct rpc_auth rpc_auth
;
83 struct gss_api_mech
*mech
;
84 enum rpc_gss_svc service
;
85 struct rpc_clnt
*client
;
86 struct dentry
*dentry
;
89 static void gss_free_ctx(struct gss_cl_ctx
*);
90 static struct rpc_pipe_ops gss_upcall_ops
;
92 static inline struct gss_cl_ctx
*
93 gss_get_ctx(struct gss_cl_ctx
*ctx
)
95 atomic_inc(&ctx
->count
);
100 gss_put_ctx(struct gss_cl_ctx
*ctx
)
102 if (atomic_dec_and_test(&ctx
->count
))
107 * called by gss_upcall_callback and gss_create_upcall in order
108 * to set the gss context. The actual exchange of an old context
109 * and a new one is protected by the inode->i_lock.
112 gss_cred_set_ctx(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
)
114 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
116 if (!test_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
))
119 rcu_assign_pointer(gss_cred
->gc_ctx
, ctx
);
120 set_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
121 smp_mb__before_clear_bit();
122 clear_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
);
126 simple_get_bytes(const void *p
, const void *end
, void *res
, size_t len
)
128 const void *q
= (const void *)((const char *)p
+ len
);
129 if (unlikely(q
> end
|| q
< p
))
130 return ERR_PTR(-EFAULT
);
135 static inline const void *
136 simple_get_netobj(const void *p
, const void *end
, struct xdr_netobj
*dest
)
141 p
= simple_get_bytes(p
, end
, &len
, sizeof(len
));
144 q
= (const void *)((const char *)p
+ len
);
145 if (unlikely(q
> end
|| q
< p
))
146 return ERR_PTR(-EFAULT
);
147 dest
->data
= kmemdup(p
, len
, GFP_KERNEL
);
148 if (unlikely(dest
->data
== NULL
))
149 return ERR_PTR(-ENOMEM
);
154 static struct gss_cl_ctx
*
155 gss_cred_get_ctx(struct rpc_cred
*cred
)
157 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
158 struct gss_cl_ctx
*ctx
= NULL
;
161 if (gss_cred
->gc_ctx
)
162 ctx
= gss_get_ctx(gss_cred
->gc_ctx
);
167 static struct gss_cl_ctx
*
168 gss_alloc_context(void)
170 struct gss_cl_ctx
*ctx
;
172 ctx
= kzalloc(sizeof(*ctx
), GFP_KERNEL
);
174 ctx
->gc_proc
= RPC_GSS_PROC_DATA
;
175 ctx
->gc_seq
= 1; /* NetApp 6.4R1 doesn't accept seq. no. 0 */
176 spin_lock_init(&ctx
->gc_seq_lock
);
177 atomic_set(&ctx
->count
,1);
182 #define GSSD_MIN_TIMEOUT (60 * 60)
184 gss_fill_context(const void *p
, const void *end
, struct gss_cl_ctx
*ctx
, struct gss_api_mech
*gm
)
188 unsigned int timeout
;
192 /* First unsigned int gives the lifetime (in seconds) of the cred */
193 p
= simple_get_bytes(p
, end
, &timeout
, sizeof(timeout
));
197 timeout
= GSSD_MIN_TIMEOUT
;
198 ctx
->gc_expiry
= jiffies
+ (unsigned long)timeout
* HZ
* 3 / 4;
199 /* Sequence number window. Determines the maximum number of simultaneous requests */
200 p
= simple_get_bytes(p
, end
, &window_size
, sizeof(window_size
));
203 ctx
->gc_win
= window_size
;
204 /* gssd signals an error by passing ctx->gc_win = 0: */
205 if (ctx
->gc_win
== 0) {
206 /* in which case, p points to an error code which we ignore */
207 p
= ERR_PTR(-EACCES
);
210 /* copy the opaque wire context */
211 p
= simple_get_netobj(p
, end
, &ctx
->gc_wire_ctx
);
214 /* import the opaque security context */
215 p
= simple_get_bytes(p
, end
, &seclen
, sizeof(seclen
));
218 q
= (const void *)((const char *)p
+ seclen
);
219 if (unlikely(q
> end
|| q
< p
)) {
220 p
= ERR_PTR(-EFAULT
);
223 ret
= gss_import_sec_context(p
, seclen
, gm
, &ctx
->gc_gss_ctx
);
230 dprintk("RPC: gss_fill_context returning %ld\n", -PTR_ERR(p
));
235 struct gss_upcall_msg
{
238 struct rpc_pipe_msg msg
;
239 struct list_head list
;
240 struct gss_auth
*auth
;
241 struct rpc_wait_queue rpc_waitqueue
;
242 wait_queue_head_t waitqueue
;
243 struct gss_cl_ctx
*ctx
;
247 gss_release_msg(struct gss_upcall_msg
*gss_msg
)
249 if (!atomic_dec_and_test(&gss_msg
->count
))
251 BUG_ON(!list_empty(&gss_msg
->list
));
252 if (gss_msg
->ctx
!= NULL
)
253 gss_put_ctx(gss_msg
->ctx
);
254 rpc_destroy_wait_queue(&gss_msg
->rpc_waitqueue
);
258 static struct gss_upcall_msg
*
259 __gss_find_upcall(struct rpc_inode
*rpci
, uid_t uid
)
261 struct gss_upcall_msg
*pos
;
262 list_for_each_entry(pos
, &rpci
->in_downcall
, list
) {
265 atomic_inc(&pos
->count
);
266 dprintk("RPC: gss_find_upcall found msg %p\n", pos
);
269 dprintk("RPC: gss_find_upcall found nothing\n");
273 /* Try to add a upcall to the pipefs queue.
274 * If an upcall owned by our uid already exists, then we return a reference
275 * to that upcall instead of adding the new upcall.
277 static inline struct gss_upcall_msg
*
278 gss_add_msg(struct gss_auth
*gss_auth
, struct gss_upcall_msg
*gss_msg
)
280 struct inode
*inode
= gss_auth
->dentry
->d_inode
;
281 struct rpc_inode
*rpci
= RPC_I(inode
);
282 struct gss_upcall_msg
*old
;
284 spin_lock(&inode
->i_lock
);
285 old
= __gss_find_upcall(rpci
, gss_msg
->uid
);
287 atomic_inc(&gss_msg
->count
);
288 list_add(&gss_msg
->list
, &rpci
->in_downcall
);
291 spin_unlock(&inode
->i_lock
);
296 __gss_unhash_msg(struct gss_upcall_msg
*gss_msg
)
298 list_del_init(&gss_msg
->list
);
299 rpc_wake_up_status(&gss_msg
->rpc_waitqueue
, gss_msg
->msg
.errno
);
300 wake_up_all(&gss_msg
->waitqueue
);
301 atomic_dec(&gss_msg
->count
);
305 gss_unhash_msg(struct gss_upcall_msg
*gss_msg
)
307 struct gss_auth
*gss_auth
= gss_msg
->auth
;
308 struct inode
*inode
= gss_auth
->dentry
->d_inode
;
310 if (list_empty(&gss_msg
->list
))
312 spin_lock(&inode
->i_lock
);
313 if (!list_empty(&gss_msg
->list
))
314 __gss_unhash_msg(gss_msg
);
315 spin_unlock(&inode
->i_lock
);
319 gss_upcall_callback(struct rpc_task
*task
)
321 struct gss_cred
*gss_cred
= container_of(task
->tk_msg
.rpc_cred
,
322 struct gss_cred
, gc_base
);
323 struct gss_upcall_msg
*gss_msg
= gss_cred
->gc_upcall
;
324 struct inode
*inode
= gss_msg
->auth
->dentry
->d_inode
;
326 spin_lock(&inode
->i_lock
);
328 gss_cred_set_ctx(task
->tk_msg
.rpc_cred
, gss_msg
->ctx
);
330 task
->tk_status
= gss_msg
->msg
.errno
;
331 gss_cred
->gc_upcall
= NULL
;
332 rpc_wake_up_status(&gss_msg
->rpc_waitqueue
, gss_msg
->msg
.errno
);
333 spin_unlock(&inode
->i_lock
);
334 gss_release_msg(gss_msg
);
337 static inline struct gss_upcall_msg
*
338 gss_alloc_msg(struct gss_auth
*gss_auth
, uid_t uid
)
340 struct gss_upcall_msg
*gss_msg
;
342 gss_msg
= kzalloc(sizeof(*gss_msg
), GFP_KERNEL
);
343 if (gss_msg
!= NULL
) {
344 INIT_LIST_HEAD(&gss_msg
->list
);
345 rpc_init_wait_queue(&gss_msg
->rpc_waitqueue
, "RPCSEC_GSS upcall waitq");
346 init_waitqueue_head(&gss_msg
->waitqueue
);
347 atomic_set(&gss_msg
->count
, 1);
348 gss_msg
->msg
.data
= &gss_msg
->uid
;
349 gss_msg
->msg
.len
= sizeof(gss_msg
->uid
);
351 gss_msg
->auth
= gss_auth
;
356 static struct gss_upcall_msg
*
357 gss_setup_upcall(struct rpc_clnt
*clnt
, struct gss_auth
*gss_auth
, struct rpc_cred
*cred
)
359 struct gss_cred
*gss_cred
= container_of(cred
,
360 struct gss_cred
, gc_base
);
361 struct gss_upcall_msg
*gss_new
, *gss_msg
;
362 uid_t uid
= cred
->cr_uid
;
364 /* Special case: rpc.gssd assumes that uid == 0 implies machine creds */
365 if (gss_cred
->gc_machine_cred
!= 0)
368 gss_new
= gss_alloc_msg(gss_auth
, uid
);
370 return ERR_PTR(-ENOMEM
);
371 gss_msg
= gss_add_msg(gss_auth
, gss_new
);
372 if (gss_msg
== gss_new
) {
373 int res
= rpc_queue_upcall(gss_auth
->dentry
->d_inode
, &gss_new
->msg
);
375 gss_unhash_msg(gss_new
);
376 gss_msg
= ERR_PTR(res
);
379 gss_release_msg(gss_new
);
384 gss_refresh_upcall(struct rpc_task
*task
)
386 struct rpc_cred
*cred
= task
->tk_msg
.rpc_cred
;
387 struct gss_auth
*gss_auth
= container_of(cred
->cr_auth
,
388 struct gss_auth
, rpc_auth
);
389 struct gss_cred
*gss_cred
= container_of(cred
,
390 struct gss_cred
, gc_base
);
391 struct gss_upcall_msg
*gss_msg
;
392 struct inode
*inode
= gss_auth
->dentry
->d_inode
;
395 dprintk("RPC: %5u gss_refresh_upcall for uid %u\n", task
->tk_pid
,
397 gss_msg
= gss_setup_upcall(task
->tk_client
, gss_auth
, cred
);
398 if (IS_ERR(gss_msg
)) {
399 err
= PTR_ERR(gss_msg
);
402 spin_lock(&inode
->i_lock
);
403 if (gss_cred
->gc_upcall
!= NULL
)
404 rpc_sleep_on(&gss_cred
->gc_upcall
->rpc_waitqueue
, task
, NULL
);
405 else if (gss_msg
->ctx
!= NULL
) {
406 gss_cred_set_ctx(task
->tk_msg
.rpc_cred
, gss_msg
->ctx
);
407 gss_cred
->gc_upcall
= NULL
;
408 rpc_wake_up_status(&gss_msg
->rpc_waitqueue
, gss_msg
->msg
.errno
);
409 } else if (gss_msg
->msg
.errno
>= 0) {
410 task
->tk_timeout
= 0;
411 gss_cred
->gc_upcall
= gss_msg
;
412 /* gss_upcall_callback will release the reference to gss_upcall_msg */
413 atomic_inc(&gss_msg
->count
);
414 rpc_sleep_on(&gss_msg
->rpc_waitqueue
, task
, gss_upcall_callback
);
416 err
= gss_msg
->msg
.errno
;
417 spin_unlock(&inode
->i_lock
);
418 gss_release_msg(gss_msg
);
420 dprintk("RPC: %5u gss_refresh_upcall for uid %u result %d\n",
421 task
->tk_pid
, cred
->cr_uid
, err
);
426 gss_create_upcall(struct gss_auth
*gss_auth
, struct gss_cred
*gss_cred
)
428 struct inode
*inode
= gss_auth
->dentry
->d_inode
;
429 struct rpc_cred
*cred
= &gss_cred
->gc_base
;
430 struct gss_upcall_msg
*gss_msg
;
434 dprintk("RPC: gss_upcall for uid %u\n", cred
->cr_uid
);
435 gss_msg
= gss_setup_upcall(gss_auth
->client
, gss_auth
, cred
);
436 if (IS_ERR(gss_msg
)) {
437 err
= PTR_ERR(gss_msg
);
441 prepare_to_wait(&gss_msg
->waitqueue
, &wait
, TASK_INTERRUPTIBLE
);
442 spin_lock(&inode
->i_lock
);
443 if (gss_msg
->ctx
!= NULL
|| gss_msg
->msg
.errno
< 0) {
446 spin_unlock(&inode
->i_lock
);
454 gss_cred_set_ctx(cred
, gss_msg
->ctx
);
456 err
= gss_msg
->msg
.errno
;
457 spin_unlock(&inode
->i_lock
);
459 finish_wait(&gss_msg
->waitqueue
, &wait
);
460 gss_release_msg(gss_msg
);
462 dprintk("RPC: gss_create_upcall for uid %u result %d\n",
468 gss_pipe_upcall(struct file
*filp
, struct rpc_pipe_msg
*msg
,
469 char __user
*dst
, size_t buflen
)
471 char *data
= (char *)msg
->data
+ msg
->copied
;
472 size_t mlen
= min(msg
->len
, buflen
);
475 left
= copy_to_user(dst
, data
, mlen
);
477 msg
->errno
= -EFAULT
;
487 #define MSG_BUF_MAXSIZE 1024
490 gss_pipe_downcall(struct file
*filp
, const char __user
*src
, size_t mlen
)
494 struct rpc_clnt
*clnt
;
495 struct gss_upcall_msg
*gss_msg
;
496 struct inode
*inode
= filp
->f_path
.dentry
->d_inode
;
497 struct gss_cl_ctx
*ctx
;
499 ssize_t err
= -EFBIG
;
501 if (mlen
> MSG_BUF_MAXSIZE
)
504 buf
= kmalloc(mlen
, GFP_KERNEL
);
508 clnt
= RPC_I(inode
)->private;
510 if (copy_from_user(buf
, src
, mlen
))
513 end
= (const void *)((char *)buf
+ mlen
);
514 p
= simple_get_bytes(buf
, end
, &uid
, sizeof(uid
));
521 ctx
= gss_alloc_context();
526 /* Find a matching upcall */
527 spin_lock(&inode
->i_lock
);
528 gss_msg
= __gss_find_upcall(RPC_I(inode
), uid
);
529 if (gss_msg
== NULL
) {
530 spin_unlock(&inode
->i_lock
);
533 list_del_init(&gss_msg
->list
);
534 spin_unlock(&inode
->i_lock
);
536 p
= gss_fill_context(p
, end
, ctx
, gss_msg
->auth
->mech
);
539 gss_msg
->msg
.errno
= (err
== -EAGAIN
) ? -EAGAIN
: -EACCES
;
540 goto err_release_msg
;
542 gss_msg
->ctx
= gss_get_ctx(ctx
);
546 spin_lock(&inode
->i_lock
);
547 __gss_unhash_msg(gss_msg
);
548 spin_unlock(&inode
->i_lock
);
549 gss_release_msg(gss_msg
);
555 dprintk("RPC: gss_pipe_downcall returning %Zd\n", err
);
560 gss_pipe_release(struct inode
*inode
)
562 struct rpc_inode
*rpci
= RPC_I(inode
);
563 struct gss_upcall_msg
*gss_msg
;
565 spin_lock(&inode
->i_lock
);
566 while (!list_empty(&rpci
->in_downcall
)) {
568 gss_msg
= list_entry(rpci
->in_downcall
.next
,
569 struct gss_upcall_msg
, list
);
570 gss_msg
->msg
.errno
= -EPIPE
;
571 atomic_inc(&gss_msg
->count
);
572 __gss_unhash_msg(gss_msg
);
573 spin_unlock(&inode
->i_lock
);
574 gss_release_msg(gss_msg
);
575 spin_lock(&inode
->i_lock
);
577 spin_unlock(&inode
->i_lock
);
581 gss_pipe_destroy_msg(struct rpc_pipe_msg
*msg
)
583 struct gss_upcall_msg
*gss_msg
= container_of(msg
, struct gss_upcall_msg
, msg
);
584 static unsigned long ratelimit
;
586 if (msg
->errno
< 0) {
587 dprintk("RPC: gss_pipe_destroy_msg releasing msg %p\n",
589 atomic_inc(&gss_msg
->count
);
590 gss_unhash_msg(gss_msg
);
591 if (msg
->errno
== -ETIMEDOUT
) {
592 unsigned long now
= jiffies
;
593 if (time_after(now
, ratelimit
)) {
594 printk(KERN_WARNING
"RPC: AUTH_GSS upcall timed out.\n"
595 "Please check user daemon is running!\n");
596 ratelimit
= now
+ 15*HZ
;
599 gss_release_msg(gss_msg
);
604 * NOTE: we have the opportunity to use different
605 * parameters based on the input flavor (which must be a pseudoflavor)
607 static struct rpc_auth
*
608 gss_create(struct rpc_clnt
*clnt
, rpc_authflavor_t flavor
)
610 struct gss_auth
*gss_auth
;
611 struct rpc_auth
* auth
;
612 int err
= -ENOMEM
; /* XXX? */
614 dprintk("RPC: creating GSS authenticator for client %p\n", clnt
);
616 if (!try_module_get(THIS_MODULE
))
618 if (!(gss_auth
= kmalloc(sizeof(*gss_auth
), GFP_KERNEL
)))
620 gss_auth
->client
= clnt
;
622 gss_auth
->mech
= gss_mech_get_by_pseudoflavor(flavor
);
623 if (!gss_auth
->mech
) {
624 printk(KERN_WARNING
"%s: Pseudoflavor %d not found!\n",
628 gss_auth
->service
= gss_pseudoflavor_to_service(gss_auth
->mech
, flavor
);
629 if (gss_auth
->service
== 0)
631 auth
= &gss_auth
->rpc_auth
;
632 auth
->au_cslack
= GSS_CRED_SLACK
>> 2;
633 auth
->au_rslack
= GSS_VERF_SLACK
>> 2;
634 auth
->au_ops
= &authgss_ops
;
635 auth
->au_flavor
= flavor
;
636 atomic_set(&auth
->au_count
, 1);
637 kref_init(&gss_auth
->kref
);
639 gss_auth
->dentry
= rpc_mkpipe(clnt
->cl_dentry
, gss_auth
->mech
->gm_name
,
640 clnt
, &gss_upcall_ops
, RPC_PIPE_WAIT_FOR_OPEN
);
641 if (IS_ERR(gss_auth
->dentry
)) {
642 err
= PTR_ERR(gss_auth
->dentry
);
646 err
= rpcauth_init_credcache(auth
);
648 goto err_unlink_pipe
;
652 rpc_unlink(gss_auth
->dentry
);
654 gss_mech_put(gss_auth
->mech
);
658 module_put(THIS_MODULE
);
663 gss_free(struct gss_auth
*gss_auth
)
665 rpc_unlink(gss_auth
->dentry
);
666 gss_auth
->dentry
= NULL
;
667 gss_mech_put(gss_auth
->mech
);
670 module_put(THIS_MODULE
);
674 gss_free_callback(struct kref
*kref
)
676 struct gss_auth
*gss_auth
= container_of(kref
, struct gss_auth
, kref
);
682 gss_destroy(struct rpc_auth
*auth
)
684 struct gss_auth
*gss_auth
;
686 dprintk("RPC: destroying GSS authenticator %p flavor %d\n",
687 auth
, auth
->au_flavor
);
689 rpcauth_destroy_credcache(auth
);
691 gss_auth
= container_of(auth
, struct gss_auth
, rpc_auth
);
692 kref_put(&gss_auth
->kref
, gss_free_callback
);
696 * gss_destroying_context will cause the RPCSEC_GSS to send a NULL RPC call
697 * to the server with the GSS control procedure field set to
698 * RPC_GSS_PROC_DESTROY. This should normally cause the server to release
699 * all RPCSEC_GSS state associated with that context.
702 gss_destroying_context(struct rpc_cred
*cred
)
704 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
705 struct gss_auth
*gss_auth
= container_of(cred
->cr_auth
, struct gss_auth
, rpc_auth
);
706 struct rpc_task
*task
;
708 if (gss_cred
->gc_ctx
== NULL
||
709 test_and_clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
) == 0)
712 gss_cred
->gc_ctx
->gc_proc
= RPC_GSS_PROC_DESTROY
;
713 cred
->cr_ops
= &gss_nullops
;
715 /* Take a reference to ensure the cred will be destroyed either
716 * by the RPC call or by the put_rpccred() below */
719 task
= rpc_call_null(gss_auth
->client
, cred
, RPC_TASK_ASYNC
|RPC_TASK_SOFT
);
727 /* gss_destroy_cred (and gss_free_ctx) are used to clean up after failure
728 * to create a new cred or context, so they check that things have been
729 * allocated before freeing them. */
731 gss_do_free_ctx(struct gss_cl_ctx
*ctx
)
733 dprintk("RPC: gss_free_ctx\n");
735 kfree(ctx
->gc_wire_ctx
.data
);
740 gss_free_ctx_callback(struct rcu_head
*head
)
742 struct gss_cl_ctx
*ctx
= container_of(head
, struct gss_cl_ctx
, gc_rcu
);
743 gss_do_free_ctx(ctx
);
747 gss_free_ctx(struct gss_cl_ctx
*ctx
)
749 struct gss_ctx
*gc_gss_ctx
;
751 gc_gss_ctx
= rcu_dereference(ctx
->gc_gss_ctx
);
752 rcu_assign_pointer(ctx
->gc_gss_ctx
, NULL
);
753 call_rcu(&ctx
->gc_rcu
, gss_free_ctx_callback
);
755 gss_delete_sec_context(&gc_gss_ctx
);
759 gss_free_cred(struct gss_cred
*gss_cred
)
761 dprintk("RPC: gss_free_cred %p\n", gss_cred
);
766 gss_free_cred_callback(struct rcu_head
*head
)
768 struct gss_cred
*gss_cred
= container_of(head
, struct gss_cred
, gc_base
.cr_rcu
);
769 gss_free_cred(gss_cred
);
773 gss_destroy_cred(struct rpc_cred
*cred
)
775 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
, gc_base
);
776 struct gss_auth
*gss_auth
= container_of(cred
->cr_auth
, struct gss_auth
, rpc_auth
);
777 struct gss_cl_ctx
*ctx
= gss_cred
->gc_ctx
;
779 if (gss_destroying_context(cred
))
781 rcu_assign_pointer(gss_cred
->gc_ctx
, NULL
);
782 call_rcu(&cred
->cr_rcu
, gss_free_cred_callback
);
785 kref_put(&gss_auth
->kref
, gss_free_callback
);
789 * Lookup RPCSEC_GSS cred for the current process
791 static struct rpc_cred
*
792 gss_lookup_cred(struct rpc_auth
*auth
, struct auth_cred
*acred
, int flags
)
794 return rpcauth_lookup_credcache(auth
, acred
, flags
);
797 static struct rpc_cred
*
798 gss_create_cred(struct rpc_auth
*auth
, struct auth_cred
*acred
, int flags
)
800 struct gss_auth
*gss_auth
= container_of(auth
, struct gss_auth
, rpc_auth
);
801 struct gss_cred
*cred
= NULL
;
804 dprintk("RPC: gss_create_cred for uid %d, flavor %d\n",
805 acred
->uid
, auth
->au_flavor
);
807 if (!(cred
= kzalloc(sizeof(*cred
), GFP_KERNEL
)))
810 rpcauth_init_cred(&cred
->gc_base
, acred
, auth
, &gss_credops
);
812 * Note: in order to force a call to call_refresh(), we deliberately
813 * fail to flag the credential as RPCAUTH_CRED_UPTODATE.
815 cred
->gc_base
.cr_flags
= 1UL << RPCAUTH_CRED_NEW
;
816 cred
->gc_service
= gss_auth
->service
;
817 cred
->gc_machine_cred
= acred
->machine_cred
;
818 kref_get(&gss_auth
->kref
);
819 return &cred
->gc_base
;
822 dprintk("RPC: gss_create_cred failed with error %d\n", err
);
827 gss_cred_init(struct rpc_auth
*auth
, struct rpc_cred
*cred
)
829 struct gss_auth
*gss_auth
= container_of(auth
, struct gss_auth
, rpc_auth
);
830 struct gss_cred
*gss_cred
= container_of(cred
,struct gss_cred
, gc_base
);
834 err
= gss_create_upcall(gss_auth
, gss_cred
);
835 } while (err
== -EAGAIN
);
840 gss_match(struct auth_cred
*acred
, struct rpc_cred
*rc
, int flags
)
842 struct gss_cred
*gss_cred
= container_of(rc
, struct gss_cred
, gc_base
);
844 if (test_bit(RPCAUTH_CRED_NEW
, &rc
->cr_flags
))
846 /* Don't match with creds that have expired. */
847 if (time_after(jiffies
, gss_cred
->gc_ctx
->gc_expiry
))
849 if (!test_bit(RPCAUTH_CRED_UPTODATE
, &rc
->cr_flags
))
852 if (acred
->machine_cred
!= gss_cred
->gc_machine_cred
)
854 return (rc
->cr_uid
== acred
->uid
);
858 * Marshal credentials.
859 * Maybe we should keep a cached credential for performance reasons.
862 gss_marshal(struct rpc_task
*task
, __be32
*p
)
864 struct rpc_cred
*cred
= task
->tk_msg
.rpc_cred
;
865 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
,
867 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
869 struct rpc_rqst
*req
= task
->tk_rqstp
;
871 struct xdr_netobj mic
;
873 struct xdr_buf verf_buf
;
875 dprintk("RPC: %5u gss_marshal\n", task
->tk_pid
);
877 *p
++ = htonl(RPC_AUTH_GSS
);
880 spin_lock(&ctx
->gc_seq_lock
);
881 req
->rq_seqno
= ctx
->gc_seq
++;
882 spin_unlock(&ctx
->gc_seq_lock
);
884 *p
++ = htonl((u32
) RPC_GSS_VERSION
);
885 *p
++ = htonl((u32
) ctx
->gc_proc
);
886 *p
++ = htonl((u32
) req
->rq_seqno
);
887 *p
++ = htonl((u32
) gss_cred
->gc_service
);
888 p
= xdr_encode_netobj(p
, &ctx
->gc_wire_ctx
);
889 *cred_len
= htonl((p
- (cred_len
+ 1)) << 2);
891 /* We compute the checksum for the verifier over the xdr-encoded bytes
892 * starting with the xid and ending at the end of the credential: */
893 iov
.iov_base
= xprt_skip_transport_header(task
->tk_xprt
,
894 req
->rq_snd_buf
.head
[0].iov_base
);
895 iov
.iov_len
= (u8
*)p
- (u8
*)iov
.iov_base
;
896 xdr_buf_from_iov(&iov
, &verf_buf
);
898 /* set verifier flavor*/
899 *p
++ = htonl(RPC_AUTH_GSS
);
901 mic
.data
= (u8
*)(p
+ 1);
902 maj_stat
= gss_get_mic(ctx
->gc_gss_ctx
, &verf_buf
, &mic
);
903 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
) {
904 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
905 } else if (maj_stat
!= 0) {
906 printk("gss_marshal: gss_get_mic FAILED (%d)\n", maj_stat
);
909 p
= xdr_encode_opaque(p
, NULL
, mic
.len
);
917 static int gss_renew_cred(struct rpc_task
*task
)
919 struct rpc_cred
*oldcred
= task
->tk_msg
.rpc_cred
;
920 struct gss_cred
*gss_cred
= container_of(oldcred
,
923 struct rpc_auth
*auth
= oldcred
->cr_auth
;
924 struct auth_cred acred
= {
925 .uid
= oldcred
->cr_uid
,
926 .machine_cred
= gss_cred
->gc_machine_cred
,
928 struct rpc_cred
*new;
930 new = gss_lookup_cred(auth
, &acred
, RPCAUTH_LOOKUP_NEW
);
933 task
->tk_msg
.rpc_cred
= new;
934 put_rpccred(oldcred
);
939 * Refresh credentials. XXX - finish
942 gss_refresh(struct rpc_task
*task
)
944 struct rpc_cred
*cred
= task
->tk_msg
.rpc_cred
;
947 if (!test_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
) &&
948 !test_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
)) {
949 ret
= gss_renew_cred(task
);
952 cred
= task
->tk_msg
.rpc_cred
;
955 if (test_bit(RPCAUTH_CRED_NEW
, &cred
->cr_flags
))
956 ret
= gss_refresh_upcall(task
);
961 /* Dummy refresh routine: used only when destroying the context */
963 gss_refresh_null(struct rpc_task
*task
)
969 gss_validate(struct rpc_task
*task
, __be32
*p
)
971 struct rpc_cred
*cred
= task
->tk_msg
.rpc_cred
;
972 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
975 struct xdr_buf verf_buf
;
976 struct xdr_netobj mic
;
980 dprintk("RPC: %5u gss_validate\n", task
->tk_pid
);
983 if ((len
= ntohl(*p
++)) > RPC_MAX_AUTH_SIZE
)
985 if (flav
!= RPC_AUTH_GSS
)
987 seq
= htonl(task
->tk_rqstp
->rq_seqno
);
989 iov
.iov_len
= sizeof(seq
);
990 xdr_buf_from_iov(&iov
, &verf_buf
);
994 maj_stat
= gss_verify_mic(ctx
->gc_gss_ctx
, &verf_buf
, &mic
);
995 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
996 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
998 dprintk("RPC: %5u gss_validate: gss_verify_mic returned "
999 "error 0x%08x\n", task
->tk_pid
, maj_stat
);
1002 /* We leave it to unwrap to calculate au_rslack. For now we just
1003 * calculate the length of the verifier: */
1004 cred
->cr_auth
->au_verfsize
= XDR_QUADLEN(len
) + 2;
1006 dprintk("RPC: %5u gss_validate: gss_verify_mic succeeded.\n",
1008 return p
+ XDR_QUADLEN(len
);
1011 dprintk("RPC: %5u gss_validate failed.\n", task
->tk_pid
);
1016 gss_wrap_req_integ(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
,
1017 kxdrproc_t encode
, struct rpc_rqst
*rqstp
, __be32
*p
, void *obj
)
1019 struct xdr_buf
*snd_buf
= &rqstp
->rq_snd_buf
;
1020 struct xdr_buf integ_buf
;
1021 __be32
*integ_len
= NULL
;
1022 struct xdr_netobj mic
;
1030 offset
= (u8
*)p
- (u8
*)snd_buf
->head
[0].iov_base
;
1031 *p
++ = htonl(rqstp
->rq_seqno
);
1033 status
= rpc_call_xdrproc(encode
, rqstp
, p
, obj
);
1037 if (xdr_buf_subsegment(snd_buf
, &integ_buf
,
1038 offset
, snd_buf
->len
- offset
))
1040 *integ_len
= htonl(integ_buf
.len
);
1042 /* guess whether we're in the head or the tail: */
1043 if (snd_buf
->page_len
|| snd_buf
->tail
[0].iov_len
)
1044 iov
= snd_buf
->tail
;
1046 iov
= snd_buf
->head
;
1047 p
= iov
->iov_base
+ iov
->iov_len
;
1048 mic
.data
= (u8
*)(p
+ 1);
1050 maj_stat
= gss_get_mic(ctx
->gc_gss_ctx
, &integ_buf
, &mic
);
1051 status
= -EIO
; /* XXX? */
1052 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1053 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1056 q
= xdr_encode_opaque(p
, NULL
, mic
.len
);
1058 offset
= (u8
*)q
- (u8
*)p
;
1059 iov
->iov_len
+= offset
;
1060 snd_buf
->len
+= offset
;
1065 priv_release_snd_buf(struct rpc_rqst
*rqstp
)
1069 for (i
=0; i
< rqstp
->rq_enc_pages_num
; i
++)
1070 __free_page(rqstp
->rq_enc_pages
[i
]);
1071 kfree(rqstp
->rq_enc_pages
);
1075 alloc_enc_pages(struct rpc_rqst
*rqstp
)
1077 struct xdr_buf
*snd_buf
= &rqstp
->rq_snd_buf
;
1080 if (snd_buf
->page_len
== 0) {
1081 rqstp
->rq_enc_pages_num
= 0;
1085 first
= snd_buf
->page_base
>> PAGE_CACHE_SHIFT
;
1086 last
= (snd_buf
->page_base
+ snd_buf
->page_len
- 1) >> PAGE_CACHE_SHIFT
;
1087 rqstp
->rq_enc_pages_num
= last
- first
+ 1 + 1;
1089 = kmalloc(rqstp
->rq_enc_pages_num
* sizeof(struct page
*),
1091 if (!rqstp
->rq_enc_pages
)
1093 for (i
=0; i
< rqstp
->rq_enc_pages_num
; i
++) {
1094 rqstp
->rq_enc_pages
[i
] = alloc_page(GFP_NOFS
);
1095 if (rqstp
->rq_enc_pages
[i
] == NULL
)
1098 rqstp
->rq_release_snd_buf
= priv_release_snd_buf
;
1101 for (i
--; i
>= 0; i
--) {
1102 __free_page(rqstp
->rq_enc_pages
[i
]);
1109 gss_wrap_req_priv(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
,
1110 kxdrproc_t encode
, struct rpc_rqst
*rqstp
, __be32
*p
, void *obj
)
1112 struct xdr_buf
*snd_buf
= &rqstp
->rq_snd_buf
;
1117 struct page
**inpages
;
1124 offset
= (u8
*)p
- (u8
*)snd_buf
->head
[0].iov_base
;
1125 *p
++ = htonl(rqstp
->rq_seqno
);
1127 status
= rpc_call_xdrproc(encode
, rqstp
, p
, obj
);
1131 status
= alloc_enc_pages(rqstp
);
1134 first
= snd_buf
->page_base
>> PAGE_CACHE_SHIFT
;
1135 inpages
= snd_buf
->pages
+ first
;
1136 snd_buf
->pages
= rqstp
->rq_enc_pages
;
1137 snd_buf
->page_base
-= first
<< PAGE_CACHE_SHIFT
;
1138 /* Give the tail its own page, in case we need extra space in the
1139 * head when wrapping: */
1140 if (snd_buf
->page_len
|| snd_buf
->tail
[0].iov_len
) {
1141 tmp
= page_address(rqstp
->rq_enc_pages
[rqstp
->rq_enc_pages_num
- 1]);
1142 memcpy(tmp
, snd_buf
->tail
[0].iov_base
, snd_buf
->tail
[0].iov_len
);
1143 snd_buf
->tail
[0].iov_base
= tmp
;
1145 maj_stat
= gss_wrap(ctx
->gc_gss_ctx
, offset
, snd_buf
, inpages
);
1146 /* RPC_SLACK_SPACE should prevent this ever happening: */
1147 BUG_ON(snd_buf
->len
> snd_buf
->buflen
);
1149 /* We're assuming that when GSS_S_CONTEXT_EXPIRED, the encryption was
1150 * done anyway, so it's safe to put the request on the wire: */
1151 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1152 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1156 *opaque_len
= htonl(snd_buf
->len
- offset
);
1157 /* guess whether we're in the head or the tail: */
1158 if (snd_buf
->page_len
|| snd_buf
->tail
[0].iov_len
)
1159 iov
= snd_buf
->tail
;
1161 iov
= snd_buf
->head
;
1162 p
= iov
->iov_base
+ iov
->iov_len
;
1163 pad
= 3 - ((snd_buf
->len
- offset
- 1) & 3);
1165 iov
->iov_len
+= pad
;
1166 snd_buf
->len
+= pad
;
1172 gss_wrap_req(struct rpc_task
*task
,
1173 kxdrproc_t encode
, void *rqstp
, __be32
*p
, void *obj
)
1175 struct rpc_cred
*cred
= task
->tk_msg
.rpc_cred
;
1176 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
,
1178 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
1181 dprintk("RPC: %5u gss_wrap_req\n", task
->tk_pid
);
1182 if (ctx
->gc_proc
!= RPC_GSS_PROC_DATA
) {
1183 /* The spec seems a little ambiguous here, but I think that not
1184 * wrapping context destruction requests makes the most sense.
1186 status
= rpc_call_xdrproc(encode
, rqstp
, p
, obj
);
1189 switch (gss_cred
->gc_service
) {
1190 case RPC_GSS_SVC_NONE
:
1191 status
= rpc_call_xdrproc(encode
, rqstp
, p
, obj
);
1193 case RPC_GSS_SVC_INTEGRITY
:
1194 status
= gss_wrap_req_integ(cred
, ctx
, encode
,
1197 case RPC_GSS_SVC_PRIVACY
:
1198 status
= gss_wrap_req_priv(cred
, ctx
, encode
,
1204 dprintk("RPC: %5u gss_wrap_req returning %d\n", task
->tk_pid
, status
);
1209 gss_unwrap_resp_integ(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
,
1210 struct rpc_rqst
*rqstp
, __be32
**p
)
1212 struct xdr_buf
*rcv_buf
= &rqstp
->rq_rcv_buf
;
1213 struct xdr_buf integ_buf
;
1214 struct xdr_netobj mic
;
1215 u32 data_offset
, mic_offset
;
1220 integ_len
= ntohl(*(*p
)++);
1223 data_offset
= (u8
*)(*p
) - (u8
*)rcv_buf
->head
[0].iov_base
;
1224 mic_offset
= integ_len
+ data_offset
;
1225 if (mic_offset
> rcv_buf
->len
)
1227 if (ntohl(*(*p
)++) != rqstp
->rq_seqno
)
1230 if (xdr_buf_subsegment(rcv_buf
, &integ_buf
, data_offset
,
1231 mic_offset
- data_offset
))
1234 if (xdr_buf_read_netobj(rcv_buf
, &mic
, mic_offset
))
1237 maj_stat
= gss_verify_mic(ctx
->gc_gss_ctx
, &integ_buf
, &mic
);
1238 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1239 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1240 if (maj_stat
!= GSS_S_COMPLETE
)
1246 gss_unwrap_resp_priv(struct rpc_cred
*cred
, struct gss_cl_ctx
*ctx
,
1247 struct rpc_rqst
*rqstp
, __be32
**p
)
1249 struct xdr_buf
*rcv_buf
= &rqstp
->rq_rcv_buf
;
1255 opaque_len
= ntohl(*(*p
)++);
1256 offset
= (u8
*)(*p
) - (u8
*)rcv_buf
->head
[0].iov_base
;
1257 if (offset
+ opaque_len
> rcv_buf
->len
)
1259 /* remove padding: */
1260 rcv_buf
->len
= offset
+ opaque_len
;
1262 maj_stat
= gss_unwrap(ctx
->gc_gss_ctx
, offset
, rcv_buf
);
1263 if (maj_stat
== GSS_S_CONTEXT_EXPIRED
)
1264 clear_bit(RPCAUTH_CRED_UPTODATE
, &cred
->cr_flags
);
1265 if (maj_stat
!= GSS_S_COMPLETE
)
1267 if (ntohl(*(*p
)++) != rqstp
->rq_seqno
)
1275 gss_unwrap_resp(struct rpc_task
*task
,
1276 kxdrproc_t decode
, void *rqstp
, __be32
*p
, void *obj
)
1278 struct rpc_cred
*cred
= task
->tk_msg
.rpc_cred
;
1279 struct gss_cred
*gss_cred
= container_of(cred
, struct gss_cred
,
1281 struct gss_cl_ctx
*ctx
= gss_cred_get_ctx(cred
);
1283 struct kvec
*head
= ((struct rpc_rqst
*)rqstp
)->rq_rcv_buf
.head
;
1284 int savedlen
= head
->iov_len
;
1287 if (ctx
->gc_proc
!= RPC_GSS_PROC_DATA
)
1289 switch (gss_cred
->gc_service
) {
1290 case RPC_GSS_SVC_NONE
:
1292 case RPC_GSS_SVC_INTEGRITY
:
1293 status
= gss_unwrap_resp_integ(cred
, ctx
, rqstp
, &p
);
1297 case RPC_GSS_SVC_PRIVACY
:
1298 status
= gss_unwrap_resp_priv(cred
, ctx
, rqstp
, &p
);
1303 /* take into account extra slack for integrity and privacy cases: */
1304 cred
->cr_auth
->au_rslack
= cred
->cr_auth
->au_verfsize
+ (p
- savedp
)
1305 + (savedlen
- head
->iov_len
);
1307 status
= rpc_call_xdrproc(decode
, rqstp
, p
, obj
);
1310 dprintk("RPC: %5u gss_unwrap_resp returning %d\n", task
->tk_pid
,
1315 static const struct rpc_authops authgss_ops
= {
1316 .owner
= THIS_MODULE
,
1317 .au_flavor
= RPC_AUTH_GSS
,
1318 .au_name
= "RPCSEC_GSS",
1319 .create
= gss_create
,
1320 .destroy
= gss_destroy
,
1321 .lookup_cred
= gss_lookup_cred
,
1322 .crcreate
= gss_create_cred
1325 static const struct rpc_credops gss_credops
= {
1326 .cr_name
= "AUTH_GSS",
1327 .crdestroy
= gss_destroy_cred
,
1328 .cr_init
= gss_cred_init
,
1329 .crbind
= rpcauth_generic_bind_cred
,
1330 .crmatch
= gss_match
,
1331 .crmarshal
= gss_marshal
,
1332 .crrefresh
= gss_refresh
,
1333 .crvalidate
= gss_validate
,
1334 .crwrap_req
= gss_wrap_req
,
1335 .crunwrap_resp
= gss_unwrap_resp
,
1338 static const struct rpc_credops gss_nullops
= {
1339 .cr_name
= "AUTH_GSS",
1340 .crdestroy
= gss_destroy_cred
,
1341 .crbind
= rpcauth_generic_bind_cred
,
1342 .crmatch
= gss_match
,
1343 .crmarshal
= gss_marshal
,
1344 .crrefresh
= gss_refresh_null
,
1345 .crvalidate
= gss_validate
,
1346 .crwrap_req
= gss_wrap_req
,
1347 .crunwrap_resp
= gss_unwrap_resp
,
1350 static struct rpc_pipe_ops gss_upcall_ops
= {
1351 .upcall
= gss_pipe_upcall
,
1352 .downcall
= gss_pipe_downcall
,
1353 .destroy_msg
= gss_pipe_destroy_msg
,
1354 .release_pipe
= gss_pipe_release
,
1358 * Initialize RPCSEC_GSS module
1360 static int __init
init_rpcsec_gss(void)
1364 err
= rpcauth_register(&authgss_ops
);
1367 err
= gss_svc_init();
1369 goto out_unregister
;
1372 rpcauth_unregister(&authgss_ops
);
1377 static void __exit
exit_rpcsec_gss(void)
1380 rpcauth_unregister(&authgss_ops
);
1383 MODULE_LICENSE("GPL");
1384 module_init(init_rpcsec_gss
)
1385 module_exit(exit_rpcsec_gss
)