dccp: Registration routines for changing feature values
[deliverable/linux.git] / net / dccp / proto.c
CommitLineData
7c657876
ACM
1/*
2 * net/dccp/proto.c
3 *
4 * An implementation of the DCCP protocol
5 * Arnaldo Carvalho de Melo <acme@conectiva.com.br>
6 *
7 * This program is free software; you can redistribute it and/or modify it
8 * under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
10 */
11
7c657876
ACM
12#include <linux/dccp.h>
13#include <linux/module.h>
14#include <linux/types.h>
15#include <linux/sched.h>
16#include <linux/kernel.h>
17#include <linux/skbuff.h>
18#include <linux/netdevice.h>
19#include <linux/in.h>
20#include <linux/if_arp.h>
21#include <linux/init.h>
22#include <linux/random.h>
23#include <net/checksum.h>
24
14c85021 25#include <net/inet_sock.h>
7c657876
ACM
26#include <net/sock.h>
27#include <net/xfrm.h>
28
6273172e 29#include <asm/ioctls.h>
7c657876
ACM
30#include <linux/spinlock.h>
31#include <linux/timer.h>
32#include <linux/delay.h>
33#include <linux/poll.h>
7c657876
ACM
34
35#include "ccid.h"
36#include "dccp.h"
afe00251 37#include "feat.h"
7c657876 38
ba89966c 39DEFINE_SNMP_STAT(struct dccp_mib, dccp_statistics) __read_mostly;
7c657876 40
f21e68ca
ACM
41EXPORT_SYMBOL_GPL(dccp_statistics);
42
7c657876
ACM
43atomic_t dccp_orphan_count = ATOMIC_INIT(0);
44
f21e68ca
ACM
45EXPORT_SYMBOL_GPL(dccp_orphan_count);
46
075ae866
ACM
47struct inet_hashinfo __cacheline_aligned dccp_hashinfo = {
48 .lhash_lock = RW_LOCK_UNLOCKED,
49 .lhash_users = ATOMIC_INIT(0),
50 .lhash_wait = __WAIT_QUEUE_HEAD_INITIALIZER(dccp_hashinfo.lhash_wait),
51};
52
53EXPORT_SYMBOL_GPL(dccp_hashinfo);
54
b1308dc0
IM
55/* the maximum queue length for tx in packets. 0 is no limit */
56int sysctl_dccp_tx_qlen __read_mostly = 5;
57
c25a18ba
ACM
58void dccp_set_state(struct sock *sk, const int state)
59{
60 const int oldstate = sk->sk_state;
61
f11135a3 62 dccp_pr_debug("%s(%p) %s --> %s\n", dccp_role(sk), sk,
c25a18ba
ACM
63 dccp_state_name(oldstate), dccp_state_name(state));
64 WARN_ON(state == oldstate);
65
66 switch (state) {
67 case DCCP_OPEN:
68 if (oldstate != DCCP_OPEN)
69 DCCP_INC_STATS(DCCP_MIB_CURRESTAB);
70 break;
71
72 case DCCP_CLOSED:
0c869620
GR
73 if (oldstate == DCCP_OPEN || oldstate == DCCP_ACTIVE_CLOSEREQ ||
74 oldstate == DCCP_CLOSING)
c25a18ba
ACM
75 DCCP_INC_STATS(DCCP_MIB_ESTABRESETS);
76
77 sk->sk_prot->unhash(sk);
78 if (inet_csk(sk)->icsk_bind_hash != NULL &&
79 !(sk->sk_userlocks & SOCK_BINDPORT_LOCK))
ab1e0a13 80 inet_put_port(sk);
c25a18ba
ACM
81 /* fall through */
82 default:
83 if (oldstate == DCCP_OPEN)
84 DCCP_DEC_STATS(DCCP_MIB_CURRESTAB);
85 }
86
87 /* Change state AFTER socket is unhashed to avoid closed
88 * socket sitting in hash tables.
89 */
90 sk->sk_state = state;
91}
92
93EXPORT_SYMBOL_GPL(dccp_set_state);
94
0c869620
GR
95static void dccp_finish_passive_close(struct sock *sk)
96{
97 switch (sk->sk_state) {
98 case DCCP_PASSIVE_CLOSE:
99 /* Node (client or server) has received Close packet. */
100 dccp_send_reset(sk, DCCP_RESET_CODE_CLOSED);
101 dccp_set_state(sk, DCCP_CLOSED);
102 break;
103 case DCCP_PASSIVE_CLOSEREQ:
104 /*
105 * Client received CloseReq. We set the `active' flag so that
106 * dccp_send_close() retransmits the Close as per RFC 4340, 8.3.
107 */
108 dccp_send_close(sk, 1);
109 dccp_set_state(sk, DCCP_CLOSING);
110 }
111}
112
c25a18ba
ACM
113void dccp_done(struct sock *sk)
114{
115 dccp_set_state(sk, DCCP_CLOSED);
116 dccp_clear_xmit_timers(sk);
117
118 sk->sk_shutdown = SHUTDOWN_MASK;
119
120 if (!sock_flag(sk, SOCK_DEAD))
121 sk->sk_state_change(sk);
122 else
123 inet_csk_destroy_sock(sk);
124}
125
126EXPORT_SYMBOL_GPL(dccp_done);
127
7c657876
ACM
128const char *dccp_packet_name(const int type)
129{
130 static const char *dccp_packet_names[] = {
131 [DCCP_PKT_REQUEST] = "REQUEST",
132 [DCCP_PKT_RESPONSE] = "RESPONSE",
133 [DCCP_PKT_DATA] = "DATA",
134 [DCCP_PKT_ACK] = "ACK",
135 [DCCP_PKT_DATAACK] = "DATAACK",
136 [DCCP_PKT_CLOSEREQ] = "CLOSEREQ",
137 [DCCP_PKT_CLOSE] = "CLOSE",
138 [DCCP_PKT_RESET] = "RESET",
139 [DCCP_PKT_SYNC] = "SYNC",
140 [DCCP_PKT_SYNCACK] = "SYNCACK",
141 };
142
143 if (type >= DCCP_NR_PKT_TYPES)
144 return "INVALID";
145 else
146 return dccp_packet_names[type];
147}
148
149EXPORT_SYMBOL_GPL(dccp_packet_name);
150
151const char *dccp_state_name(const int state)
152{
153 static char *dccp_state_names[] = {
f11135a3
GR
154 [DCCP_OPEN] = "OPEN",
155 [DCCP_REQUESTING] = "REQUESTING",
156 [DCCP_PARTOPEN] = "PARTOPEN",
157 [DCCP_LISTEN] = "LISTEN",
158 [DCCP_RESPOND] = "RESPOND",
159 [DCCP_CLOSING] = "CLOSING",
160 [DCCP_ACTIVE_CLOSEREQ] = "CLOSEREQ",
161 [DCCP_PASSIVE_CLOSE] = "PASSIVE_CLOSE",
162 [DCCP_PASSIVE_CLOSEREQ] = "PASSIVE_CLOSEREQ",
163 [DCCP_TIME_WAIT] = "TIME_WAIT",
164 [DCCP_CLOSED] = "CLOSED",
7c657876
ACM
165 };
166
167 if (state >= DCCP_MAX_STATES)
168 return "INVALID STATE!";
169 else
170 return dccp_state_names[state];
171}
172
173EXPORT_SYMBOL_GPL(dccp_state_name);
174
72478873 175int dccp_init_sock(struct sock *sk, const __u8 ctl_sock_initialized)
3e0fadc5
ACM
176{
177 struct dccp_sock *dp = dccp_sk(sk);
a4bf3902 178 struct dccp_minisock *dmsk = dccp_msk(sk);
3e0fadc5 179 struct inet_connection_sock *icsk = inet_csk(sk);
3e0fadc5 180
a4bf3902 181 dccp_minisock_init(&dp->dccps_minisock);
3e0fadc5 182
e18d7a98
ACM
183 icsk->icsk_rto = DCCP_TIMEOUT_INIT;
184 icsk->icsk_syn_retries = sysctl_dccp_request_retries;
185 sk->sk_state = DCCP_CLOSED;
186 sk->sk_write_space = dccp_write_space;
187 icsk->icsk_sync_mss = dccp_sync_mss;
188 dp->dccps_mss_cache = 536;
189 dp->dccps_rate_last = jiffies;
190 dp->dccps_role = DCCP_ROLE_UNDEFINED;
191 dp->dccps_service = DCCP_SERVICE_CODE_IS_ABSENT;
192 dp->dccps_l_ack_ratio = dp->dccps_r_ack_ratio = 1;
193
194 dccp_init_xmit_timers(sk);
195
828755ce 196 INIT_LIST_HEAD(&dp->dccps_featneg);
3e0fadc5
ACM
197 /*
198 * FIXME: We're hardcoding the CCID, and doing this at this point makes
199 * the listening (master) sock get CCID control blocks, which is not
200 * necessary, but for now, to not mess with the test userspace apps,
201 * lets leave it here, later the real solution is to do this in a
202 * setsockopt(CCIDs-I-want/accept). -acme
203 */
72478873 204 if (likely(ctl_sock_initialized)) {
86349c8d 205 int rc = dccp_feat_init(sk);
3e0fadc5
ACM
206
207 if (rc)
208 return rc;
209
a4bf3902 210 if (dmsk->dccpms_send_ack_vector) {
3e0fadc5
ACM
211 dp->dccps_hc_rx_ackvec = dccp_ackvec_alloc(GFP_KERNEL);
212 if (dp->dccps_hc_rx_ackvec == NULL)
213 return -ENOMEM;
214 }
a4bf3902
ACM
215 dp->dccps_hc_rx_ccid = ccid_hc_rx_new(dmsk->dccpms_rx_ccid,
216 sk, GFP_KERNEL);
217 dp->dccps_hc_tx_ccid = ccid_hc_tx_new(dmsk->dccpms_tx_ccid,
218 sk, GFP_KERNEL);
8109b02b 219 if (unlikely(dp->dccps_hc_rx_ccid == NULL ||
3e0fadc5
ACM
220 dp->dccps_hc_tx_ccid == NULL)) {
221 ccid_hc_rx_delete(dp->dccps_hc_rx_ccid, sk);
222 ccid_hc_tx_delete(dp->dccps_hc_tx_ccid, sk);
a4bf3902 223 if (dmsk->dccpms_send_ack_vector) {
3e0fadc5
ACM
224 dccp_ackvec_free(dp->dccps_hc_rx_ackvec);
225 dp->dccps_hc_rx_ackvec = NULL;
226 }
227 dp->dccps_hc_rx_ccid = dp->dccps_hc_tx_ccid = NULL;
228 return -ENOMEM;
229 }
230 } else {
231 /* control socket doesn't need feat nego */
a4bf3902
ACM
232 INIT_LIST_HEAD(&dmsk->dccpms_pending);
233 INIT_LIST_HEAD(&dmsk->dccpms_conf);
3e0fadc5
ACM
234 }
235
3e0fadc5
ACM
236 return 0;
237}
238
239EXPORT_SYMBOL_GPL(dccp_init_sock);
240
7d06b2e0 241void dccp_destroy_sock(struct sock *sk)
3e0fadc5
ACM
242{
243 struct dccp_sock *dp = dccp_sk(sk);
8ca0d17b 244 struct dccp_minisock *dmsk = dccp_msk(sk);
3e0fadc5
ACM
245
246 /*
247 * DCCP doesn't use sk_write_queue, just sk_send_head
248 * for retransmissions
249 */
250 if (sk->sk_send_head != NULL) {
251 kfree_skb(sk->sk_send_head);
252 sk->sk_send_head = NULL;
253 }
254
255 /* Clean up a referenced DCCP bind bucket. */
256 if (inet_csk(sk)->icsk_bind_hash != NULL)
ab1e0a13 257 inet_put_port(sk);
3e0fadc5
ACM
258
259 kfree(dp->dccps_service_list);
260 dp->dccps_service_list = NULL;
261
8ca0d17b 262 if (dmsk->dccpms_send_ack_vector) {
3e0fadc5
ACM
263 dccp_ackvec_free(dp->dccps_hc_rx_ackvec);
264 dp->dccps_hc_rx_ackvec = NULL;
265 }
266 ccid_hc_rx_delete(dp->dccps_hc_rx_ccid, sk);
267 ccid_hc_tx_delete(dp->dccps_hc_tx_ccid, sk);
268 dp->dccps_hc_rx_ccid = dp->dccps_hc_tx_ccid = NULL;
269
270 /* clean up feature negotiation state */
70208383 271 dccp_feat_list_purge(&dp->dccps_featneg);
3e0fadc5
ACM
272}
273
274EXPORT_SYMBOL_GPL(dccp_destroy_sock);
275
72a3effa 276static inline int dccp_listen_start(struct sock *sk, int backlog)
7c657876 277{
67e6b629
ACM
278 struct dccp_sock *dp = dccp_sk(sk);
279
280 dp->dccps_role = DCCP_ROLE_LISTEN;
72a3effa 281 return inet_csk_listen_start(sk, backlog);
7c657876
ACM
282}
283
ce865a61
GR
284static inline int dccp_need_reset(int state)
285{
286 return state != DCCP_CLOSED && state != DCCP_LISTEN &&
287 state != DCCP_REQUESTING;
288}
289
7c657876
ACM
290int dccp_disconnect(struct sock *sk, int flags)
291{
292 struct inet_connection_sock *icsk = inet_csk(sk);
293 struct inet_sock *inet = inet_sk(sk);
294 int err = 0;
295 const int old_state = sk->sk_state;
296
297 if (old_state != DCCP_CLOSED)
298 dccp_set_state(sk, DCCP_CLOSED);
299
ce865a61
GR
300 /*
301 * This corresponds to the ABORT function of RFC793, sec. 3.8
302 * TCP uses a RST segment, DCCP a Reset packet with Code 2, "Aborted".
303 */
7c657876
ACM
304 if (old_state == DCCP_LISTEN) {
305 inet_csk_listen_stop(sk);
ce865a61
GR
306 } else if (dccp_need_reset(old_state)) {
307 dccp_send_reset(sk, DCCP_RESET_CODE_ABORTED);
308 sk->sk_err = ECONNRESET;
7c657876
ACM
309 } else if (old_state == DCCP_REQUESTING)
310 sk->sk_err = ECONNRESET;
311
312 dccp_clear_xmit_timers(sk);
48816322 313
7c657876 314 __skb_queue_purge(&sk->sk_receive_queue);
48816322 315 __skb_queue_purge(&sk->sk_write_queue);
7c657876
ACM
316 if (sk->sk_send_head != NULL) {
317 __kfree_skb(sk->sk_send_head);
318 sk->sk_send_head = NULL;
319 }
320
321 inet->dport = 0;
322
323 if (!(sk->sk_userlocks & SOCK_BINDADDR_LOCK))
324 inet_reset_saddr(sk);
325
326 sk->sk_shutdown = 0;
327 sock_reset_flag(sk, SOCK_DONE);
328
329 icsk->icsk_backoff = 0;
330 inet_csk_delack_init(sk);
331 __sk_dst_reset(sk);
332
547b792c 333 WARN_ON(inet->num && !icsk->icsk_bind_hash);
7c657876
ACM
334
335 sk->sk_error_report(sk);
336 return err;
337}
338
f21e68ca
ACM
339EXPORT_SYMBOL_GPL(dccp_disconnect);
340
331968bd
ACM
341/*
342 * Wait for a DCCP event.
343 *
344 * Note that we don't need to lock the socket, as the upper poll layers
345 * take care of normal races (between the test and the event) and we don't
346 * go look at any of the socket buffers directly.
347 */
f21e68ca
ACM
348unsigned int dccp_poll(struct file *file, struct socket *sock,
349 poll_table *wait)
331968bd
ACM
350{
351 unsigned int mask;
352 struct sock *sk = sock->sk;
353
354 poll_wait(file, sk->sk_sleep, wait);
355 if (sk->sk_state == DCCP_LISTEN)
356 return inet_csk_listen_poll(sk);
357
358 /* Socket is not locked. We are protected from async events
359 by poll logic and correct handling of state changes
360 made by another threads is impossible in any case.
361 */
362
363 mask = 0;
364 if (sk->sk_err)
365 mask = POLLERR;
366
367 if (sk->sk_shutdown == SHUTDOWN_MASK || sk->sk_state == DCCP_CLOSED)
368 mask |= POLLHUP;
369 if (sk->sk_shutdown & RCV_SHUTDOWN)
f348d70a 370 mask |= POLLIN | POLLRDNORM | POLLRDHUP;
331968bd
ACM
371
372 /* Connected? */
373 if ((1 << sk->sk_state) & ~(DCCPF_REQUESTING | DCCPF_RESPOND)) {
374 if (atomic_read(&sk->sk_rmem_alloc) > 0)
375 mask |= POLLIN | POLLRDNORM;
376
377 if (!(sk->sk_shutdown & SEND_SHUTDOWN)) {
378 if (sk_stream_wspace(sk) >= sk_stream_min_wspace(sk)) {
379 mask |= POLLOUT | POLLWRNORM;
380 } else { /* send SIGIO later */
381 set_bit(SOCK_ASYNC_NOSPACE,
382 &sk->sk_socket->flags);
383 set_bit(SOCK_NOSPACE, &sk->sk_socket->flags);
384
385 /* Race breaker. If space is freed after
386 * wspace test but before the flags are set,
387 * IO signal will be lost.
388 */
389 if (sk_stream_wspace(sk) >= sk_stream_min_wspace(sk))
390 mask |= POLLOUT | POLLWRNORM;
391 }
392 }
393 }
394 return mask;
395}
396
f21e68ca
ACM
397EXPORT_SYMBOL_GPL(dccp_poll);
398
7c657876
ACM
399int dccp_ioctl(struct sock *sk, int cmd, unsigned long arg)
400{
6273172e
ACM
401 int rc = -ENOTCONN;
402
403 lock_sock(sk);
404
405 if (sk->sk_state == DCCP_LISTEN)
406 goto out;
407
408 switch (cmd) {
409 case SIOCINQ: {
410 struct sk_buff *skb;
411 unsigned long amount = 0;
412
413 skb = skb_peek(&sk->sk_receive_queue);
414 if (skb != NULL) {
415 /*
416 * We will only return the amount of this packet since
417 * that is all that will be read.
418 */
419 amount = skb->len;
420 }
421 rc = put_user(amount, (int __user *)arg);
422 }
423 break;
424 default:
425 rc = -ENOIOCTLCMD;
426 break;
427 }
428out:
429 release_sock(sk);
430 return rc;
7c657876
ACM
431}
432
f21e68ca
ACM
433EXPORT_SYMBOL_GPL(dccp_ioctl);
434
60fe62e7 435static int dccp_setsockopt_service(struct sock *sk, const __be32 service,
67e6b629
ACM
436 char __user *optval, int optlen)
437{
438 struct dccp_sock *dp = dccp_sk(sk);
439 struct dccp_service_list *sl = NULL;
440
8109b02b 441 if (service == DCCP_SERVICE_INVALID_VALUE ||
67e6b629
ACM
442 optlen > DCCP_SERVICE_LIST_MAX_LEN * sizeof(u32))
443 return -EINVAL;
444
445 if (optlen > sizeof(service)) {
446 sl = kmalloc(optlen, GFP_KERNEL);
447 if (sl == NULL)
448 return -ENOMEM;
449
450 sl->dccpsl_nr = optlen / sizeof(u32) - 1;
451 if (copy_from_user(sl->dccpsl_list,
452 optval + sizeof(service),
453 optlen - sizeof(service)) ||
454 dccp_list_has_service(sl, DCCP_SERVICE_INVALID_VALUE)) {
455 kfree(sl);
456 return -EFAULT;
457 }
458 }
459
460 lock_sock(sk);
461 dp->dccps_service = service;
462
a51482bd 463 kfree(dp->dccps_service_list);
67e6b629
ACM
464
465 dp->dccps_service_list = sl;
466 release_sock(sk);
467 return 0;
468}
469
afe00251
AB
470/* byte 1 is feature. the rest is the preference list */
471static int dccp_setsockopt_change(struct sock *sk, int type,
472 struct dccp_so_feat __user *optval)
473{
474 struct dccp_so_feat opt;
475 u8 *val;
476 int rc;
477
478 if (copy_from_user(&opt, optval, sizeof(opt)))
479 return -EFAULT;
3e8a0a55
ACM
480 /*
481 * rfc4340: 6.1. Change Options
482 */
483 if (opt.dccpsf_len < 1)
484 return -EINVAL;
afe00251
AB
485
486 val = kmalloc(opt.dccpsf_len, GFP_KERNEL);
487 if (!val)
488 return -ENOMEM;
489
490 if (copy_from_user(val, opt.dccpsf_val, opt.dccpsf_len)) {
491 rc = -EFAULT;
492 goto out_free_val;
493 }
494
8ca0d17b
ACM
495 rc = dccp_feat_change(dccp_msk(sk), type, opt.dccpsf_feat,
496 val, opt.dccpsf_len, GFP_KERNEL);
afe00251
AB
497 if (rc)
498 goto out_free_val;
499
500out:
501 return rc;
502
503out_free_val:
504 kfree(val);
505 goto out;
506}
507
3fdadf7d
DM
508static int do_dccp_setsockopt(struct sock *sk, int level, int optname,
509 char __user *optval, int optlen)
7c657876 510{
09dbc389
GR
511 struct dccp_sock *dp = dccp_sk(sk);
512 int val, err = 0;
7c657876 513
a84ffe43
ACM
514 if (optlen < sizeof(int))
515 return -EINVAL;
516
517 if (get_user(val, (int __user *)optval))
518 return -EFAULT;
519
67e6b629
ACM
520 if (optname == DCCP_SOCKOPT_SERVICE)
521 return dccp_setsockopt_service(sk, val, optval, optlen);
a84ffe43 522
67e6b629 523 lock_sock(sk);
a84ffe43
ACM
524 switch (optname) {
525 case DCCP_SOCKOPT_PACKET_SIZE:
5aed3243 526 DCCP_WARN("sockopt(PACKET_SIZE) is deprecated: fix your app\n");
841bac1d 527 err = 0;
a84ffe43 528 break;
afe00251
AB
529 case DCCP_SOCKOPT_CHANGE_L:
530 if (optlen != sizeof(struct dccp_so_feat))
531 err = -EINVAL;
532 else
533 err = dccp_setsockopt_change(sk, DCCPO_CHANGE_L,
c9eaf173 534 (struct dccp_so_feat __user *)
afe00251
AB
535 optval);
536 break;
afe00251
AB
537 case DCCP_SOCKOPT_CHANGE_R:
538 if (optlen != sizeof(struct dccp_so_feat))
539 err = -EINVAL;
540 else
541 err = dccp_setsockopt_change(sk, DCCPO_CHANGE_R,
9faefb6d 542 (struct dccp_so_feat __user *)
afe00251
AB
543 optval);
544 break;
b8599d20
GR
545 case DCCP_SOCKOPT_SERVER_TIMEWAIT:
546 if (dp->dccps_role != DCCP_ROLE_SERVER)
547 err = -EOPNOTSUPP;
548 else
549 dp->dccps_server_timewait = (val != 0);
550 break;
6f4e5fff
GR
551 case DCCP_SOCKOPT_SEND_CSCOV: /* sender side, RFC 4340, sec. 9.2 */
552 if (val < 0 || val > 15)
553 err = -EINVAL;
554 else
555 dp->dccps_pcslen = val;
556 break;
557 case DCCP_SOCKOPT_RECV_CSCOV: /* receiver side, RFC 4340 sec. 9.2.1 */
558 if (val < 0 || val > 15)
559 err = -EINVAL;
560 else {
561 dp->dccps_pcrlen = val;
562 /* FIXME: add feature negotiation,
563 * ChangeL(MinimumChecksumCoverage, val) */
564 }
565 break;
a84ffe43
ACM
566 default:
567 err = -ENOPROTOOPT;
568 break;
569 }
6f4e5fff 570
a84ffe43
ACM
571 release_sock(sk);
572 return err;
7c657876
ACM
573}
574
3fdadf7d
DM
575int dccp_setsockopt(struct sock *sk, int level, int optname,
576 char __user *optval, int optlen)
577{
578 if (level != SOL_DCCP)
579 return inet_csk(sk)->icsk_af_ops->setsockopt(sk, level,
580 optname, optval,
581 optlen);
582 return do_dccp_setsockopt(sk, level, optname, optval, optlen);
583}
543d9cfe 584
f21e68ca
ACM
585EXPORT_SYMBOL_GPL(dccp_setsockopt);
586
3fdadf7d
DM
587#ifdef CONFIG_COMPAT
588int compat_dccp_setsockopt(struct sock *sk, int level, int optname,
543d9cfe 589 char __user *optval, int optlen)
3fdadf7d 590{
dec73ff0
ACM
591 if (level != SOL_DCCP)
592 return inet_csk_compat_setsockopt(sk, level, optname,
593 optval, optlen);
3fdadf7d
DM
594 return do_dccp_setsockopt(sk, level, optname, optval, optlen);
595}
543d9cfe 596
3fdadf7d
DM
597EXPORT_SYMBOL_GPL(compat_dccp_setsockopt);
598#endif
599
67e6b629 600static int dccp_getsockopt_service(struct sock *sk, int len,
60fe62e7 601 __be32 __user *optval,
67e6b629
ACM
602 int __user *optlen)
603{
604 const struct dccp_sock *dp = dccp_sk(sk);
605 const struct dccp_service_list *sl;
606 int err = -ENOENT, slen = 0, total_len = sizeof(u32);
607
608 lock_sock(sk);
67e6b629
ACM
609 if ((sl = dp->dccps_service_list) != NULL) {
610 slen = sl->dccpsl_nr * sizeof(u32);
611 total_len += slen;
612 }
613
614 err = -EINVAL;
615 if (total_len > len)
616 goto out;
617
618 err = 0;
619 if (put_user(total_len, optlen) ||
620 put_user(dp->dccps_service, optval) ||
621 (sl != NULL && copy_to_user(optval + 1, sl->dccpsl_list, slen)))
622 err = -EFAULT;
623out:
624 release_sock(sk);
625 return err;
626}
627
3fdadf7d 628static int do_dccp_getsockopt(struct sock *sk, int level, int optname,
a1d3a355 629 char __user *optval, int __user *optlen)
7c657876 630{
a84ffe43
ACM
631 struct dccp_sock *dp;
632 int val, len;
7c657876 633
a84ffe43
ACM
634 if (get_user(len, optlen))
635 return -EFAULT;
636
39ebc027 637 if (len < (int)sizeof(int))
a84ffe43
ACM
638 return -EINVAL;
639
640 dp = dccp_sk(sk);
641
642 switch (optname) {
643 case DCCP_SOCKOPT_PACKET_SIZE:
5aed3243 644 DCCP_WARN("sockopt(PACKET_SIZE) is deprecated: fix your app\n");
841bac1d 645 return 0;
88f964db
ACM
646 case DCCP_SOCKOPT_SERVICE:
647 return dccp_getsockopt_service(sk, len,
60fe62e7 648 (__be32 __user *)optval, optlen);
7c559a9e
GR
649 case DCCP_SOCKOPT_GET_CUR_MPS:
650 val = dp->dccps_mss_cache;
7c559a9e 651 break;
b8599d20
GR
652 case DCCP_SOCKOPT_SERVER_TIMEWAIT:
653 val = dp->dccps_server_timewait;
b8599d20 654 break;
6f4e5fff
GR
655 case DCCP_SOCKOPT_SEND_CSCOV:
656 val = dp->dccps_pcslen;
657 break;
658 case DCCP_SOCKOPT_RECV_CSCOV:
659 val = dp->dccps_pcrlen;
660 break;
88f964db
ACM
661 case 128 ... 191:
662 return ccid_hc_rx_getsockopt(dp->dccps_hc_rx_ccid, sk, optname,
663 len, (u32 __user *)optval, optlen);
664 case 192 ... 255:
665 return ccid_hc_tx_getsockopt(dp->dccps_hc_tx_ccid, sk, optname,
666 len, (u32 __user *)optval, optlen);
a84ffe43
ACM
667 default:
668 return -ENOPROTOOPT;
669 }
670
79133506 671 len = sizeof(val);
a84ffe43
ACM
672 if (put_user(len, optlen) || copy_to_user(optval, &val, len))
673 return -EFAULT;
674
675 return 0;
7c657876
ACM
676}
677
3fdadf7d
DM
678int dccp_getsockopt(struct sock *sk, int level, int optname,
679 char __user *optval, int __user *optlen)
680{
681 if (level != SOL_DCCP)
682 return inet_csk(sk)->icsk_af_ops->getsockopt(sk, level,
683 optname, optval,
684 optlen);
685 return do_dccp_getsockopt(sk, level, optname, optval, optlen);
686}
543d9cfe 687
f21e68ca
ACM
688EXPORT_SYMBOL_GPL(dccp_getsockopt);
689
3fdadf7d
DM
690#ifdef CONFIG_COMPAT
691int compat_dccp_getsockopt(struct sock *sk, int level, int optname,
543d9cfe 692 char __user *optval, int __user *optlen)
3fdadf7d 693{
dec73ff0
ACM
694 if (level != SOL_DCCP)
695 return inet_csk_compat_getsockopt(sk, level, optname,
696 optval, optlen);
3fdadf7d
DM
697 return do_dccp_getsockopt(sk, level, optname, optval, optlen);
698}
543d9cfe 699
3fdadf7d
DM
700EXPORT_SYMBOL_GPL(compat_dccp_getsockopt);
701#endif
702
7c657876
ACM
703int dccp_sendmsg(struct kiocb *iocb, struct sock *sk, struct msghdr *msg,
704 size_t len)
705{
706 const struct dccp_sock *dp = dccp_sk(sk);
707 const int flags = msg->msg_flags;
708 const int noblock = flags & MSG_DONTWAIT;
709 struct sk_buff *skb;
710 int rc, size;
711 long timeo;
712
713 if (len > dp->dccps_mss_cache)
714 return -EMSGSIZE;
715
716 lock_sock(sk);
b1308dc0
IM
717
718 if (sysctl_dccp_tx_qlen &&
719 (sk->sk_write_queue.qlen >= sysctl_dccp_tx_qlen)) {
720 rc = -EAGAIN;
721 goto out_release;
722 }
723
27258ee5 724 timeo = sock_sndtimeo(sk, noblock);
7c657876
ACM
725
726 /*
727 * We have to use sk_stream_wait_connect here to set sk_write_pending,
728 * so that the trick in dccp_rcv_request_sent_state_process.
729 */
730 /* Wait for a connection to finish. */
cecd8d0e 731 if ((1 << sk->sk_state) & ~(DCCPF_OPEN | DCCPF_PARTOPEN))
7c657876 732 if ((rc = sk_stream_wait_connect(sk, &timeo)) != 0)
27258ee5 733 goto out_release;
7c657876
ACM
734
735 size = sk->sk_prot->max_header + len;
736 release_sock(sk);
737 skb = sock_alloc_send_skb(sk, size, noblock, &rc);
738 lock_sock(sk);
7c657876
ACM
739 if (skb == NULL)
740 goto out_release;
741
742 skb_reserve(skb, sk->sk_prot->max_header);
743 rc = memcpy_fromiovec(skb_put(skb, len), msg->msg_iov, len);
27258ee5
ACM
744 if (rc != 0)
745 goto out_discard;
746
97e5848d
IM
747 skb_queue_tail(&sk->sk_write_queue, skb);
748 dccp_write_xmit(sk,0);
7c657876
ACM
749out_release:
750 release_sock(sk);
751 return rc ? : len;
27258ee5
ACM
752out_discard:
753 kfree_skb(skb);
7c657876 754 goto out_release;
7c657876
ACM
755}
756
f21e68ca
ACM
757EXPORT_SYMBOL_GPL(dccp_sendmsg);
758
7c657876
ACM
759int dccp_recvmsg(struct kiocb *iocb, struct sock *sk, struct msghdr *msg,
760 size_t len, int nonblock, int flags, int *addr_len)
761{
762 const struct dccp_hdr *dh;
7c657876
ACM
763 long timeo;
764
765 lock_sock(sk);
766
531669a0
ACM
767 if (sk->sk_state == DCCP_LISTEN) {
768 len = -ENOTCONN;
7c657876 769 goto out;
7c657876 770 }
7c657876 771
531669a0 772 timeo = sock_rcvtimeo(sk, nonblock);
7c657876
ACM
773
774 do {
531669a0 775 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue);
7c657876 776
531669a0
ACM
777 if (skb == NULL)
778 goto verify_sock_status;
7c657876 779
531669a0 780 dh = dccp_hdr(skb);
7c657876 781
0c869620
GR
782 switch (dh->dccph_type) {
783 case DCCP_PKT_DATA:
784 case DCCP_PKT_DATAACK:
531669a0 785 goto found_ok_skb;
7c657876 786
0c869620
GR
787 case DCCP_PKT_CLOSE:
788 case DCCP_PKT_CLOSEREQ:
789 if (!(flags & MSG_PEEK))
790 dccp_finish_passive_close(sk);
791 /* fall through */
792 case DCCP_PKT_RESET:
793 dccp_pr_debug("found fin (%s) ok!\n",
794 dccp_packet_name(dh->dccph_type));
531669a0
ACM
795 len = 0;
796 goto found_fin_ok;
0c869620
GR
797 default:
798 dccp_pr_debug("packet_type=%s\n",
799 dccp_packet_name(dh->dccph_type));
800 sk_eat_skb(sk, skb, 0);
531669a0 801 }
531669a0
ACM
802verify_sock_status:
803 if (sock_flag(sk, SOCK_DONE)) {
804 len = 0;
7c657876 805 break;
531669a0 806 }
7c657876 807
531669a0
ACM
808 if (sk->sk_err) {
809 len = sock_error(sk);
810 break;
811 }
7c657876 812
531669a0
ACM
813 if (sk->sk_shutdown & RCV_SHUTDOWN) {
814 len = 0;
815 break;
816 }
7c657876 817
531669a0
ACM
818 if (sk->sk_state == DCCP_CLOSED) {
819 if (!sock_flag(sk, SOCK_DONE)) {
820 /* This occurs when user tries to read
821 * from never connected socket.
822 */
823 len = -ENOTCONN;
7c657876
ACM
824 break;
825 }
531669a0
ACM
826 len = 0;
827 break;
7c657876
ACM
828 }
829
531669a0
ACM
830 if (!timeo) {
831 len = -EAGAIN;
832 break;
833 }
7c657876 834
531669a0
ACM
835 if (signal_pending(current)) {
836 len = sock_intr_errno(timeo);
837 break;
838 }
7c657876 839
531669a0 840 sk_wait_data(sk, &timeo);
7c657876 841 continue;
7c657876 842 found_ok_skb:
531669a0
ACM
843 if (len > skb->len)
844 len = skb->len;
845 else if (len < skb->len)
846 msg->msg_flags |= MSG_TRUNC;
847
848 if (skb_copy_datagram_iovec(skb, 0, msg->msg_iov, len)) {
849 /* Exception. Bailout! */
850 len = -EFAULT;
851 break;
7c657876 852 }
7c657876
ACM
853 found_fin_ok:
854 if (!(flags & MSG_PEEK))
624d1164 855 sk_eat_skb(sk, skb, 0);
7c657876 856 break;
531669a0 857 } while (1);
7c657876
ACM
858out:
859 release_sock(sk);
531669a0 860 return len;
7c657876
ACM
861}
862
f21e68ca
ACM
863EXPORT_SYMBOL_GPL(dccp_recvmsg);
864
865int inet_dccp_listen(struct socket *sock, int backlog)
7c657876
ACM
866{
867 struct sock *sk = sock->sk;
868 unsigned char old_state;
869 int err;
870
871 lock_sock(sk);
872
873 err = -EINVAL;
874 if (sock->state != SS_UNCONNECTED || sock->type != SOCK_DCCP)
875 goto out;
876
877 old_state = sk->sk_state;
878 if (!((1 << old_state) & (DCCPF_CLOSED | DCCPF_LISTEN)))
879 goto out;
880
881 /* Really, if the socket is already in listen state
882 * we can only allow the backlog to be adjusted.
883 */
884 if (old_state != DCCP_LISTEN) {
885 /*
886 * FIXME: here it probably should be sk->sk_prot->listen_start
887 * see tcp_listen_start
888 */
72a3effa 889 err = dccp_listen_start(sk, backlog);
7c657876
ACM
890 if (err)
891 goto out;
892 }
893 sk->sk_max_ack_backlog = backlog;
894 err = 0;
895
896out:
897 release_sock(sk);
898 return err;
899}
900
f21e68ca
ACM
901EXPORT_SYMBOL_GPL(inet_dccp_listen);
902
0c869620 903static void dccp_terminate_connection(struct sock *sk)
7c657876 904{
0c869620 905 u8 next_state = DCCP_CLOSED;
7c657876 906
0c869620
GR
907 switch (sk->sk_state) {
908 case DCCP_PASSIVE_CLOSE:
909 case DCCP_PASSIVE_CLOSEREQ:
910 dccp_finish_passive_close(sk);
911 break;
912 case DCCP_PARTOPEN:
913 dccp_pr_debug("Stop PARTOPEN timer (%p)\n", sk);
914 inet_csk_clear_xmit_timer(sk, ICSK_TIME_DACK);
915 /* fall through */
916 case DCCP_OPEN:
917 dccp_send_close(sk, 1);
7c657876 918
b8599d20
GR
919 if (dccp_sk(sk)->dccps_role == DCCP_ROLE_SERVER &&
920 !dccp_sk(sk)->dccps_server_timewait)
0c869620
GR
921 next_state = DCCP_ACTIVE_CLOSEREQ;
922 else
923 next_state = DCCP_CLOSING;
924 /* fall through */
925 default:
926 dccp_set_state(sk, next_state);
927 }
7c657876
ACM
928}
929
930void dccp_close(struct sock *sk, long timeout)
931{
97e5848d 932 struct dccp_sock *dp = dccp_sk(sk);
7c657876 933 struct sk_buff *skb;
d83bd95b 934 u32 data_was_unread = 0;
134af346 935 int state;
7c657876
ACM
936
937 lock_sock(sk);
938
939 sk->sk_shutdown = SHUTDOWN_MASK;
940
941 if (sk->sk_state == DCCP_LISTEN) {
942 dccp_set_state(sk, DCCP_CLOSED);
943
944 /* Special case. */
945 inet_csk_listen_stop(sk);
946
947 goto adjudge_to_death;
948 }
949
97e5848d
IM
950 sk_stop_timer(sk, &dp->dccps_xmit_timer);
951
7c657876
ACM
952 /*
953 * We need to flush the recv. buffs. We do this only on the
954 * descriptor close, not protocol-sourced closes, because the
955 *reader process may not have drained the data yet!
956 */
7c657876 957 while ((skb = __skb_dequeue(&sk->sk_receive_queue)) != NULL) {
d83bd95b 958 data_was_unread += skb->len;
7c657876
ACM
959 __kfree_skb(skb);
960 }
961
d83bd95b
GR
962 if (data_was_unread) {
963 /* Unread data was tossed, send an appropriate Reset Code */
964 DCCP_WARN("DCCP: ABORT -- %u bytes unread\n", data_was_unread);
965 dccp_send_reset(sk, DCCP_RESET_CODE_ABORTED);
966 dccp_set_state(sk, DCCP_CLOSED);
967 } else if (sock_flag(sk, SOCK_LINGER) && !sk->sk_lingertime) {
7c657876
ACM
968 /* Check zero linger _after_ checking for unread data. */
969 sk->sk_prot->disconnect(sk, 0);
0c869620
GR
970 } else if (sk->sk_state != DCCP_CLOSED) {
971 dccp_terminate_connection(sk);
7c657876
ACM
972 }
973
974 sk_stream_wait_close(sk, timeout);
975
976adjudge_to_death:
134af346
HX
977 state = sk->sk_state;
978 sock_hold(sk);
979 sock_orphan(sk);
980 atomic_inc(sk->sk_prot->orphan_count);
981
7ad07e7c
ACM
982 /*
983 * It is the last release_sock in its life. It will remove backlog.
984 */
7c657876
ACM
985 release_sock(sk);
986 /*
987 * Now socket is owned by kernel and we acquire BH lock
988 * to finish close. No need to check for user refs.
989 */
990 local_bh_disable();
991 bh_lock_sock(sk);
547b792c 992 WARN_ON(sock_owned_by_user(sk));
7c657876 993
134af346
HX
994 /* Have we already been destroyed by a softirq or backlog? */
995 if (state != DCCP_CLOSED && sk->sk_state == DCCP_CLOSED)
996 goto out;
7ad07e7c 997
7c657876
ACM
998 if (sk->sk_state == DCCP_CLOSED)
999 inet_csk_destroy_sock(sk);
1000
1001 /* Otherwise, socket is reprieved until protocol close. */
1002
134af346 1003out:
7c657876
ACM
1004 bh_unlock_sock(sk);
1005 local_bh_enable();
1006 sock_put(sk);
1007}
1008
f21e68ca
ACM
1009EXPORT_SYMBOL_GPL(dccp_close);
1010
7c657876
ACM
1011void dccp_shutdown(struct sock *sk, int how)
1012{
8e8c71f1 1013 dccp_pr_debug("called shutdown(%x)\n", how);
7c657876
ACM
1014}
1015
f21e68ca
ACM
1016EXPORT_SYMBOL_GPL(dccp_shutdown);
1017
24e8b7e4 1018static inline int dccp_mib_init(void)
7c657876 1019{
24e8b7e4 1020 return snmp_mib_init((void**)dccp_statistics, sizeof(struct dccp_mib));
7c657876
ACM
1021}
1022
24e8b7e4 1023static inline void dccp_mib_exit(void)
46f09ffa 1024{
24e8b7e4 1025 snmp_mib_free((void**)dccp_statistics);
46f09ffa
ACM
1026}
1027
7c657876
ACM
1028static int thash_entries;
1029module_param(thash_entries, int, 0444);
1030MODULE_PARM_DESC(thash_entries, "Number of ehash buckets");
1031
a1d3a355 1032#ifdef CONFIG_IP_DCCP_DEBUG
7c657876 1033int dccp_debug;
43264991 1034module_param(dccp_debug, bool, 0644);
7c657876 1035MODULE_PARM_DESC(dccp_debug, "Enable debug messages");
f21e68ca
ACM
1036
1037EXPORT_SYMBOL_GPL(dccp_debug);
a1d3a355 1038#endif
7c657876
ACM
1039
1040static int __init dccp_init(void)
1041{
1042 unsigned long goal;
1043 int ehash_order, bhash_order, i;
b61fafc4 1044 int rc = -ENOBUFS;
7c657876 1045
028b0275
PM
1046 BUILD_BUG_ON(sizeof(struct dccp_skb_cb) >
1047 FIELD_SIZEOF(struct sk_buff, cb));
1048
7690af3f
ACM
1049 dccp_hashinfo.bind_bucket_cachep =
1050 kmem_cache_create("dccp_bind_bucket",
1051 sizeof(struct inet_bind_bucket), 0,
20c2df83 1052 SLAB_HWCACHE_ALIGN, NULL);
7c657876 1053 if (!dccp_hashinfo.bind_bucket_cachep)
b61fafc4 1054 goto out;
7c657876
ACM
1055
1056 /*
1057 * Size and allocate the main established and bind bucket
1058 * hash tables.
1059 *
1060 * The methodology is similar to that of the buffer cache.
1061 */
1062 if (num_physpages >= (128 * 1024))
1063 goal = num_physpages >> (21 - PAGE_SHIFT);
1064 else
1065 goal = num_physpages >> (23 - PAGE_SHIFT);
1066
1067 if (thash_entries)
7690af3f
ACM
1068 goal = (thash_entries *
1069 sizeof(struct inet_ehash_bucket)) >> PAGE_SHIFT;
7c657876
ACM
1070 for (ehash_order = 0; (1UL << ehash_order) < goal; ehash_order++)
1071 ;
1072 do {
1073 dccp_hashinfo.ehash_size = (1UL << ehash_order) * PAGE_SIZE /
1074 sizeof(struct inet_ehash_bucket);
7690af3f
ACM
1075 while (dccp_hashinfo.ehash_size &
1076 (dccp_hashinfo.ehash_size - 1))
7c657876
ACM
1077 dccp_hashinfo.ehash_size--;
1078 dccp_hashinfo.ehash = (struct inet_ehash_bucket *)
1079 __get_free_pages(GFP_ATOMIC, ehash_order);
1080 } while (!dccp_hashinfo.ehash && --ehash_order > 0);
1081
1082 if (!dccp_hashinfo.ehash) {
59348b19 1083 DCCP_CRIT("Failed to allocate DCCP established hash table");
7c657876
ACM
1084 goto out_free_bind_bucket_cachep;
1085 }
1086
dbca9b27 1087 for (i = 0; i < dccp_hashinfo.ehash_size; i++) {
7c657876 1088 INIT_HLIST_HEAD(&dccp_hashinfo.ehash[i].chain);
dbca9b27 1089 INIT_HLIST_HEAD(&dccp_hashinfo.ehash[i].twchain);
7c657876
ACM
1090 }
1091
230140cf
ED
1092 if (inet_ehash_locks_alloc(&dccp_hashinfo))
1093 goto out_free_dccp_ehash;
1094
7c657876
ACM
1095 bhash_order = ehash_order;
1096
1097 do {
1098 dccp_hashinfo.bhash_size = (1UL << bhash_order) * PAGE_SIZE /
1099 sizeof(struct inet_bind_hashbucket);
7690af3f
ACM
1100 if ((dccp_hashinfo.bhash_size > (64 * 1024)) &&
1101 bhash_order > 0)
7c657876
ACM
1102 continue;
1103 dccp_hashinfo.bhash = (struct inet_bind_hashbucket *)
1104 __get_free_pages(GFP_ATOMIC, bhash_order);
1105 } while (!dccp_hashinfo.bhash && --bhash_order >= 0);
1106
1107 if (!dccp_hashinfo.bhash) {
59348b19 1108 DCCP_CRIT("Failed to allocate DCCP bind hash table");
230140cf 1109 goto out_free_dccp_locks;
7c657876
ACM
1110 }
1111
1112 for (i = 0; i < dccp_hashinfo.bhash_size; i++) {
1113 spin_lock_init(&dccp_hashinfo.bhash[i].lock);
1114 INIT_HLIST_HEAD(&dccp_hashinfo.bhash[i].chain);
1115 }
1116
46f09ffa 1117 rc = dccp_mib_init();
fa23e2ec 1118 if (rc)
7c657876
ACM
1119 goto out_free_dccp_bhash;
1120
9b07ef5d 1121 rc = dccp_ackvec_init();
7c657876 1122 if (rc)
b61fafc4 1123 goto out_free_dccp_mib;
9b07ef5d 1124
e55d912f 1125 rc = dccp_sysctl_init();
9b07ef5d
ACM
1126 if (rc)
1127 goto out_ackvec_exit;
4c70f383
GR
1128
1129 dccp_timestamping_init();
7c657876
ACM
1130out:
1131 return rc;
9b07ef5d
ACM
1132out_ackvec_exit:
1133 dccp_ackvec_exit();
b61fafc4 1134out_free_dccp_mib:
46f09ffa 1135 dccp_mib_exit();
7c657876
ACM
1136out_free_dccp_bhash:
1137 free_pages((unsigned long)dccp_hashinfo.bhash, bhash_order);
1138 dccp_hashinfo.bhash = NULL;
230140cf
ED
1139out_free_dccp_locks:
1140 inet_ehash_locks_free(&dccp_hashinfo);
7c657876
ACM
1141out_free_dccp_ehash:
1142 free_pages((unsigned long)dccp_hashinfo.ehash, ehash_order);
1143 dccp_hashinfo.ehash = NULL;
1144out_free_bind_bucket_cachep:
1145 kmem_cache_destroy(dccp_hashinfo.bind_bucket_cachep);
1146 dccp_hashinfo.bind_bucket_cachep = NULL;
7c657876
ACM
1147 goto out;
1148}
1149
7c657876
ACM
1150static void __exit dccp_fini(void)
1151{
46f09ffa 1152 dccp_mib_exit();
725ba8ee
ACM
1153 free_pages((unsigned long)dccp_hashinfo.bhash,
1154 get_order(dccp_hashinfo.bhash_size *
1155 sizeof(struct inet_bind_hashbucket)));
1156 free_pages((unsigned long)dccp_hashinfo.ehash,
1157 get_order(dccp_hashinfo.ehash_size *
1158 sizeof(struct inet_ehash_bucket)));
230140cf 1159 inet_ehash_locks_free(&dccp_hashinfo);
7c657876 1160 kmem_cache_destroy(dccp_hashinfo.bind_bucket_cachep);
9b07ef5d 1161 dccp_ackvec_exit();
e55d912f 1162 dccp_sysctl_exit();
7c657876
ACM
1163}
1164
1165module_init(dccp_init);
1166module_exit(dccp_fini);
1167
7c657876
ACM
1168MODULE_LICENSE("GPL");
1169MODULE_AUTHOR("Arnaldo Carvalho de Melo <acme@conectiva.com.br>");
1170MODULE_DESCRIPTION("DCCP - Datagram Congestion Controlled Protocol");
This page took 0.515231 seconds and 5 git commands to generate.