de03fe7002d049b9726a2329231f7dee8903c5f5
[deliverable/linux.git] / net / ipv4 / devinet.c
1 /*
2 * NET3 IP device support routines.
3 *
4 * This program is free software; you can redistribute it and/or
5 * modify it under the terms of the GNU General Public License
6 * as published by the Free Software Foundation; either version
7 * 2 of the License, or (at your option) any later version.
8 *
9 * Derived from the IP parts of dev.c 1.0.19
10 * Authors: Ross Biro
11 * Fred N. van Kempen, <waltje@uWalt.NL.Mugnet.ORG>
12 * Mark Evans, <evansmp@uhura.aston.ac.uk>
13 *
14 * Additional Authors:
15 * Alan Cox, <gw4pts@gw4pts.ampr.org>
16 * Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
17 *
18 * Changes:
19 * Alexey Kuznetsov: pa_* fields are replaced with ifaddr
20 * lists.
21 * Cyrus Durgin: updated for kmod
22 * Matthias Andree: in devinet_ioctl, compare label and
23 * address (4.4BSD alias style support),
24 * fall back to comparing just the label
25 * if no match found.
26 */
27
28
29 #include <asm/uaccess.h>
30 #include <linux/bitops.h>
31 #include <linux/capability.h>
32 #include <linux/module.h>
33 #include <linux/types.h>
34 #include <linux/kernel.h>
35 #include <linux/string.h>
36 #include <linux/mm.h>
37 #include <linux/socket.h>
38 #include <linux/sockios.h>
39 #include <linux/in.h>
40 #include <linux/errno.h>
41 #include <linux/interrupt.h>
42 #include <linux/if_addr.h>
43 #include <linux/if_ether.h>
44 #include <linux/inet.h>
45 #include <linux/netdevice.h>
46 #include <linux/etherdevice.h>
47 #include <linux/skbuff.h>
48 #include <linux/init.h>
49 #include <linux/notifier.h>
50 #include <linux/inetdevice.h>
51 #include <linux/igmp.h>
52 #include <linux/slab.h>
53 #include <linux/hash.h>
54 #ifdef CONFIG_SYSCTL
55 #include <linux/sysctl.h>
56 #endif
57 #include <linux/kmod.h>
58 #include <linux/netconf.h>
59
60 #include <net/arp.h>
61 #include <net/ip.h>
62 #include <net/route.h>
63 #include <net/ip_fib.h>
64 #include <net/rtnetlink.h>
65 #include <net/net_namespace.h>
66 #include <net/addrconf.h>
67
68 #include "fib_lookup.h"
69
70 static struct ipv4_devconf ipv4_devconf = {
71 .data = {
72 [IPV4_DEVCONF_ACCEPT_REDIRECTS - 1] = 1,
73 [IPV4_DEVCONF_SEND_REDIRECTS - 1] = 1,
74 [IPV4_DEVCONF_SECURE_REDIRECTS - 1] = 1,
75 [IPV4_DEVCONF_SHARED_MEDIA - 1] = 1,
76 [IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL - 1] = 10000 /*ms*/,
77 [IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL - 1] = 1000 /*ms*/,
78 },
79 };
80
81 static struct ipv4_devconf ipv4_devconf_dflt = {
82 .data = {
83 [IPV4_DEVCONF_ACCEPT_REDIRECTS - 1] = 1,
84 [IPV4_DEVCONF_SEND_REDIRECTS - 1] = 1,
85 [IPV4_DEVCONF_SECURE_REDIRECTS - 1] = 1,
86 [IPV4_DEVCONF_SHARED_MEDIA - 1] = 1,
87 [IPV4_DEVCONF_ACCEPT_SOURCE_ROUTE - 1] = 1,
88 [IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL - 1] = 10000 /*ms*/,
89 [IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL - 1] = 1000 /*ms*/,
90 },
91 };
92
93 #define IPV4_DEVCONF_DFLT(net, attr) \
94 IPV4_DEVCONF((*net->ipv4.devconf_dflt), attr)
95
96 static const struct nla_policy ifa_ipv4_policy[IFA_MAX+1] = {
97 [IFA_LOCAL] = { .type = NLA_U32 },
98 [IFA_ADDRESS] = { .type = NLA_U32 },
99 [IFA_BROADCAST] = { .type = NLA_U32 },
100 [IFA_LABEL] = { .type = NLA_STRING, .len = IFNAMSIZ - 1 },
101 [IFA_CACHEINFO] = { .len = sizeof(struct ifa_cacheinfo) },
102 [IFA_FLAGS] = { .type = NLA_U32 },
103 };
104
105 #define IN4_ADDR_HSIZE_SHIFT 8
106 #define IN4_ADDR_HSIZE (1U << IN4_ADDR_HSIZE_SHIFT)
107
108 static struct hlist_head inet_addr_lst[IN4_ADDR_HSIZE];
109 static DEFINE_SPINLOCK(inet_addr_hash_lock);
110
111 static u32 inet_addr_hash(struct net *net, __be32 addr)
112 {
113 u32 val = (__force u32) addr ^ net_hash_mix(net);
114
115 return hash_32(val, IN4_ADDR_HSIZE_SHIFT);
116 }
117
118 static void inet_hash_insert(struct net *net, struct in_ifaddr *ifa)
119 {
120 u32 hash = inet_addr_hash(net, ifa->ifa_local);
121
122 spin_lock(&inet_addr_hash_lock);
123 hlist_add_head_rcu(&ifa->hash, &inet_addr_lst[hash]);
124 spin_unlock(&inet_addr_hash_lock);
125 }
126
127 static void inet_hash_remove(struct in_ifaddr *ifa)
128 {
129 spin_lock(&inet_addr_hash_lock);
130 hlist_del_init_rcu(&ifa->hash);
131 spin_unlock(&inet_addr_hash_lock);
132 }
133
134 /**
135 * __ip_dev_find - find the first device with a given source address.
136 * @net: the net namespace
137 * @addr: the source address
138 * @devref: if true, take a reference on the found device
139 *
140 * If a caller uses devref=false, it should be protected by RCU, or RTNL
141 */
142 struct net_device *__ip_dev_find(struct net *net, __be32 addr, bool devref)
143 {
144 u32 hash = inet_addr_hash(net, addr);
145 struct net_device *result = NULL;
146 struct in_ifaddr *ifa;
147
148 rcu_read_lock();
149 hlist_for_each_entry_rcu(ifa, &inet_addr_lst[hash], hash) {
150 if (ifa->ifa_local == addr) {
151 struct net_device *dev = ifa->ifa_dev->dev;
152
153 if (!net_eq(dev_net(dev), net))
154 continue;
155 result = dev;
156 break;
157 }
158 }
159 if (!result) {
160 struct flowi4 fl4 = { .daddr = addr };
161 struct fib_result res = { 0 };
162 struct fib_table *local;
163
164 /* Fallback to FIB local table so that communication
165 * over loopback subnets work.
166 */
167 local = fib_get_table(net, RT_TABLE_LOCAL);
168 if (local &&
169 !fib_table_lookup(local, &fl4, &res, FIB_LOOKUP_NOREF) &&
170 res.type == RTN_LOCAL)
171 result = FIB_RES_DEV(res);
172 }
173 if (result && devref)
174 dev_hold(result);
175 rcu_read_unlock();
176 return result;
177 }
178 EXPORT_SYMBOL(__ip_dev_find);
179
180 static void rtmsg_ifa(int event, struct in_ifaddr *, struct nlmsghdr *, u32);
181
182 static BLOCKING_NOTIFIER_HEAD(inetaddr_chain);
183 static void inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap,
184 int destroy);
185 #ifdef CONFIG_SYSCTL
186 static void devinet_sysctl_register(struct in_device *idev);
187 static void devinet_sysctl_unregister(struct in_device *idev);
188 #else
189 static void devinet_sysctl_register(struct in_device *idev)
190 {
191 }
192 static void devinet_sysctl_unregister(struct in_device *idev)
193 {
194 }
195 #endif
196
197 /* Locks all the inet devices. */
198
199 static struct in_ifaddr *inet_alloc_ifa(void)
200 {
201 return kzalloc(sizeof(struct in_ifaddr), GFP_KERNEL);
202 }
203
204 static void inet_rcu_free_ifa(struct rcu_head *head)
205 {
206 struct in_ifaddr *ifa = container_of(head, struct in_ifaddr, rcu_head);
207 if (ifa->ifa_dev)
208 in_dev_put(ifa->ifa_dev);
209 kfree(ifa);
210 }
211
212 static void inet_free_ifa(struct in_ifaddr *ifa)
213 {
214 call_rcu(&ifa->rcu_head, inet_rcu_free_ifa);
215 }
216
217 void in_dev_finish_destroy(struct in_device *idev)
218 {
219 struct net_device *dev = idev->dev;
220
221 WARN_ON(idev->ifa_list);
222 WARN_ON(idev->mc_list);
223 kfree(rcu_dereference_protected(idev->mc_hash, 1));
224 #ifdef NET_REFCNT_DEBUG
225 pr_debug("%s: %p=%s\n", __func__, idev, dev ? dev->name : "NIL");
226 #endif
227 dev_put(dev);
228 if (!idev->dead)
229 pr_err("Freeing alive in_device %p\n", idev);
230 else
231 kfree(idev);
232 }
233 EXPORT_SYMBOL(in_dev_finish_destroy);
234
235 static struct in_device *inetdev_init(struct net_device *dev)
236 {
237 struct in_device *in_dev;
238
239 ASSERT_RTNL();
240
241 in_dev = kzalloc(sizeof(*in_dev), GFP_KERNEL);
242 if (!in_dev)
243 goto out;
244 memcpy(&in_dev->cnf, dev_net(dev)->ipv4.devconf_dflt,
245 sizeof(in_dev->cnf));
246 in_dev->cnf.sysctl = NULL;
247 in_dev->dev = dev;
248 in_dev->arp_parms = neigh_parms_alloc(dev, &arp_tbl);
249 if (!in_dev->arp_parms)
250 goto out_kfree;
251 if (IPV4_DEVCONF(in_dev->cnf, FORWARDING))
252 dev_disable_lro(dev);
253 /* Reference in_dev->dev */
254 dev_hold(dev);
255 /* Account for reference dev->ip_ptr (below) */
256 in_dev_hold(in_dev);
257
258 devinet_sysctl_register(in_dev);
259 ip_mc_init_dev(in_dev);
260 if (dev->flags & IFF_UP)
261 ip_mc_up(in_dev);
262
263 /* we can receive as soon as ip_ptr is set -- do this last */
264 rcu_assign_pointer(dev->ip_ptr, in_dev);
265 out:
266 return in_dev;
267 out_kfree:
268 kfree(in_dev);
269 in_dev = NULL;
270 goto out;
271 }
272
273 static void in_dev_rcu_put(struct rcu_head *head)
274 {
275 struct in_device *idev = container_of(head, struct in_device, rcu_head);
276 in_dev_put(idev);
277 }
278
279 static void inetdev_destroy(struct in_device *in_dev)
280 {
281 struct in_ifaddr *ifa;
282 struct net_device *dev;
283
284 ASSERT_RTNL();
285
286 dev = in_dev->dev;
287
288 in_dev->dead = 1;
289
290 ip_mc_destroy_dev(in_dev);
291
292 while ((ifa = in_dev->ifa_list) != NULL) {
293 inet_del_ifa(in_dev, &in_dev->ifa_list, 0);
294 inet_free_ifa(ifa);
295 }
296
297 RCU_INIT_POINTER(dev->ip_ptr, NULL);
298
299 devinet_sysctl_unregister(in_dev);
300 neigh_parms_release(&arp_tbl, in_dev->arp_parms);
301 arp_ifdown(dev);
302
303 call_rcu(&in_dev->rcu_head, in_dev_rcu_put);
304 }
305
306 int inet_addr_onlink(struct in_device *in_dev, __be32 a, __be32 b)
307 {
308 rcu_read_lock();
309 for_primary_ifa(in_dev) {
310 if (inet_ifa_match(a, ifa)) {
311 if (!b || inet_ifa_match(b, ifa)) {
312 rcu_read_unlock();
313 return 1;
314 }
315 }
316 } endfor_ifa(in_dev);
317 rcu_read_unlock();
318 return 0;
319 }
320
321 static void __inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap,
322 int destroy, struct nlmsghdr *nlh, u32 portid)
323 {
324 struct in_ifaddr *promote = NULL;
325 struct in_ifaddr *ifa, *ifa1 = *ifap;
326 struct in_ifaddr *last_prim = in_dev->ifa_list;
327 struct in_ifaddr *prev_prom = NULL;
328 int do_promote = IN_DEV_PROMOTE_SECONDARIES(in_dev);
329
330 ASSERT_RTNL();
331
332 /* 1. Deleting primary ifaddr forces deletion all secondaries
333 * unless alias promotion is set
334 **/
335
336 if (!(ifa1->ifa_flags & IFA_F_SECONDARY)) {
337 struct in_ifaddr **ifap1 = &ifa1->ifa_next;
338
339 while ((ifa = *ifap1) != NULL) {
340 if (!(ifa->ifa_flags & IFA_F_SECONDARY) &&
341 ifa1->ifa_scope <= ifa->ifa_scope)
342 last_prim = ifa;
343
344 if (!(ifa->ifa_flags & IFA_F_SECONDARY) ||
345 ifa1->ifa_mask != ifa->ifa_mask ||
346 !inet_ifa_match(ifa1->ifa_address, ifa)) {
347 ifap1 = &ifa->ifa_next;
348 prev_prom = ifa;
349 continue;
350 }
351
352 if (!do_promote) {
353 inet_hash_remove(ifa);
354 *ifap1 = ifa->ifa_next;
355
356 rtmsg_ifa(RTM_DELADDR, ifa, nlh, portid);
357 blocking_notifier_call_chain(&inetaddr_chain,
358 NETDEV_DOWN, ifa);
359 inet_free_ifa(ifa);
360 } else {
361 promote = ifa;
362 break;
363 }
364 }
365 }
366
367 /* On promotion all secondaries from subnet are changing
368 * the primary IP, we must remove all their routes silently
369 * and later to add them back with new prefsrc. Do this
370 * while all addresses are on the device list.
371 */
372 for (ifa = promote; ifa; ifa = ifa->ifa_next) {
373 if (ifa1->ifa_mask == ifa->ifa_mask &&
374 inet_ifa_match(ifa1->ifa_address, ifa))
375 fib_del_ifaddr(ifa, ifa1);
376 }
377
378 /* 2. Unlink it */
379
380 *ifap = ifa1->ifa_next;
381 inet_hash_remove(ifa1);
382
383 /* 3. Announce address deletion */
384
385 /* Send message first, then call notifier.
386 At first sight, FIB update triggered by notifier
387 will refer to already deleted ifaddr, that could confuse
388 netlink listeners. It is not true: look, gated sees
389 that route deleted and if it still thinks that ifaddr
390 is valid, it will try to restore deleted routes... Grr.
391 So that, this order is correct.
392 */
393 rtmsg_ifa(RTM_DELADDR, ifa1, nlh, portid);
394 blocking_notifier_call_chain(&inetaddr_chain, NETDEV_DOWN, ifa1);
395
396 if (promote) {
397 struct in_ifaddr *next_sec = promote->ifa_next;
398
399 if (prev_prom) {
400 prev_prom->ifa_next = promote->ifa_next;
401 promote->ifa_next = last_prim->ifa_next;
402 last_prim->ifa_next = promote;
403 }
404
405 promote->ifa_flags &= ~IFA_F_SECONDARY;
406 rtmsg_ifa(RTM_NEWADDR, promote, nlh, portid);
407 blocking_notifier_call_chain(&inetaddr_chain,
408 NETDEV_UP, promote);
409 for (ifa = next_sec; ifa; ifa = ifa->ifa_next) {
410 if (ifa1->ifa_mask != ifa->ifa_mask ||
411 !inet_ifa_match(ifa1->ifa_address, ifa))
412 continue;
413 fib_add_ifaddr(ifa);
414 }
415
416 }
417 if (destroy)
418 inet_free_ifa(ifa1);
419 }
420
421 static void inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap,
422 int destroy)
423 {
424 __inet_del_ifa(in_dev, ifap, destroy, NULL, 0);
425 }
426
427 static void check_lifetime(struct work_struct *work);
428
429 static DECLARE_DELAYED_WORK(check_lifetime_work, check_lifetime);
430
431 static int __inet_insert_ifa(struct in_ifaddr *ifa, struct nlmsghdr *nlh,
432 u32 portid)
433 {
434 struct in_device *in_dev = ifa->ifa_dev;
435 struct in_ifaddr *ifa1, **ifap, **last_primary;
436
437 ASSERT_RTNL();
438
439 if (!ifa->ifa_local) {
440 inet_free_ifa(ifa);
441 return 0;
442 }
443
444 ifa->ifa_flags &= ~IFA_F_SECONDARY;
445 last_primary = &in_dev->ifa_list;
446
447 for (ifap = &in_dev->ifa_list; (ifa1 = *ifap) != NULL;
448 ifap = &ifa1->ifa_next) {
449 if (!(ifa1->ifa_flags & IFA_F_SECONDARY) &&
450 ifa->ifa_scope <= ifa1->ifa_scope)
451 last_primary = &ifa1->ifa_next;
452 if (ifa1->ifa_mask == ifa->ifa_mask &&
453 inet_ifa_match(ifa1->ifa_address, ifa)) {
454 if (ifa1->ifa_local == ifa->ifa_local) {
455 inet_free_ifa(ifa);
456 return -EEXIST;
457 }
458 if (ifa1->ifa_scope != ifa->ifa_scope) {
459 inet_free_ifa(ifa);
460 return -EINVAL;
461 }
462 ifa->ifa_flags |= IFA_F_SECONDARY;
463 }
464 }
465
466 if (!(ifa->ifa_flags & IFA_F_SECONDARY)) {
467 net_srandom(ifa->ifa_local);
468 ifap = last_primary;
469 }
470
471 ifa->ifa_next = *ifap;
472 *ifap = ifa;
473
474 inet_hash_insert(dev_net(in_dev->dev), ifa);
475
476 cancel_delayed_work(&check_lifetime_work);
477 schedule_delayed_work(&check_lifetime_work, 0);
478
479 /* Send message first, then call notifier.
480 Notifier will trigger FIB update, so that
481 listeners of netlink will know about new ifaddr */
482 rtmsg_ifa(RTM_NEWADDR, ifa, nlh, portid);
483 blocking_notifier_call_chain(&inetaddr_chain, NETDEV_UP, ifa);
484
485 return 0;
486 }
487
488 static int inet_insert_ifa(struct in_ifaddr *ifa)
489 {
490 return __inet_insert_ifa(ifa, NULL, 0);
491 }
492
493 static int inet_set_ifa(struct net_device *dev, struct in_ifaddr *ifa)
494 {
495 struct in_device *in_dev = __in_dev_get_rtnl(dev);
496
497 ASSERT_RTNL();
498
499 if (!in_dev) {
500 inet_free_ifa(ifa);
501 return -ENOBUFS;
502 }
503 ipv4_devconf_setall(in_dev);
504 neigh_parms_data_state_setall(in_dev->arp_parms);
505 if (ifa->ifa_dev != in_dev) {
506 WARN_ON(ifa->ifa_dev);
507 in_dev_hold(in_dev);
508 ifa->ifa_dev = in_dev;
509 }
510 if (ipv4_is_loopback(ifa->ifa_local))
511 ifa->ifa_scope = RT_SCOPE_HOST;
512 return inet_insert_ifa(ifa);
513 }
514
515 /* Caller must hold RCU or RTNL :
516 * We dont take a reference on found in_device
517 */
518 struct in_device *inetdev_by_index(struct net *net, int ifindex)
519 {
520 struct net_device *dev;
521 struct in_device *in_dev = NULL;
522
523 rcu_read_lock();
524 dev = dev_get_by_index_rcu(net, ifindex);
525 if (dev)
526 in_dev = rcu_dereference_rtnl(dev->ip_ptr);
527 rcu_read_unlock();
528 return in_dev;
529 }
530 EXPORT_SYMBOL(inetdev_by_index);
531
532 /* Called only from RTNL semaphored context. No locks. */
533
534 struct in_ifaddr *inet_ifa_byprefix(struct in_device *in_dev, __be32 prefix,
535 __be32 mask)
536 {
537 ASSERT_RTNL();
538
539 for_primary_ifa(in_dev) {
540 if (ifa->ifa_mask == mask && inet_ifa_match(prefix, ifa))
541 return ifa;
542 } endfor_ifa(in_dev);
543 return NULL;
544 }
545
546 static int inet_rtm_deladdr(struct sk_buff *skb, struct nlmsghdr *nlh)
547 {
548 struct net *net = sock_net(skb->sk);
549 struct nlattr *tb[IFA_MAX+1];
550 struct in_device *in_dev;
551 struct ifaddrmsg *ifm;
552 struct in_ifaddr *ifa, **ifap;
553 int err = -EINVAL;
554
555 ASSERT_RTNL();
556
557 err = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_ipv4_policy);
558 if (err < 0)
559 goto errout;
560
561 ifm = nlmsg_data(nlh);
562 in_dev = inetdev_by_index(net, ifm->ifa_index);
563 if (in_dev == NULL) {
564 err = -ENODEV;
565 goto errout;
566 }
567
568 for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL;
569 ifap = &ifa->ifa_next) {
570 if (tb[IFA_LOCAL] &&
571 ifa->ifa_local != nla_get_be32(tb[IFA_LOCAL]))
572 continue;
573
574 if (tb[IFA_LABEL] && nla_strcmp(tb[IFA_LABEL], ifa->ifa_label))
575 continue;
576
577 if (tb[IFA_ADDRESS] &&
578 (ifm->ifa_prefixlen != ifa->ifa_prefixlen ||
579 !inet_ifa_match(nla_get_be32(tb[IFA_ADDRESS]), ifa)))
580 continue;
581
582 __inet_del_ifa(in_dev, ifap, 1, nlh, NETLINK_CB(skb).portid);
583 return 0;
584 }
585
586 err = -EADDRNOTAVAIL;
587 errout:
588 return err;
589 }
590
591 #define INFINITY_LIFE_TIME 0xFFFFFFFF
592
593 static void check_lifetime(struct work_struct *work)
594 {
595 unsigned long now, next, next_sec, next_sched;
596 struct in_ifaddr *ifa;
597 struct hlist_node *n;
598 int i;
599
600 now = jiffies;
601 next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY);
602
603 for (i = 0; i < IN4_ADDR_HSIZE; i++) {
604 bool change_needed = false;
605
606 rcu_read_lock();
607 hlist_for_each_entry_rcu(ifa, &inet_addr_lst[i], hash) {
608 unsigned long age;
609
610 if (ifa->ifa_flags & IFA_F_PERMANENT)
611 continue;
612
613 /* We try to batch several events at once. */
614 age = (now - ifa->ifa_tstamp +
615 ADDRCONF_TIMER_FUZZ_MINUS) / HZ;
616
617 if (ifa->ifa_valid_lft != INFINITY_LIFE_TIME &&
618 age >= ifa->ifa_valid_lft) {
619 change_needed = true;
620 } else if (ifa->ifa_preferred_lft ==
621 INFINITY_LIFE_TIME) {
622 continue;
623 } else if (age >= ifa->ifa_preferred_lft) {
624 if (time_before(ifa->ifa_tstamp +
625 ifa->ifa_valid_lft * HZ, next))
626 next = ifa->ifa_tstamp +
627 ifa->ifa_valid_lft * HZ;
628
629 if (!(ifa->ifa_flags & IFA_F_DEPRECATED))
630 change_needed = true;
631 } else if (time_before(ifa->ifa_tstamp +
632 ifa->ifa_preferred_lft * HZ,
633 next)) {
634 next = ifa->ifa_tstamp +
635 ifa->ifa_preferred_lft * HZ;
636 }
637 }
638 rcu_read_unlock();
639 if (!change_needed)
640 continue;
641 rtnl_lock();
642 hlist_for_each_entry_safe(ifa, n, &inet_addr_lst[i], hash) {
643 unsigned long age;
644
645 if (ifa->ifa_flags & IFA_F_PERMANENT)
646 continue;
647
648 /* We try to batch several events at once. */
649 age = (now - ifa->ifa_tstamp +
650 ADDRCONF_TIMER_FUZZ_MINUS) / HZ;
651
652 if (ifa->ifa_valid_lft != INFINITY_LIFE_TIME &&
653 age >= ifa->ifa_valid_lft) {
654 struct in_ifaddr **ifap;
655
656 for (ifap = &ifa->ifa_dev->ifa_list;
657 *ifap != NULL; ifap = &(*ifap)->ifa_next) {
658 if (*ifap == ifa) {
659 inet_del_ifa(ifa->ifa_dev,
660 ifap, 1);
661 break;
662 }
663 }
664 } else if (ifa->ifa_preferred_lft !=
665 INFINITY_LIFE_TIME &&
666 age >= ifa->ifa_preferred_lft &&
667 !(ifa->ifa_flags & IFA_F_DEPRECATED)) {
668 ifa->ifa_flags |= IFA_F_DEPRECATED;
669 rtmsg_ifa(RTM_NEWADDR, ifa, NULL, 0);
670 }
671 }
672 rtnl_unlock();
673 }
674
675 next_sec = round_jiffies_up(next);
676 next_sched = next;
677
678 /* If rounded timeout is accurate enough, accept it. */
679 if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ))
680 next_sched = next_sec;
681
682 now = jiffies;
683 /* And minimum interval is ADDRCONF_TIMER_FUZZ_MAX. */
684 if (time_before(next_sched, now + ADDRCONF_TIMER_FUZZ_MAX))
685 next_sched = now + ADDRCONF_TIMER_FUZZ_MAX;
686
687 schedule_delayed_work(&check_lifetime_work, next_sched - now);
688 }
689
690 static void set_ifa_lifetime(struct in_ifaddr *ifa, __u32 valid_lft,
691 __u32 prefered_lft)
692 {
693 unsigned long timeout;
694
695 ifa->ifa_flags &= ~(IFA_F_PERMANENT | IFA_F_DEPRECATED);
696
697 timeout = addrconf_timeout_fixup(valid_lft, HZ);
698 if (addrconf_finite_timeout(timeout))
699 ifa->ifa_valid_lft = timeout;
700 else
701 ifa->ifa_flags |= IFA_F_PERMANENT;
702
703 timeout = addrconf_timeout_fixup(prefered_lft, HZ);
704 if (addrconf_finite_timeout(timeout)) {
705 if (timeout == 0)
706 ifa->ifa_flags |= IFA_F_DEPRECATED;
707 ifa->ifa_preferred_lft = timeout;
708 }
709 ifa->ifa_tstamp = jiffies;
710 if (!ifa->ifa_cstamp)
711 ifa->ifa_cstamp = ifa->ifa_tstamp;
712 }
713
714 static struct in_ifaddr *rtm_to_ifaddr(struct net *net, struct nlmsghdr *nlh,
715 __u32 *pvalid_lft, __u32 *pprefered_lft)
716 {
717 struct nlattr *tb[IFA_MAX+1];
718 struct in_ifaddr *ifa;
719 struct ifaddrmsg *ifm;
720 struct net_device *dev;
721 struct in_device *in_dev;
722 int err;
723
724 err = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_ipv4_policy);
725 if (err < 0)
726 goto errout;
727
728 ifm = nlmsg_data(nlh);
729 err = -EINVAL;
730 if (ifm->ifa_prefixlen > 32 || tb[IFA_LOCAL] == NULL)
731 goto errout;
732
733 dev = __dev_get_by_index(net, ifm->ifa_index);
734 err = -ENODEV;
735 if (dev == NULL)
736 goto errout;
737
738 in_dev = __in_dev_get_rtnl(dev);
739 err = -ENOBUFS;
740 if (in_dev == NULL)
741 goto errout;
742
743 ifa = inet_alloc_ifa();
744 if (ifa == NULL)
745 /*
746 * A potential indev allocation can be left alive, it stays
747 * assigned to its device and is destroy with it.
748 */
749 goto errout;
750
751 ipv4_devconf_setall(in_dev);
752 neigh_parms_data_state_setall(in_dev->arp_parms);
753 in_dev_hold(in_dev);
754
755 if (tb[IFA_ADDRESS] == NULL)
756 tb[IFA_ADDRESS] = tb[IFA_LOCAL];
757
758 INIT_HLIST_NODE(&ifa->hash);
759 ifa->ifa_prefixlen = ifm->ifa_prefixlen;
760 ifa->ifa_mask = inet_make_mask(ifm->ifa_prefixlen);
761 ifa->ifa_flags = tb[IFA_FLAGS] ? nla_get_u32(tb[IFA_FLAGS]) :
762 ifm->ifa_flags;
763 ifa->ifa_scope = ifm->ifa_scope;
764 ifa->ifa_dev = in_dev;
765
766 ifa->ifa_local = nla_get_be32(tb[IFA_LOCAL]);
767 ifa->ifa_address = nla_get_be32(tb[IFA_ADDRESS]);
768
769 if (tb[IFA_BROADCAST])
770 ifa->ifa_broadcast = nla_get_be32(tb[IFA_BROADCAST]);
771
772 if (tb[IFA_LABEL])
773 nla_strlcpy(ifa->ifa_label, tb[IFA_LABEL], IFNAMSIZ);
774 else
775 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
776
777 if (tb[IFA_CACHEINFO]) {
778 struct ifa_cacheinfo *ci;
779
780 ci = nla_data(tb[IFA_CACHEINFO]);
781 if (!ci->ifa_valid || ci->ifa_prefered > ci->ifa_valid) {
782 err = -EINVAL;
783 goto errout_free;
784 }
785 *pvalid_lft = ci->ifa_valid;
786 *pprefered_lft = ci->ifa_prefered;
787 }
788
789 return ifa;
790
791 errout_free:
792 inet_free_ifa(ifa);
793 errout:
794 return ERR_PTR(err);
795 }
796
797 static struct in_ifaddr *find_matching_ifa(struct in_ifaddr *ifa)
798 {
799 struct in_device *in_dev = ifa->ifa_dev;
800 struct in_ifaddr *ifa1, **ifap;
801
802 if (!ifa->ifa_local)
803 return NULL;
804
805 for (ifap = &in_dev->ifa_list; (ifa1 = *ifap) != NULL;
806 ifap = &ifa1->ifa_next) {
807 if (ifa1->ifa_mask == ifa->ifa_mask &&
808 inet_ifa_match(ifa1->ifa_address, ifa) &&
809 ifa1->ifa_local == ifa->ifa_local)
810 return ifa1;
811 }
812 return NULL;
813 }
814
815 static int inet_rtm_newaddr(struct sk_buff *skb, struct nlmsghdr *nlh)
816 {
817 struct net *net = sock_net(skb->sk);
818 struct in_ifaddr *ifa;
819 struct in_ifaddr *ifa_existing;
820 __u32 valid_lft = INFINITY_LIFE_TIME;
821 __u32 prefered_lft = INFINITY_LIFE_TIME;
822
823 ASSERT_RTNL();
824
825 ifa = rtm_to_ifaddr(net, nlh, &valid_lft, &prefered_lft);
826 if (IS_ERR(ifa))
827 return PTR_ERR(ifa);
828
829 ifa_existing = find_matching_ifa(ifa);
830 if (!ifa_existing) {
831 /* It would be best to check for !NLM_F_CREATE here but
832 * userspace alreay relies on not having to provide this.
833 */
834 set_ifa_lifetime(ifa, valid_lft, prefered_lft);
835 return __inet_insert_ifa(ifa, nlh, NETLINK_CB(skb).portid);
836 } else {
837 inet_free_ifa(ifa);
838
839 if (nlh->nlmsg_flags & NLM_F_EXCL ||
840 !(nlh->nlmsg_flags & NLM_F_REPLACE))
841 return -EEXIST;
842 ifa = ifa_existing;
843 set_ifa_lifetime(ifa, valid_lft, prefered_lft);
844 cancel_delayed_work(&check_lifetime_work);
845 schedule_delayed_work(&check_lifetime_work, 0);
846 rtmsg_ifa(RTM_NEWADDR, ifa, nlh, NETLINK_CB(skb).portid);
847 blocking_notifier_call_chain(&inetaddr_chain, NETDEV_UP, ifa);
848 }
849 return 0;
850 }
851
852 /*
853 * Determine a default network mask, based on the IP address.
854 */
855
856 static int inet_abc_len(__be32 addr)
857 {
858 int rc = -1; /* Something else, probably a multicast. */
859
860 if (ipv4_is_zeronet(addr))
861 rc = 0;
862 else {
863 __u32 haddr = ntohl(addr);
864
865 if (IN_CLASSA(haddr))
866 rc = 8;
867 else if (IN_CLASSB(haddr))
868 rc = 16;
869 else if (IN_CLASSC(haddr))
870 rc = 24;
871 }
872
873 return rc;
874 }
875
876
877 int devinet_ioctl(struct net *net, unsigned int cmd, void __user *arg)
878 {
879 struct ifreq ifr;
880 struct sockaddr_in sin_orig;
881 struct sockaddr_in *sin = (struct sockaddr_in *)&ifr.ifr_addr;
882 struct in_device *in_dev;
883 struct in_ifaddr **ifap = NULL;
884 struct in_ifaddr *ifa = NULL;
885 struct net_device *dev;
886 char *colon;
887 int ret = -EFAULT;
888 int tryaddrmatch = 0;
889
890 /*
891 * Fetch the caller's info block into kernel space
892 */
893
894 if (copy_from_user(&ifr, arg, sizeof(struct ifreq)))
895 goto out;
896 ifr.ifr_name[IFNAMSIZ - 1] = 0;
897
898 /* save original address for comparison */
899 memcpy(&sin_orig, sin, sizeof(*sin));
900
901 colon = strchr(ifr.ifr_name, ':');
902 if (colon)
903 *colon = 0;
904
905 dev_load(net, ifr.ifr_name);
906
907 switch (cmd) {
908 case SIOCGIFADDR: /* Get interface address */
909 case SIOCGIFBRDADDR: /* Get the broadcast address */
910 case SIOCGIFDSTADDR: /* Get the destination address */
911 case SIOCGIFNETMASK: /* Get the netmask for the interface */
912 /* Note that these ioctls will not sleep,
913 so that we do not impose a lock.
914 One day we will be forced to put shlock here (I mean SMP)
915 */
916 tryaddrmatch = (sin_orig.sin_family == AF_INET);
917 memset(sin, 0, sizeof(*sin));
918 sin->sin_family = AF_INET;
919 break;
920
921 case SIOCSIFFLAGS:
922 ret = -EPERM;
923 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
924 goto out;
925 break;
926 case SIOCSIFADDR: /* Set interface address (and family) */
927 case SIOCSIFBRDADDR: /* Set the broadcast address */
928 case SIOCSIFDSTADDR: /* Set the destination address */
929 case SIOCSIFNETMASK: /* Set the netmask for the interface */
930 ret = -EPERM;
931 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
932 goto out;
933 ret = -EINVAL;
934 if (sin->sin_family != AF_INET)
935 goto out;
936 break;
937 default:
938 ret = -EINVAL;
939 goto out;
940 }
941
942 rtnl_lock();
943
944 ret = -ENODEV;
945 dev = __dev_get_by_name(net, ifr.ifr_name);
946 if (!dev)
947 goto done;
948
949 if (colon)
950 *colon = ':';
951
952 in_dev = __in_dev_get_rtnl(dev);
953 if (in_dev) {
954 if (tryaddrmatch) {
955 /* Matthias Andree */
956 /* compare label and address (4.4BSD style) */
957 /* note: we only do this for a limited set of ioctls
958 and only if the original address family was AF_INET.
959 This is checked above. */
960 for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL;
961 ifap = &ifa->ifa_next) {
962 if (!strcmp(ifr.ifr_name, ifa->ifa_label) &&
963 sin_orig.sin_addr.s_addr ==
964 ifa->ifa_local) {
965 break; /* found */
966 }
967 }
968 }
969 /* we didn't get a match, maybe the application is
970 4.3BSD-style and passed in junk so we fall back to
971 comparing just the label */
972 if (!ifa) {
973 for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL;
974 ifap = &ifa->ifa_next)
975 if (!strcmp(ifr.ifr_name, ifa->ifa_label))
976 break;
977 }
978 }
979
980 ret = -EADDRNOTAVAIL;
981 if (!ifa && cmd != SIOCSIFADDR && cmd != SIOCSIFFLAGS)
982 goto done;
983
984 switch (cmd) {
985 case SIOCGIFADDR: /* Get interface address */
986 sin->sin_addr.s_addr = ifa->ifa_local;
987 goto rarok;
988
989 case SIOCGIFBRDADDR: /* Get the broadcast address */
990 sin->sin_addr.s_addr = ifa->ifa_broadcast;
991 goto rarok;
992
993 case SIOCGIFDSTADDR: /* Get the destination address */
994 sin->sin_addr.s_addr = ifa->ifa_address;
995 goto rarok;
996
997 case SIOCGIFNETMASK: /* Get the netmask for the interface */
998 sin->sin_addr.s_addr = ifa->ifa_mask;
999 goto rarok;
1000
1001 case SIOCSIFFLAGS:
1002 if (colon) {
1003 ret = -EADDRNOTAVAIL;
1004 if (!ifa)
1005 break;
1006 ret = 0;
1007 if (!(ifr.ifr_flags & IFF_UP))
1008 inet_del_ifa(in_dev, ifap, 1);
1009 break;
1010 }
1011 ret = dev_change_flags(dev, ifr.ifr_flags);
1012 break;
1013
1014 case SIOCSIFADDR: /* Set interface address (and family) */
1015 ret = -EINVAL;
1016 if (inet_abc_len(sin->sin_addr.s_addr) < 0)
1017 break;
1018
1019 if (!ifa) {
1020 ret = -ENOBUFS;
1021 ifa = inet_alloc_ifa();
1022 if (!ifa)
1023 break;
1024 INIT_HLIST_NODE(&ifa->hash);
1025 if (colon)
1026 memcpy(ifa->ifa_label, ifr.ifr_name, IFNAMSIZ);
1027 else
1028 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1029 } else {
1030 ret = 0;
1031 if (ifa->ifa_local == sin->sin_addr.s_addr)
1032 break;
1033 inet_del_ifa(in_dev, ifap, 0);
1034 ifa->ifa_broadcast = 0;
1035 ifa->ifa_scope = 0;
1036 }
1037
1038 ifa->ifa_address = ifa->ifa_local = sin->sin_addr.s_addr;
1039
1040 if (!(dev->flags & IFF_POINTOPOINT)) {
1041 ifa->ifa_prefixlen = inet_abc_len(ifa->ifa_address);
1042 ifa->ifa_mask = inet_make_mask(ifa->ifa_prefixlen);
1043 if ((dev->flags & IFF_BROADCAST) &&
1044 ifa->ifa_prefixlen < 31)
1045 ifa->ifa_broadcast = ifa->ifa_address |
1046 ~ifa->ifa_mask;
1047 } else {
1048 ifa->ifa_prefixlen = 32;
1049 ifa->ifa_mask = inet_make_mask(32);
1050 }
1051 set_ifa_lifetime(ifa, INFINITY_LIFE_TIME, INFINITY_LIFE_TIME);
1052 ret = inet_set_ifa(dev, ifa);
1053 break;
1054
1055 case SIOCSIFBRDADDR: /* Set the broadcast address */
1056 ret = 0;
1057 if (ifa->ifa_broadcast != sin->sin_addr.s_addr) {
1058 inet_del_ifa(in_dev, ifap, 0);
1059 ifa->ifa_broadcast = sin->sin_addr.s_addr;
1060 inet_insert_ifa(ifa);
1061 }
1062 break;
1063
1064 case SIOCSIFDSTADDR: /* Set the destination address */
1065 ret = 0;
1066 if (ifa->ifa_address == sin->sin_addr.s_addr)
1067 break;
1068 ret = -EINVAL;
1069 if (inet_abc_len(sin->sin_addr.s_addr) < 0)
1070 break;
1071 ret = 0;
1072 inet_del_ifa(in_dev, ifap, 0);
1073 ifa->ifa_address = sin->sin_addr.s_addr;
1074 inet_insert_ifa(ifa);
1075 break;
1076
1077 case SIOCSIFNETMASK: /* Set the netmask for the interface */
1078
1079 /*
1080 * The mask we set must be legal.
1081 */
1082 ret = -EINVAL;
1083 if (bad_mask(sin->sin_addr.s_addr, 0))
1084 break;
1085 ret = 0;
1086 if (ifa->ifa_mask != sin->sin_addr.s_addr) {
1087 __be32 old_mask = ifa->ifa_mask;
1088 inet_del_ifa(in_dev, ifap, 0);
1089 ifa->ifa_mask = sin->sin_addr.s_addr;
1090 ifa->ifa_prefixlen = inet_mask_len(ifa->ifa_mask);
1091
1092 /* See if current broadcast address matches
1093 * with current netmask, then recalculate
1094 * the broadcast address. Otherwise it's a
1095 * funny address, so don't touch it since
1096 * the user seems to know what (s)he's doing...
1097 */
1098 if ((dev->flags & IFF_BROADCAST) &&
1099 (ifa->ifa_prefixlen < 31) &&
1100 (ifa->ifa_broadcast ==
1101 (ifa->ifa_local|~old_mask))) {
1102 ifa->ifa_broadcast = (ifa->ifa_local |
1103 ~sin->sin_addr.s_addr);
1104 }
1105 inet_insert_ifa(ifa);
1106 }
1107 break;
1108 }
1109 done:
1110 rtnl_unlock();
1111 out:
1112 return ret;
1113 rarok:
1114 rtnl_unlock();
1115 ret = copy_to_user(arg, &ifr, sizeof(struct ifreq)) ? -EFAULT : 0;
1116 goto out;
1117 }
1118
1119 static int inet_gifconf(struct net_device *dev, char __user *buf, int len)
1120 {
1121 struct in_device *in_dev = __in_dev_get_rtnl(dev);
1122 struct in_ifaddr *ifa;
1123 struct ifreq ifr;
1124 int done = 0;
1125
1126 if (!in_dev)
1127 goto out;
1128
1129 for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) {
1130 if (!buf) {
1131 done += sizeof(ifr);
1132 continue;
1133 }
1134 if (len < (int) sizeof(ifr))
1135 break;
1136 memset(&ifr, 0, sizeof(struct ifreq));
1137 strcpy(ifr.ifr_name, ifa->ifa_label);
1138
1139 (*(struct sockaddr_in *)&ifr.ifr_addr).sin_family = AF_INET;
1140 (*(struct sockaddr_in *)&ifr.ifr_addr).sin_addr.s_addr =
1141 ifa->ifa_local;
1142
1143 if (copy_to_user(buf, &ifr, sizeof(struct ifreq))) {
1144 done = -EFAULT;
1145 break;
1146 }
1147 buf += sizeof(struct ifreq);
1148 len -= sizeof(struct ifreq);
1149 done += sizeof(struct ifreq);
1150 }
1151 out:
1152 return done;
1153 }
1154
1155 __be32 inet_select_addr(const struct net_device *dev, __be32 dst, int scope)
1156 {
1157 __be32 addr = 0;
1158 struct in_device *in_dev;
1159 struct net *net = dev_net(dev);
1160
1161 rcu_read_lock();
1162 in_dev = __in_dev_get_rcu(dev);
1163 if (!in_dev)
1164 goto no_in_dev;
1165
1166 for_primary_ifa(in_dev) {
1167 if (ifa->ifa_scope > scope)
1168 continue;
1169 if (!dst || inet_ifa_match(dst, ifa)) {
1170 addr = ifa->ifa_local;
1171 break;
1172 }
1173 if (!addr)
1174 addr = ifa->ifa_local;
1175 } endfor_ifa(in_dev);
1176
1177 if (addr)
1178 goto out_unlock;
1179 no_in_dev:
1180
1181 /* Not loopback addresses on loopback should be preferred
1182 in this case. It is importnat that lo is the first interface
1183 in dev_base list.
1184 */
1185 for_each_netdev_rcu(net, dev) {
1186 in_dev = __in_dev_get_rcu(dev);
1187 if (!in_dev)
1188 continue;
1189
1190 for_primary_ifa(in_dev) {
1191 if (ifa->ifa_scope != RT_SCOPE_LINK &&
1192 ifa->ifa_scope <= scope) {
1193 addr = ifa->ifa_local;
1194 goto out_unlock;
1195 }
1196 } endfor_ifa(in_dev);
1197 }
1198 out_unlock:
1199 rcu_read_unlock();
1200 return addr;
1201 }
1202 EXPORT_SYMBOL(inet_select_addr);
1203
1204 static __be32 confirm_addr_indev(struct in_device *in_dev, __be32 dst,
1205 __be32 local, int scope)
1206 {
1207 int same = 0;
1208 __be32 addr = 0;
1209
1210 for_ifa(in_dev) {
1211 if (!addr &&
1212 (local == ifa->ifa_local || !local) &&
1213 ifa->ifa_scope <= scope) {
1214 addr = ifa->ifa_local;
1215 if (same)
1216 break;
1217 }
1218 if (!same) {
1219 same = (!local || inet_ifa_match(local, ifa)) &&
1220 (!dst || inet_ifa_match(dst, ifa));
1221 if (same && addr) {
1222 if (local || !dst)
1223 break;
1224 /* Is the selected addr into dst subnet? */
1225 if (inet_ifa_match(addr, ifa))
1226 break;
1227 /* No, then can we use new local src? */
1228 if (ifa->ifa_scope <= scope) {
1229 addr = ifa->ifa_local;
1230 break;
1231 }
1232 /* search for large dst subnet for addr */
1233 same = 0;
1234 }
1235 }
1236 } endfor_ifa(in_dev);
1237
1238 return same ? addr : 0;
1239 }
1240
1241 /*
1242 * Confirm that local IP address exists using wildcards:
1243 * - net: netns to check, cannot be NULL
1244 * - in_dev: only on this interface, NULL=any interface
1245 * - dst: only in the same subnet as dst, 0=any dst
1246 * - local: address, 0=autoselect the local address
1247 * - scope: maximum allowed scope value for the local address
1248 */
1249 __be32 inet_confirm_addr(struct net *net, struct in_device *in_dev,
1250 __be32 dst, __be32 local, int scope)
1251 {
1252 __be32 addr = 0;
1253 struct net_device *dev;
1254
1255 if (in_dev != NULL)
1256 return confirm_addr_indev(in_dev, dst, local, scope);
1257
1258 rcu_read_lock();
1259 for_each_netdev_rcu(net, dev) {
1260 in_dev = __in_dev_get_rcu(dev);
1261 if (in_dev) {
1262 addr = confirm_addr_indev(in_dev, dst, local, scope);
1263 if (addr)
1264 break;
1265 }
1266 }
1267 rcu_read_unlock();
1268
1269 return addr;
1270 }
1271 EXPORT_SYMBOL(inet_confirm_addr);
1272
1273 /*
1274 * Device notifier
1275 */
1276
1277 int register_inetaddr_notifier(struct notifier_block *nb)
1278 {
1279 return blocking_notifier_chain_register(&inetaddr_chain, nb);
1280 }
1281 EXPORT_SYMBOL(register_inetaddr_notifier);
1282
1283 int unregister_inetaddr_notifier(struct notifier_block *nb)
1284 {
1285 return blocking_notifier_chain_unregister(&inetaddr_chain, nb);
1286 }
1287 EXPORT_SYMBOL(unregister_inetaddr_notifier);
1288
1289 /* Rename ifa_labels for a device name change. Make some effort to preserve
1290 * existing alias numbering and to create unique labels if possible.
1291 */
1292 static void inetdev_changename(struct net_device *dev, struct in_device *in_dev)
1293 {
1294 struct in_ifaddr *ifa;
1295 int named = 0;
1296
1297 for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) {
1298 char old[IFNAMSIZ], *dot;
1299
1300 memcpy(old, ifa->ifa_label, IFNAMSIZ);
1301 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1302 if (named++ == 0)
1303 goto skip;
1304 dot = strchr(old, ':');
1305 if (dot == NULL) {
1306 sprintf(old, ":%d", named);
1307 dot = old;
1308 }
1309 if (strlen(dot) + strlen(dev->name) < IFNAMSIZ)
1310 strcat(ifa->ifa_label, dot);
1311 else
1312 strcpy(ifa->ifa_label + (IFNAMSIZ - strlen(dot) - 1), dot);
1313 skip:
1314 rtmsg_ifa(RTM_NEWADDR, ifa, NULL, 0);
1315 }
1316 }
1317
1318 static bool inetdev_valid_mtu(unsigned int mtu)
1319 {
1320 return mtu >= 68;
1321 }
1322
1323 static void inetdev_send_gratuitous_arp(struct net_device *dev,
1324 struct in_device *in_dev)
1325
1326 {
1327 struct in_ifaddr *ifa;
1328
1329 for (ifa = in_dev->ifa_list; ifa;
1330 ifa = ifa->ifa_next) {
1331 arp_send(ARPOP_REQUEST, ETH_P_ARP,
1332 ifa->ifa_local, dev,
1333 ifa->ifa_local, NULL,
1334 dev->dev_addr, NULL);
1335 }
1336 }
1337
1338 /* Called only under RTNL semaphore */
1339
1340 static int inetdev_event(struct notifier_block *this, unsigned long event,
1341 void *ptr)
1342 {
1343 struct net_device *dev = netdev_notifier_info_to_dev(ptr);
1344 struct in_device *in_dev = __in_dev_get_rtnl(dev);
1345
1346 ASSERT_RTNL();
1347
1348 if (!in_dev) {
1349 if (event == NETDEV_REGISTER) {
1350 in_dev = inetdev_init(dev);
1351 if (!in_dev)
1352 return notifier_from_errno(-ENOMEM);
1353 if (dev->flags & IFF_LOOPBACK) {
1354 IN_DEV_CONF_SET(in_dev, NOXFRM, 1);
1355 IN_DEV_CONF_SET(in_dev, NOPOLICY, 1);
1356 }
1357 } else if (event == NETDEV_CHANGEMTU) {
1358 /* Re-enabling IP */
1359 if (inetdev_valid_mtu(dev->mtu))
1360 in_dev = inetdev_init(dev);
1361 }
1362 goto out;
1363 }
1364
1365 switch (event) {
1366 case NETDEV_REGISTER:
1367 pr_debug("%s: bug\n", __func__);
1368 RCU_INIT_POINTER(dev->ip_ptr, NULL);
1369 break;
1370 case NETDEV_UP:
1371 if (!inetdev_valid_mtu(dev->mtu))
1372 break;
1373 if (dev->flags & IFF_LOOPBACK) {
1374 struct in_ifaddr *ifa = inet_alloc_ifa();
1375
1376 if (ifa) {
1377 INIT_HLIST_NODE(&ifa->hash);
1378 ifa->ifa_local =
1379 ifa->ifa_address = htonl(INADDR_LOOPBACK);
1380 ifa->ifa_prefixlen = 8;
1381 ifa->ifa_mask = inet_make_mask(8);
1382 in_dev_hold(in_dev);
1383 ifa->ifa_dev = in_dev;
1384 ifa->ifa_scope = RT_SCOPE_HOST;
1385 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ);
1386 set_ifa_lifetime(ifa, INFINITY_LIFE_TIME,
1387 INFINITY_LIFE_TIME);
1388 inet_insert_ifa(ifa);
1389 }
1390 }
1391 ip_mc_up(in_dev);
1392 /* fall through */
1393 case NETDEV_CHANGEADDR:
1394 if (!IN_DEV_ARP_NOTIFY(in_dev))
1395 break;
1396 /* fall through */
1397 case NETDEV_NOTIFY_PEERS:
1398 /* Send gratuitous ARP to notify of link change */
1399 inetdev_send_gratuitous_arp(dev, in_dev);
1400 break;
1401 case NETDEV_DOWN:
1402 ip_mc_down(in_dev);
1403 break;
1404 case NETDEV_PRE_TYPE_CHANGE:
1405 ip_mc_unmap(in_dev);
1406 break;
1407 case NETDEV_POST_TYPE_CHANGE:
1408 ip_mc_remap(in_dev);
1409 break;
1410 case NETDEV_CHANGEMTU:
1411 if (inetdev_valid_mtu(dev->mtu))
1412 break;
1413 /* disable IP when MTU is not enough */
1414 case NETDEV_UNREGISTER:
1415 inetdev_destroy(in_dev);
1416 break;
1417 case NETDEV_CHANGENAME:
1418 /* Do not notify about label change, this event is
1419 * not interesting to applications using netlink.
1420 */
1421 inetdev_changename(dev, in_dev);
1422
1423 devinet_sysctl_unregister(in_dev);
1424 devinet_sysctl_register(in_dev);
1425 break;
1426 }
1427 out:
1428 return NOTIFY_DONE;
1429 }
1430
1431 static struct notifier_block ip_netdev_notifier = {
1432 .notifier_call = inetdev_event,
1433 };
1434
1435 static size_t inet_nlmsg_size(void)
1436 {
1437 return NLMSG_ALIGN(sizeof(struct ifaddrmsg))
1438 + nla_total_size(4) /* IFA_ADDRESS */
1439 + nla_total_size(4) /* IFA_LOCAL */
1440 + nla_total_size(4) /* IFA_BROADCAST */
1441 + nla_total_size(IFNAMSIZ) /* IFA_LABEL */
1442 + nla_total_size(4); /* IFA_FLAGS */
1443 }
1444
1445 static inline u32 cstamp_delta(unsigned long cstamp)
1446 {
1447 return (cstamp - INITIAL_JIFFIES) * 100UL / HZ;
1448 }
1449
1450 static int put_cacheinfo(struct sk_buff *skb, unsigned long cstamp,
1451 unsigned long tstamp, u32 preferred, u32 valid)
1452 {
1453 struct ifa_cacheinfo ci;
1454
1455 ci.cstamp = cstamp_delta(cstamp);
1456 ci.tstamp = cstamp_delta(tstamp);
1457 ci.ifa_prefered = preferred;
1458 ci.ifa_valid = valid;
1459
1460 return nla_put(skb, IFA_CACHEINFO, sizeof(ci), &ci);
1461 }
1462
1463 static int inet_fill_ifaddr(struct sk_buff *skb, struct in_ifaddr *ifa,
1464 u32 portid, u32 seq, int event, unsigned int flags)
1465 {
1466 struct ifaddrmsg *ifm;
1467 struct nlmsghdr *nlh;
1468 u32 preferred, valid;
1469
1470 nlh = nlmsg_put(skb, portid, seq, event, sizeof(*ifm), flags);
1471 if (nlh == NULL)
1472 return -EMSGSIZE;
1473
1474 ifm = nlmsg_data(nlh);
1475 ifm->ifa_family = AF_INET;
1476 ifm->ifa_prefixlen = ifa->ifa_prefixlen;
1477 ifm->ifa_flags = ifa->ifa_flags;
1478 ifm->ifa_scope = ifa->ifa_scope;
1479 ifm->ifa_index = ifa->ifa_dev->dev->ifindex;
1480
1481 if (!(ifm->ifa_flags & IFA_F_PERMANENT)) {
1482 preferred = ifa->ifa_preferred_lft;
1483 valid = ifa->ifa_valid_lft;
1484 if (preferred != INFINITY_LIFE_TIME) {
1485 long tval = (jiffies - ifa->ifa_tstamp) / HZ;
1486
1487 if (preferred > tval)
1488 preferred -= tval;
1489 else
1490 preferred = 0;
1491 if (valid != INFINITY_LIFE_TIME) {
1492 if (valid > tval)
1493 valid -= tval;
1494 else
1495 valid = 0;
1496 }
1497 }
1498 } else {
1499 preferred = INFINITY_LIFE_TIME;
1500 valid = INFINITY_LIFE_TIME;
1501 }
1502 if ((ifa->ifa_address &&
1503 nla_put_be32(skb, IFA_ADDRESS, ifa->ifa_address)) ||
1504 (ifa->ifa_local &&
1505 nla_put_be32(skb, IFA_LOCAL, ifa->ifa_local)) ||
1506 (ifa->ifa_broadcast &&
1507 nla_put_be32(skb, IFA_BROADCAST, ifa->ifa_broadcast)) ||
1508 (ifa->ifa_label[0] &&
1509 nla_put_string(skb, IFA_LABEL, ifa->ifa_label)) ||
1510 nla_put_u32(skb, IFA_FLAGS, ifa->ifa_flags) ||
1511 put_cacheinfo(skb, ifa->ifa_cstamp, ifa->ifa_tstamp,
1512 preferred, valid))
1513 goto nla_put_failure;
1514
1515 return nlmsg_end(skb, nlh);
1516
1517 nla_put_failure:
1518 nlmsg_cancel(skb, nlh);
1519 return -EMSGSIZE;
1520 }
1521
1522 static int inet_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb)
1523 {
1524 struct net *net = sock_net(skb->sk);
1525 int h, s_h;
1526 int idx, s_idx;
1527 int ip_idx, s_ip_idx;
1528 struct net_device *dev;
1529 struct in_device *in_dev;
1530 struct in_ifaddr *ifa;
1531 struct hlist_head *head;
1532
1533 s_h = cb->args[0];
1534 s_idx = idx = cb->args[1];
1535 s_ip_idx = ip_idx = cb->args[2];
1536
1537 for (h = s_h; h < NETDEV_HASHENTRIES; h++, s_idx = 0) {
1538 idx = 0;
1539 head = &net->dev_index_head[h];
1540 rcu_read_lock();
1541 cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^
1542 net->dev_base_seq;
1543 hlist_for_each_entry_rcu(dev, head, index_hlist) {
1544 if (idx < s_idx)
1545 goto cont;
1546 if (h > s_h || idx > s_idx)
1547 s_ip_idx = 0;
1548 in_dev = __in_dev_get_rcu(dev);
1549 if (!in_dev)
1550 goto cont;
1551
1552 for (ifa = in_dev->ifa_list, ip_idx = 0; ifa;
1553 ifa = ifa->ifa_next, ip_idx++) {
1554 if (ip_idx < s_ip_idx)
1555 continue;
1556 if (inet_fill_ifaddr(skb, ifa,
1557 NETLINK_CB(cb->skb).portid,
1558 cb->nlh->nlmsg_seq,
1559 RTM_NEWADDR, NLM_F_MULTI) <= 0) {
1560 rcu_read_unlock();
1561 goto done;
1562 }
1563 nl_dump_check_consistent(cb, nlmsg_hdr(skb));
1564 }
1565 cont:
1566 idx++;
1567 }
1568 rcu_read_unlock();
1569 }
1570
1571 done:
1572 cb->args[0] = h;
1573 cb->args[1] = idx;
1574 cb->args[2] = ip_idx;
1575
1576 return skb->len;
1577 }
1578
1579 static void rtmsg_ifa(int event, struct in_ifaddr *ifa, struct nlmsghdr *nlh,
1580 u32 portid)
1581 {
1582 struct sk_buff *skb;
1583 u32 seq = nlh ? nlh->nlmsg_seq : 0;
1584 int err = -ENOBUFS;
1585 struct net *net;
1586
1587 net = dev_net(ifa->ifa_dev->dev);
1588 skb = nlmsg_new(inet_nlmsg_size(), GFP_KERNEL);
1589 if (skb == NULL)
1590 goto errout;
1591
1592 err = inet_fill_ifaddr(skb, ifa, portid, seq, event, 0);
1593 if (err < 0) {
1594 /* -EMSGSIZE implies BUG in inet_nlmsg_size() */
1595 WARN_ON(err == -EMSGSIZE);
1596 kfree_skb(skb);
1597 goto errout;
1598 }
1599 rtnl_notify(skb, net, portid, RTNLGRP_IPV4_IFADDR, nlh, GFP_KERNEL);
1600 return;
1601 errout:
1602 if (err < 0)
1603 rtnl_set_sk_err(net, RTNLGRP_IPV4_IFADDR, err);
1604 }
1605
1606 static size_t inet_get_link_af_size(const struct net_device *dev)
1607 {
1608 struct in_device *in_dev = rcu_dereference_rtnl(dev->ip_ptr);
1609
1610 if (!in_dev)
1611 return 0;
1612
1613 return nla_total_size(IPV4_DEVCONF_MAX * 4); /* IFLA_INET_CONF */
1614 }
1615
1616 static int inet_fill_link_af(struct sk_buff *skb, const struct net_device *dev)
1617 {
1618 struct in_device *in_dev = rcu_dereference_rtnl(dev->ip_ptr);
1619 struct nlattr *nla;
1620 int i;
1621
1622 if (!in_dev)
1623 return -ENODATA;
1624
1625 nla = nla_reserve(skb, IFLA_INET_CONF, IPV4_DEVCONF_MAX * 4);
1626 if (nla == NULL)
1627 return -EMSGSIZE;
1628
1629 for (i = 0; i < IPV4_DEVCONF_MAX; i++)
1630 ((u32 *) nla_data(nla))[i] = in_dev->cnf.data[i];
1631
1632 return 0;
1633 }
1634
1635 static const struct nla_policy inet_af_policy[IFLA_INET_MAX+1] = {
1636 [IFLA_INET_CONF] = { .type = NLA_NESTED },
1637 };
1638
1639 static int inet_validate_link_af(const struct net_device *dev,
1640 const struct nlattr *nla)
1641 {
1642 struct nlattr *a, *tb[IFLA_INET_MAX+1];
1643 int err, rem;
1644
1645 if (dev && !__in_dev_get_rtnl(dev))
1646 return -EAFNOSUPPORT;
1647
1648 err = nla_parse_nested(tb, IFLA_INET_MAX, nla, inet_af_policy);
1649 if (err < 0)
1650 return err;
1651
1652 if (tb[IFLA_INET_CONF]) {
1653 nla_for_each_nested(a, tb[IFLA_INET_CONF], rem) {
1654 int cfgid = nla_type(a);
1655
1656 if (nla_len(a) < 4)
1657 return -EINVAL;
1658
1659 if (cfgid <= 0 || cfgid > IPV4_DEVCONF_MAX)
1660 return -EINVAL;
1661 }
1662 }
1663
1664 return 0;
1665 }
1666
1667 static int inet_set_link_af(struct net_device *dev, const struct nlattr *nla)
1668 {
1669 struct in_device *in_dev = __in_dev_get_rtnl(dev);
1670 struct nlattr *a, *tb[IFLA_INET_MAX+1];
1671 int rem;
1672
1673 if (!in_dev)
1674 return -EAFNOSUPPORT;
1675
1676 if (nla_parse_nested(tb, IFLA_INET_MAX, nla, NULL) < 0)
1677 BUG();
1678
1679 if (tb[IFLA_INET_CONF]) {
1680 nla_for_each_nested(a, tb[IFLA_INET_CONF], rem)
1681 ipv4_devconf_set(in_dev, nla_type(a), nla_get_u32(a));
1682 }
1683
1684 return 0;
1685 }
1686
1687 static int inet_netconf_msgsize_devconf(int type)
1688 {
1689 int size = NLMSG_ALIGN(sizeof(struct netconfmsg))
1690 + nla_total_size(4); /* NETCONFA_IFINDEX */
1691
1692 /* type -1 is used for ALL */
1693 if (type == -1 || type == NETCONFA_FORWARDING)
1694 size += nla_total_size(4);
1695 if (type == -1 || type == NETCONFA_RP_FILTER)
1696 size += nla_total_size(4);
1697 if (type == -1 || type == NETCONFA_MC_FORWARDING)
1698 size += nla_total_size(4);
1699 if (type == -1 || type == NETCONFA_PROXY_ARP)
1700 size += nla_total_size(4);
1701
1702 return size;
1703 }
1704
1705 static int inet_netconf_fill_devconf(struct sk_buff *skb, int ifindex,
1706 struct ipv4_devconf *devconf, u32 portid,
1707 u32 seq, int event, unsigned int flags,
1708 int type)
1709 {
1710 struct nlmsghdr *nlh;
1711 struct netconfmsg *ncm;
1712
1713 nlh = nlmsg_put(skb, portid, seq, event, sizeof(struct netconfmsg),
1714 flags);
1715 if (nlh == NULL)
1716 return -EMSGSIZE;
1717
1718 ncm = nlmsg_data(nlh);
1719 ncm->ncm_family = AF_INET;
1720
1721 if (nla_put_s32(skb, NETCONFA_IFINDEX, ifindex) < 0)
1722 goto nla_put_failure;
1723
1724 /* type -1 is used for ALL */
1725 if ((type == -1 || type == NETCONFA_FORWARDING) &&
1726 nla_put_s32(skb, NETCONFA_FORWARDING,
1727 IPV4_DEVCONF(*devconf, FORWARDING)) < 0)
1728 goto nla_put_failure;
1729 if ((type == -1 || type == NETCONFA_RP_FILTER) &&
1730 nla_put_s32(skb, NETCONFA_RP_FILTER,
1731 IPV4_DEVCONF(*devconf, RP_FILTER)) < 0)
1732 goto nla_put_failure;
1733 if ((type == -1 || type == NETCONFA_MC_FORWARDING) &&
1734 nla_put_s32(skb, NETCONFA_MC_FORWARDING,
1735 IPV4_DEVCONF(*devconf, MC_FORWARDING)) < 0)
1736 goto nla_put_failure;
1737 if ((type == -1 || type == NETCONFA_PROXY_ARP) &&
1738 nla_put_s32(skb, NETCONFA_PROXY_ARP,
1739 IPV4_DEVCONF(*devconf, PROXY_ARP)) < 0)
1740 goto nla_put_failure;
1741
1742 return nlmsg_end(skb, nlh);
1743
1744 nla_put_failure:
1745 nlmsg_cancel(skb, nlh);
1746 return -EMSGSIZE;
1747 }
1748
1749 void inet_netconf_notify_devconf(struct net *net, int type, int ifindex,
1750 struct ipv4_devconf *devconf)
1751 {
1752 struct sk_buff *skb;
1753 int err = -ENOBUFS;
1754
1755 skb = nlmsg_new(inet_netconf_msgsize_devconf(type), GFP_ATOMIC);
1756 if (skb == NULL)
1757 goto errout;
1758
1759 err = inet_netconf_fill_devconf(skb, ifindex, devconf, 0, 0,
1760 RTM_NEWNETCONF, 0, type);
1761 if (err < 0) {
1762 /* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */
1763 WARN_ON(err == -EMSGSIZE);
1764 kfree_skb(skb);
1765 goto errout;
1766 }
1767 rtnl_notify(skb, net, 0, RTNLGRP_IPV4_NETCONF, NULL, GFP_ATOMIC);
1768 return;
1769 errout:
1770 if (err < 0)
1771 rtnl_set_sk_err(net, RTNLGRP_IPV4_NETCONF, err);
1772 }
1773
1774 static const struct nla_policy devconf_ipv4_policy[NETCONFA_MAX+1] = {
1775 [NETCONFA_IFINDEX] = { .len = sizeof(int) },
1776 [NETCONFA_FORWARDING] = { .len = sizeof(int) },
1777 [NETCONFA_RP_FILTER] = { .len = sizeof(int) },
1778 [NETCONFA_PROXY_ARP] = { .len = sizeof(int) },
1779 };
1780
1781 static int inet_netconf_get_devconf(struct sk_buff *in_skb,
1782 struct nlmsghdr *nlh)
1783 {
1784 struct net *net = sock_net(in_skb->sk);
1785 struct nlattr *tb[NETCONFA_MAX+1];
1786 struct netconfmsg *ncm;
1787 struct sk_buff *skb;
1788 struct ipv4_devconf *devconf;
1789 struct in_device *in_dev;
1790 struct net_device *dev;
1791 int ifindex;
1792 int err;
1793
1794 err = nlmsg_parse(nlh, sizeof(*ncm), tb, NETCONFA_MAX,
1795 devconf_ipv4_policy);
1796 if (err < 0)
1797 goto errout;
1798
1799 err = EINVAL;
1800 if (!tb[NETCONFA_IFINDEX])
1801 goto errout;
1802
1803 ifindex = nla_get_s32(tb[NETCONFA_IFINDEX]);
1804 switch (ifindex) {
1805 case NETCONFA_IFINDEX_ALL:
1806 devconf = net->ipv4.devconf_all;
1807 break;
1808 case NETCONFA_IFINDEX_DEFAULT:
1809 devconf = net->ipv4.devconf_dflt;
1810 break;
1811 default:
1812 dev = __dev_get_by_index(net, ifindex);
1813 if (dev == NULL)
1814 goto errout;
1815 in_dev = __in_dev_get_rtnl(dev);
1816 if (in_dev == NULL)
1817 goto errout;
1818 devconf = &in_dev->cnf;
1819 break;
1820 }
1821
1822 err = -ENOBUFS;
1823 skb = nlmsg_new(inet_netconf_msgsize_devconf(-1), GFP_ATOMIC);
1824 if (skb == NULL)
1825 goto errout;
1826
1827 err = inet_netconf_fill_devconf(skb, ifindex, devconf,
1828 NETLINK_CB(in_skb).portid,
1829 nlh->nlmsg_seq, RTM_NEWNETCONF, 0,
1830 -1);
1831 if (err < 0) {
1832 /* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */
1833 WARN_ON(err == -EMSGSIZE);
1834 kfree_skb(skb);
1835 goto errout;
1836 }
1837 err = rtnl_unicast(skb, net, NETLINK_CB(in_skb).portid);
1838 errout:
1839 return err;
1840 }
1841
1842 static int inet_netconf_dump_devconf(struct sk_buff *skb,
1843 struct netlink_callback *cb)
1844 {
1845 struct net *net = sock_net(skb->sk);
1846 int h, s_h;
1847 int idx, s_idx;
1848 struct net_device *dev;
1849 struct in_device *in_dev;
1850 struct hlist_head *head;
1851
1852 s_h = cb->args[0];
1853 s_idx = idx = cb->args[1];
1854
1855 for (h = s_h; h < NETDEV_HASHENTRIES; h++, s_idx = 0) {
1856 idx = 0;
1857 head = &net->dev_index_head[h];
1858 rcu_read_lock();
1859 cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^
1860 net->dev_base_seq;
1861 hlist_for_each_entry_rcu(dev, head, index_hlist) {
1862 if (idx < s_idx)
1863 goto cont;
1864 in_dev = __in_dev_get_rcu(dev);
1865 if (!in_dev)
1866 goto cont;
1867
1868 if (inet_netconf_fill_devconf(skb, dev->ifindex,
1869 &in_dev->cnf,
1870 NETLINK_CB(cb->skb).portid,
1871 cb->nlh->nlmsg_seq,
1872 RTM_NEWNETCONF,
1873 NLM_F_MULTI,
1874 -1) <= 0) {
1875 rcu_read_unlock();
1876 goto done;
1877 }
1878 nl_dump_check_consistent(cb, nlmsg_hdr(skb));
1879 cont:
1880 idx++;
1881 }
1882 rcu_read_unlock();
1883 }
1884 if (h == NETDEV_HASHENTRIES) {
1885 if (inet_netconf_fill_devconf(skb, NETCONFA_IFINDEX_ALL,
1886 net->ipv4.devconf_all,
1887 NETLINK_CB(cb->skb).portid,
1888 cb->nlh->nlmsg_seq,
1889 RTM_NEWNETCONF, NLM_F_MULTI,
1890 -1) <= 0)
1891 goto done;
1892 else
1893 h++;
1894 }
1895 if (h == NETDEV_HASHENTRIES + 1) {
1896 if (inet_netconf_fill_devconf(skb, NETCONFA_IFINDEX_DEFAULT,
1897 net->ipv4.devconf_dflt,
1898 NETLINK_CB(cb->skb).portid,
1899 cb->nlh->nlmsg_seq,
1900 RTM_NEWNETCONF, NLM_F_MULTI,
1901 -1) <= 0)
1902 goto done;
1903 else
1904 h++;
1905 }
1906 done:
1907 cb->args[0] = h;
1908 cb->args[1] = idx;
1909
1910 return skb->len;
1911 }
1912
1913 #ifdef CONFIG_SYSCTL
1914
1915 static void devinet_copy_dflt_conf(struct net *net, int i)
1916 {
1917 struct net_device *dev;
1918
1919 rcu_read_lock();
1920 for_each_netdev_rcu(net, dev) {
1921 struct in_device *in_dev;
1922
1923 in_dev = __in_dev_get_rcu(dev);
1924 if (in_dev && !test_bit(i, in_dev->cnf.state))
1925 in_dev->cnf.data[i] = net->ipv4.devconf_dflt->data[i];
1926 }
1927 rcu_read_unlock();
1928 }
1929
1930 /* called with RTNL locked */
1931 static void inet_forward_change(struct net *net)
1932 {
1933 struct net_device *dev;
1934 int on = IPV4_DEVCONF_ALL(net, FORWARDING);
1935
1936 IPV4_DEVCONF_ALL(net, ACCEPT_REDIRECTS) = !on;
1937 IPV4_DEVCONF_DFLT(net, FORWARDING) = on;
1938 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
1939 NETCONFA_IFINDEX_ALL,
1940 net->ipv4.devconf_all);
1941 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
1942 NETCONFA_IFINDEX_DEFAULT,
1943 net->ipv4.devconf_dflt);
1944
1945 for_each_netdev(net, dev) {
1946 struct in_device *in_dev;
1947 if (on)
1948 dev_disable_lro(dev);
1949 rcu_read_lock();
1950 in_dev = __in_dev_get_rcu(dev);
1951 if (in_dev) {
1952 IN_DEV_CONF_SET(in_dev, FORWARDING, on);
1953 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
1954 dev->ifindex, &in_dev->cnf);
1955 }
1956 rcu_read_unlock();
1957 }
1958 }
1959
1960 static int devinet_conf_ifindex(struct net *net, struct ipv4_devconf *cnf)
1961 {
1962 if (cnf == net->ipv4.devconf_dflt)
1963 return NETCONFA_IFINDEX_DEFAULT;
1964 else if (cnf == net->ipv4.devconf_all)
1965 return NETCONFA_IFINDEX_ALL;
1966 else {
1967 struct in_device *idev
1968 = container_of(cnf, struct in_device, cnf);
1969 return idev->dev->ifindex;
1970 }
1971 }
1972
1973 static int devinet_conf_proc(struct ctl_table *ctl, int write,
1974 void __user *buffer,
1975 size_t *lenp, loff_t *ppos)
1976 {
1977 int old_value = *(int *)ctl->data;
1978 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
1979 int new_value = *(int *)ctl->data;
1980
1981 if (write) {
1982 struct ipv4_devconf *cnf = ctl->extra1;
1983 struct net *net = ctl->extra2;
1984 int i = (int *)ctl->data - cnf->data;
1985 int ifindex;
1986
1987 set_bit(i, cnf->state);
1988
1989 if (cnf == net->ipv4.devconf_dflt)
1990 devinet_copy_dflt_conf(net, i);
1991 if (i == IPV4_DEVCONF_ACCEPT_LOCAL - 1 ||
1992 i == IPV4_DEVCONF_ROUTE_LOCALNET - 1)
1993 if ((new_value == 0) && (old_value != 0))
1994 rt_cache_flush(net);
1995
1996 if (i == IPV4_DEVCONF_RP_FILTER - 1 &&
1997 new_value != old_value) {
1998 ifindex = devinet_conf_ifindex(net, cnf);
1999 inet_netconf_notify_devconf(net, NETCONFA_RP_FILTER,
2000 ifindex, cnf);
2001 }
2002 if (i == IPV4_DEVCONF_PROXY_ARP - 1 &&
2003 new_value != old_value) {
2004 ifindex = devinet_conf_ifindex(net, cnf);
2005 inet_netconf_notify_devconf(net, NETCONFA_PROXY_ARP,
2006 ifindex, cnf);
2007 }
2008 }
2009
2010 return ret;
2011 }
2012
2013 static int devinet_sysctl_forward(struct ctl_table *ctl, int write,
2014 void __user *buffer,
2015 size_t *lenp, loff_t *ppos)
2016 {
2017 int *valp = ctl->data;
2018 int val = *valp;
2019 loff_t pos = *ppos;
2020 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2021
2022 if (write && *valp != val) {
2023 struct net *net = ctl->extra2;
2024
2025 if (valp != &IPV4_DEVCONF_DFLT(net, FORWARDING)) {
2026 if (!rtnl_trylock()) {
2027 /* Restore the original values before restarting */
2028 *valp = val;
2029 *ppos = pos;
2030 return restart_syscall();
2031 }
2032 if (valp == &IPV4_DEVCONF_ALL(net, FORWARDING)) {
2033 inet_forward_change(net);
2034 } else {
2035 struct ipv4_devconf *cnf = ctl->extra1;
2036 struct in_device *idev =
2037 container_of(cnf, struct in_device, cnf);
2038 if (*valp)
2039 dev_disable_lro(idev->dev);
2040 inet_netconf_notify_devconf(net,
2041 NETCONFA_FORWARDING,
2042 idev->dev->ifindex,
2043 cnf);
2044 }
2045 rtnl_unlock();
2046 rt_cache_flush(net);
2047 } else
2048 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING,
2049 NETCONFA_IFINDEX_DEFAULT,
2050 net->ipv4.devconf_dflt);
2051 }
2052
2053 return ret;
2054 }
2055
2056 static int ipv4_doint_and_flush(struct ctl_table *ctl, int write,
2057 void __user *buffer,
2058 size_t *lenp, loff_t *ppos)
2059 {
2060 int *valp = ctl->data;
2061 int val = *valp;
2062 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2063 struct net *net = ctl->extra2;
2064
2065 if (write && *valp != val)
2066 rt_cache_flush(net);
2067
2068 return ret;
2069 }
2070
2071 #define DEVINET_SYSCTL_ENTRY(attr, name, mval, proc) \
2072 { \
2073 .procname = name, \
2074 .data = ipv4_devconf.data + \
2075 IPV4_DEVCONF_ ## attr - 1, \
2076 .maxlen = sizeof(int), \
2077 .mode = mval, \
2078 .proc_handler = proc, \
2079 .extra1 = &ipv4_devconf, \
2080 }
2081
2082 #define DEVINET_SYSCTL_RW_ENTRY(attr, name) \
2083 DEVINET_SYSCTL_ENTRY(attr, name, 0644, devinet_conf_proc)
2084
2085 #define DEVINET_SYSCTL_RO_ENTRY(attr, name) \
2086 DEVINET_SYSCTL_ENTRY(attr, name, 0444, devinet_conf_proc)
2087
2088 #define DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, proc) \
2089 DEVINET_SYSCTL_ENTRY(attr, name, 0644, proc)
2090
2091 #define DEVINET_SYSCTL_FLUSHING_ENTRY(attr, name) \
2092 DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, ipv4_doint_and_flush)
2093
2094 static struct devinet_sysctl_table {
2095 struct ctl_table_header *sysctl_header;
2096 struct ctl_table devinet_vars[__IPV4_DEVCONF_MAX];
2097 } devinet_sysctl = {
2098 .devinet_vars = {
2099 DEVINET_SYSCTL_COMPLEX_ENTRY(FORWARDING, "forwarding",
2100 devinet_sysctl_forward),
2101 DEVINET_SYSCTL_RO_ENTRY(MC_FORWARDING, "mc_forwarding"),
2102
2103 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_REDIRECTS, "accept_redirects"),
2104 DEVINET_SYSCTL_RW_ENTRY(SECURE_REDIRECTS, "secure_redirects"),
2105 DEVINET_SYSCTL_RW_ENTRY(SHARED_MEDIA, "shared_media"),
2106 DEVINET_SYSCTL_RW_ENTRY(RP_FILTER, "rp_filter"),
2107 DEVINET_SYSCTL_RW_ENTRY(SEND_REDIRECTS, "send_redirects"),
2108 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_SOURCE_ROUTE,
2109 "accept_source_route"),
2110 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_LOCAL, "accept_local"),
2111 DEVINET_SYSCTL_RW_ENTRY(SRC_VMARK, "src_valid_mark"),
2112 DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP, "proxy_arp"),
2113 DEVINET_SYSCTL_RW_ENTRY(MEDIUM_ID, "medium_id"),
2114 DEVINET_SYSCTL_RW_ENTRY(BOOTP_RELAY, "bootp_relay"),
2115 DEVINET_SYSCTL_RW_ENTRY(LOG_MARTIANS, "log_martians"),
2116 DEVINET_SYSCTL_RW_ENTRY(TAG, "tag"),
2117 DEVINET_SYSCTL_RW_ENTRY(ARPFILTER, "arp_filter"),
2118 DEVINET_SYSCTL_RW_ENTRY(ARP_ANNOUNCE, "arp_announce"),
2119 DEVINET_SYSCTL_RW_ENTRY(ARP_IGNORE, "arp_ignore"),
2120 DEVINET_SYSCTL_RW_ENTRY(ARP_ACCEPT, "arp_accept"),
2121 DEVINET_SYSCTL_RW_ENTRY(ARP_NOTIFY, "arp_notify"),
2122 DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP_PVLAN, "proxy_arp_pvlan"),
2123 DEVINET_SYSCTL_RW_ENTRY(FORCE_IGMP_VERSION,
2124 "force_igmp_version"),
2125 DEVINET_SYSCTL_RW_ENTRY(IGMPV2_UNSOLICITED_REPORT_INTERVAL,
2126 "igmpv2_unsolicited_report_interval"),
2127 DEVINET_SYSCTL_RW_ENTRY(IGMPV3_UNSOLICITED_REPORT_INTERVAL,
2128 "igmpv3_unsolicited_report_interval"),
2129
2130 DEVINET_SYSCTL_FLUSHING_ENTRY(NOXFRM, "disable_xfrm"),
2131 DEVINET_SYSCTL_FLUSHING_ENTRY(NOPOLICY, "disable_policy"),
2132 DEVINET_SYSCTL_FLUSHING_ENTRY(PROMOTE_SECONDARIES,
2133 "promote_secondaries"),
2134 DEVINET_SYSCTL_FLUSHING_ENTRY(ROUTE_LOCALNET,
2135 "route_localnet"),
2136 },
2137 };
2138
2139 static int __devinet_sysctl_register(struct net *net, char *dev_name,
2140 struct ipv4_devconf *p)
2141 {
2142 int i;
2143 struct devinet_sysctl_table *t;
2144 char path[sizeof("net/ipv4/conf/") + IFNAMSIZ];
2145
2146 t = kmemdup(&devinet_sysctl, sizeof(*t), GFP_KERNEL);
2147 if (!t)
2148 goto out;
2149
2150 for (i = 0; i < ARRAY_SIZE(t->devinet_vars) - 1; i++) {
2151 t->devinet_vars[i].data += (char *)p - (char *)&ipv4_devconf;
2152 t->devinet_vars[i].extra1 = p;
2153 t->devinet_vars[i].extra2 = net;
2154 }
2155
2156 snprintf(path, sizeof(path), "net/ipv4/conf/%s", dev_name);
2157
2158 t->sysctl_header = register_net_sysctl(net, path, t->devinet_vars);
2159 if (!t->sysctl_header)
2160 goto free;
2161
2162 p->sysctl = t;
2163 return 0;
2164
2165 free:
2166 kfree(t);
2167 out:
2168 return -ENOBUFS;
2169 }
2170
2171 static void __devinet_sysctl_unregister(struct ipv4_devconf *cnf)
2172 {
2173 struct devinet_sysctl_table *t = cnf->sysctl;
2174
2175 if (t == NULL)
2176 return;
2177
2178 cnf->sysctl = NULL;
2179 unregister_net_sysctl_table(t->sysctl_header);
2180 kfree(t);
2181 }
2182
2183 static void devinet_sysctl_register(struct in_device *idev)
2184 {
2185 neigh_sysctl_register(idev->dev, idev->arp_parms, NULL);
2186 __devinet_sysctl_register(dev_net(idev->dev), idev->dev->name,
2187 &idev->cnf);
2188 }
2189
2190 static void devinet_sysctl_unregister(struct in_device *idev)
2191 {
2192 __devinet_sysctl_unregister(&idev->cnf);
2193 neigh_sysctl_unregister(idev->arp_parms);
2194 }
2195
2196 static struct ctl_table ctl_forward_entry[] = {
2197 {
2198 .procname = "ip_forward",
2199 .data = &ipv4_devconf.data[
2200 IPV4_DEVCONF_FORWARDING - 1],
2201 .maxlen = sizeof(int),
2202 .mode = 0644,
2203 .proc_handler = devinet_sysctl_forward,
2204 .extra1 = &ipv4_devconf,
2205 .extra2 = &init_net,
2206 },
2207 { },
2208 };
2209 #endif
2210
2211 static __net_init int devinet_init_net(struct net *net)
2212 {
2213 int err;
2214 struct ipv4_devconf *all, *dflt;
2215 #ifdef CONFIG_SYSCTL
2216 struct ctl_table *tbl = ctl_forward_entry;
2217 struct ctl_table_header *forw_hdr;
2218 #endif
2219
2220 err = -ENOMEM;
2221 all = &ipv4_devconf;
2222 dflt = &ipv4_devconf_dflt;
2223
2224 if (!net_eq(net, &init_net)) {
2225 all = kmemdup(all, sizeof(ipv4_devconf), GFP_KERNEL);
2226 if (all == NULL)
2227 goto err_alloc_all;
2228
2229 dflt = kmemdup(dflt, sizeof(ipv4_devconf_dflt), GFP_KERNEL);
2230 if (dflt == NULL)
2231 goto err_alloc_dflt;
2232
2233 #ifdef CONFIG_SYSCTL
2234 tbl = kmemdup(tbl, sizeof(ctl_forward_entry), GFP_KERNEL);
2235 if (tbl == NULL)
2236 goto err_alloc_ctl;
2237
2238 tbl[0].data = &all->data[IPV4_DEVCONF_FORWARDING - 1];
2239 tbl[0].extra1 = all;
2240 tbl[0].extra2 = net;
2241 #endif
2242 }
2243
2244 #ifdef CONFIG_SYSCTL
2245 err = __devinet_sysctl_register(net, "all", all);
2246 if (err < 0)
2247 goto err_reg_all;
2248
2249 err = __devinet_sysctl_register(net, "default", dflt);
2250 if (err < 0)
2251 goto err_reg_dflt;
2252
2253 err = -ENOMEM;
2254 forw_hdr = register_net_sysctl(net, "net/ipv4", tbl);
2255 if (forw_hdr == NULL)
2256 goto err_reg_ctl;
2257 net->ipv4.forw_hdr = forw_hdr;
2258 #endif
2259
2260 net->ipv4.devconf_all = all;
2261 net->ipv4.devconf_dflt = dflt;
2262 return 0;
2263
2264 #ifdef CONFIG_SYSCTL
2265 err_reg_ctl:
2266 __devinet_sysctl_unregister(dflt);
2267 err_reg_dflt:
2268 __devinet_sysctl_unregister(all);
2269 err_reg_all:
2270 if (tbl != ctl_forward_entry)
2271 kfree(tbl);
2272 err_alloc_ctl:
2273 #endif
2274 if (dflt != &ipv4_devconf_dflt)
2275 kfree(dflt);
2276 err_alloc_dflt:
2277 if (all != &ipv4_devconf)
2278 kfree(all);
2279 err_alloc_all:
2280 return err;
2281 }
2282
2283 static __net_exit void devinet_exit_net(struct net *net)
2284 {
2285 #ifdef CONFIG_SYSCTL
2286 struct ctl_table *tbl;
2287
2288 tbl = net->ipv4.forw_hdr->ctl_table_arg;
2289 unregister_net_sysctl_table(net->ipv4.forw_hdr);
2290 __devinet_sysctl_unregister(net->ipv4.devconf_dflt);
2291 __devinet_sysctl_unregister(net->ipv4.devconf_all);
2292 kfree(tbl);
2293 #endif
2294 kfree(net->ipv4.devconf_dflt);
2295 kfree(net->ipv4.devconf_all);
2296 }
2297
2298 static __net_initdata struct pernet_operations devinet_ops = {
2299 .init = devinet_init_net,
2300 .exit = devinet_exit_net,
2301 };
2302
2303 static struct rtnl_af_ops inet_af_ops = {
2304 .family = AF_INET,
2305 .fill_link_af = inet_fill_link_af,
2306 .get_link_af_size = inet_get_link_af_size,
2307 .validate_link_af = inet_validate_link_af,
2308 .set_link_af = inet_set_link_af,
2309 };
2310
2311 void __init devinet_init(void)
2312 {
2313 int i;
2314
2315 for (i = 0; i < IN4_ADDR_HSIZE; i++)
2316 INIT_HLIST_HEAD(&inet_addr_lst[i]);
2317
2318 register_pernet_subsys(&devinet_ops);
2319
2320 register_gifconf(PF_INET, inet_gifconf);
2321 register_netdevice_notifier(&ip_netdev_notifier);
2322
2323 schedule_delayed_work(&check_lifetime_work, 0);
2324
2325 rtnl_af_register(&inet_af_ops);
2326
2327 rtnl_register(PF_INET, RTM_NEWADDR, inet_rtm_newaddr, NULL, NULL);
2328 rtnl_register(PF_INET, RTM_DELADDR, inet_rtm_deladdr, NULL, NULL);
2329 rtnl_register(PF_INET, RTM_GETADDR, NULL, inet_dump_ifaddr, NULL);
2330 rtnl_register(PF_INET, RTM_GETNETCONF, inet_netconf_get_devconf,
2331 inet_netconf_dump_devconf, NULL);
2332 }
2333
This page took 0.123431 seconds and 4 git commands to generate.