ip_gre: make ipgre_tunnel_xmit() not parse network header as IP unconditionally
[deliverable/linux.git] / net / ipv4 / ip_gre.c
1 /*
2 * Linux NET3: GRE over IP protocol decoder.
3 *
4 * Authors: Alexey Kuznetsov (kuznet@ms2.inr.ac.ru)
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 *
11 */
12
13 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
14
15 #include <linux/capability.h>
16 #include <linux/module.h>
17 #include <linux/types.h>
18 #include <linux/kernel.h>
19 #include <linux/slab.h>
20 #include <asm/uaccess.h>
21 #include <linux/skbuff.h>
22 #include <linux/netdevice.h>
23 #include <linux/in.h>
24 #include <linux/tcp.h>
25 #include <linux/udp.h>
26 #include <linux/if_arp.h>
27 #include <linux/mroute.h>
28 #include <linux/init.h>
29 #include <linux/in6.h>
30 #include <linux/inetdevice.h>
31 #include <linux/igmp.h>
32 #include <linux/netfilter_ipv4.h>
33 #include <linux/etherdevice.h>
34 #include <linux/if_ether.h>
35
36 #include <net/sock.h>
37 #include <net/ip.h>
38 #include <net/icmp.h>
39 #include <net/protocol.h>
40 #include <net/ipip.h>
41 #include <net/arp.h>
42 #include <net/checksum.h>
43 #include <net/dsfield.h>
44 #include <net/inet_ecn.h>
45 #include <net/xfrm.h>
46 #include <net/net_namespace.h>
47 #include <net/netns/generic.h>
48 #include <net/rtnetlink.h>
49 #include <net/gre.h>
50
51 #if IS_ENABLED(CONFIG_IPV6)
52 #include <net/ipv6.h>
53 #include <net/ip6_fib.h>
54 #include <net/ip6_route.h>
55 #endif
56
57 /*
58 Problems & solutions
59 --------------------
60
61 1. The most important issue is detecting local dead loops.
62 They would cause complete host lockup in transmit, which
63 would be "resolved" by stack overflow or, if queueing is enabled,
64 with infinite looping in net_bh.
65
66 We cannot track such dead loops during route installation,
67 it is infeasible task. The most general solutions would be
68 to keep skb->encapsulation counter (sort of local ttl),
69 and silently drop packet when it expires. It is a good
70 solution, but it supposes maintaining new variable in ALL
71 skb, even if no tunneling is used.
72
73 Current solution: xmit_recursion breaks dead loops. This is a percpu
74 counter, since when we enter the first ndo_xmit(), cpu migration is
75 forbidden. We force an exit if this counter reaches RECURSION_LIMIT
76
77 2. Networking dead loops would not kill routers, but would really
78 kill network. IP hop limit plays role of "t->recursion" in this case,
79 if we copy it from packet being encapsulated to upper header.
80 It is very good solution, but it introduces two problems:
81
82 - Routing protocols, using packets with ttl=1 (OSPF, RIP2),
83 do not work over tunnels.
84 - traceroute does not work. I planned to relay ICMP from tunnel,
85 so that this problem would be solved and traceroute output
86 would even more informative. This idea appeared to be wrong:
87 only Linux complies to rfc1812 now (yes, guys, Linux is the only
88 true router now :-)), all routers (at least, in neighbourhood of mine)
89 return only 8 bytes of payload. It is the end.
90
91 Hence, if we want that OSPF worked or traceroute said something reasonable,
92 we should search for another solution.
93
94 One of them is to parse packet trying to detect inner encapsulation
95 made by our node. It is difficult or even impossible, especially,
96 taking into account fragmentation. TO be short, ttl is not solution at all.
97
98 Current solution: The solution was UNEXPECTEDLY SIMPLE.
99 We force DF flag on tunnels with preconfigured hop limit,
100 that is ALL. :-) Well, it does not remove the problem completely,
101 but exponential growth of network traffic is changed to linear
102 (branches, that exceed pmtu are pruned) and tunnel mtu
103 rapidly degrades to value <68, where looping stops.
104 Yes, it is not good if there exists a router in the loop,
105 which does not force DF, even when encapsulating packets have DF set.
106 But it is not our problem! Nobody could accuse us, we made
107 all that we could make. Even if it is your gated who injected
108 fatal route to network, even if it were you who configured
109 fatal static route: you are innocent. :-)
110
111
112
113 3. Really, ipv4/ipip.c, ipv4/ip_gre.c and ipv6/sit.c contain
114 practically identical code. It would be good to glue them
115 together, but it is not very evident, how to make them modular.
116 sit is integral part of IPv6, ipip and gre are naturally modular.
117 We could extract common parts (hash table, ioctl etc)
118 to a separate module (ip_tunnel.c).
119
120 Alexey Kuznetsov.
121 */
122
123 static bool log_ecn_error = true;
124 module_param(log_ecn_error, bool, 0644);
125 MODULE_PARM_DESC(log_ecn_error, "Log packets received with corrupted ECN");
126
127 static struct rtnl_link_ops ipgre_link_ops __read_mostly;
128 static int ipgre_tunnel_init(struct net_device *dev);
129 static void ipgre_tunnel_setup(struct net_device *dev);
130 static int ipgre_tunnel_bind_dev(struct net_device *dev);
131
132 /* Fallback tunnel: no source, no destination, no key, no options */
133
134 #define HASH_SIZE 16
135
136 static int ipgre_net_id __read_mostly;
137 struct ipgre_net {
138 struct ip_tunnel __rcu *tunnels[4][HASH_SIZE];
139
140 struct net_device *fb_tunnel_dev;
141 };
142
143 /* Tunnel hash table */
144
145 /*
146 4 hash tables:
147
148 3: (remote,local)
149 2: (remote,*)
150 1: (*,local)
151 0: (*,*)
152
153 We require exact key match i.e. if a key is present in packet
154 it will match only tunnel with the same key; if it is not present,
155 it will match only keyless tunnel.
156
157 All keysless packets, if not matched configured keyless tunnels
158 will match fallback tunnel.
159 */
160
161 #define HASH(addr) (((__force u32)addr^((__force u32)addr>>4))&0xF)
162
163 #define tunnels_r_l tunnels[3]
164 #define tunnels_r tunnels[2]
165 #define tunnels_l tunnels[1]
166 #define tunnels_wc tunnels[0]
167
168 static struct rtnl_link_stats64 *ipgre_get_stats64(struct net_device *dev,
169 struct rtnl_link_stats64 *tot)
170 {
171 int i;
172
173 for_each_possible_cpu(i) {
174 const struct pcpu_tstats *tstats = per_cpu_ptr(dev->tstats, i);
175 u64 rx_packets, rx_bytes, tx_packets, tx_bytes;
176 unsigned int start;
177
178 do {
179 start = u64_stats_fetch_begin_bh(&tstats->syncp);
180 rx_packets = tstats->rx_packets;
181 tx_packets = tstats->tx_packets;
182 rx_bytes = tstats->rx_bytes;
183 tx_bytes = tstats->tx_bytes;
184 } while (u64_stats_fetch_retry_bh(&tstats->syncp, start));
185
186 tot->rx_packets += rx_packets;
187 tot->tx_packets += tx_packets;
188 tot->rx_bytes += rx_bytes;
189 tot->tx_bytes += tx_bytes;
190 }
191
192 tot->multicast = dev->stats.multicast;
193 tot->rx_crc_errors = dev->stats.rx_crc_errors;
194 tot->rx_fifo_errors = dev->stats.rx_fifo_errors;
195 tot->rx_length_errors = dev->stats.rx_length_errors;
196 tot->rx_frame_errors = dev->stats.rx_frame_errors;
197 tot->rx_errors = dev->stats.rx_errors;
198
199 tot->tx_fifo_errors = dev->stats.tx_fifo_errors;
200 tot->tx_carrier_errors = dev->stats.tx_carrier_errors;
201 tot->tx_dropped = dev->stats.tx_dropped;
202 tot->tx_aborted_errors = dev->stats.tx_aborted_errors;
203 tot->tx_errors = dev->stats.tx_errors;
204
205 return tot;
206 }
207
208 /* Does key in tunnel parameters match packet */
209 static bool ipgre_key_match(const struct ip_tunnel_parm *p,
210 __be16 flags, __be32 key)
211 {
212 if (p->i_flags & GRE_KEY) {
213 if (flags & GRE_KEY)
214 return key == p->i_key;
215 else
216 return false; /* key expected, none present */
217 } else
218 return !(flags & GRE_KEY);
219 }
220
221 /* Given src, dst and key, find appropriate for input tunnel. */
222
223 static struct ip_tunnel *ipgre_tunnel_lookup(struct net_device *dev,
224 __be32 remote, __be32 local,
225 __be16 flags, __be32 key,
226 __be16 gre_proto)
227 {
228 struct net *net = dev_net(dev);
229 int link = dev->ifindex;
230 unsigned int h0 = HASH(remote);
231 unsigned int h1 = HASH(key);
232 struct ip_tunnel *t, *cand = NULL;
233 struct ipgre_net *ign = net_generic(net, ipgre_net_id);
234 int dev_type = (gre_proto == htons(ETH_P_TEB)) ?
235 ARPHRD_ETHER : ARPHRD_IPGRE;
236 int score, cand_score = 4;
237
238 for_each_ip_tunnel_rcu(t, ign->tunnels_r_l[h0 ^ h1]) {
239 if (local != t->parms.iph.saddr ||
240 remote != t->parms.iph.daddr ||
241 !(t->dev->flags & IFF_UP))
242 continue;
243
244 if (!ipgre_key_match(&t->parms, flags, key))
245 continue;
246
247 if (t->dev->type != ARPHRD_IPGRE &&
248 t->dev->type != dev_type)
249 continue;
250
251 score = 0;
252 if (t->parms.link != link)
253 score |= 1;
254 if (t->dev->type != dev_type)
255 score |= 2;
256 if (score == 0)
257 return t;
258
259 if (score < cand_score) {
260 cand = t;
261 cand_score = score;
262 }
263 }
264
265 for_each_ip_tunnel_rcu(t, ign->tunnels_r[h0 ^ h1]) {
266 if (remote != t->parms.iph.daddr ||
267 !(t->dev->flags & IFF_UP))
268 continue;
269
270 if (!ipgre_key_match(&t->parms, flags, key))
271 continue;
272
273 if (t->dev->type != ARPHRD_IPGRE &&
274 t->dev->type != dev_type)
275 continue;
276
277 score = 0;
278 if (t->parms.link != link)
279 score |= 1;
280 if (t->dev->type != dev_type)
281 score |= 2;
282 if (score == 0)
283 return t;
284
285 if (score < cand_score) {
286 cand = t;
287 cand_score = score;
288 }
289 }
290
291 for_each_ip_tunnel_rcu(t, ign->tunnels_l[h1]) {
292 if ((local != t->parms.iph.saddr &&
293 (local != t->parms.iph.daddr ||
294 !ipv4_is_multicast(local))) ||
295 !(t->dev->flags & IFF_UP))
296 continue;
297
298 if (!ipgre_key_match(&t->parms, flags, key))
299 continue;
300
301 if (t->dev->type != ARPHRD_IPGRE &&
302 t->dev->type != dev_type)
303 continue;
304
305 score = 0;
306 if (t->parms.link != link)
307 score |= 1;
308 if (t->dev->type != dev_type)
309 score |= 2;
310 if (score == 0)
311 return t;
312
313 if (score < cand_score) {
314 cand = t;
315 cand_score = score;
316 }
317 }
318
319 for_each_ip_tunnel_rcu(t, ign->tunnels_wc[h1]) {
320 if (t->parms.i_key != key ||
321 !(t->dev->flags & IFF_UP))
322 continue;
323
324 if (t->dev->type != ARPHRD_IPGRE &&
325 t->dev->type != dev_type)
326 continue;
327
328 score = 0;
329 if (t->parms.link != link)
330 score |= 1;
331 if (t->dev->type != dev_type)
332 score |= 2;
333 if (score == 0)
334 return t;
335
336 if (score < cand_score) {
337 cand = t;
338 cand_score = score;
339 }
340 }
341
342 if (cand != NULL)
343 return cand;
344
345 dev = ign->fb_tunnel_dev;
346 if (dev->flags & IFF_UP)
347 return netdev_priv(dev);
348
349 return NULL;
350 }
351
352 static struct ip_tunnel __rcu **__ipgre_bucket(struct ipgre_net *ign,
353 struct ip_tunnel_parm *parms)
354 {
355 __be32 remote = parms->iph.daddr;
356 __be32 local = parms->iph.saddr;
357 __be32 key = parms->i_key;
358 unsigned int h = HASH(key);
359 int prio = 0;
360
361 if (local)
362 prio |= 1;
363 if (remote && !ipv4_is_multicast(remote)) {
364 prio |= 2;
365 h ^= HASH(remote);
366 }
367
368 return &ign->tunnels[prio][h];
369 }
370
371 static inline struct ip_tunnel __rcu **ipgre_bucket(struct ipgre_net *ign,
372 struct ip_tunnel *t)
373 {
374 return __ipgre_bucket(ign, &t->parms);
375 }
376
377 static void ipgre_tunnel_link(struct ipgre_net *ign, struct ip_tunnel *t)
378 {
379 struct ip_tunnel __rcu **tp = ipgre_bucket(ign, t);
380
381 rcu_assign_pointer(t->next, rtnl_dereference(*tp));
382 rcu_assign_pointer(*tp, t);
383 }
384
385 static void ipgre_tunnel_unlink(struct ipgre_net *ign, struct ip_tunnel *t)
386 {
387 struct ip_tunnel __rcu **tp;
388 struct ip_tunnel *iter;
389
390 for (tp = ipgre_bucket(ign, t);
391 (iter = rtnl_dereference(*tp)) != NULL;
392 tp = &iter->next) {
393 if (t == iter) {
394 rcu_assign_pointer(*tp, t->next);
395 break;
396 }
397 }
398 }
399
400 static struct ip_tunnel *ipgre_tunnel_find(struct net *net,
401 struct ip_tunnel_parm *parms,
402 int type)
403 {
404 __be32 remote = parms->iph.daddr;
405 __be32 local = parms->iph.saddr;
406 __be32 key = parms->i_key;
407 int link = parms->link;
408 struct ip_tunnel *t;
409 struct ip_tunnel __rcu **tp;
410 struct ipgre_net *ign = net_generic(net, ipgre_net_id);
411
412 for (tp = __ipgre_bucket(ign, parms);
413 (t = rtnl_dereference(*tp)) != NULL;
414 tp = &t->next)
415 if (local == t->parms.iph.saddr &&
416 remote == t->parms.iph.daddr &&
417 key == t->parms.i_key &&
418 link == t->parms.link &&
419 type == t->dev->type)
420 break;
421
422 return t;
423 }
424
425 static struct ip_tunnel *ipgre_tunnel_locate(struct net *net,
426 struct ip_tunnel_parm *parms, int create)
427 {
428 struct ip_tunnel *t, *nt;
429 struct net_device *dev;
430 char name[IFNAMSIZ];
431 struct ipgre_net *ign = net_generic(net, ipgre_net_id);
432
433 t = ipgre_tunnel_find(net, parms, ARPHRD_IPGRE);
434 if (t || !create)
435 return t;
436
437 if (parms->name[0])
438 strlcpy(name, parms->name, IFNAMSIZ);
439 else
440 strcpy(name, "gre%d");
441
442 dev = alloc_netdev(sizeof(*t), name, ipgre_tunnel_setup);
443 if (!dev)
444 return NULL;
445
446 dev_net_set(dev, net);
447
448 nt = netdev_priv(dev);
449 nt->parms = *parms;
450 dev->rtnl_link_ops = &ipgre_link_ops;
451
452 dev->mtu = ipgre_tunnel_bind_dev(dev);
453
454 if (register_netdevice(dev) < 0)
455 goto failed_free;
456
457 /* Can use a lockless transmit, unless we generate output sequences */
458 if (!(nt->parms.o_flags & GRE_SEQ))
459 dev->features |= NETIF_F_LLTX;
460
461 dev_hold(dev);
462 ipgre_tunnel_link(ign, nt);
463 return nt;
464
465 failed_free:
466 free_netdev(dev);
467 return NULL;
468 }
469
470 static void ipgre_tunnel_uninit(struct net_device *dev)
471 {
472 struct net *net = dev_net(dev);
473 struct ipgre_net *ign = net_generic(net, ipgre_net_id);
474
475 ipgre_tunnel_unlink(ign, netdev_priv(dev));
476 dev_put(dev);
477 }
478
479
480 static void ipgre_err(struct sk_buff *skb, u32 info)
481 {
482
483 /* All the routers (except for Linux) return only
484 8 bytes of packet payload. It means, that precise relaying of
485 ICMP in the real Internet is absolutely infeasible.
486
487 Moreover, Cisco "wise men" put GRE key to the third word
488 in GRE header. It makes impossible maintaining even soft state for keyed
489 GRE tunnels with enabled checksum. Tell them "thank you".
490
491 Well, I wonder, rfc1812 was written by Cisco employee,
492 what the hell these idiots break standards established
493 by themselves???
494 */
495
496 const struct iphdr *iph = (const struct iphdr *)skb->data;
497 __be16 *p = (__be16 *)(skb->data+(iph->ihl<<2));
498 int grehlen = (iph->ihl<<2) + 4;
499 const int type = icmp_hdr(skb)->type;
500 const int code = icmp_hdr(skb)->code;
501 struct ip_tunnel *t;
502 __be16 flags;
503 __be32 key = 0;
504
505 flags = p[0];
506 if (flags&(GRE_CSUM|GRE_KEY|GRE_SEQ|GRE_ROUTING|GRE_VERSION)) {
507 if (flags&(GRE_VERSION|GRE_ROUTING))
508 return;
509 if (flags&GRE_KEY) {
510 grehlen += 4;
511 if (flags&GRE_CSUM)
512 grehlen += 4;
513 }
514 }
515
516 /* If only 8 bytes returned, keyed message will be dropped here */
517 if (skb_headlen(skb) < grehlen)
518 return;
519
520 if (flags & GRE_KEY)
521 key = *(((__be32 *)p) + (grehlen / 4) - 1);
522
523 switch (type) {
524 default:
525 case ICMP_PARAMETERPROB:
526 return;
527
528 case ICMP_DEST_UNREACH:
529 switch (code) {
530 case ICMP_SR_FAILED:
531 case ICMP_PORT_UNREACH:
532 /* Impossible event. */
533 return;
534 default:
535 /* All others are translated to HOST_UNREACH.
536 rfc2003 contains "deep thoughts" about NET_UNREACH,
537 I believe they are just ether pollution. --ANK
538 */
539 break;
540 }
541 break;
542 case ICMP_TIME_EXCEEDED:
543 if (code != ICMP_EXC_TTL)
544 return;
545 break;
546
547 case ICMP_REDIRECT:
548 break;
549 }
550
551 t = ipgre_tunnel_lookup(skb->dev, iph->daddr, iph->saddr,
552 flags, key, p[1]);
553
554 if (t == NULL)
555 return;
556
557 if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) {
558 ipv4_update_pmtu(skb, dev_net(skb->dev), info,
559 t->parms.link, 0, IPPROTO_GRE, 0);
560 return;
561 }
562 if (type == ICMP_REDIRECT) {
563 ipv4_redirect(skb, dev_net(skb->dev), t->parms.link, 0,
564 IPPROTO_GRE, 0);
565 return;
566 }
567 if (t->parms.iph.daddr == 0 ||
568 ipv4_is_multicast(t->parms.iph.daddr))
569 return;
570
571 if (t->parms.iph.ttl == 0 && type == ICMP_TIME_EXCEEDED)
572 return;
573
574 if (time_before(jiffies, t->err_time + IPTUNNEL_ERR_TIMEO))
575 t->err_count++;
576 else
577 t->err_count = 1;
578 t->err_time = jiffies;
579 }
580
581 static inline u8
582 ipgre_ecn_encapsulate(u8 tos, const struct iphdr *old_iph, struct sk_buff *skb)
583 {
584 u8 inner = 0;
585 if (skb->protocol == htons(ETH_P_IP))
586 inner = old_iph->tos;
587 else if (skb->protocol == htons(ETH_P_IPV6))
588 inner = ipv6_get_dsfield((const struct ipv6hdr *)old_iph);
589 return INET_ECN_encapsulate(tos, inner);
590 }
591
592 static int ipgre_rcv(struct sk_buff *skb)
593 {
594 const struct iphdr *iph;
595 u8 *h;
596 __be16 flags;
597 __sum16 csum = 0;
598 __be32 key = 0;
599 u32 seqno = 0;
600 struct ip_tunnel *tunnel;
601 int offset = 4;
602 __be16 gre_proto;
603 int err;
604
605 if (!pskb_may_pull(skb, 16))
606 goto drop;
607
608 iph = ip_hdr(skb);
609 h = skb->data;
610 flags = *(__be16 *)h;
611
612 if (flags&(GRE_CSUM|GRE_KEY|GRE_ROUTING|GRE_SEQ|GRE_VERSION)) {
613 /* - Version must be 0.
614 - We do not support routing headers.
615 */
616 if (flags&(GRE_VERSION|GRE_ROUTING))
617 goto drop;
618
619 if (flags&GRE_CSUM) {
620 switch (skb->ip_summed) {
621 case CHECKSUM_COMPLETE:
622 csum = csum_fold(skb->csum);
623 if (!csum)
624 break;
625 /* fall through */
626 case CHECKSUM_NONE:
627 skb->csum = 0;
628 csum = __skb_checksum_complete(skb);
629 skb->ip_summed = CHECKSUM_COMPLETE;
630 }
631 offset += 4;
632 }
633 if (flags&GRE_KEY) {
634 key = *(__be32 *)(h + offset);
635 offset += 4;
636 }
637 if (flags&GRE_SEQ) {
638 seqno = ntohl(*(__be32 *)(h + offset));
639 offset += 4;
640 }
641 }
642
643 gre_proto = *(__be16 *)(h + 2);
644
645 tunnel = ipgre_tunnel_lookup(skb->dev,
646 iph->saddr, iph->daddr, flags, key,
647 gre_proto);
648 if (tunnel) {
649 struct pcpu_tstats *tstats;
650
651 secpath_reset(skb);
652
653 skb->protocol = gre_proto;
654 /* WCCP version 1 and 2 protocol decoding.
655 * - Change protocol to IP
656 * - When dealing with WCCPv2, Skip extra 4 bytes in GRE header
657 */
658 if (flags == 0 && gre_proto == htons(ETH_P_WCCP)) {
659 skb->protocol = htons(ETH_P_IP);
660 if ((*(h + offset) & 0xF0) != 0x40)
661 offset += 4;
662 }
663
664 skb->mac_header = skb->network_header;
665 __pskb_pull(skb, offset);
666 skb_postpull_rcsum(skb, skb_transport_header(skb), offset);
667 skb->pkt_type = PACKET_HOST;
668 #ifdef CONFIG_NET_IPGRE_BROADCAST
669 if (ipv4_is_multicast(iph->daddr)) {
670 /* Looped back packet, drop it! */
671 if (rt_is_output_route(skb_rtable(skb)))
672 goto drop;
673 tunnel->dev->stats.multicast++;
674 skb->pkt_type = PACKET_BROADCAST;
675 }
676 #endif
677
678 if (((flags&GRE_CSUM) && csum) ||
679 (!(flags&GRE_CSUM) && tunnel->parms.i_flags&GRE_CSUM)) {
680 tunnel->dev->stats.rx_crc_errors++;
681 tunnel->dev->stats.rx_errors++;
682 goto drop;
683 }
684 if (tunnel->parms.i_flags&GRE_SEQ) {
685 if (!(flags&GRE_SEQ) ||
686 (tunnel->i_seqno && (s32)(seqno - tunnel->i_seqno) < 0)) {
687 tunnel->dev->stats.rx_fifo_errors++;
688 tunnel->dev->stats.rx_errors++;
689 goto drop;
690 }
691 tunnel->i_seqno = seqno + 1;
692 }
693
694 /* Warning: All skb pointers will be invalidated! */
695 if (tunnel->dev->type == ARPHRD_ETHER) {
696 if (!pskb_may_pull(skb, ETH_HLEN)) {
697 tunnel->dev->stats.rx_length_errors++;
698 tunnel->dev->stats.rx_errors++;
699 goto drop;
700 }
701
702 iph = ip_hdr(skb);
703 skb->protocol = eth_type_trans(skb, tunnel->dev);
704 skb_postpull_rcsum(skb, eth_hdr(skb), ETH_HLEN);
705 }
706
707 __skb_tunnel_rx(skb, tunnel->dev);
708
709 skb_reset_network_header(skb);
710 err = IP_ECN_decapsulate(iph, skb);
711 if (unlikely(err)) {
712 if (log_ecn_error)
713 net_info_ratelimited("non-ECT from %pI4 with TOS=%#x\n",
714 &iph->saddr, iph->tos);
715 if (err > 1) {
716 ++tunnel->dev->stats.rx_frame_errors;
717 ++tunnel->dev->stats.rx_errors;
718 goto drop;
719 }
720 }
721
722 tstats = this_cpu_ptr(tunnel->dev->tstats);
723 u64_stats_update_begin(&tstats->syncp);
724 tstats->rx_packets++;
725 tstats->rx_bytes += skb->len;
726 u64_stats_update_end(&tstats->syncp);
727
728 gro_cells_receive(&tunnel->gro_cells, skb);
729 return 0;
730 }
731 icmp_send(skb, ICMP_DEST_UNREACH, ICMP_PORT_UNREACH, 0);
732
733 drop:
734 kfree_skb(skb);
735 return 0;
736 }
737
738 static netdev_tx_t ipgre_tunnel_xmit(struct sk_buff *skb, struct net_device *dev)
739 {
740 struct ip_tunnel *tunnel = netdev_priv(dev);
741 const struct iphdr *old_iph = ip_hdr(skb);
742 const struct iphdr *tiph;
743 struct flowi4 fl4;
744 u8 tos;
745 __be16 df;
746 struct rtable *rt; /* Route to the other host */
747 struct net_device *tdev; /* Device to other host */
748 struct iphdr *iph; /* Our new IP header */
749 unsigned int max_headroom; /* The extra header space needed */
750 int gre_hlen;
751 __be32 dst;
752 int mtu;
753
754 if (skb->ip_summed == CHECKSUM_PARTIAL &&
755 skb_checksum_help(skb))
756 goto tx_error;
757
758 if (dev->type == ARPHRD_ETHER)
759 IPCB(skb)->flags = 0;
760
761 if (dev->header_ops && dev->type == ARPHRD_IPGRE) {
762 gre_hlen = 0;
763 if (skb->protocol == htons(ETH_P_IP))
764 tiph = (const struct iphdr *)skb->data;
765 else
766 tiph = &tunnel->parms.iph;
767 } else {
768 gre_hlen = tunnel->hlen;
769 tiph = &tunnel->parms.iph;
770 }
771
772 if ((dst = tiph->daddr) == 0) {
773 /* NBMA tunnel */
774
775 if (skb_dst(skb) == NULL) {
776 dev->stats.tx_fifo_errors++;
777 goto tx_error;
778 }
779
780 if (skb->protocol == htons(ETH_P_IP)) {
781 rt = skb_rtable(skb);
782 dst = rt_nexthop(rt, old_iph->daddr);
783 }
784 #if IS_ENABLED(CONFIG_IPV6)
785 else if (skb->protocol == htons(ETH_P_IPV6)) {
786 const struct in6_addr *addr6;
787 struct neighbour *neigh;
788 bool do_tx_error_icmp;
789 int addr_type;
790
791 neigh = dst_neigh_lookup(skb_dst(skb), &ipv6_hdr(skb)->daddr);
792 if (neigh == NULL)
793 goto tx_error;
794
795 addr6 = (const struct in6_addr *)&neigh->primary_key;
796 addr_type = ipv6_addr_type(addr6);
797
798 if (addr_type == IPV6_ADDR_ANY) {
799 addr6 = &ipv6_hdr(skb)->daddr;
800 addr_type = ipv6_addr_type(addr6);
801 }
802
803 if ((addr_type & IPV6_ADDR_COMPATv4) == 0)
804 do_tx_error_icmp = true;
805 else {
806 do_tx_error_icmp = false;
807 dst = addr6->s6_addr32[3];
808 }
809 neigh_release(neigh);
810 if (do_tx_error_icmp)
811 goto tx_error_icmp;
812 }
813 #endif
814 else
815 goto tx_error;
816 }
817
818 tos = tiph->tos;
819 if (tos == 1) {
820 tos = 0;
821 if (skb->protocol == htons(ETH_P_IP))
822 tos = old_iph->tos;
823 else if (skb->protocol == htons(ETH_P_IPV6))
824 tos = ipv6_get_dsfield((const struct ipv6hdr *)old_iph);
825 }
826
827 rt = ip_route_output_gre(dev_net(dev), &fl4, dst, tiph->saddr,
828 tunnel->parms.o_key, RT_TOS(tos),
829 tunnel->parms.link);
830 if (IS_ERR(rt)) {
831 dev->stats.tx_carrier_errors++;
832 goto tx_error;
833 }
834 tdev = rt->dst.dev;
835
836 if (tdev == dev) {
837 ip_rt_put(rt);
838 dev->stats.collisions++;
839 goto tx_error;
840 }
841
842 df = tiph->frag_off;
843 if (df)
844 mtu = dst_mtu(&rt->dst) - dev->hard_header_len - tunnel->hlen;
845 else
846 mtu = skb_dst(skb) ? dst_mtu(skb_dst(skb)) : dev->mtu;
847
848 if (skb_dst(skb))
849 skb_dst(skb)->ops->update_pmtu(skb_dst(skb), NULL, skb, mtu);
850
851 if (skb->protocol == htons(ETH_P_IP)) {
852 df |= (old_iph->frag_off&htons(IP_DF));
853
854 if ((old_iph->frag_off&htons(IP_DF)) &&
855 mtu < ntohs(old_iph->tot_len)) {
856 icmp_send(skb, ICMP_DEST_UNREACH, ICMP_FRAG_NEEDED, htonl(mtu));
857 ip_rt_put(rt);
858 goto tx_error;
859 }
860 }
861 #if IS_ENABLED(CONFIG_IPV6)
862 else if (skb->protocol == htons(ETH_P_IPV6)) {
863 struct rt6_info *rt6 = (struct rt6_info *)skb_dst(skb);
864
865 if (rt6 && mtu < dst_mtu(skb_dst(skb)) && mtu >= IPV6_MIN_MTU) {
866 if ((tunnel->parms.iph.daddr &&
867 !ipv4_is_multicast(tunnel->parms.iph.daddr)) ||
868 rt6->rt6i_dst.plen == 128) {
869 rt6->rt6i_flags |= RTF_MODIFIED;
870 dst_metric_set(skb_dst(skb), RTAX_MTU, mtu);
871 }
872 }
873
874 if (mtu >= IPV6_MIN_MTU && mtu < skb->len - tunnel->hlen + gre_hlen) {
875 icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu);
876 ip_rt_put(rt);
877 goto tx_error;
878 }
879 }
880 #endif
881
882 if (tunnel->err_count > 0) {
883 if (time_before(jiffies,
884 tunnel->err_time + IPTUNNEL_ERR_TIMEO)) {
885 tunnel->err_count--;
886
887 dst_link_failure(skb);
888 } else
889 tunnel->err_count = 0;
890 }
891
892 max_headroom = LL_RESERVED_SPACE(tdev) + gre_hlen + rt->dst.header_len;
893
894 if (skb_headroom(skb) < max_headroom || skb_shared(skb)||
895 (skb_cloned(skb) && !skb_clone_writable(skb, 0))) {
896 struct sk_buff *new_skb = skb_realloc_headroom(skb, max_headroom);
897 if (max_headroom > dev->needed_headroom)
898 dev->needed_headroom = max_headroom;
899 if (!new_skb) {
900 ip_rt_put(rt);
901 dev->stats.tx_dropped++;
902 dev_kfree_skb(skb);
903 return NETDEV_TX_OK;
904 }
905 if (skb->sk)
906 skb_set_owner_w(new_skb, skb->sk);
907 dev_kfree_skb(skb);
908 skb = new_skb;
909 old_iph = ip_hdr(skb);
910 }
911
912 skb_reset_transport_header(skb);
913 skb_push(skb, gre_hlen);
914 skb_reset_network_header(skb);
915 memset(&(IPCB(skb)->opt), 0, sizeof(IPCB(skb)->opt));
916 IPCB(skb)->flags &= ~(IPSKB_XFRM_TUNNEL_SIZE | IPSKB_XFRM_TRANSFORMED |
917 IPSKB_REROUTED);
918 skb_dst_drop(skb);
919 skb_dst_set(skb, &rt->dst);
920
921 /*
922 * Push down and install the IPIP header.
923 */
924
925 iph = ip_hdr(skb);
926 iph->version = 4;
927 iph->ihl = sizeof(struct iphdr) >> 2;
928 iph->frag_off = df;
929 iph->protocol = IPPROTO_GRE;
930 iph->tos = ipgre_ecn_encapsulate(tos, old_iph, skb);
931 iph->daddr = fl4.daddr;
932 iph->saddr = fl4.saddr;
933
934 if ((iph->ttl = tiph->ttl) == 0) {
935 if (skb->protocol == htons(ETH_P_IP))
936 iph->ttl = old_iph->ttl;
937 #if IS_ENABLED(CONFIG_IPV6)
938 else if (skb->protocol == htons(ETH_P_IPV6))
939 iph->ttl = ((const struct ipv6hdr *)old_iph)->hop_limit;
940 #endif
941 else
942 iph->ttl = ip4_dst_hoplimit(&rt->dst);
943 }
944
945 ((__be16 *)(iph + 1))[0] = tunnel->parms.o_flags;
946 ((__be16 *)(iph + 1))[1] = (dev->type == ARPHRD_ETHER) ?
947 htons(ETH_P_TEB) : skb->protocol;
948
949 if (tunnel->parms.o_flags&(GRE_KEY|GRE_CSUM|GRE_SEQ)) {
950 __be32 *ptr = (__be32 *)(((u8 *)iph) + tunnel->hlen - 4);
951
952 if (tunnel->parms.o_flags&GRE_SEQ) {
953 ++tunnel->o_seqno;
954 *ptr = htonl(tunnel->o_seqno);
955 ptr--;
956 }
957 if (tunnel->parms.o_flags&GRE_KEY) {
958 *ptr = tunnel->parms.o_key;
959 ptr--;
960 }
961 if (tunnel->parms.o_flags&GRE_CSUM) {
962 *ptr = 0;
963 *(__sum16 *)ptr = ip_compute_csum((void *)(iph+1), skb->len - sizeof(struct iphdr));
964 }
965 }
966
967 iptunnel_xmit(skb, dev);
968 return NETDEV_TX_OK;
969
970 #if IS_ENABLED(CONFIG_IPV6)
971 tx_error_icmp:
972 dst_link_failure(skb);
973 #endif
974 tx_error:
975 dev->stats.tx_errors++;
976 dev_kfree_skb(skb);
977 return NETDEV_TX_OK;
978 }
979
980 static int ipgre_tunnel_bind_dev(struct net_device *dev)
981 {
982 struct net_device *tdev = NULL;
983 struct ip_tunnel *tunnel;
984 const struct iphdr *iph;
985 int hlen = LL_MAX_HEADER;
986 int mtu = ETH_DATA_LEN;
987 int addend = sizeof(struct iphdr) + 4;
988
989 tunnel = netdev_priv(dev);
990 iph = &tunnel->parms.iph;
991
992 /* Guess output device to choose reasonable mtu and needed_headroom */
993
994 if (iph->daddr) {
995 struct flowi4 fl4;
996 struct rtable *rt;
997
998 rt = ip_route_output_gre(dev_net(dev), &fl4,
999 iph->daddr, iph->saddr,
1000 tunnel->parms.o_key,
1001 RT_TOS(iph->tos),
1002 tunnel->parms.link);
1003 if (!IS_ERR(rt)) {
1004 tdev = rt->dst.dev;
1005 ip_rt_put(rt);
1006 }
1007
1008 if (dev->type != ARPHRD_ETHER)
1009 dev->flags |= IFF_POINTOPOINT;
1010 }
1011
1012 if (!tdev && tunnel->parms.link)
1013 tdev = __dev_get_by_index(dev_net(dev), tunnel->parms.link);
1014
1015 if (tdev) {
1016 hlen = tdev->hard_header_len + tdev->needed_headroom;
1017 mtu = tdev->mtu;
1018 }
1019 dev->iflink = tunnel->parms.link;
1020
1021 /* Precalculate GRE options length */
1022 if (tunnel->parms.o_flags&(GRE_CSUM|GRE_KEY|GRE_SEQ)) {
1023 if (tunnel->parms.o_flags&GRE_CSUM)
1024 addend += 4;
1025 if (tunnel->parms.o_flags&GRE_KEY)
1026 addend += 4;
1027 if (tunnel->parms.o_flags&GRE_SEQ)
1028 addend += 4;
1029 }
1030 dev->needed_headroom = addend + hlen;
1031 mtu -= dev->hard_header_len + addend;
1032
1033 if (mtu < 68)
1034 mtu = 68;
1035
1036 tunnel->hlen = addend;
1037
1038 return mtu;
1039 }
1040
1041 static int
1042 ipgre_tunnel_ioctl (struct net_device *dev, struct ifreq *ifr, int cmd)
1043 {
1044 int err = 0;
1045 struct ip_tunnel_parm p;
1046 struct ip_tunnel *t;
1047 struct net *net = dev_net(dev);
1048 struct ipgre_net *ign = net_generic(net, ipgre_net_id);
1049
1050 switch (cmd) {
1051 case SIOCGETTUNNEL:
1052 t = NULL;
1053 if (dev == ign->fb_tunnel_dev) {
1054 if (copy_from_user(&p, ifr->ifr_ifru.ifru_data, sizeof(p))) {
1055 err = -EFAULT;
1056 break;
1057 }
1058 t = ipgre_tunnel_locate(net, &p, 0);
1059 }
1060 if (t == NULL)
1061 t = netdev_priv(dev);
1062 memcpy(&p, &t->parms, sizeof(p));
1063 if (copy_to_user(ifr->ifr_ifru.ifru_data, &p, sizeof(p)))
1064 err = -EFAULT;
1065 break;
1066
1067 case SIOCADDTUNNEL:
1068 case SIOCCHGTUNNEL:
1069 err = -EPERM;
1070 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
1071 goto done;
1072
1073 err = -EFAULT;
1074 if (copy_from_user(&p, ifr->ifr_ifru.ifru_data, sizeof(p)))
1075 goto done;
1076
1077 err = -EINVAL;
1078 if (p.iph.version != 4 || p.iph.protocol != IPPROTO_GRE ||
1079 p.iph.ihl != 5 || (p.iph.frag_off&htons(~IP_DF)) ||
1080 ((p.i_flags|p.o_flags)&(GRE_VERSION|GRE_ROUTING)))
1081 goto done;
1082 if (p.iph.ttl)
1083 p.iph.frag_off |= htons(IP_DF);
1084
1085 if (!(p.i_flags&GRE_KEY))
1086 p.i_key = 0;
1087 if (!(p.o_flags&GRE_KEY))
1088 p.o_key = 0;
1089
1090 t = ipgre_tunnel_locate(net, &p, cmd == SIOCADDTUNNEL);
1091
1092 if (dev != ign->fb_tunnel_dev && cmd == SIOCCHGTUNNEL) {
1093 if (t != NULL) {
1094 if (t->dev != dev) {
1095 err = -EEXIST;
1096 break;
1097 }
1098 } else {
1099 unsigned int nflags = 0;
1100
1101 t = netdev_priv(dev);
1102
1103 if (ipv4_is_multicast(p.iph.daddr))
1104 nflags = IFF_BROADCAST;
1105 else if (p.iph.daddr)
1106 nflags = IFF_POINTOPOINT;
1107
1108 if ((dev->flags^nflags)&(IFF_POINTOPOINT|IFF_BROADCAST)) {
1109 err = -EINVAL;
1110 break;
1111 }
1112 ipgre_tunnel_unlink(ign, t);
1113 synchronize_net();
1114 t->parms.iph.saddr = p.iph.saddr;
1115 t->parms.iph.daddr = p.iph.daddr;
1116 t->parms.i_key = p.i_key;
1117 t->parms.o_key = p.o_key;
1118 memcpy(dev->dev_addr, &p.iph.saddr, 4);
1119 memcpy(dev->broadcast, &p.iph.daddr, 4);
1120 ipgre_tunnel_link(ign, t);
1121 netdev_state_change(dev);
1122 }
1123 }
1124
1125 if (t) {
1126 err = 0;
1127 if (cmd == SIOCCHGTUNNEL) {
1128 t->parms.iph.ttl = p.iph.ttl;
1129 t->parms.iph.tos = p.iph.tos;
1130 t->parms.iph.frag_off = p.iph.frag_off;
1131 if (t->parms.link != p.link) {
1132 t->parms.link = p.link;
1133 dev->mtu = ipgre_tunnel_bind_dev(dev);
1134 netdev_state_change(dev);
1135 }
1136 }
1137 if (copy_to_user(ifr->ifr_ifru.ifru_data, &t->parms, sizeof(p)))
1138 err = -EFAULT;
1139 } else
1140 err = (cmd == SIOCADDTUNNEL ? -ENOBUFS : -ENOENT);
1141 break;
1142
1143 case SIOCDELTUNNEL:
1144 err = -EPERM;
1145 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
1146 goto done;
1147
1148 if (dev == ign->fb_tunnel_dev) {
1149 err = -EFAULT;
1150 if (copy_from_user(&p, ifr->ifr_ifru.ifru_data, sizeof(p)))
1151 goto done;
1152 err = -ENOENT;
1153 if ((t = ipgre_tunnel_locate(net, &p, 0)) == NULL)
1154 goto done;
1155 err = -EPERM;
1156 if (t == netdev_priv(ign->fb_tunnel_dev))
1157 goto done;
1158 dev = t->dev;
1159 }
1160 unregister_netdevice(dev);
1161 err = 0;
1162 break;
1163
1164 default:
1165 err = -EINVAL;
1166 }
1167
1168 done:
1169 return err;
1170 }
1171
1172 static int ipgre_tunnel_change_mtu(struct net_device *dev, int new_mtu)
1173 {
1174 struct ip_tunnel *tunnel = netdev_priv(dev);
1175 if (new_mtu < 68 ||
1176 new_mtu > 0xFFF8 - dev->hard_header_len - tunnel->hlen)
1177 return -EINVAL;
1178 dev->mtu = new_mtu;
1179 return 0;
1180 }
1181
1182 /* Nice toy. Unfortunately, useless in real life :-)
1183 It allows to construct virtual multiprotocol broadcast "LAN"
1184 over the Internet, provided multicast routing is tuned.
1185
1186
1187 I have no idea was this bicycle invented before me,
1188 so that I had to set ARPHRD_IPGRE to a random value.
1189 I have an impression, that Cisco could make something similar,
1190 but this feature is apparently missing in IOS<=11.2(8).
1191
1192 I set up 10.66.66/24 and fec0:6666:6666::0/96 as virtual networks
1193 with broadcast 224.66.66.66. If you have access to mbone, play with me :-)
1194
1195 ping -t 255 224.66.66.66
1196
1197 If nobody answers, mbone does not work.
1198
1199 ip tunnel add Universe mode gre remote 224.66.66.66 local <Your_real_addr> ttl 255
1200 ip addr add 10.66.66.<somewhat>/24 dev Universe
1201 ifconfig Universe up
1202 ifconfig Universe add fe80::<Your_real_addr>/10
1203 ifconfig Universe add fec0:6666:6666::<Your_real_addr>/96
1204 ftp 10.66.66.66
1205 ...
1206 ftp fec0:6666:6666::193.233.7.65
1207 ...
1208
1209 */
1210
1211 static int ipgre_header(struct sk_buff *skb, struct net_device *dev,
1212 unsigned short type,
1213 const void *daddr, const void *saddr, unsigned int len)
1214 {
1215 struct ip_tunnel *t = netdev_priv(dev);
1216 struct iphdr *iph = (struct iphdr *)skb_push(skb, t->hlen);
1217 __be16 *p = (__be16 *)(iph+1);
1218
1219 memcpy(iph, &t->parms.iph, sizeof(struct iphdr));
1220 p[0] = t->parms.o_flags;
1221 p[1] = htons(type);
1222
1223 /*
1224 * Set the source hardware address.
1225 */
1226
1227 if (saddr)
1228 memcpy(&iph->saddr, saddr, 4);
1229 if (daddr)
1230 memcpy(&iph->daddr, daddr, 4);
1231 if (iph->daddr)
1232 return t->hlen;
1233
1234 return -t->hlen;
1235 }
1236
1237 static int ipgre_header_parse(const struct sk_buff *skb, unsigned char *haddr)
1238 {
1239 const struct iphdr *iph = (const struct iphdr *) skb_mac_header(skb);
1240 memcpy(haddr, &iph->saddr, 4);
1241 return 4;
1242 }
1243
1244 static const struct header_ops ipgre_header_ops = {
1245 .create = ipgre_header,
1246 .parse = ipgre_header_parse,
1247 };
1248
1249 #ifdef CONFIG_NET_IPGRE_BROADCAST
1250 static int ipgre_open(struct net_device *dev)
1251 {
1252 struct ip_tunnel *t = netdev_priv(dev);
1253
1254 if (ipv4_is_multicast(t->parms.iph.daddr)) {
1255 struct flowi4 fl4;
1256 struct rtable *rt;
1257
1258 rt = ip_route_output_gre(dev_net(dev), &fl4,
1259 t->parms.iph.daddr,
1260 t->parms.iph.saddr,
1261 t->parms.o_key,
1262 RT_TOS(t->parms.iph.tos),
1263 t->parms.link);
1264 if (IS_ERR(rt))
1265 return -EADDRNOTAVAIL;
1266 dev = rt->dst.dev;
1267 ip_rt_put(rt);
1268 if (__in_dev_get_rtnl(dev) == NULL)
1269 return -EADDRNOTAVAIL;
1270 t->mlink = dev->ifindex;
1271 ip_mc_inc_group(__in_dev_get_rtnl(dev), t->parms.iph.daddr);
1272 }
1273 return 0;
1274 }
1275
1276 static int ipgre_close(struct net_device *dev)
1277 {
1278 struct ip_tunnel *t = netdev_priv(dev);
1279
1280 if (ipv4_is_multicast(t->parms.iph.daddr) && t->mlink) {
1281 struct in_device *in_dev;
1282 in_dev = inetdev_by_index(dev_net(dev), t->mlink);
1283 if (in_dev)
1284 ip_mc_dec_group(in_dev, t->parms.iph.daddr);
1285 }
1286 return 0;
1287 }
1288
1289 #endif
1290
1291 static const struct net_device_ops ipgre_netdev_ops = {
1292 .ndo_init = ipgre_tunnel_init,
1293 .ndo_uninit = ipgre_tunnel_uninit,
1294 #ifdef CONFIG_NET_IPGRE_BROADCAST
1295 .ndo_open = ipgre_open,
1296 .ndo_stop = ipgre_close,
1297 #endif
1298 .ndo_start_xmit = ipgre_tunnel_xmit,
1299 .ndo_do_ioctl = ipgre_tunnel_ioctl,
1300 .ndo_change_mtu = ipgre_tunnel_change_mtu,
1301 .ndo_get_stats64 = ipgre_get_stats64,
1302 };
1303
1304 static void ipgre_dev_free(struct net_device *dev)
1305 {
1306 struct ip_tunnel *tunnel = netdev_priv(dev);
1307
1308 gro_cells_destroy(&tunnel->gro_cells);
1309 free_percpu(dev->tstats);
1310 free_netdev(dev);
1311 }
1312
1313 #define GRE_FEATURES (NETIF_F_SG | \
1314 NETIF_F_FRAGLIST | \
1315 NETIF_F_HIGHDMA | \
1316 NETIF_F_HW_CSUM)
1317
1318 static void ipgre_tunnel_setup(struct net_device *dev)
1319 {
1320 dev->netdev_ops = &ipgre_netdev_ops;
1321 dev->destructor = ipgre_dev_free;
1322
1323 dev->type = ARPHRD_IPGRE;
1324 dev->needed_headroom = LL_MAX_HEADER + sizeof(struct iphdr) + 4;
1325 dev->mtu = ETH_DATA_LEN - sizeof(struct iphdr) - 4;
1326 dev->flags = IFF_NOARP;
1327 dev->iflink = 0;
1328 dev->addr_len = 4;
1329 dev->features |= NETIF_F_NETNS_LOCAL;
1330 dev->priv_flags &= ~IFF_XMIT_DST_RELEASE;
1331
1332 dev->features |= GRE_FEATURES;
1333 dev->hw_features |= GRE_FEATURES;
1334 }
1335
1336 static int ipgre_tunnel_init(struct net_device *dev)
1337 {
1338 struct ip_tunnel *tunnel;
1339 struct iphdr *iph;
1340 int err;
1341
1342 tunnel = netdev_priv(dev);
1343 iph = &tunnel->parms.iph;
1344
1345 tunnel->dev = dev;
1346 strcpy(tunnel->parms.name, dev->name);
1347
1348 memcpy(dev->dev_addr, &tunnel->parms.iph.saddr, 4);
1349 memcpy(dev->broadcast, &tunnel->parms.iph.daddr, 4);
1350
1351 if (iph->daddr) {
1352 #ifdef CONFIG_NET_IPGRE_BROADCAST
1353 if (ipv4_is_multicast(iph->daddr)) {
1354 if (!iph->saddr)
1355 return -EINVAL;
1356 dev->flags = IFF_BROADCAST;
1357 dev->header_ops = &ipgre_header_ops;
1358 }
1359 #endif
1360 } else
1361 dev->header_ops = &ipgre_header_ops;
1362
1363 dev->tstats = alloc_percpu(struct pcpu_tstats);
1364 if (!dev->tstats)
1365 return -ENOMEM;
1366
1367 err = gro_cells_init(&tunnel->gro_cells, dev);
1368 if (err) {
1369 free_percpu(dev->tstats);
1370 return err;
1371 }
1372
1373 return 0;
1374 }
1375
1376 static void ipgre_fb_tunnel_init(struct net_device *dev)
1377 {
1378 struct ip_tunnel *tunnel = netdev_priv(dev);
1379 struct iphdr *iph = &tunnel->parms.iph;
1380
1381 tunnel->dev = dev;
1382 strcpy(tunnel->parms.name, dev->name);
1383
1384 iph->version = 4;
1385 iph->protocol = IPPROTO_GRE;
1386 iph->ihl = 5;
1387 tunnel->hlen = sizeof(struct iphdr) + 4;
1388
1389 dev_hold(dev);
1390 }
1391
1392
1393 static const struct gre_protocol ipgre_protocol = {
1394 .handler = ipgre_rcv,
1395 .err_handler = ipgre_err,
1396 };
1397
1398 static void ipgre_destroy_tunnels(struct ipgre_net *ign, struct list_head *head)
1399 {
1400 int prio;
1401
1402 for (prio = 0; prio < 4; prio++) {
1403 int h;
1404 for (h = 0; h < HASH_SIZE; h++) {
1405 struct ip_tunnel *t;
1406
1407 t = rtnl_dereference(ign->tunnels[prio][h]);
1408
1409 while (t != NULL) {
1410 unregister_netdevice_queue(t->dev, head);
1411 t = rtnl_dereference(t->next);
1412 }
1413 }
1414 }
1415 }
1416
1417 static int __net_init ipgre_init_net(struct net *net)
1418 {
1419 struct ipgre_net *ign = net_generic(net, ipgre_net_id);
1420 int err;
1421
1422 ign->fb_tunnel_dev = alloc_netdev(sizeof(struct ip_tunnel), "gre0",
1423 ipgre_tunnel_setup);
1424 if (!ign->fb_tunnel_dev) {
1425 err = -ENOMEM;
1426 goto err_alloc_dev;
1427 }
1428 dev_net_set(ign->fb_tunnel_dev, net);
1429
1430 ipgre_fb_tunnel_init(ign->fb_tunnel_dev);
1431 ign->fb_tunnel_dev->rtnl_link_ops = &ipgre_link_ops;
1432
1433 if ((err = register_netdev(ign->fb_tunnel_dev)))
1434 goto err_reg_dev;
1435
1436 rcu_assign_pointer(ign->tunnels_wc[0],
1437 netdev_priv(ign->fb_tunnel_dev));
1438 return 0;
1439
1440 err_reg_dev:
1441 ipgre_dev_free(ign->fb_tunnel_dev);
1442 err_alloc_dev:
1443 return err;
1444 }
1445
1446 static void __net_exit ipgre_exit_net(struct net *net)
1447 {
1448 struct ipgre_net *ign;
1449 LIST_HEAD(list);
1450
1451 ign = net_generic(net, ipgre_net_id);
1452 rtnl_lock();
1453 ipgre_destroy_tunnels(ign, &list);
1454 unregister_netdevice_many(&list);
1455 rtnl_unlock();
1456 }
1457
1458 static struct pernet_operations ipgre_net_ops = {
1459 .init = ipgre_init_net,
1460 .exit = ipgre_exit_net,
1461 .id = &ipgre_net_id,
1462 .size = sizeof(struct ipgre_net),
1463 };
1464
1465 static int ipgre_tunnel_validate(struct nlattr *tb[], struct nlattr *data[])
1466 {
1467 __be16 flags;
1468
1469 if (!data)
1470 return 0;
1471
1472 flags = 0;
1473 if (data[IFLA_GRE_IFLAGS])
1474 flags |= nla_get_be16(data[IFLA_GRE_IFLAGS]);
1475 if (data[IFLA_GRE_OFLAGS])
1476 flags |= nla_get_be16(data[IFLA_GRE_OFLAGS]);
1477 if (flags & (GRE_VERSION|GRE_ROUTING))
1478 return -EINVAL;
1479
1480 return 0;
1481 }
1482
1483 static int ipgre_tap_validate(struct nlattr *tb[], struct nlattr *data[])
1484 {
1485 __be32 daddr;
1486
1487 if (tb[IFLA_ADDRESS]) {
1488 if (nla_len(tb[IFLA_ADDRESS]) != ETH_ALEN)
1489 return -EINVAL;
1490 if (!is_valid_ether_addr(nla_data(tb[IFLA_ADDRESS])))
1491 return -EADDRNOTAVAIL;
1492 }
1493
1494 if (!data)
1495 goto out;
1496
1497 if (data[IFLA_GRE_REMOTE]) {
1498 memcpy(&daddr, nla_data(data[IFLA_GRE_REMOTE]), 4);
1499 if (!daddr)
1500 return -EINVAL;
1501 }
1502
1503 out:
1504 return ipgre_tunnel_validate(tb, data);
1505 }
1506
1507 static void ipgre_netlink_parms(struct nlattr *data[],
1508 struct ip_tunnel_parm *parms)
1509 {
1510 memset(parms, 0, sizeof(*parms));
1511
1512 parms->iph.protocol = IPPROTO_GRE;
1513
1514 if (!data)
1515 return;
1516
1517 if (data[IFLA_GRE_LINK])
1518 parms->link = nla_get_u32(data[IFLA_GRE_LINK]);
1519
1520 if (data[IFLA_GRE_IFLAGS])
1521 parms->i_flags = nla_get_be16(data[IFLA_GRE_IFLAGS]);
1522
1523 if (data[IFLA_GRE_OFLAGS])
1524 parms->o_flags = nla_get_be16(data[IFLA_GRE_OFLAGS]);
1525
1526 if (data[IFLA_GRE_IKEY])
1527 parms->i_key = nla_get_be32(data[IFLA_GRE_IKEY]);
1528
1529 if (data[IFLA_GRE_OKEY])
1530 parms->o_key = nla_get_be32(data[IFLA_GRE_OKEY]);
1531
1532 if (data[IFLA_GRE_LOCAL])
1533 parms->iph.saddr = nla_get_be32(data[IFLA_GRE_LOCAL]);
1534
1535 if (data[IFLA_GRE_REMOTE])
1536 parms->iph.daddr = nla_get_be32(data[IFLA_GRE_REMOTE]);
1537
1538 if (data[IFLA_GRE_TTL])
1539 parms->iph.ttl = nla_get_u8(data[IFLA_GRE_TTL]);
1540
1541 if (data[IFLA_GRE_TOS])
1542 parms->iph.tos = nla_get_u8(data[IFLA_GRE_TOS]);
1543
1544 if (!data[IFLA_GRE_PMTUDISC] || nla_get_u8(data[IFLA_GRE_PMTUDISC]))
1545 parms->iph.frag_off = htons(IP_DF);
1546 }
1547
1548 static int ipgre_tap_init(struct net_device *dev)
1549 {
1550 struct ip_tunnel *tunnel;
1551
1552 tunnel = netdev_priv(dev);
1553
1554 tunnel->dev = dev;
1555 strcpy(tunnel->parms.name, dev->name);
1556
1557 ipgre_tunnel_bind_dev(dev);
1558
1559 dev->tstats = alloc_percpu(struct pcpu_tstats);
1560 if (!dev->tstats)
1561 return -ENOMEM;
1562
1563 return 0;
1564 }
1565
1566 static const struct net_device_ops ipgre_tap_netdev_ops = {
1567 .ndo_init = ipgre_tap_init,
1568 .ndo_uninit = ipgre_tunnel_uninit,
1569 .ndo_start_xmit = ipgre_tunnel_xmit,
1570 .ndo_set_mac_address = eth_mac_addr,
1571 .ndo_validate_addr = eth_validate_addr,
1572 .ndo_change_mtu = ipgre_tunnel_change_mtu,
1573 .ndo_get_stats64 = ipgre_get_stats64,
1574 };
1575
1576 static void ipgre_tap_setup(struct net_device *dev)
1577 {
1578
1579 ether_setup(dev);
1580
1581 dev->netdev_ops = &ipgre_tap_netdev_ops;
1582 dev->destructor = ipgre_dev_free;
1583
1584 dev->iflink = 0;
1585 dev->features |= NETIF_F_NETNS_LOCAL;
1586 }
1587
1588 static int ipgre_newlink(struct net *src_net, struct net_device *dev, struct nlattr *tb[],
1589 struct nlattr *data[])
1590 {
1591 struct ip_tunnel *nt;
1592 struct net *net = dev_net(dev);
1593 struct ipgre_net *ign = net_generic(net, ipgre_net_id);
1594 int mtu;
1595 int err;
1596
1597 nt = netdev_priv(dev);
1598 ipgre_netlink_parms(data, &nt->parms);
1599
1600 if (ipgre_tunnel_find(net, &nt->parms, dev->type))
1601 return -EEXIST;
1602
1603 if (dev->type == ARPHRD_ETHER && !tb[IFLA_ADDRESS])
1604 eth_hw_addr_random(dev);
1605
1606 mtu = ipgre_tunnel_bind_dev(dev);
1607 if (!tb[IFLA_MTU])
1608 dev->mtu = mtu;
1609
1610 /* Can use a lockless transmit, unless we generate output sequences */
1611 if (!(nt->parms.o_flags & GRE_SEQ))
1612 dev->features |= NETIF_F_LLTX;
1613
1614 err = register_netdevice(dev);
1615 if (err)
1616 goto out;
1617
1618 dev_hold(dev);
1619 ipgre_tunnel_link(ign, nt);
1620
1621 out:
1622 return err;
1623 }
1624
1625 static int ipgre_changelink(struct net_device *dev, struct nlattr *tb[],
1626 struct nlattr *data[])
1627 {
1628 struct ip_tunnel *t, *nt;
1629 struct net *net = dev_net(dev);
1630 struct ipgre_net *ign = net_generic(net, ipgre_net_id);
1631 struct ip_tunnel_parm p;
1632 int mtu;
1633
1634 if (dev == ign->fb_tunnel_dev)
1635 return -EINVAL;
1636
1637 nt = netdev_priv(dev);
1638 ipgre_netlink_parms(data, &p);
1639
1640 t = ipgre_tunnel_locate(net, &p, 0);
1641
1642 if (t) {
1643 if (t->dev != dev)
1644 return -EEXIST;
1645 } else {
1646 t = nt;
1647
1648 if (dev->type != ARPHRD_ETHER) {
1649 unsigned int nflags = 0;
1650
1651 if (ipv4_is_multicast(p.iph.daddr))
1652 nflags = IFF_BROADCAST;
1653 else if (p.iph.daddr)
1654 nflags = IFF_POINTOPOINT;
1655
1656 if ((dev->flags ^ nflags) &
1657 (IFF_POINTOPOINT | IFF_BROADCAST))
1658 return -EINVAL;
1659 }
1660
1661 ipgre_tunnel_unlink(ign, t);
1662 t->parms.iph.saddr = p.iph.saddr;
1663 t->parms.iph.daddr = p.iph.daddr;
1664 t->parms.i_key = p.i_key;
1665 if (dev->type != ARPHRD_ETHER) {
1666 memcpy(dev->dev_addr, &p.iph.saddr, 4);
1667 memcpy(dev->broadcast, &p.iph.daddr, 4);
1668 }
1669 ipgre_tunnel_link(ign, t);
1670 netdev_state_change(dev);
1671 }
1672
1673 t->parms.o_key = p.o_key;
1674 t->parms.iph.ttl = p.iph.ttl;
1675 t->parms.iph.tos = p.iph.tos;
1676 t->parms.iph.frag_off = p.iph.frag_off;
1677
1678 if (t->parms.link != p.link) {
1679 t->parms.link = p.link;
1680 mtu = ipgre_tunnel_bind_dev(dev);
1681 if (!tb[IFLA_MTU])
1682 dev->mtu = mtu;
1683 netdev_state_change(dev);
1684 }
1685
1686 return 0;
1687 }
1688
1689 static size_t ipgre_get_size(const struct net_device *dev)
1690 {
1691 return
1692 /* IFLA_GRE_LINK */
1693 nla_total_size(4) +
1694 /* IFLA_GRE_IFLAGS */
1695 nla_total_size(2) +
1696 /* IFLA_GRE_OFLAGS */
1697 nla_total_size(2) +
1698 /* IFLA_GRE_IKEY */
1699 nla_total_size(4) +
1700 /* IFLA_GRE_OKEY */
1701 nla_total_size(4) +
1702 /* IFLA_GRE_LOCAL */
1703 nla_total_size(4) +
1704 /* IFLA_GRE_REMOTE */
1705 nla_total_size(4) +
1706 /* IFLA_GRE_TTL */
1707 nla_total_size(1) +
1708 /* IFLA_GRE_TOS */
1709 nla_total_size(1) +
1710 /* IFLA_GRE_PMTUDISC */
1711 nla_total_size(1) +
1712 0;
1713 }
1714
1715 static int ipgre_fill_info(struct sk_buff *skb, const struct net_device *dev)
1716 {
1717 struct ip_tunnel *t = netdev_priv(dev);
1718 struct ip_tunnel_parm *p = &t->parms;
1719
1720 if (nla_put_u32(skb, IFLA_GRE_LINK, p->link) ||
1721 nla_put_be16(skb, IFLA_GRE_IFLAGS, p->i_flags) ||
1722 nla_put_be16(skb, IFLA_GRE_OFLAGS, p->o_flags) ||
1723 nla_put_be32(skb, IFLA_GRE_IKEY, p->i_key) ||
1724 nla_put_be32(skb, IFLA_GRE_OKEY, p->o_key) ||
1725 nla_put_be32(skb, IFLA_GRE_LOCAL, p->iph.saddr) ||
1726 nla_put_be32(skb, IFLA_GRE_REMOTE, p->iph.daddr) ||
1727 nla_put_u8(skb, IFLA_GRE_TTL, p->iph.ttl) ||
1728 nla_put_u8(skb, IFLA_GRE_TOS, p->iph.tos) ||
1729 nla_put_u8(skb, IFLA_GRE_PMTUDISC,
1730 !!(p->iph.frag_off & htons(IP_DF))))
1731 goto nla_put_failure;
1732 return 0;
1733
1734 nla_put_failure:
1735 return -EMSGSIZE;
1736 }
1737
1738 static const struct nla_policy ipgre_policy[IFLA_GRE_MAX + 1] = {
1739 [IFLA_GRE_LINK] = { .type = NLA_U32 },
1740 [IFLA_GRE_IFLAGS] = { .type = NLA_U16 },
1741 [IFLA_GRE_OFLAGS] = { .type = NLA_U16 },
1742 [IFLA_GRE_IKEY] = { .type = NLA_U32 },
1743 [IFLA_GRE_OKEY] = { .type = NLA_U32 },
1744 [IFLA_GRE_LOCAL] = { .len = FIELD_SIZEOF(struct iphdr, saddr) },
1745 [IFLA_GRE_REMOTE] = { .len = FIELD_SIZEOF(struct iphdr, daddr) },
1746 [IFLA_GRE_TTL] = { .type = NLA_U8 },
1747 [IFLA_GRE_TOS] = { .type = NLA_U8 },
1748 [IFLA_GRE_PMTUDISC] = { .type = NLA_U8 },
1749 };
1750
1751 static struct rtnl_link_ops ipgre_link_ops __read_mostly = {
1752 .kind = "gre",
1753 .maxtype = IFLA_GRE_MAX,
1754 .policy = ipgre_policy,
1755 .priv_size = sizeof(struct ip_tunnel),
1756 .setup = ipgre_tunnel_setup,
1757 .validate = ipgre_tunnel_validate,
1758 .newlink = ipgre_newlink,
1759 .changelink = ipgre_changelink,
1760 .get_size = ipgre_get_size,
1761 .fill_info = ipgre_fill_info,
1762 };
1763
1764 static struct rtnl_link_ops ipgre_tap_ops __read_mostly = {
1765 .kind = "gretap",
1766 .maxtype = IFLA_GRE_MAX,
1767 .policy = ipgre_policy,
1768 .priv_size = sizeof(struct ip_tunnel),
1769 .setup = ipgre_tap_setup,
1770 .validate = ipgre_tap_validate,
1771 .newlink = ipgre_newlink,
1772 .changelink = ipgre_changelink,
1773 .get_size = ipgre_get_size,
1774 .fill_info = ipgre_fill_info,
1775 };
1776
1777 /*
1778 * And now the modules code and kernel interface.
1779 */
1780
1781 static int __init ipgre_init(void)
1782 {
1783 int err;
1784
1785 pr_info("GRE over IPv4 tunneling driver\n");
1786
1787 err = register_pernet_device(&ipgre_net_ops);
1788 if (err < 0)
1789 return err;
1790
1791 err = gre_add_protocol(&ipgre_protocol, GREPROTO_CISCO);
1792 if (err < 0) {
1793 pr_info("%s: can't add protocol\n", __func__);
1794 goto add_proto_failed;
1795 }
1796
1797 err = rtnl_link_register(&ipgre_link_ops);
1798 if (err < 0)
1799 goto rtnl_link_failed;
1800
1801 err = rtnl_link_register(&ipgre_tap_ops);
1802 if (err < 0)
1803 goto tap_ops_failed;
1804
1805 out:
1806 return err;
1807
1808 tap_ops_failed:
1809 rtnl_link_unregister(&ipgre_link_ops);
1810 rtnl_link_failed:
1811 gre_del_protocol(&ipgre_protocol, GREPROTO_CISCO);
1812 add_proto_failed:
1813 unregister_pernet_device(&ipgre_net_ops);
1814 goto out;
1815 }
1816
1817 static void __exit ipgre_fini(void)
1818 {
1819 rtnl_link_unregister(&ipgre_tap_ops);
1820 rtnl_link_unregister(&ipgre_link_ops);
1821 if (gre_del_protocol(&ipgre_protocol, GREPROTO_CISCO) < 0)
1822 pr_info("%s: can't remove protocol\n", __func__);
1823 unregister_pernet_device(&ipgre_net_ops);
1824 }
1825
1826 module_init(ipgre_init);
1827 module_exit(ipgre_fini);
1828 MODULE_LICENSE("GPL");
1829 MODULE_ALIAS_RTNL_LINK("gre");
1830 MODULE_ALIAS_RTNL_LINK("gretap");
1831 MODULE_ALIAS_NETDEV("gre0");
This page took 0.130134 seconds and 5 git commands to generate.