2 * Linux NET3: GRE over IP protocol decoder.
4 * Authors: Alexey Kuznetsov (kuznet@ms2.inr.ac.ru)
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
13 #include <linux/capability.h>
14 #include <linux/module.h>
15 #include <linux/types.h>
16 #include <linux/kernel.h>
17 #include <asm/uaccess.h>
18 #include <linux/skbuff.h>
19 #include <linux/netdevice.h>
21 #include <linux/tcp.h>
22 #include <linux/udp.h>
23 #include <linux/if_arp.h>
24 #include <linux/mroute.h>
25 #include <linux/init.h>
26 #include <linux/in6.h>
27 #include <linux/inetdevice.h>
28 #include <linux/igmp.h>
29 #include <linux/netfilter_ipv4.h>
30 #include <linux/if_ether.h>
35 #include <net/protocol.h>
38 #include <net/checksum.h>
39 #include <net/dsfield.h>
40 #include <net/inet_ecn.h>
45 #include <net/ip6_fib.h>
46 #include <net/ip6_route.h>
53 1. The most important issue is detecting local dead loops.
54 They would cause complete host lockup in transmit, which
55 would be "resolved" by stack overflow or, if queueing is enabled,
56 with infinite looping in net_bh.
58 We cannot track such dead loops during route installation,
59 it is infeasible task. The most general solutions would be
60 to keep skb->encapsulation counter (sort of local ttl),
61 and silently drop packet when it expires. It is the best
62 solution, but it supposes maintaing new variable in ALL
63 skb, even if no tunneling is used.
65 Current solution: t->recursion lock breaks dead loops. It looks
66 like dev->tbusy flag, but I preferred new variable, because
67 the semantics is different. One day, when hard_start_xmit
68 will be multithreaded we will have to use skb->encapsulation.
72 2. Networking dead loops would not kill routers, but would really
73 kill network. IP hop limit plays role of "t->recursion" in this case,
74 if we copy it from packet being encapsulated to upper header.
75 It is very good solution, but it introduces two problems:
77 - Routing protocols, using packets with ttl=1 (OSPF, RIP2),
78 do not work over tunnels.
79 - traceroute does not work. I planned to relay ICMP from tunnel,
80 so that this problem would be solved and traceroute output
81 would even more informative. This idea appeared to be wrong:
82 only Linux complies to rfc1812 now (yes, guys, Linux is the only
83 true router now :-)), all routers (at least, in neighbourhood of mine)
84 return only 8 bytes of payload. It is the end.
86 Hence, if we want that OSPF worked or traceroute said something reasonable,
87 we should search for another solution.
89 One of them is to parse packet trying to detect inner encapsulation
90 made by our node. It is difficult or even impossible, especially,
91 taking into account fragmentation. TO be short, tt is not solution at all.
93 Current solution: The solution was UNEXPECTEDLY SIMPLE.
94 We force DF flag on tunnels with preconfigured hop limit,
95 that is ALL. :-) Well, it does not remove the problem completely,
96 but exponential growth of network traffic is changed to linear
97 (branches, that exceed pmtu are pruned) and tunnel mtu
98 fastly degrades to value <68, where looping stops.
99 Yes, it is not good if there exists a router in the loop,
100 which does not force DF, even when encapsulating packets have DF set.
101 But it is not our problem! Nobody could accuse us, we made
102 all that we could make. Even if it is your gated who injected
103 fatal route to network, even if it were you who configured
104 fatal static route: you are innocent. :-)
108 3. Really, ipv4/ipip.c, ipv4/ip_gre.c and ipv6/sit.c contain
109 practically identical code. It would be good to glue them
110 together, but it is not very evident, how to make them modular.
111 sit is integral part of IPv6, ipip and gre are naturally modular.
112 We could extract common parts (hash table, ioctl etc)
113 to a separate module (ip_tunnel.c).
118 static int ipgre_tunnel_init(struct net_device *dev);
119 static void ipgre_tunnel_setup(struct net_device *dev);
121 /* Fallback tunnel: no source, no destination, no key, no options */
123 static int ipgre_fb_tunnel_init(struct net_device *dev);
125 static struct net_device *ipgre_fb_tunnel_dev;
127 /* Tunnel hash table */
137 We require exact key match i.e. if a key is present in packet
138 it will match only tunnel with the same key; if it is not present,
139 it will match only keyless tunnel.
141 All keysless packets, if not matched configured keyless tunnels
142 will match fallback tunnel.
146 #define HASH(addr) (((__force u32)addr^((__force u32)addr>>4))&0xF)
148 static struct ip_tunnel *tunnels[4][HASH_SIZE];
150 #define tunnels_r_l (tunnels[3])
151 #define tunnels_r (tunnels[2])
152 #define tunnels_l (tunnels[1])
153 #define tunnels_wc (tunnels[0])
155 static DEFINE_RWLOCK(ipgre_lock);
157 /* Given src, dst and key, find appropriate for input tunnel. */
159 static struct ip_tunnel * ipgre_tunnel_lookup(__be32 remote, __be32 local, __be32 key)
161 unsigned h0 = HASH(remote);
162 unsigned h1 = HASH(key);
165 for (t = tunnels_r_l[h0^h1]; t; t = t->next) {
166 if (local == t->parms.iph.saddr && remote == t->parms.iph.daddr) {
167 if (t->parms.i_key == key && (t->dev->flags&IFF_UP))
171 for (t = tunnels_r[h0^h1]; t; t = t->next) {
172 if (remote == t->parms.iph.daddr) {
173 if (t->parms.i_key == key && (t->dev->flags&IFF_UP))
177 for (t = tunnels_l[h1]; t; t = t->next) {
178 if (local == t->parms.iph.saddr ||
179 (local == t->parms.iph.daddr &&
180 ipv4_is_multicast(local))) {
181 if (t->parms.i_key == key && (t->dev->flags&IFF_UP))
185 for (t = tunnels_wc[h1]; t; t = t->next) {
186 if (t->parms.i_key == key && (t->dev->flags&IFF_UP))
190 if (ipgre_fb_tunnel_dev->flags&IFF_UP)
191 return netdev_priv(ipgre_fb_tunnel_dev);
195 static struct ip_tunnel **__ipgre_bucket(struct ip_tunnel_parm *parms)
197 __be32 remote = parms->iph.daddr;
198 __be32 local = parms->iph.saddr;
199 __be32 key = parms->i_key;
200 unsigned h = HASH(key);
205 if (remote && !ipv4_is_multicast(remote)) {
210 return &tunnels[prio][h];
213 static inline struct ip_tunnel **ipgre_bucket(struct ip_tunnel *t)
215 return __ipgre_bucket(&t->parms);
218 static void ipgre_tunnel_link(struct ip_tunnel *t)
220 struct ip_tunnel **tp = ipgre_bucket(t);
223 write_lock_bh(&ipgre_lock);
225 write_unlock_bh(&ipgre_lock);
228 static void ipgre_tunnel_unlink(struct ip_tunnel *t)
230 struct ip_tunnel **tp;
232 for (tp = ipgre_bucket(t); *tp; tp = &(*tp)->next) {
234 write_lock_bh(&ipgre_lock);
236 write_unlock_bh(&ipgre_lock);
242 static struct ip_tunnel * ipgre_tunnel_locate(struct ip_tunnel_parm *parms, int create)
244 __be32 remote = parms->iph.daddr;
245 __be32 local = parms->iph.saddr;
246 __be32 key = parms->i_key;
247 struct ip_tunnel *t, **tp, *nt;
248 struct net_device *dev;
251 for (tp = __ipgre_bucket(parms); (t = *tp) != NULL; tp = &t->next) {
252 if (local == t->parms.iph.saddr && remote == t->parms.iph.daddr) {
253 if (key == t->parms.i_key)
261 strlcpy(name, parms->name, IFNAMSIZ);
264 for (i=1; i<100; i++) {
265 sprintf(name, "gre%d", i);
266 if (__dev_get_by_name(&init_net, name) == NULL)
273 dev = alloc_netdev(sizeof(*t), name, ipgre_tunnel_setup);
277 dev->init = ipgre_tunnel_init;
278 nt = netdev_priv(dev);
281 if (register_netdevice(dev) < 0) {
287 ipgre_tunnel_link(nt);
294 static void ipgre_tunnel_uninit(struct net_device *dev)
296 ipgre_tunnel_unlink(netdev_priv(dev));
301 static void ipgre_err(struct sk_buff *skb, u32 info)
303 #ifndef I_WISH_WORLD_WERE_PERFECT
305 /* It is not :-( All the routers (except for Linux) return only
306 8 bytes of packet payload. It means, that precise relaying of
307 ICMP in the real Internet is absolutely infeasible.
309 Moreover, Cisco "wise men" put GRE key to the third word
310 in GRE header. It makes impossible maintaining even soft state for keyed
311 GRE tunnels with enabled checksum. Tell them "thank you".
313 Well, I wonder, rfc1812 was written by Cisco employee,
314 what the hell these idiots break standrads established
318 struct iphdr *iph = (struct iphdr*)skb->data;
319 __be16 *p = (__be16*)(skb->data+(iph->ihl<<2));
320 int grehlen = (iph->ihl<<2) + 4;
321 const int type = icmp_hdr(skb)->type;
322 const int code = icmp_hdr(skb)->code;
327 if (flags&(GRE_CSUM|GRE_KEY|GRE_SEQ|GRE_ROUTING|GRE_VERSION)) {
328 if (flags&(GRE_VERSION|GRE_ROUTING))
337 /* If only 8 bytes returned, keyed message will be dropped here */
338 if (skb_headlen(skb) < grehlen)
343 case ICMP_PARAMETERPROB:
346 case ICMP_DEST_UNREACH:
349 case ICMP_PORT_UNREACH:
350 /* Impossible event. */
352 case ICMP_FRAG_NEEDED:
353 /* Soft state for pmtu is maintained by IP core. */
356 /* All others are translated to HOST_UNREACH.
357 rfc2003 contains "deep thoughts" about NET_UNREACH,
358 I believe they are just ether pollution. --ANK
363 case ICMP_TIME_EXCEEDED:
364 if (code != ICMP_EXC_TTL)
369 read_lock(&ipgre_lock);
370 t = ipgre_tunnel_lookup(iph->daddr, iph->saddr, (flags&GRE_KEY) ? *(((__be32*)p) + (grehlen>>2) - 1) : 0);
371 if (t == NULL || t->parms.iph.daddr == 0 ||
372 ipv4_is_multicast(t->parms.iph.daddr))
375 if (t->parms.iph.ttl == 0 && type == ICMP_TIME_EXCEEDED)
378 if (jiffies - t->err_time < IPTUNNEL_ERR_TIMEO)
382 t->err_time = jiffies;
384 read_unlock(&ipgre_lock);
387 struct iphdr *iph = (struct iphdr*)dp;
389 __be16 *p = (__be16*)(dp+(iph->ihl<<2));
390 const int type = icmp_hdr(skb)->type;
391 const int code = icmp_hdr(skb)->code;
397 int grehlen = (iph->ihl<<2) + 4;
398 struct sk_buff *skb2;
402 if (p[1] != htons(ETH_P_IP))
406 if (flags&(GRE_CSUM|GRE_KEY|GRE_SEQ|GRE_ROUTING|GRE_VERSION)) {
407 if (flags&(GRE_VERSION|GRE_ROUTING))
416 if (len < grehlen + sizeof(struct iphdr))
418 eiph = (struct iphdr*)(dp + grehlen);
423 case ICMP_PARAMETERPROB:
424 n = ntohl(icmp_hdr(skb)->un.gateway) >> 24;
425 if (n < (iph->ihl<<2))
428 /* So... This guy found something strange INSIDE encapsulated
429 packet. Well, he is fool, but what can we do ?
431 rel_type = ICMP_PARAMETERPROB;
433 rel_info = htonl(n << 24);
436 case ICMP_DEST_UNREACH:
439 case ICMP_PORT_UNREACH:
440 /* Impossible event. */
442 case ICMP_FRAG_NEEDED:
443 /* And it is the only really necessary thing :-) */
444 n = ntohs(icmp_hdr(skb)->un.frag.mtu);
448 /* BSD 4.2 MORE DOES NOT EXIST IN NATURE. */
449 if (n > ntohs(eiph->tot_len))
454 /* All others are translated to HOST_UNREACH.
455 rfc2003 contains "deep thoughts" about NET_UNREACH,
456 I believe, it is just ether pollution. --ANK
458 rel_type = ICMP_DEST_UNREACH;
459 rel_code = ICMP_HOST_UNREACH;
463 case ICMP_TIME_EXCEEDED:
464 if (code != ICMP_EXC_TTL)
469 /* Prepare fake skb to feed it to icmp_send */
470 skb2 = skb_clone(skb, GFP_ATOMIC);
473 dst_release(skb2->dst);
475 skb_pull(skb2, skb->data - (u8*)eiph);
476 skb_reset_network_header(skb2);
478 /* Try to guess incoming interface */
479 memset(&fl, 0, sizeof(fl));
480 fl.fl4_dst = eiph->saddr;
481 fl.fl4_tos = RT_TOS(eiph->tos);
482 fl.proto = IPPROTO_GRE;
483 if (ip_route_output_key(&init_net, &rt, &fl)) {
487 skb2->dev = rt->u.dst.dev;
489 /* route "incoming" packet */
490 if (rt->rt_flags&RTCF_LOCAL) {
493 fl.fl4_dst = eiph->daddr;
494 fl.fl4_src = eiph->saddr;
495 fl.fl4_tos = eiph->tos;
496 if (ip_route_output_key(&init_net, &rt, &fl) ||
497 rt->u.dst.dev->type != ARPHRD_IPGRE) {
504 if (ip_route_input(skb2, eiph->daddr, eiph->saddr, eiph->tos, skb2->dev) ||
505 skb2->dst->dev->type != ARPHRD_IPGRE) {
511 /* change mtu on this route */
512 if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) {
513 if (n > dst_mtu(skb2->dst)) {
517 skb2->dst->ops->update_pmtu(skb2->dst, n);
518 } else if (type == ICMP_TIME_EXCEEDED) {
519 struct ip_tunnel *t = netdev_priv(skb2->dev);
520 if (t->parms.iph.ttl) {
521 rel_type = ICMP_DEST_UNREACH;
522 rel_code = ICMP_HOST_UNREACH;
526 icmp_send(skb2, rel_type, rel_code, rel_info);
531 static inline void ipgre_ecn_decapsulate(struct iphdr *iph, struct sk_buff *skb)
533 if (INET_ECN_is_ce(iph->tos)) {
534 if (skb->protocol == htons(ETH_P_IP)) {
535 IP_ECN_set_ce(ip_hdr(skb));
536 } else if (skb->protocol == htons(ETH_P_IPV6)) {
537 IP6_ECN_set_ce(ipv6_hdr(skb));
543 ipgre_ecn_encapsulate(u8 tos, struct iphdr *old_iph, struct sk_buff *skb)
546 if (skb->protocol == htons(ETH_P_IP))
547 inner = old_iph->tos;
548 else if (skb->protocol == htons(ETH_P_IPV6))
549 inner = ipv6_get_dsfield((struct ipv6hdr *)old_iph);
550 return INET_ECN_encapsulate(tos, inner);
553 static int ipgre_rcv(struct sk_buff *skb)
561 struct ip_tunnel *tunnel;
564 if (!pskb_may_pull(skb, 16))
571 if (flags&(GRE_CSUM|GRE_KEY|GRE_ROUTING|GRE_SEQ|GRE_VERSION)) {
572 /* - Version must be 0.
573 - We do not support routing headers.
575 if (flags&(GRE_VERSION|GRE_ROUTING))
578 if (flags&GRE_CSUM) {
579 switch (skb->ip_summed) {
580 case CHECKSUM_COMPLETE:
581 csum = csum_fold(skb->csum);
587 csum = __skb_checksum_complete(skb);
588 skb->ip_summed = CHECKSUM_COMPLETE;
593 key = *(__be32*)(h + offset);
597 seqno = ntohl(*(__be32*)(h + offset));
602 read_lock(&ipgre_lock);
603 if ((tunnel = ipgre_tunnel_lookup(iph->saddr, iph->daddr, key)) != NULL) {
606 skb->protocol = *(__be16*)(h + 2);
607 /* WCCP version 1 and 2 protocol decoding.
608 * - Change protocol to IP
609 * - When dealing with WCCPv2, Skip extra 4 bytes in GRE header
612 skb->protocol == htons(ETH_P_WCCP)) {
613 skb->protocol = htons(ETH_P_IP);
614 if ((*(h + offset) & 0xF0) != 0x40)
618 skb->mac_header = skb->network_header;
619 __pskb_pull(skb, offset);
620 skb_reset_network_header(skb);
621 skb_postpull_rcsum(skb, skb_transport_header(skb), offset);
622 skb->pkt_type = PACKET_HOST;
623 #ifdef CONFIG_NET_IPGRE_BROADCAST
624 if (ipv4_is_multicast(iph->daddr)) {
625 /* Looped back packet, drop it! */
626 if (((struct rtable*)skb->dst)->fl.iif == 0)
628 tunnel->stat.multicast++;
629 skb->pkt_type = PACKET_BROADCAST;
633 if (((flags&GRE_CSUM) && csum) ||
634 (!(flags&GRE_CSUM) && tunnel->parms.i_flags&GRE_CSUM)) {
635 tunnel->stat.rx_crc_errors++;
636 tunnel->stat.rx_errors++;
639 if (tunnel->parms.i_flags&GRE_SEQ) {
640 if (!(flags&GRE_SEQ) ||
641 (tunnel->i_seqno && (s32)(seqno - tunnel->i_seqno) < 0)) {
642 tunnel->stat.rx_fifo_errors++;
643 tunnel->stat.rx_errors++;
646 tunnel->i_seqno = seqno + 1;
648 tunnel->stat.rx_packets++;
649 tunnel->stat.rx_bytes += skb->len;
650 skb->dev = tunnel->dev;
651 dst_release(skb->dst);
654 ipgre_ecn_decapsulate(iph, skb);
656 read_unlock(&ipgre_lock);
659 icmp_send(skb, ICMP_DEST_UNREACH, ICMP_PORT_UNREACH, 0);
662 read_unlock(&ipgre_lock);
668 static int ipgre_tunnel_xmit(struct sk_buff *skb, struct net_device *dev)
670 struct ip_tunnel *tunnel = netdev_priv(dev);
671 struct net_device_stats *stats = &tunnel->stat;
672 struct iphdr *old_iph = ip_hdr(skb);
676 struct rtable *rt; /* Route to the other host */
677 struct net_device *tdev; /* Device to other host */
678 struct iphdr *iph; /* Our new IP header */
679 unsigned int max_headroom; /* The extra header space needed */
684 if (tunnel->recursion++) {
685 tunnel->stat.collisions++;
689 if (dev->header_ops) {
691 tiph = (struct iphdr*)skb->data;
693 gre_hlen = tunnel->hlen;
694 tiph = &tunnel->parms.iph;
697 if ((dst = tiph->daddr) == 0) {
700 if (skb->dst == NULL) {
701 tunnel->stat.tx_fifo_errors++;
705 if (skb->protocol == htons(ETH_P_IP)) {
706 rt = (struct rtable*)skb->dst;
707 if ((dst = rt->rt_gateway) == 0)
711 else if (skb->protocol == htons(ETH_P_IPV6)) {
712 struct in6_addr *addr6;
714 struct neighbour *neigh = skb->dst->neighbour;
719 addr6 = (struct in6_addr*)&neigh->primary_key;
720 addr_type = ipv6_addr_type(addr6);
722 if (addr_type == IPV6_ADDR_ANY) {
723 addr6 = &ipv6_hdr(skb)->daddr;
724 addr_type = ipv6_addr_type(addr6);
727 if ((addr_type & IPV6_ADDR_COMPATv4) == 0)
730 dst = addr6->s6_addr32[3];
739 if (skb->protocol == htons(ETH_P_IP))
745 struct flowi fl = { .oif = tunnel->parms.link,
748 .saddr = tiph->saddr,
749 .tos = RT_TOS(tos) } },
750 .proto = IPPROTO_GRE };
751 if (ip_route_output_key(&init_net, &rt, &fl)) {
752 tunnel->stat.tx_carrier_errors++;
756 tdev = rt->u.dst.dev;
760 tunnel->stat.collisions++;
766 mtu = dst_mtu(&rt->u.dst) - tunnel->hlen;
768 mtu = skb->dst ? dst_mtu(skb->dst) : dev->mtu;
771 skb->dst->ops->update_pmtu(skb->dst, mtu);
773 if (skb->protocol == htons(ETH_P_IP)) {
774 df |= (old_iph->frag_off&htons(IP_DF));
776 if ((old_iph->frag_off&htons(IP_DF)) &&
777 mtu < ntohs(old_iph->tot_len)) {
778 icmp_send(skb, ICMP_DEST_UNREACH, ICMP_FRAG_NEEDED, htonl(mtu));
784 else if (skb->protocol == htons(ETH_P_IPV6)) {
785 struct rt6_info *rt6 = (struct rt6_info*)skb->dst;
787 if (rt6 && mtu < dst_mtu(skb->dst) && mtu >= IPV6_MIN_MTU) {
788 if ((tunnel->parms.iph.daddr &&
789 !ipv4_is_multicast(tunnel->parms.iph.daddr)) ||
790 rt6->rt6i_dst.plen == 128) {
791 rt6->rt6i_flags |= RTF_MODIFIED;
792 skb->dst->metrics[RTAX_MTU-1] = mtu;
796 if (mtu >= IPV6_MIN_MTU && mtu < skb->len - tunnel->hlen + gre_hlen) {
797 icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, dev);
804 if (tunnel->err_count > 0) {
805 if (jiffies - tunnel->err_time < IPTUNNEL_ERR_TIMEO) {
808 dst_link_failure(skb);
810 tunnel->err_count = 0;
813 max_headroom = LL_RESERVED_SPACE(tdev) + gre_hlen;
815 if (skb_headroom(skb) < max_headroom || skb_shared(skb)||
816 (skb_cloned(skb) && !skb_clone_writable(skb, 0))) {
817 struct sk_buff *new_skb = skb_realloc_headroom(skb, max_headroom);
826 skb_set_owner_w(new_skb, skb->sk);
829 old_iph = ip_hdr(skb);
832 skb->transport_header = skb->network_header;
833 skb_push(skb, gre_hlen);
834 skb_reset_network_header(skb);
835 memset(&(IPCB(skb)->opt), 0, sizeof(IPCB(skb)->opt));
836 IPCB(skb)->flags &= ~(IPSKB_XFRM_TUNNEL_SIZE | IPSKB_XFRM_TRANSFORMED |
838 dst_release(skb->dst);
839 skb->dst = &rt->u.dst;
842 * Push down and install the IPIP header.
847 iph->ihl = sizeof(struct iphdr) >> 2;
849 iph->protocol = IPPROTO_GRE;
850 iph->tos = ipgre_ecn_encapsulate(tos, old_iph, skb);
851 iph->daddr = rt->rt_dst;
852 iph->saddr = rt->rt_src;
854 if ((iph->ttl = tiph->ttl) == 0) {
855 if (skb->protocol == htons(ETH_P_IP))
856 iph->ttl = old_iph->ttl;
858 else if (skb->protocol == htons(ETH_P_IPV6))
859 iph->ttl = ((struct ipv6hdr*)old_iph)->hop_limit;
862 iph->ttl = dst_metric(&rt->u.dst, RTAX_HOPLIMIT);
865 ((__be16*)(iph+1))[0] = tunnel->parms.o_flags;
866 ((__be16*)(iph+1))[1] = skb->protocol;
868 if (tunnel->parms.o_flags&(GRE_KEY|GRE_CSUM|GRE_SEQ)) {
869 __be32 *ptr = (__be32*)(((u8*)iph) + tunnel->hlen - 4);
871 if (tunnel->parms.o_flags&GRE_SEQ) {
873 *ptr = htonl(tunnel->o_seqno);
876 if (tunnel->parms.o_flags&GRE_KEY) {
877 *ptr = tunnel->parms.o_key;
880 if (tunnel->parms.o_flags&GRE_CSUM) {
882 *(__sum16*)ptr = ip_compute_csum((void*)(iph+1), skb->len - sizeof(struct iphdr));
893 dst_link_failure(skb);
902 static void ipgre_tunnel_bind_dev(struct net_device *dev)
904 struct net_device *tdev = NULL;
905 struct ip_tunnel *tunnel;
907 int hlen = LL_MAX_HEADER;
908 int mtu = ETH_DATA_LEN;
909 int addend = sizeof(struct iphdr) + 4;
911 tunnel = netdev_priv(dev);
912 iph = &tunnel->parms.iph;
914 /* Guess output device to choose reasonable mtu and hard_header_len */
917 struct flowi fl = { .oif = tunnel->parms.link,
919 { .daddr = iph->daddr,
921 .tos = RT_TOS(iph->tos) } },
922 .proto = IPPROTO_GRE };
924 if (!ip_route_output_key(&init_net, &rt, &fl)) {
925 tdev = rt->u.dst.dev;
928 dev->flags |= IFF_POINTOPOINT;
931 if (!tdev && tunnel->parms.link)
932 tdev = __dev_get_by_index(&init_net, tunnel->parms.link);
935 hlen = tdev->hard_header_len;
938 dev->iflink = tunnel->parms.link;
940 /* Precalculate GRE options length */
941 if (tunnel->parms.o_flags&(GRE_CSUM|GRE_KEY|GRE_SEQ)) {
942 if (tunnel->parms.o_flags&GRE_CSUM)
944 if (tunnel->parms.o_flags&GRE_KEY)
946 if (tunnel->parms.o_flags&GRE_SEQ)
949 dev->hard_header_len = hlen + addend;
950 dev->mtu = mtu - addend;
951 tunnel->hlen = addend;
956 ipgre_tunnel_ioctl (struct net_device *dev, struct ifreq *ifr, int cmd)
959 struct ip_tunnel_parm p;
965 if (dev == ipgre_fb_tunnel_dev) {
966 if (copy_from_user(&p, ifr->ifr_ifru.ifru_data, sizeof(p))) {
970 t = ipgre_tunnel_locate(&p, 0);
973 t = netdev_priv(dev);
974 memcpy(&p, &t->parms, sizeof(p));
975 if (copy_to_user(ifr->ifr_ifru.ifru_data, &p, sizeof(p)))
982 if (!capable(CAP_NET_ADMIN))
986 if (copy_from_user(&p, ifr->ifr_ifru.ifru_data, sizeof(p)))
990 if (p.iph.version != 4 || p.iph.protocol != IPPROTO_GRE ||
991 p.iph.ihl != 5 || (p.iph.frag_off&htons(~IP_DF)) ||
992 ((p.i_flags|p.o_flags)&(GRE_VERSION|GRE_ROUTING)))
995 p.iph.frag_off |= htons(IP_DF);
997 if (!(p.i_flags&GRE_KEY))
999 if (!(p.o_flags&GRE_KEY))
1002 t = ipgre_tunnel_locate(&p, cmd == SIOCADDTUNNEL);
1004 if (dev != ipgre_fb_tunnel_dev && cmd == SIOCCHGTUNNEL) {
1006 if (t->dev != dev) {
1013 t = netdev_priv(dev);
1015 if (ipv4_is_multicast(p.iph.daddr))
1016 nflags = IFF_BROADCAST;
1017 else if (p.iph.daddr)
1018 nflags = IFF_POINTOPOINT;
1020 if ((dev->flags^nflags)&(IFF_POINTOPOINT|IFF_BROADCAST)) {
1024 ipgre_tunnel_unlink(t);
1025 t->parms.iph.saddr = p.iph.saddr;
1026 t->parms.iph.daddr = p.iph.daddr;
1027 t->parms.i_key = p.i_key;
1028 t->parms.o_key = p.o_key;
1029 memcpy(dev->dev_addr, &p.iph.saddr, 4);
1030 memcpy(dev->broadcast, &p.iph.daddr, 4);
1031 ipgre_tunnel_link(t);
1032 netdev_state_change(dev);
1038 if (cmd == SIOCCHGTUNNEL) {
1039 t->parms.iph.ttl = p.iph.ttl;
1040 t->parms.iph.tos = p.iph.tos;
1041 t->parms.iph.frag_off = p.iph.frag_off;
1042 if (t->parms.link != p.link) {
1043 t->parms.link = p.link;
1044 ipgre_tunnel_bind_dev(dev);
1045 netdev_state_change(dev);
1048 if (copy_to_user(ifr->ifr_ifru.ifru_data, &t->parms, sizeof(p)))
1051 err = (cmd == SIOCADDTUNNEL ? -ENOBUFS : -ENOENT);
1056 if (!capable(CAP_NET_ADMIN))
1059 if (dev == ipgre_fb_tunnel_dev) {
1061 if (copy_from_user(&p, ifr->ifr_ifru.ifru_data, sizeof(p)))
1064 if ((t = ipgre_tunnel_locate(&p, 0)) == NULL)
1067 if (t == netdev_priv(ipgre_fb_tunnel_dev))
1071 unregister_netdevice(dev);
1083 static struct net_device_stats *ipgre_tunnel_get_stats(struct net_device *dev)
1085 return &(((struct ip_tunnel*)netdev_priv(dev))->stat);
1088 static int ipgre_tunnel_change_mtu(struct net_device *dev, int new_mtu)
1090 struct ip_tunnel *tunnel = netdev_priv(dev);
1091 if (new_mtu < 68 || new_mtu > 0xFFF8 - tunnel->hlen)
1097 /* Nice toy. Unfortunately, useless in real life :-)
1098 It allows to construct virtual multiprotocol broadcast "LAN"
1099 over the Internet, provided multicast routing is tuned.
1102 I have no idea was this bicycle invented before me,
1103 so that I had to set ARPHRD_IPGRE to a random value.
1104 I have an impression, that Cisco could make something similar,
1105 but this feature is apparently missing in IOS<=11.2(8).
1107 I set up 10.66.66/24 and fec0:6666:6666::0/96 as virtual networks
1108 with broadcast 224.66.66.66. If you have access to mbone, play with me :-)
1110 ping -t 255 224.66.66.66
1112 If nobody answers, mbone does not work.
1114 ip tunnel add Universe mode gre remote 224.66.66.66 local <Your_real_addr> ttl 255
1115 ip addr add 10.66.66.<somewhat>/24 dev Universe
1116 ifconfig Universe up
1117 ifconfig Universe add fe80::<Your_real_addr>/10
1118 ifconfig Universe add fec0:6666:6666::<Your_real_addr>/96
1121 ftp fec0:6666:6666::193.233.7.65
1126 static int ipgre_header(struct sk_buff *skb, struct net_device *dev,
1127 unsigned short type,
1128 const void *daddr, const void *saddr, unsigned len)
1130 struct ip_tunnel *t = netdev_priv(dev);
1131 struct iphdr *iph = (struct iphdr *)skb_push(skb, t->hlen);
1132 __be16 *p = (__be16*)(iph+1);
1134 memcpy(iph, &t->parms.iph, sizeof(struct iphdr));
1135 p[0] = t->parms.o_flags;
1139 * Set the source hardware address.
1143 memcpy(&iph->saddr, saddr, 4);
1146 memcpy(&iph->daddr, daddr, 4);
1149 if (iph->daddr && !ipv4_is_multicast(iph->daddr))
1155 static int ipgre_header_parse(const struct sk_buff *skb, unsigned char *haddr)
1157 struct iphdr *iph = (struct iphdr*) skb_mac_header(skb);
1158 memcpy(haddr, &iph->saddr, 4);
1162 static const struct header_ops ipgre_header_ops = {
1163 .create = ipgre_header,
1164 .parse = ipgre_header_parse,
1167 #ifdef CONFIG_NET_IPGRE_BROADCAST
1168 static int ipgre_open(struct net_device *dev)
1170 struct ip_tunnel *t = netdev_priv(dev);
1172 if (ipv4_is_multicast(t->parms.iph.daddr)) {
1173 struct flowi fl = { .oif = t->parms.link,
1175 { .daddr = t->parms.iph.daddr,
1176 .saddr = t->parms.iph.saddr,
1177 .tos = RT_TOS(t->parms.iph.tos) } },
1178 .proto = IPPROTO_GRE };
1180 if (ip_route_output_key(&init_net, &rt, &fl))
1181 return -EADDRNOTAVAIL;
1182 dev = rt->u.dst.dev;
1184 if (__in_dev_get_rtnl(dev) == NULL)
1185 return -EADDRNOTAVAIL;
1186 t->mlink = dev->ifindex;
1187 ip_mc_inc_group(__in_dev_get_rtnl(dev), t->parms.iph.daddr);
1192 static int ipgre_close(struct net_device *dev)
1194 struct ip_tunnel *t = netdev_priv(dev);
1195 if (ipv4_is_multicast(t->parms.iph.daddr) && t->mlink) {
1196 struct in_device *in_dev;
1197 in_dev = inetdev_by_index(dev->nd_net, t->mlink);
1199 ip_mc_dec_group(in_dev, t->parms.iph.daddr);
1208 static void ipgre_tunnel_setup(struct net_device *dev)
1210 dev->uninit = ipgre_tunnel_uninit;
1211 dev->destructor = free_netdev;
1212 dev->hard_start_xmit = ipgre_tunnel_xmit;
1213 dev->get_stats = ipgre_tunnel_get_stats;
1214 dev->do_ioctl = ipgre_tunnel_ioctl;
1215 dev->change_mtu = ipgre_tunnel_change_mtu;
1217 dev->type = ARPHRD_IPGRE;
1218 dev->hard_header_len = LL_MAX_HEADER + sizeof(struct iphdr) + 4;
1219 dev->mtu = ETH_DATA_LEN - sizeof(struct iphdr) - 4;
1220 dev->flags = IFF_NOARP;
1225 static int ipgre_tunnel_init(struct net_device *dev)
1227 struct ip_tunnel *tunnel;
1230 tunnel = netdev_priv(dev);
1231 iph = &tunnel->parms.iph;
1234 strcpy(tunnel->parms.name, dev->name);
1236 memcpy(dev->dev_addr, &tunnel->parms.iph.saddr, 4);
1237 memcpy(dev->broadcast, &tunnel->parms.iph.daddr, 4);
1239 ipgre_tunnel_bind_dev(dev);
1242 #ifdef CONFIG_NET_IPGRE_BROADCAST
1243 if (ipv4_is_multicast(iph->daddr)) {
1246 dev->flags = IFF_BROADCAST;
1247 dev->header_ops = &ipgre_header_ops;
1248 dev->open = ipgre_open;
1249 dev->stop = ipgre_close;
1253 dev->header_ops = &ipgre_header_ops;
1258 static int __init ipgre_fb_tunnel_init(struct net_device *dev)
1260 struct ip_tunnel *tunnel = netdev_priv(dev);
1261 struct iphdr *iph = &tunnel->parms.iph;
1264 strcpy(tunnel->parms.name, dev->name);
1267 iph->protocol = IPPROTO_GRE;
1269 tunnel->hlen = sizeof(struct iphdr) + 4;
1272 tunnels_wc[0] = tunnel;
1277 static struct net_protocol ipgre_protocol = {
1278 .handler = ipgre_rcv,
1279 .err_handler = ipgre_err,
1284 * And now the modules code and kernel interface.
1287 static int __init ipgre_init(void)
1291 printk(KERN_INFO "GRE over IPv4 tunneling driver\n");
1293 if (inet_add_protocol(&ipgre_protocol, IPPROTO_GRE) < 0) {
1294 printk(KERN_INFO "ipgre init: can't add protocol\n");
1298 ipgre_fb_tunnel_dev = alloc_netdev(sizeof(struct ip_tunnel), "gre0",
1299 ipgre_tunnel_setup);
1300 if (!ipgre_fb_tunnel_dev) {
1305 ipgre_fb_tunnel_dev->init = ipgre_fb_tunnel_init;
1307 if ((err = register_netdev(ipgre_fb_tunnel_dev)))
1312 free_netdev(ipgre_fb_tunnel_dev);
1314 inet_del_protocol(&ipgre_protocol, IPPROTO_GRE);
1318 static void __exit ipgre_destroy_tunnels(void)
1322 for (prio = 0; prio < 4; prio++) {
1324 for (h = 0; h < HASH_SIZE; h++) {
1325 struct ip_tunnel *t;
1326 while ((t = tunnels[prio][h]) != NULL)
1327 unregister_netdevice(t->dev);
1332 static void __exit ipgre_fini(void)
1334 if (inet_del_protocol(&ipgre_protocol, IPPROTO_GRE) < 0)
1335 printk(KERN_INFO "ipgre close: can't remove protocol\n");
1338 ipgre_destroy_tunnels();
1342 module_init(ipgre_init);
1343 module_exit(ipgre_fini);
1344 MODULE_LICENSE("GPL");