]>
Commit | Line | Data |
---|---|---|
be6b635c AB |
1 | /* |
2 | * Copyright (c) 2011 Patrick McHardy <kaber@trash.net> | |
3 | * | |
4 | * This program is free software; you can redistribute it and/or modify | |
5 | * it under the terms of the GNU General Public License version 2 as | |
6 | * published by the Free Software Foundation. | |
7 | * | |
8 | * Based on Rusty Russell's IPv6 MASQUERADE target. Development of IPv6 | |
9 | * NAT funded by Astaro. | |
10 | */ | |
11 | ||
12 | #include <linux/kernel.h> | |
13 | #include <linux/module.h> | |
14 | #include <linux/atomic.h> | |
15 | #include <linux/netdevice.h> | |
16 | #include <linux/ipv6.h> | |
17 | #include <linux/netfilter.h> | |
18 | #include <linux/netfilter_ipv6.h> | |
19 | #include <net/netfilter/nf_nat.h> | |
20 | #include <net/addrconf.h> | |
21 | #include <net/ipv6.h> | |
22 | #include <net/netfilter/ipv6/nf_nat_masquerade.h> | |
23 | ||
d93c6258 FW |
24 | #define MAX_WORK_COUNT 16 |
25 | ||
26 | static atomic_t v6_worker_count; | |
27 | ||
be6b635c AB |
28 | unsigned int |
29 | nf_nat_masquerade_ipv6(struct sk_buff *skb, const struct nf_nat_range *range, | |
30 | const struct net_device *out) | |
31 | { | |
32 | enum ip_conntrack_info ctinfo; | |
ff459018 | 33 | struct nf_conn_nat *nat; |
be6b635c AB |
34 | struct in6_addr src; |
35 | struct nf_conn *ct; | |
36 | struct nf_nat_range newrange; | |
37 | ||
38 | ct = nf_ct_get(skb, &ctinfo); | |
39 | NF_CT_ASSERT(ct && (ctinfo == IP_CT_NEW || ctinfo == IP_CT_RELATED || | |
40 | ctinfo == IP_CT_RELATED_REPLY)); | |
41 | ||
0a031ac5 | 42 | if (ipv6_dev_get_saddr(nf_ct_net(ct), out, |
be6b635c AB |
43 | &ipv6_hdr(skb)->daddr, 0, &src) < 0) |
44 | return NF_DROP; | |
45 | ||
ff459018 FW |
46 | nat = nf_ct_nat_ext_add(ct); |
47 | if (nat) | |
48 | nat->masq_index = out->ifindex; | |
be6b635c AB |
49 | |
50 | newrange.flags = range->flags | NF_NAT_RANGE_MAP_IPS; | |
51 | newrange.min_addr.in6 = src; | |
52 | newrange.max_addr.in6 = src; | |
53 | newrange.min_proto = range->min_proto; | |
54 | newrange.max_proto = range->max_proto; | |
55 | ||
56 | return nf_nat_setup_info(ct, &newrange, NF_NAT_MANIP_SRC); | |
57 | } | |
58 | EXPORT_SYMBOL_GPL(nf_nat_masquerade_ipv6); | |
59 | ||
60 | static int device_cmp(struct nf_conn *ct, void *ifindex) | |
61 | { | |
62 | const struct nf_conn_nat *nat = nfct_nat(ct); | |
63 | ||
64 | if (!nat) | |
65 | return 0; | |
66 | if (nf_ct_l3num(ct) != NFPROTO_IPV6) | |
67 | return 0; | |
68 | return nat->masq_index == (int)(long)ifindex; | |
69 | } | |
70 | ||
71 | static int masq_device_event(struct notifier_block *this, | |
72 | unsigned long event, void *ptr) | |
73 | { | |
74 | const struct net_device *dev = netdev_notifier_info_to_dev(ptr); | |
75 | struct net *net = dev_net(dev); | |
76 | ||
77 | if (event == NETDEV_DOWN) | |
78 | nf_ct_iterate_cleanup(net, device_cmp, | |
79 | (void *)(long)dev->ifindex, 0, 0); | |
80 | ||
81 | return NOTIFY_DONE; | |
82 | } | |
83 | ||
84 | static struct notifier_block masq_dev_notifier = { | |
85 | .notifier_call = masq_device_event, | |
86 | }; | |
87 | ||
d93c6258 FW |
88 | struct masq_dev_work { |
89 | struct work_struct work; | |
90 | struct net *net; | |
91 | int ifindex; | |
92 | }; | |
93 | ||
94 | static void iterate_cleanup_work(struct work_struct *work) | |
95 | { | |
96 | struct masq_dev_work *w; | |
97 | long index; | |
98 | ||
99 | w = container_of(work, struct masq_dev_work, work); | |
100 | ||
101 | index = w->ifindex; | |
102 | nf_ct_iterate_cleanup(w->net, device_cmp, (void *)index, 0, 0); | |
103 | ||
104 | put_net(w->net); | |
105 | kfree(w); | |
106 | atomic_dec(&v6_worker_count); | |
107 | module_put(THIS_MODULE); | |
108 | } | |
109 | ||
110 | /* ipv6 inet notifier is an atomic notifier, i.e. we cannot | |
111 | * schedule. | |
112 | * | |
113 | * Unfortunately, nf_ct_iterate_cleanup can run for a long | |
114 | * time if there are lots of conntracks and the system | |
115 | * handles high softirq load, so it frequently calls cond_resched | |
116 | * while iterating the conntrack table. | |
117 | * | |
118 | * So we defer nf_ct_iterate_cleanup walk to the system workqueue. | |
119 | * | |
120 | * As we can have 'a lot' of inet_events (depending on amount | |
121 | * of ipv6 addresses being deleted), we also need to add an upper | |
122 | * limit to the number of queued work items. | |
123 | */ | |
be6b635c AB |
124 | static int masq_inet_event(struct notifier_block *this, |
125 | unsigned long event, void *ptr) | |
126 | { | |
127 | struct inet6_ifaddr *ifa = ptr; | |
d93c6258 FW |
128 | const struct net_device *dev; |
129 | struct masq_dev_work *w; | |
130 | struct net *net; | |
131 | ||
132 | if (event != NETDEV_DOWN || | |
133 | atomic_read(&v6_worker_count) >= MAX_WORK_COUNT) | |
134 | return NOTIFY_DONE; | |
135 | ||
136 | dev = ifa->idev->dev; | |
137 | net = maybe_get_net(dev_net(dev)); | |
138 | if (!net) | |
139 | return NOTIFY_DONE; | |
be6b635c | 140 | |
d93c6258 FW |
141 | if (!try_module_get(THIS_MODULE)) |
142 | goto err_module; | |
143 | ||
144 | w = kmalloc(sizeof(*w), GFP_ATOMIC); | |
145 | if (w) { | |
146 | atomic_inc(&v6_worker_count); | |
147 | ||
148 | INIT_WORK(&w->work, iterate_cleanup_work); | |
149 | w->ifindex = dev->ifindex; | |
150 | w->net = net; | |
151 | schedule_work(&w->work); | |
152 | ||
153 | return NOTIFY_DONE; | |
154 | } | |
155 | ||
156 | module_put(THIS_MODULE); | |
157 | err_module: | |
158 | put_net(net); | |
159 | return NOTIFY_DONE; | |
be6b635c AB |
160 | } |
161 | ||
162 | static struct notifier_block masq_inet_notifier = { | |
163 | .notifier_call = masq_inet_event, | |
164 | }; | |
165 | ||
166 | static atomic_t masquerade_notifier_refcount = ATOMIC_INIT(0); | |
167 | ||
168 | void nf_nat_masquerade_ipv6_register_notifier(void) | |
169 | { | |
170 | /* check if the notifier is already set */ | |
171 | if (atomic_inc_return(&masquerade_notifier_refcount) > 1) | |
172 | return; | |
173 | ||
174 | register_netdevice_notifier(&masq_dev_notifier); | |
175 | register_inet6addr_notifier(&masq_inet_notifier); | |
176 | } | |
177 | EXPORT_SYMBOL_GPL(nf_nat_masquerade_ipv6_register_notifier); | |
178 | ||
179 | void nf_nat_masquerade_ipv6_unregister_notifier(void) | |
180 | { | |
181 | /* check if the notifier still has clients */ | |
182 | if (atomic_dec_return(&masquerade_notifier_refcount) > 0) | |
183 | return; | |
184 | ||
185 | unregister_inet6addr_notifier(&masq_inet_notifier); | |
186 | unregister_netdevice_notifier(&masq_dev_notifier); | |
187 | } | |
188 | EXPORT_SYMBOL_GPL(nf_nat_masquerade_ipv6_unregister_notifier); | |
189 | ||
190 | MODULE_LICENSE("GPL"); | |
191 | MODULE_AUTHOR("Patrick McHardy <kaber@trash.net>"); |