]> git.proxmox.com Git - mirror_ubuntu-zesty-kernel.git/blame - net/ipv6/tcp_ipv6.c
genetlink: Optimize and one bug fix in genl_generate_id()
[mirror_ubuntu-zesty-kernel.git] / net / ipv6 / tcp_ipv6.c
CommitLineData
1da177e4
LT
1/*
2 * TCP over IPv6
1ab1457c 3 * Linux INET6 implementation
1da177e4
LT
4 *
5 * Authors:
1ab1457c 6 * Pedro Roque <roque@di.fc.ul.pt>
1da177e4 7 *
1ab1457c 8 * Based on:
1da177e4
LT
9 * linux/net/ipv4/tcp.c
10 * linux/net/ipv4/tcp_input.c
11 * linux/net/ipv4/tcp_output.c
12 *
13 * Fixes:
14 * Hideaki YOSHIFUJI : sin6_scope_id support
15 * YOSHIFUJI Hideaki @USAGI and: Support IPV6_V6ONLY socket option, which
16 * Alexey Kuznetsov allow both IPv4 and IPv6 sockets to bind
17 * a single port at the same time.
18 * YOSHIFUJI Hideaki @USAGI: convert /proc/net/tcp6 to seq_file.
19 *
20 * This program is free software; you can redistribute it and/or
21 * modify it under the terms of the GNU General Public License
22 * as published by the Free Software Foundation; either version
23 * 2 of the License, or (at your option) any later version.
24 */
25
eb4dea58 26#include <linux/bottom_half.h>
1da177e4 27#include <linux/module.h>
1da177e4
LT
28#include <linux/errno.h>
29#include <linux/types.h>
30#include <linux/socket.h>
31#include <linux/sockios.h>
32#include <linux/net.h>
33#include <linux/jiffies.h>
34#include <linux/in.h>
35#include <linux/in6.h>
36#include <linux/netdevice.h>
37#include <linux/init.h>
38#include <linux/jhash.h>
39#include <linux/ipsec.h>
40#include <linux/times.h>
41
42#include <linux/ipv6.h>
43#include <linux/icmpv6.h>
44#include <linux/random.h>
45
46#include <net/tcp.h>
47#include <net/ndisc.h>
5324a040 48#include <net/inet6_hashtables.h>
8129765a 49#include <net/inet6_connection_sock.h>
1da177e4
LT
50#include <net/ipv6.h>
51#include <net/transp_v6.h>
52#include <net/addrconf.h>
53#include <net/ip6_route.h>
54#include <net/ip6_checksum.h>
55#include <net/inet_ecn.h>
56#include <net/protocol.h>
57#include <net/xfrm.h>
1da177e4
LT
58#include <net/snmp.h>
59#include <net/dsfield.h>
6d6ee43e 60#include <net/timewait_sock.h>
18134bed 61#include <net/netdma.h>
3d58b5fa 62#include <net/inet_common.h>
1da177e4
LT
63
64#include <asm/uaccess.h>
65
66#include <linux/proc_fs.h>
67#include <linux/seq_file.h>
68
cfb6eeb4
YH
69#include <linux/crypto.h>
70#include <linux/scatterlist.h>
71
cfb6eeb4 72static void tcp_v6_send_reset(struct sock *sk, struct sk_buff *skb);
6edafaaf
GJ
73static void tcp_v6_reqsk_send_ack(struct sock *sk, struct sk_buff *skb,
74 struct request_sock *req);
1da177e4
LT
75
76static int tcp_v6_do_rcv(struct sock *sk, struct sk_buff *skb);
1da177e4 77
3b401a81
SH
78static const struct inet_connection_sock_af_ops ipv6_mapped;
79static const struct inet_connection_sock_af_ops ipv6_specific;
a928630a 80#ifdef CONFIG_TCP_MD5SIG
b2e4b3de
SH
81static const struct tcp_sock_af_ops tcp_sock_ipv6_specific;
82static const struct tcp_sock_af_ops tcp_sock_ipv6_mapped_specific;
9501f972
YH
83#else
84static struct tcp_md5sig_key *tcp_v6_md5_do_lookup(struct sock *sk,
85 struct in6_addr *addr)
86{
87 return NULL;
88}
a928630a 89#endif
1da177e4 90
1da177e4
LT
91static void tcp_v6_hash(struct sock *sk)
92{
93 if (sk->sk_state != TCP_CLOSE) {
8292a17a 94 if (inet_csk(sk)->icsk_af_ops == &ipv6_mapped) {
1da177e4
LT
95 tcp_prot.hash(sk);
96 return;
97 }
98 local_bh_disable();
ab1e0a13 99 __inet6_hash(sk);
1da177e4
LT
100 local_bh_enable();
101 }
102}
103
684f2176 104static __inline__ __sum16 tcp_v6_check(int len,
1ab1457c
YH
105 struct in6_addr *saddr,
106 struct in6_addr *daddr,
868c86bc 107 __wsum base)
1da177e4
LT
108{
109 return csum_ipv6_magic(saddr, daddr, len, IPPROTO_TCP, base);
110}
111
a94f723d 112static __u32 tcp_v6_init_sequence(struct sk_buff *skb)
1da177e4 113{
0660e03f
ACM
114 return secure_tcpv6_sequence_number(ipv6_hdr(skb)->daddr.s6_addr32,
115 ipv6_hdr(skb)->saddr.s6_addr32,
aa8223c7
ACM
116 tcp_hdr(skb)->dest,
117 tcp_hdr(skb)->source);
1da177e4
LT
118}
119
1ab1457c 120static int tcp_v6_connect(struct sock *sk, struct sockaddr *uaddr,
1da177e4
LT
121 int addr_len)
122{
123 struct sockaddr_in6 *usin = (struct sockaddr_in6 *) uaddr;
1ab1457c 124 struct inet_sock *inet = inet_sk(sk);
d83d8461 125 struct inet_connection_sock *icsk = inet_csk(sk);
1da177e4
LT
126 struct ipv6_pinfo *np = inet6_sk(sk);
127 struct tcp_sock *tp = tcp_sk(sk);
128 struct in6_addr *saddr = NULL, *final_p = NULL, final;
129 struct flowi fl;
130 struct dst_entry *dst;
131 int addr_type;
132 int err;
133
1ab1457c 134 if (addr_len < SIN6_LEN_RFC2133)
1da177e4
LT
135 return -EINVAL;
136
1ab1457c 137 if (usin->sin6_family != AF_INET6)
1da177e4
LT
138 return(-EAFNOSUPPORT);
139
140 memset(&fl, 0, sizeof(fl));
141
142 if (np->sndflow) {
143 fl.fl6_flowlabel = usin->sin6_flowinfo&IPV6_FLOWINFO_MASK;
144 IP6_ECN_flow_init(fl.fl6_flowlabel);
145 if (fl.fl6_flowlabel&IPV6_FLOWLABEL_MASK) {
146 struct ip6_flowlabel *flowlabel;
147 flowlabel = fl6_sock_lookup(sk, fl.fl6_flowlabel);
148 if (flowlabel == NULL)
149 return -EINVAL;
150 ipv6_addr_copy(&usin->sin6_addr, &flowlabel->dst);
151 fl6_sock_release(flowlabel);
152 }
153 }
154
155 /*
1ab1457c
YH
156 * connect() to INADDR_ANY means loopback (BSD'ism).
157 */
158
159 if(ipv6_addr_any(&usin->sin6_addr))
160 usin->sin6_addr.s6_addr[15] = 0x1;
1da177e4
LT
161
162 addr_type = ipv6_addr_type(&usin->sin6_addr);
163
164 if(addr_type & IPV6_ADDR_MULTICAST)
165 return -ENETUNREACH;
166
167 if (addr_type&IPV6_ADDR_LINKLOCAL) {
168 if (addr_len >= sizeof(struct sockaddr_in6) &&
169 usin->sin6_scope_id) {
170 /* If interface is set while binding, indices
171 * must coincide.
172 */
173 if (sk->sk_bound_dev_if &&
174 sk->sk_bound_dev_if != usin->sin6_scope_id)
175 return -EINVAL;
176
177 sk->sk_bound_dev_if = usin->sin6_scope_id;
178 }
179
180 /* Connect to link-local address requires an interface */
181 if (!sk->sk_bound_dev_if)
182 return -EINVAL;
183 }
184
185 if (tp->rx_opt.ts_recent_stamp &&
186 !ipv6_addr_equal(&np->daddr, &usin->sin6_addr)) {
187 tp->rx_opt.ts_recent = 0;
188 tp->rx_opt.ts_recent_stamp = 0;
189 tp->write_seq = 0;
190 }
191
192 ipv6_addr_copy(&np->daddr, &usin->sin6_addr);
193 np->flow_label = fl.fl6_flowlabel;
194
195 /*
196 * TCP over IPv4
197 */
198
199 if (addr_type == IPV6_ADDR_MAPPED) {
d83d8461 200 u32 exthdrlen = icsk->icsk_ext_hdr_len;
1da177e4
LT
201 struct sockaddr_in sin;
202
203 SOCK_DEBUG(sk, "connect: ipv4 mapped\n");
204
205 if (__ipv6_only_sock(sk))
206 return -ENETUNREACH;
207
208 sin.sin_family = AF_INET;
209 sin.sin_port = usin->sin6_port;
210 sin.sin_addr.s_addr = usin->sin6_addr.s6_addr32[3];
211
d83d8461 212 icsk->icsk_af_ops = &ipv6_mapped;
1da177e4 213 sk->sk_backlog_rcv = tcp_v4_do_rcv;
cfb6eeb4
YH
214#ifdef CONFIG_TCP_MD5SIG
215 tp->af_specific = &tcp_sock_ipv6_mapped_specific;
216#endif
1da177e4
LT
217
218 err = tcp_v4_connect(sk, (struct sockaddr *)&sin, sizeof(sin));
219
220 if (err) {
d83d8461
ACM
221 icsk->icsk_ext_hdr_len = exthdrlen;
222 icsk->icsk_af_ops = &ipv6_specific;
1da177e4 223 sk->sk_backlog_rcv = tcp_v6_do_rcv;
cfb6eeb4
YH
224#ifdef CONFIG_TCP_MD5SIG
225 tp->af_specific = &tcp_sock_ipv6_specific;
226#endif
1da177e4
LT
227 goto failure;
228 } else {
b301e82c
BH
229 ipv6_addr_set_v4mapped(inet->saddr, &np->saddr);
230 ipv6_addr_set_v4mapped(inet->rcv_saddr, &np->rcv_saddr);
1da177e4
LT
231 }
232
233 return err;
234 }
235
236 if (!ipv6_addr_any(&np->rcv_saddr))
237 saddr = &np->rcv_saddr;
238
239 fl.proto = IPPROTO_TCP;
240 ipv6_addr_copy(&fl.fl6_dst, &np->daddr);
241 ipv6_addr_copy(&fl.fl6_src,
242 (saddr ? saddr : &np->saddr));
243 fl.oif = sk->sk_bound_dev_if;
51953d5b 244 fl.mark = sk->sk_mark;
1da177e4
LT
245 fl.fl_ip_dport = usin->sin6_port;
246 fl.fl_ip_sport = inet->sport;
247
248 if (np->opt && np->opt->srcrt) {
249 struct rt0_hdr *rt0 = (struct rt0_hdr *)np->opt->srcrt;
250 ipv6_addr_copy(&final, &fl.fl6_dst);
251 ipv6_addr_copy(&fl.fl6_dst, rt0->addr);
252 final_p = &final;
253 }
254
beb8d13b
VY
255 security_sk_classify_flow(sk, &fl);
256
1da177e4
LT
257 err = ip6_dst_lookup(sk, &dst, &fl);
258 if (err)
259 goto failure;
260 if (final_p)
261 ipv6_addr_copy(&fl.fl6_dst, final_p);
262
52479b62
AD
263 err = __xfrm_lookup(sock_net(sk), &dst, &fl, sk, XFRM_LOOKUP_WAIT);
264 if (err < 0) {
14e50e57
DM
265 if (err == -EREMOTE)
266 err = ip6_dst_blackhole(sk, &dst, &fl);
267 if (err < 0)
268 goto failure;
269 }
1da177e4
LT
270
271 if (saddr == NULL) {
272 saddr = &fl.fl6_src;
273 ipv6_addr_copy(&np->rcv_saddr, saddr);
274 }
275
276 /* set the source address */
277 ipv6_addr_copy(&np->saddr, saddr);
278 inet->rcv_saddr = LOOPBACK4_IPV6;
279
f83ef8c0 280 sk->sk_gso_type = SKB_GSO_TCPV6;
8e1ef0a9 281 __ip6_dst_store(sk, dst, NULL, NULL);
1da177e4 282
d83d8461 283 icsk->icsk_ext_hdr_len = 0;
1da177e4 284 if (np->opt)
d83d8461
ACM
285 icsk->icsk_ext_hdr_len = (np->opt->opt_flen +
286 np->opt->opt_nflen);
1da177e4
LT
287
288 tp->rx_opt.mss_clamp = IPV6_MIN_MTU - sizeof(struct tcphdr) - sizeof(struct ipv6hdr);
289
290 inet->dport = usin->sin6_port;
291
292 tcp_set_state(sk, TCP_SYN_SENT);
d8313f5c 293 err = inet6_hash_connect(&tcp_death_row, sk);
1da177e4
LT
294 if (err)
295 goto late_failure;
296
297 if (!tp->write_seq)
298 tp->write_seq = secure_tcpv6_sequence_number(np->saddr.s6_addr32,
299 np->daddr.s6_addr32,
300 inet->sport,
301 inet->dport);
302
303 err = tcp_connect(sk);
304 if (err)
305 goto late_failure;
306
307 return 0;
308
309late_failure:
310 tcp_set_state(sk, TCP_CLOSE);
311 __sk_dst_reset(sk);
312failure:
313 inet->dport = 0;
314 sk->sk_route_caps = 0;
315 return err;
316}
317
318static void tcp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt,
d5fdd6ba 319 u8 type, u8 code, int offset, __be32 info)
1da177e4
LT
320{
321 struct ipv6hdr *hdr = (struct ipv6hdr*)skb->data;
505cbfc5 322 const struct tcphdr *th = (struct tcphdr *)(skb->data+offset);
1da177e4
LT
323 struct ipv6_pinfo *np;
324 struct sock *sk;
325 int err;
1ab1457c 326 struct tcp_sock *tp;
1da177e4 327 __u32 seq;
ca12a1a4 328 struct net *net = dev_net(skb->dev);
1da177e4 329
ca12a1a4 330 sk = inet6_lookup(net, &tcp_hashinfo, &hdr->daddr,
d86e0dac 331 th->dest, &hdr->saddr, th->source, skb->dev->ifindex);
1da177e4
LT
332
333 if (sk == NULL) {
e41b5368
DL
334 ICMP6_INC_STATS_BH(net, __in6_dev_get(skb->dev),
335 ICMP6_MIB_INERRORS);
1da177e4
LT
336 return;
337 }
338
339 if (sk->sk_state == TCP_TIME_WAIT) {
9469c7b4 340 inet_twsk_put(inet_twsk(sk));
1da177e4
LT
341 return;
342 }
343
344 bh_lock_sock(sk);
345 if (sock_owned_by_user(sk))
de0744af 346 NET_INC_STATS_BH(net, LINUX_MIB_LOCKDROPPEDICMPS);
1da177e4
LT
347
348 if (sk->sk_state == TCP_CLOSE)
349 goto out;
350
351 tp = tcp_sk(sk);
1ab1457c 352 seq = ntohl(th->seq);
1da177e4
LT
353 if (sk->sk_state != TCP_LISTEN &&
354 !between(seq, tp->snd_una, tp->snd_nxt)) {
de0744af 355 NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS);
1da177e4
LT
356 goto out;
357 }
358
359 np = inet6_sk(sk);
360
361 if (type == ICMPV6_PKT_TOOBIG) {
362 struct dst_entry *dst = NULL;
363
364 if (sock_owned_by_user(sk))
365 goto out;
366 if ((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE))
367 goto out;
368
369 /* icmp should have updated the destination cache entry */
370 dst = __sk_dst_check(sk, np->dst_cookie);
371
372 if (dst == NULL) {
373 struct inet_sock *inet = inet_sk(sk);
374 struct flowi fl;
375
376 /* BUGGG_FUTURE: Again, it is not clear how
377 to handle rthdr case. Ignore this complexity
378 for now.
379 */
380 memset(&fl, 0, sizeof(fl));
381 fl.proto = IPPROTO_TCP;
382 ipv6_addr_copy(&fl.fl6_dst, &np->daddr);
383 ipv6_addr_copy(&fl.fl6_src, &np->saddr);
384 fl.oif = sk->sk_bound_dev_if;
51953d5b 385 fl.mark = sk->sk_mark;
1da177e4
LT
386 fl.fl_ip_dport = inet->dport;
387 fl.fl_ip_sport = inet->sport;
beb8d13b 388 security_skb_classify_flow(skb, &fl);
1da177e4
LT
389
390 if ((err = ip6_dst_lookup(sk, &dst, &fl))) {
391 sk->sk_err_soft = -err;
392 goto out;
393 }
394
52479b62 395 if ((err = xfrm_lookup(net, &dst, &fl, sk, 0)) < 0) {
1da177e4
LT
396 sk->sk_err_soft = -err;
397 goto out;
398 }
399
400 } else
401 dst_hold(dst);
402
d83d8461 403 if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst)) {
1da177e4
LT
404 tcp_sync_mss(sk, dst_mtu(dst));
405 tcp_simple_retransmit(sk);
406 } /* else let the usual retransmit timer handle it */
407 dst_release(dst);
408 goto out;
409 }
410
411 icmpv6_err_convert(type, code, &err);
412
60236fdd 413 /* Might be for an request_sock */
1da177e4 414 switch (sk->sk_state) {
60236fdd 415 struct request_sock *req, **prev;
1da177e4
LT
416 case TCP_LISTEN:
417 if (sock_owned_by_user(sk))
418 goto out;
419
8129765a
ACM
420 req = inet6_csk_search_req(sk, &prev, th->dest, &hdr->daddr,
421 &hdr->saddr, inet6_iif(skb));
1da177e4
LT
422 if (!req)
423 goto out;
424
425 /* ICMPs are not backlogged, hence we cannot get
426 * an established socket here.
427 */
547b792c 428 WARN_ON(req->sk != NULL);
1da177e4 429
2e6599cb 430 if (seq != tcp_rsk(req)->snt_isn) {
de0744af 431 NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS);
1da177e4
LT
432 goto out;
433 }
434
463c84b9 435 inet_csk_reqsk_queue_drop(sk, req, prev);
1da177e4
LT
436 goto out;
437
438 case TCP_SYN_SENT:
439 case TCP_SYN_RECV: /* Cannot happen.
1ab1457c 440 It can, it SYNs are crossed. --ANK */
1da177e4 441 if (!sock_owned_by_user(sk)) {
1da177e4
LT
442 sk->sk_err = err;
443 sk->sk_error_report(sk); /* Wake people up to see the error (see connect in sock.c) */
444
445 tcp_done(sk);
446 } else
447 sk->sk_err_soft = err;
448 goto out;
449 }
450
451 if (!sock_owned_by_user(sk) && np->recverr) {
452 sk->sk_err = err;
453 sk->sk_error_report(sk);
454 } else
455 sk->sk_err_soft = err;
456
457out:
458 bh_unlock_sock(sk);
459 sock_put(sk);
460}
461
462
fd80eb94 463static int tcp_v6_send_synack(struct sock *sk, struct request_sock *req)
1da177e4 464{
ca304b61 465 struct inet6_request_sock *treq = inet6_rsk(req);
1da177e4
LT
466 struct ipv6_pinfo *np = inet6_sk(sk);
467 struct sk_buff * skb;
468 struct ipv6_txoptions *opt = NULL;
469 struct in6_addr * final_p = NULL, final;
470 struct flowi fl;
fd80eb94 471 struct dst_entry *dst;
1da177e4
LT
472 int err = -1;
473
474 memset(&fl, 0, sizeof(fl));
475 fl.proto = IPPROTO_TCP;
2e6599cb
ACM
476 ipv6_addr_copy(&fl.fl6_dst, &treq->rmt_addr);
477 ipv6_addr_copy(&fl.fl6_src, &treq->loc_addr);
1da177e4 478 fl.fl6_flowlabel = 0;
2e6599cb 479 fl.oif = treq->iif;
51953d5b 480 fl.mark = sk->sk_mark;
2e6599cb 481 fl.fl_ip_dport = inet_rsk(req)->rmt_port;
fd507037 482 fl.fl_ip_sport = inet_rsk(req)->loc_port;
4237c75c 483 security_req_classify_flow(req, &fl);
1da177e4 484
fd80eb94
DL
485 opt = np->opt;
486 if (opt && opt->srcrt) {
487 struct rt0_hdr *rt0 = (struct rt0_hdr *) opt->srcrt;
488 ipv6_addr_copy(&final, &fl.fl6_dst);
489 ipv6_addr_copy(&fl.fl6_dst, rt0->addr);
490 final_p = &final;
1da177e4
LT
491 }
492
fd80eb94
DL
493 err = ip6_dst_lookup(sk, &dst, &fl);
494 if (err)
495 goto done;
496 if (final_p)
497 ipv6_addr_copy(&fl.fl6_dst, final_p);
52479b62 498 if ((err = xfrm_lookup(sock_net(sk), &dst, &fl, sk, 0)) < 0)
fd80eb94
DL
499 goto done;
500
1da177e4
LT
501 skb = tcp_make_synack(sk, dst, req);
502 if (skb) {
aa8223c7 503 struct tcphdr *th = tcp_hdr(skb);
1da177e4 504
684f2176 505 th->check = tcp_v6_check(skb->len,
2e6599cb 506 &treq->loc_addr, &treq->rmt_addr,
07f0757a 507 csum_partial(th, skb->len, skb->csum));
1da177e4 508
2e6599cb 509 ipv6_addr_copy(&fl.fl6_dst, &treq->rmt_addr);
1da177e4 510 err = ip6_xmit(sk, skb, &fl, opt, 0);
b9df3cb8 511 err = net_xmit_eval(err);
1da177e4
LT
512 }
513
514done:
1ab1457c 515 if (opt && opt != np->opt)
1da177e4 516 sock_kfree_s(sk, opt, opt->tot_len);
78b91042 517 dst_release(dst);
1da177e4
LT
518 return err;
519}
520
c6aefafb
GG
521static inline void syn_flood_warning(struct sk_buff *skb)
522{
523#ifdef CONFIG_SYN_COOKIES
524 if (sysctl_tcp_syncookies)
525 printk(KERN_INFO
526 "TCPv6: Possible SYN flooding on port %d. "
527 "Sending cookies.\n", ntohs(tcp_hdr(skb)->dest));
528 else
529#endif
530 printk(KERN_INFO
531 "TCPv6: Possible SYN flooding on port %d. "
532 "Dropping request.\n", ntohs(tcp_hdr(skb)->dest));
533}
534
60236fdd 535static void tcp_v6_reqsk_destructor(struct request_sock *req)
1da177e4 536{
800d55f1 537 kfree_skb(inet6_rsk(req)->pktopts);
1da177e4
LT
538}
539
cfb6eeb4
YH
540#ifdef CONFIG_TCP_MD5SIG
541static struct tcp_md5sig_key *tcp_v6_md5_do_lookup(struct sock *sk,
542 struct in6_addr *addr)
543{
544 struct tcp_sock *tp = tcp_sk(sk);
545 int i;
546
547 BUG_ON(tp == NULL);
548
549 if (!tp->md5sig_info || !tp->md5sig_info->entries6)
550 return NULL;
551
552 for (i = 0; i < tp->md5sig_info->entries6; i++) {
caad295f 553 if (ipv6_addr_equal(&tp->md5sig_info->keys6[i].addr, addr))
f8ab18d2 554 return &tp->md5sig_info->keys6[i].base;
cfb6eeb4
YH
555 }
556 return NULL;
557}
558
559static struct tcp_md5sig_key *tcp_v6_md5_lookup(struct sock *sk,
560 struct sock *addr_sk)
561{
562 return tcp_v6_md5_do_lookup(sk, &inet6_sk(addr_sk)->daddr);
563}
564
565static struct tcp_md5sig_key *tcp_v6_reqsk_md5_lookup(struct sock *sk,
566 struct request_sock *req)
567{
568 return tcp_v6_md5_do_lookup(sk, &inet6_rsk(req)->rmt_addr);
569}
570
571static int tcp_v6_md5_do_add(struct sock *sk, struct in6_addr *peer,
572 char *newkey, u8 newkeylen)
573{
574 /* Add key to the list */
b0a713e9 575 struct tcp_md5sig_key *key;
cfb6eeb4
YH
576 struct tcp_sock *tp = tcp_sk(sk);
577 struct tcp6_md5sig_key *keys;
578
b0a713e9 579 key = tcp_v6_md5_do_lookup(sk, peer);
cfb6eeb4
YH
580 if (key) {
581 /* modify existing entry - just update that one */
b0a713e9
MD
582 kfree(key->key);
583 key->key = newkey;
584 key->keylen = newkeylen;
cfb6eeb4
YH
585 } else {
586 /* reallocate new list if current one is full. */
587 if (!tp->md5sig_info) {
588 tp->md5sig_info = kzalloc(sizeof(*tp->md5sig_info), GFP_ATOMIC);
589 if (!tp->md5sig_info) {
590 kfree(newkey);
591 return -ENOMEM;
592 }
3d7dbeac 593 sk->sk_route_caps &= ~NETIF_F_GSO_MASK;
cfb6eeb4 594 }
aa133076 595 if (tcp_alloc_md5sig_pool(sk) == NULL) {
aacbe8c8
YH
596 kfree(newkey);
597 return -ENOMEM;
598 }
cfb6eeb4
YH
599 if (tp->md5sig_info->alloced6 == tp->md5sig_info->entries6) {
600 keys = kmalloc((sizeof (tp->md5sig_info->keys6[0]) *
601 (tp->md5sig_info->entries6 + 1)), GFP_ATOMIC);
602
603 if (!keys) {
604 tcp_free_md5sig_pool();
605 kfree(newkey);
606 return -ENOMEM;
607 }
608
609 if (tp->md5sig_info->entries6)
610 memmove(keys, tp->md5sig_info->keys6,
611 (sizeof (tp->md5sig_info->keys6[0]) *
612 tp->md5sig_info->entries6));
613
614 kfree(tp->md5sig_info->keys6);
615 tp->md5sig_info->keys6 = keys;
616 tp->md5sig_info->alloced6++;
617 }
618
619 ipv6_addr_copy(&tp->md5sig_info->keys6[tp->md5sig_info->entries6].addr,
620 peer);
f8ab18d2
DM
621 tp->md5sig_info->keys6[tp->md5sig_info->entries6].base.key = newkey;
622 tp->md5sig_info->keys6[tp->md5sig_info->entries6].base.keylen = newkeylen;
cfb6eeb4
YH
623
624 tp->md5sig_info->entries6++;
625 }
626 return 0;
627}
628
629static int tcp_v6_md5_add_func(struct sock *sk, struct sock *addr_sk,
630 u8 *newkey, __u8 newkeylen)
631{
632 return tcp_v6_md5_do_add(sk, &inet6_sk(addr_sk)->daddr,
633 newkey, newkeylen);
634}
635
636static int tcp_v6_md5_do_del(struct sock *sk, struct in6_addr *peer)
637{
638 struct tcp_sock *tp = tcp_sk(sk);
639 int i;
640
641 for (i = 0; i < tp->md5sig_info->entries6; i++) {
caad295f 642 if (ipv6_addr_equal(&tp->md5sig_info->keys6[i].addr, peer)) {
cfb6eeb4 643 /* Free the key */
f8ab18d2 644 kfree(tp->md5sig_info->keys6[i].base.key);
cfb6eeb4
YH
645 tp->md5sig_info->entries6--;
646
647 if (tp->md5sig_info->entries6 == 0) {
648 kfree(tp->md5sig_info->keys6);
649 tp->md5sig_info->keys6 = NULL;
ca983cef 650 tp->md5sig_info->alloced6 = 0;
cfb6eeb4
YH
651 } else {
652 /* shrink the database */
653 if (tp->md5sig_info->entries6 != i)
654 memmove(&tp->md5sig_info->keys6[i],
655 &tp->md5sig_info->keys6[i+1],
656 (tp->md5sig_info->entries6 - i)
657 * sizeof (tp->md5sig_info->keys6[0]));
658 }
77adefdc
YH
659 tcp_free_md5sig_pool();
660 return 0;
cfb6eeb4
YH
661 }
662 }
663 return -ENOENT;
664}
665
666static void tcp_v6_clear_md5_list (struct sock *sk)
667{
668 struct tcp_sock *tp = tcp_sk(sk);
669 int i;
670
671 if (tp->md5sig_info->entries6) {
672 for (i = 0; i < tp->md5sig_info->entries6; i++)
f8ab18d2 673 kfree(tp->md5sig_info->keys6[i].base.key);
cfb6eeb4
YH
674 tp->md5sig_info->entries6 = 0;
675 tcp_free_md5sig_pool();
676 }
677
678 kfree(tp->md5sig_info->keys6);
679 tp->md5sig_info->keys6 = NULL;
680 tp->md5sig_info->alloced6 = 0;
681
682 if (tp->md5sig_info->entries4) {
683 for (i = 0; i < tp->md5sig_info->entries4; i++)
f8ab18d2 684 kfree(tp->md5sig_info->keys4[i].base.key);
cfb6eeb4
YH
685 tp->md5sig_info->entries4 = 0;
686 tcp_free_md5sig_pool();
687 }
688
689 kfree(tp->md5sig_info->keys4);
690 tp->md5sig_info->keys4 = NULL;
691 tp->md5sig_info->alloced4 = 0;
692}
693
694static int tcp_v6_parse_md5_keys (struct sock *sk, char __user *optval,
695 int optlen)
696{
697 struct tcp_md5sig cmd;
698 struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)&cmd.tcpm_addr;
699 u8 *newkey;
700
701 if (optlen < sizeof(cmd))
702 return -EINVAL;
703
704 if (copy_from_user(&cmd, optval, sizeof(cmd)))
705 return -EFAULT;
706
707 if (sin6->sin6_family != AF_INET6)
708 return -EINVAL;
709
710 if (!cmd.tcpm_keylen) {
711 if (!tcp_sk(sk)->md5sig_info)
712 return -ENOENT;
e773e4fa 713 if (ipv6_addr_v4mapped(&sin6->sin6_addr))
cfb6eeb4
YH
714 return tcp_v4_md5_do_del(sk, sin6->sin6_addr.s6_addr32[3]);
715 return tcp_v6_md5_do_del(sk, &sin6->sin6_addr);
716 }
717
718 if (cmd.tcpm_keylen > TCP_MD5SIG_MAXKEYLEN)
719 return -EINVAL;
720
721 if (!tcp_sk(sk)->md5sig_info) {
722 struct tcp_sock *tp = tcp_sk(sk);
723 struct tcp_md5sig_info *p;
724
725 p = kzalloc(sizeof(struct tcp_md5sig_info), GFP_KERNEL);
726 if (!p)
727 return -ENOMEM;
728
729 tp->md5sig_info = p;
3d7dbeac 730 sk->sk_route_caps &= ~NETIF_F_GSO_MASK;
cfb6eeb4
YH
731 }
732
af879cc7 733 newkey = kmemdup(cmd.tcpm_key, cmd.tcpm_keylen, GFP_KERNEL);
cfb6eeb4
YH
734 if (!newkey)
735 return -ENOMEM;
e773e4fa 736 if (ipv6_addr_v4mapped(&sin6->sin6_addr)) {
cfb6eeb4
YH
737 return tcp_v4_md5_do_add(sk, sin6->sin6_addr.s6_addr32[3],
738 newkey, cmd.tcpm_keylen);
739 }
740 return tcp_v6_md5_do_add(sk, &sin6->sin6_addr, newkey, cmd.tcpm_keylen);
741}
742
49a72dfb
AL
743static int tcp_v6_md5_hash_pseudoheader(struct tcp_md5sig_pool *hp,
744 struct in6_addr *daddr,
745 struct in6_addr *saddr, int nbytes)
cfb6eeb4 746{
cfb6eeb4 747 struct tcp6_pseudohdr *bp;
49a72dfb 748 struct scatterlist sg;
8d26d76d 749
cfb6eeb4 750 bp = &hp->md5_blk.ip6;
cfb6eeb4
YH
751 /* 1. TCP pseudo-header (RFC2460) */
752 ipv6_addr_copy(&bp->saddr, saddr);
753 ipv6_addr_copy(&bp->daddr, daddr);
49a72dfb 754 bp->protocol = cpu_to_be32(IPPROTO_TCP);
00b1304c 755 bp->len = cpu_to_be32(nbytes);
cfb6eeb4 756
49a72dfb
AL
757 sg_init_one(&sg, bp, sizeof(*bp));
758 return crypto_hash_update(&hp->md5_desc, &sg, sizeof(*bp));
759}
c7da57a1 760
49a72dfb
AL
761static int tcp_v6_md5_hash_hdr(char *md5_hash, struct tcp_md5sig_key *key,
762 struct in6_addr *daddr, struct in6_addr *saddr,
763 struct tcphdr *th)
764{
765 struct tcp_md5sig_pool *hp;
766 struct hash_desc *desc;
767
768 hp = tcp_get_md5sig_pool();
769 if (!hp)
770 goto clear_hash_noput;
771 desc = &hp->md5_desc;
772
773 if (crypto_hash_init(desc))
774 goto clear_hash;
775 if (tcp_v6_md5_hash_pseudoheader(hp, daddr, saddr, th->doff << 2))
776 goto clear_hash;
777 if (tcp_md5_hash_header(hp, th))
778 goto clear_hash;
779 if (tcp_md5_hash_key(hp, key))
780 goto clear_hash;
781 if (crypto_hash_final(desc, md5_hash))
cfb6eeb4 782 goto clear_hash;
cfb6eeb4 783
cfb6eeb4 784 tcp_put_md5sig_pool();
cfb6eeb4 785 return 0;
49a72dfb 786
cfb6eeb4
YH
787clear_hash:
788 tcp_put_md5sig_pool();
789clear_hash_noput:
790 memset(md5_hash, 0, 16);
49a72dfb 791 return 1;
cfb6eeb4
YH
792}
793
49a72dfb
AL
794static int tcp_v6_md5_hash_skb(char *md5_hash, struct tcp_md5sig_key *key,
795 struct sock *sk, struct request_sock *req,
796 struct sk_buff *skb)
cfb6eeb4
YH
797{
798 struct in6_addr *saddr, *daddr;
49a72dfb
AL
799 struct tcp_md5sig_pool *hp;
800 struct hash_desc *desc;
801 struct tcphdr *th = tcp_hdr(skb);
cfb6eeb4
YH
802
803 if (sk) {
804 saddr = &inet6_sk(sk)->saddr;
805 daddr = &inet6_sk(sk)->daddr;
49a72dfb 806 } else if (req) {
cfb6eeb4
YH
807 saddr = &inet6_rsk(req)->loc_addr;
808 daddr = &inet6_rsk(req)->rmt_addr;
49a72dfb
AL
809 } else {
810 struct ipv6hdr *ip6h = ipv6_hdr(skb);
811 saddr = &ip6h->saddr;
812 daddr = &ip6h->daddr;
cfb6eeb4 813 }
49a72dfb
AL
814
815 hp = tcp_get_md5sig_pool();
816 if (!hp)
817 goto clear_hash_noput;
818 desc = &hp->md5_desc;
819
820 if (crypto_hash_init(desc))
821 goto clear_hash;
822
823 if (tcp_v6_md5_hash_pseudoheader(hp, daddr, saddr, skb->len))
824 goto clear_hash;
825 if (tcp_md5_hash_header(hp, th))
826 goto clear_hash;
827 if (tcp_md5_hash_skb_data(hp, skb, th->doff << 2))
828 goto clear_hash;
829 if (tcp_md5_hash_key(hp, key))
830 goto clear_hash;
831 if (crypto_hash_final(desc, md5_hash))
832 goto clear_hash;
833
834 tcp_put_md5sig_pool();
835 return 0;
836
837clear_hash:
838 tcp_put_md5sig_pool();
839clear_hash_noput:
840 memset(md5_hash, 0, 16);
841 return 1;
cfb6eeb4
YH
842}
843
844static int tcp_v6_inbound_md5_hash (struct sock *sk, struct sk_buff *skb)
845{
846 __u8 *hash_location = NULL;
847 struct tcp_md5sig_key *hash_expected;
0660e03f 848 struct ipv6hdr *ip6h = ipv6_hdr(skb);
aa8223c7 849 struct tcphdr *th = tcp_hdr(skb);
cfb6eeb4 850 int genhash;
cfb6eeb4
YH
851 u8 newhash[16];
852
853 hash_expected = tcp_v6_md5_do_lookup(sk, &ip6h->saddr);
7d5d5525 854 hash_location = tcp_parse_md5sig_option(th);
cfb6eeb4 855
785957d3
DM
856 /* We've parsed the options - do we have a hash? */
857 if (!hash_expected && !hash_location)
858 return 0;
859
860 if (hash_expected && !hash_location) {
861 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPMD5NOTFOUND);
cfb6eeb4
YH
862 return 1;
863 }
864
785957d3
DM
865 if (!hash_expected && hash_location) {
866 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPMD5UNEXPECTED);
cfb6eeb4
YH
867 return 1;
868 }
869
870 /* check the signature */
49a72dfb
AL
871 genhash = tcp_v6_md5_hash_skb(newhash,
872 hash_expected,
873 NULL, NULL, skb);
874
cfb6eeb4
YH
875 if (genhash || memcmp(hash_location, newhash, 16) != 0) {
876 if (net_ratelimit()) {
5b095d98 877 printk(KERN_INFO "MD5 Hash %s for (%pI6, %u)->(%pI6, %u)\n",
cfb6eeb4 878 genhash ? "failed" : "mismatch",
0c6ce78a
HH
879 &ip6h->saddr, ntohs(th->source),
880 &ip6h->daddr, ntohs(th->dest));
cfb6eeb4
YH
881 }
882 return 1;
883 }
884 return 0;
885}
886#endif
887
c6aefafb 888struct request_sock_ops tcp6_request_sock_ops __read_mostly = {
1da177e4 889 .family = AF_INET6,
2e6599cb 890 .obj_size = sizeof(struct tcp6_request_sock),
1da177e4 891 .rtx_syn_ack = tcp_v6_send_synack,
60236fdd
ACM
892 .send_ack = tcp_v6_reqsk_send_ack,
893 .destructor = tcp_v6_reqsk_destructor,
1da177e4
LT
894 .send_reset = tcp_v6_send_reset
895};
896
cfb6eeb4 897#ifdef CONFIG_TCP_MD5SIG
b2e4b3de 898static const struct tcp_request_sock_ops tcp_request_sock_ipv6_ops = {
cfb6eeb4 899 .md5_lookup = tcp_v6_reqsk_md5_lookup,
e3afe7b7 900 .calc_md5_hash = tcp_v6_md5_hash_skb,
cfb6eeb4 901};
b6332e6c 902#endif
cfb6eeb4 903
6d6ee43e
ACM
904static struct timewait_sock_ops tcp6_timewait_sock_ops = {
905 .twsk_obj_size = sizeof(struct tcp6_timewait_sock),
906 .twsk_unique = tcp_twsk_unique,
cfb6eeb4 907 .twsk_destructor= tcp_twsk_destructor,
6d6ee43e
ACM
908};
909
8292a17a 910static void tcp_v6_send_check(struct sock *sk, int len, struct sk_buff *skb)
1da177e4
LT
911{
912 struct ipv6_pinfo *np = inet6_sk(sk);
aa8223c7 913 struct tcphdr *th = tcp_hdr(skb);
1da177e4 914
84fa7933 915 if (skb->ip_summed == CHECKSUM_PARTIAL) {
1da177e4 916 th->check = ~csum_ipv6_magic(&np->saddr, &np->daddr, len, IPPROTO_TCP, 0);
663ead3b 917 skb->csum_start = skb_transport_header(skb) - skb->head;
ff1dcadb 918 skb->csum_offset = offsetof(struct tcphdr, check);
1da177e4 919 } else {
1ab1457c 920 th->check = csum_ipv6_magic(&np->saddr, &np->daddr, len, IPPROTO_TCP,
07f0757a 921 csum_partial(th, th->doff<<2,
1da177e4
LT
922 skb->csum));
923 }
924}
925
a430a43d
HX
926static int tcp_v6_gso_send_check(struct sk_buff *skb)
927{
928 struct ipv6hdr *ipv6h;
929 struct tcphdr *th;
930
931 if (!pskb_may_pull(skb, sizeof(*th)))
932 return -EINVAL;
933
0660e03f 934 ipv6h = ipv6_hdr(skb);
aa8223c7 935 th = tcp_hdr(skb);
a430a43d
HX
936
937 th->check = 0;
938 th->check = ~csum_ipv6_magic(&ipv6h->saddr, &ipv6h->daddr, skb->len,
939 IPPROTO_TCP, 0);
663ead3b 940 skb->csum_start = skb_transport_header(skb) - skb->head;
ff1dcadb 941 skb->csum_offset = offsetof(struct tcphdr, check);
84fa7933 942 skb->ip_summed = CHECKSUM_PARTIAL;
a430a43d
HX
943 return 0;
944}
1da177e4 945
36990673
HX
946static struct sk_buff **tcp6_gro_receive(struct sk_buff **head,
947 struct sk_buff *skb)
684f2176 948{
36e7b1b8 949 struct ipv6hdr *iph = skb_gro_network_header(skb);
684f2176
HX
950
951 switch (skb->ip_summed) {
952 case CHECKSUM_COMPLETE:
86911732 953 if (!tcp_v6_check(skb_gro_len(skb), &iph->saddr, &iph->daddr,
684f2176
HX
954 skb->csum)) {
955 skb->ip_summed = CHECKSUM_UNNECESSARY;
956 break;
957 }
958
959 /* fall through */
960 case CHECKSUM_NONE:
961 NAPI_GRO_CB(skb)->flush = 1;
962 return NULL;
963 }
964
965 return tcp_gro_receive(head, skb);
966}
684f2176 967
36990673 968static int tcp6_gro_complete(struct sk_buff *skb)
684f2176
HX
969{
970 struct ipv6hdr *iph = ipv6_hdr(skb);
971 struct tcphdr *th = tcp_hdr(skb);
972
973 th->check = ~tcp_v6_check(skb->len - skb_transport_offset(skb),
974 &iph->saddr, &iph->daddr, 0);
975 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6;
976
977 return tcp_gro_complete(skb);
978}
684f2176 979
626e264d
IJ
980static void tcp_v6_send_response(struct sk_buff *skb, u32 seq, u32 ack, u32 win,
981 u32 ts, struct tcp_md5sig_key *key, int rst)
1da177e4 982{
aa8223c7 983 struct tcphdr *th = tcp_hdr(skb), *t1;
1da177e4
LT
984 struct sk_buff *buff;
985 struct flowi fl;
adf30907 986 struct net *net = dev_net(skb_dst(skb)->dev);
e5047992 987 struct sock *ctl_sk = net->ipv6.tcp_sk;
77c676da 988 unsigned int tot_len = sizeof(struct tcphdr);
adf30907 989 struct dst_entry *dst;
81ada62d 990 __be32 *topt;
1da177e4 991
626e264d
IJ
992 if (ts)
993 tot_len += TCPOLEN_TSTAMP_ALIGNED;
cfb6eeb4 994#ifdef CONFIG_TCP_MD5SIG
cfb6eeb4
YH
995 if (key)
996 tot_len += TCPOLEN_MD5SIG_ALIGNED;
997#endif
998
cfb6eeb4 999 buff = alloc_skb(MAX_HEADER + sizeof(struct ipv6hdr) + tot_len,
1da177e4 1000 GFP_ATOMIC);
1ab1457c
YH
1001 if (buff == NULL)
1002 return;
1da177e4 1003
cfb6eeb4 1004 skb_reserve(buff, MAX_HEADER + sizeof(struct ipv6hdr) + tot_len);
1da177e4 1005
cfb6eeb4 1006 t1 = (struct tcphdr *) skb_push(buff, tot_len);
a8fdf2b3 1007 skb_reset_transport_header(skb);
1da177e4
LT
1008
1009 /* Swap the send and the receive. */
1010 memset(t1, 0, sizeof(*t1));
1011 t1->dest = th->source;
1012 t1->source = th->dest;
cfb6eeb4 1013 t1->doff = tot_len / 4;
626e264d
IJ
1014 t1->seq = htonl(seq);
1015 t1->ack_seq = htonl(ack);
1016 t1->ack = !rst || !th->ack;
1017 t1->rst = rst;
1018 t1->window = htons(win);
1da177e4 1019
81ada62d
IJ
1020 topt = (__be32 *)(t1 + 1);
1021
626e264d
IJ
1022 if (ts) {
1023 *topt++ = htonl((TCPOPT_NOP << 24) | (TCPOPT_NOP << 16) |
1024 (TCPOPT_TIMESTAMP << 8) | TCPOLEN_TIMESTAMP);
1025 *topt++ = htonl(tcp_time_stamp);
1026 *topt++ = htonl(ts);
1027 }
1028
cfb6eeb4
YH
1029#ifdef CONFIG_TCP_MD5SIG
1030 if (key) {
81ada62d
IJ
1031 *topt++ = htonl((TCPOPT_NOP << 24) | (TCPOPT_NOP << 16) |
1032 (TCPOPT_MD5SIG << 8) | TCPOLEN_MD5SIG);
1033 tcp_v6_md5_hash_hdr((__u8 *)topt, key,
78e645cb
IJ
1034 &ipv6_hdr(skb)->saddr,
1035 &ipv6_hdr(skb)->daddr, t1);
cfb6eeb4
YH
1036 }
1037#endif
1038
07f0757a 1039 buff->csum = csum_partial(t1, tot_len, 0);
1da177e4
LT
1040
1041 memset(&fl, 0, sizeof(fl));
0660e03f
ACM
1042 ipv6_addr_copy(&fl.fl6_dst, &ipv6_hdr(skb)->saddr);
1043 ipv6_addr_copy(&fl.fl6_src, &ipv6_hdr(skb)->daddr);
1da177e4
LT
1044
1045 t1->check = csum_ipv6_magic(&fl.fl6_src, &fl.fl6_dst,
52cd5750 1046 tot_len, IPPROTO_TCP,
1da177e4
LT
1047 buff->csum);
1048
1049 fl.proto = IPPROTO_TCP;
505cbfc5 1050 fl.oif = inet6_iif(skb);
1da177e4
LT
1051 fl.fl_ip_dport = t1->dest;
1052 fl.fl_ip_sport = t1->source;
beb8d13b 1053 security_skb_classify_flow(skb, &fl);
1da177e4 1054
c20121ae
DL
1055 /* Pass a socket to ip6_dst_lookup either it is for RST
1056 * Underlying function will use this to retrieve the network
1057 * namespace
1058 */
adf30907
ED
1059 if (!ip6_dst_lookup(ctl_sk, &dst, &fl)) {
1060 if (xfrm_lookup(net, &dst, &fl, NULL, 0) >= 0) {
1061 skb_dst_set(buff, dst);
e5047992 1062 ip6_xmit(ctl_sk, buff, &fl, NULL, 0);
63231bdd 1063 TCP_INC_STATS_BH(net, TCP_MIB_OUTSEGS);
626e264d
IJ
1064 if (rst)
1065 TCP_INC_STATS_BH(net, TCP_MIB_OUTRSTS);
1da177e4 1066 return;
ecc51b6d 1067 }
1da177e4
LT
1068 }
1069
1070 kfree_skb(buff);
1071}
1072
626e264d 1073static void tcp_v6_send_reset(struct sock *sk, struct sk_buff *skb)
1da177e4 1074{
626e264d
IJ
1075 struct tcphdr *th = tcp_hdr(skb);
1076 u32 seq = 0, ack_seq = 0;
fa3e5b4e 1077 struct tcp_md5sig_key *key = NULL;
1da177e4 1078
626e264d 1079 if (th->rst)
1da177e4
LT
1080 return;
1081
626e264d
IJ
1082 if (!ipv6_unicast_destination(skb))
1083 return;
1da177e4 1084
cfb6eeb4 1085#ifdef CONFIG_TCP_MD5SIG
626e264d
IJ
1086 if (sk)
1087 key = tcp_v6_md5_do_lookup(sk, &ipv6_hdr(skb)->daddr);
cfb6eeb4
YH
1088#endif
1089
626e264d
IJ
1090 if (th->ack)
1091 seq = ntohl(th->ack_seq);
1092 else
1093 ack_seq = ntohl(th->seq) + th->syn + th->fin + skb->len -
1094 (th->doff << 2);
1da177e4 1095
626e264d
IJ
1096 tcp_v6_send_response(skb, seq, ack_seq, 0, 0, key, 1);
1097}
1da177e4 1098
626e264d
IJ
1099static void tcp_v6_send_ack(struct sk_buff *skb, u32 seq, u32 ack, u32 win, u32 ts,
1100 struct tcp_md5sig_key *key)
1101{
1102 tcp_v6_send_response(skb, seq, ack, win, ts, key, 0);
1da177e4
LT
1103}
1104
1105static void tcp_v6_timewait_ack(struct sock *sk, struct sk_buff *skb)
1106{
8feaf0c0 1107 struct inet_timewait_sock *tw = inet_twsk(sk);
cfb6eeb4 1108 struct tcp_timewait_sock *tcptw = tcp_twsk(sk);
1da177e4 1109
9501f972 1110 tcp_v6_send_ack(skb, tcptw->tw_snd_nxt, tcptw->tw_rcv_nxt,
8feaf0c0 1111 tcptw->tw_rcv_wnd >> tw->tw_rcv_wscale,
9501f972 1112 tcptw->tw_ts_recent, tcp_twsk_md5_key(tcptw));
1da177e4 1113
8feaf0c0 1114 inet_twsk_put(tw);
1da177e4
LT
1115}
1116
6edafaaf
GJ
1117static void tcp_v6_reqsk_send_ack(struct sock *sk, struct sk_buff *skb,
1118 struct request_sock *req)
1da177e4 1119{
9501f972 1120 tcp_v6_send_ack(skb, tcp_rsk(req)->snt_isn + 1, tcp_rsk(req)->rcv_isn + 1, req->rcv_wnd, req->ts_recent,
6edafaaf 1121 tcp_v6_md5_do_lookup(sk, &ipv6_hdr(skb)->daddr));
1da177e4
LT
1122}
1123
1124
1125static struct sock *tcp_v6_hnd_req(struct sock *sk,struct sk_buff *skb)
1126{
60236fdd 1127 struct request_sock *req, **prev;
aa8223c7 1128 const struct tcphdr *th = tcp_hdr(skb);
1da177e4
LT
1129 struct sock *nsk;
1130
1131 /* Find possible connection requests. */
8129765a 1132 req = inet6_csk_search_req(sk, &prev, th->source,
0660e03f
ACM
1133 &ipv6_hdr(skb)->saddr,
1134 &ipv6_hdr(skb)->daddr, inet6_iif(skb));
1da177e4
LT
1135 if (req)
1136 return tcp_check_req(sk, skb, req, prev);
1137
3b1e0a65 1138 nsk = __inet6_lookup_established(sock_net(sk), &tcp_hashinfo,
d86e0dac
PE
1139 &ipv6_hdr(skb)->saddr, th->source,
1140 &ipv6_hdr(skb)->daddr, ntohs(th->dest), inet6_iif(skb));
1da177e4
LT
1141
1142 if (nsk) {
1143 if (nsk->sk_state != TCP_TIME_WAIT) {
1144 bh_lock_sock(nsk);
1145 return nsk;
1146 }
9469c7b4 1147 inet_twsk_put(inet_twsk(nsk));
1da177e4
LT
1148 return NULL;
1149 }
1150
c6aefafb 1151#ifdef CONFIG_SYN_COOKIES
1da177e4 1152 if (!th->rst && !th->syn && th->ack)
c6aefafb 1153 sk = cookie_v6_check(sk, skb);
1da177e4
LT
1154#endif
1155 return sk;
1156}
1157
1da177e4
LT
1158/* FIXME: this is substantially similar to the ipv4 code.
1159 * Can some kind of merge be done? -- erics
1160 */
1161static int tcp_v6_conn_request(struct sock *sk, struct sk_buff *skb)
1162{
ca304b61 1163 struct inet6_request_sock *treq;
1da177e4
LT
1164 struct ipv6_pinfo *np = inet6_sk(sk);
1165 struct tcp_options_received tmp_opt;
1166 struct tcp_sock *tp = tcp_sk(sk);
60236fdd 1167 struct request_sock *req = NULL;
1da177e4 1168 __u32 isn = TCP_SKB_CB(skb)->when;
c6aefafb
GG
1169#ifdef CONFIG_SYN_COOKIES
1170 int want_cookie = 0;
1171#else
1172#define want_cookie 0
1173#endif
1da177e4
LT
1174
1175 if (skb->protocol == htons(ETH_P_IP))
1176 return tcp_v4_conn_request(sk, skb);
1177
1178 if (!ipv6_unicast_destination(skb))
1ab1457c 1179 goto drop;
1da177e4 1180
463c84b9 1181 if (inet_csk_reqsk_queue_is_full(sk) && !isn) {
1da177e4 1182 if (net_ratelimit())
c6aefafb
GG
1183 syn_flood_warning(skb);
1184#ifdef CONFIG_SYN_COOKIES
1185 if (sysctl_tcp_syncookies)
1186 want_cookie = 1;
1187 else
1188#endif
1ab1457c 1189 goto drop;
1da177e4
LT
1190 }
1191
463c84b9 1192 if (sk_acceptq_is_full(sk) && inet_csk_reqsk_queue_young(sk) > 1)
1da177e4
LT
1193 goto drop;
1194
ca304b61 1195 req = inet6_reqsk_alloc(&tcp6_request_sock_ops);
1da177e4
LT
1196 if (req == NULL)
1197 goto drop;
1198
cfb6eeb4
YH
1199#ifdef CONFIG_TCP_MD5SIG
1200 tcp_rsk(req)->af_specific = &tcp_request_sock_ipv6_ops;
1201#endif
1202
1da177e4
LT
1203 tcp_clear_options(&tmp_opt);
1204 tmp_opt.mss_clamp = IPV6_MIN_MTU - sizeof(struct tcphdr) - sizeof(struct ipv6hdr);
1205 tmp_opt.user_mss = tp->rx_opt.user_mss;
1206
1207 tcp_parse_options(skb, &tmp_opt, 0);
1208
4dfc2817 1209 if (want_cookie && !tmp_opt.saw_tstamp)
c6aefafb 1210 tcp_clear_options(&tmp_opt);
c6aefafb 1211
1da177e4
LT
1212 tmp_opt.tstamp_ok = tmp_opt.saw_tstamp;
1213 tcp_openreq_init(req, &tmp_opt, skb);
1214
ca304b61 1215 treq = inet6_rsk(req);
0660e03f
ACM
1216 ipv6_addr_copy(&treq->rmt_addr, &ipv6_hdr(skb)->saddr);
1217 ipv6_addr_copy(&treq->loc_addr, &ipv6_hdr(skb)->daddr);
c6aefafb
GG
1218 if (!want_cookie)
1219 TCP_ECN_create_request(req, tcp_hdr(skb));
1220
1221 if (want_cookie) {
1222 isn = cookie_v6_init_sequence(sk, skb, &req->mss);
4dfc2817 1223 req->cookie_ts = tmp_opt.tstamp_ok;
c6aefafb
GG
1224 } else if (!isn) {
1225 if (ipv6_opt_accepted(sk, skb) ||
1226 np->rxopt.bits.rxinfo || np->rxopt.bits.rxoinfo ||
1227 np->rxopt.bits.rxhlim || np->rxopt.bits.rxohlim) {
1228 atomic_inc(&skb->users);
1229 treq->pktopts = skb;
1230 }
1231 treq->iif = sk->sk_bound_dev_if;
1da177e4 1232
c6aefafb
GG
1233 /* So that link locals have meaning */
1234 if (!sk->sk_bound_dev_if &&
1235 ipv6_addr_type(&treq->rmt_addr) & IPV6_ADDR_LINKLOCAL)
1236 treq->iif = inet6_iif(skb);
1da177e4 1237
a94f723d 1238 isn = tcp_v6_init_sequence(skb);
c6aefafb 1239 }
1da177e4 1240
2e6599cb 1241 tcp_rsk(req)->snt_isn = isn;
1da177e4 1242
4237c75c
VY
1243 security_inet_conn_request(sk, skb, req);
1244
fd80eb94 1245 if (tcp_v6_send_synack(sk, req))
1da177e4
LT
1246 goto drop;
1247
c6aefafb
GG
1248 if (!want_cookie) {
1249 inet6_csk_reqsk_queue_hash_add(sk, req, TCP_TIMEOUT_INIT);
1250 return 0;
1251 }
1da177e4
LT
1252
1253drop:
1254 if (req)
60236fdd 1255 reqsk_free(req);
1da177e4 1256
1da177e4
LT
1257 return 0; /* don't send reset */
1258}
1259
1260static struct sock * tcp_v6_syn_recv_sock(struct sock *sk, struct sk_buff *skb,
60236fdd 1261 struct request_sock *req,
1da177e4
LT
1262 struct dst_entry *dst)
1263{
78d15e82 1264 struct inet6_request_sock *treq;
1da177e4
LT
1265 struct ipv6_pinfo *newnp, *np = inet6_sk(sk);
1266 struct tcp6_sock *newtcp6sk;
1267 struct inet_sock *newinet;
1268 struct tcp_sock *newtp;
1269 struct sock *newsk;
1270 struct ipv6_txoptions *opt;
cfb6eeb4
YH
1271#ifdef CONFIG_TCP_MD5SIG
1272 struct tcp_md5sig_key *key;
1273#endif
1da177e4
LT
1274
1275 if (skb->protocol == htons(ETH_P_IP)) {
1276 /*
1277 * v6 mapped
1278 */
1279
1280 newsk = tcp_v4_syn_recv_sock(sk, skb, req, dst);
1281
1ab1457c 1282 if (newsk == NULL)
1da177e4
LT
1283 return NULL;
1284
1285 newtcp6sk = (struct tcp6_sock *)newsk;
1286 inet_sk(newsk)->pinet6 = &newtcp6sk->inet6;
1287
1288 newinet = inet_sk(newsk);
1289 newnp = inet6_sk(newsk);
1290 newtp = tcp_sk(newsk);
1291
1292 memcpy(newnp, np, sizeof(struct ipv6_pinfo));
1293
b301e82c 1294 ipv6_addr_set_v4mapped(newinet->daddr, &newnp->daddr);
1da177e4 1295
b301e82c 1296 ipv6_addr_set_v4mapped(newinet->saddr, &newnp->saddr);
1da177e4
LT
1297
1298 ipv6_addr_copy(&newnp->rcv_saddr, &newnp->saddr);
1299
8292a17a 1300 inet_csk(newsk)->icsk_af_ops = &ipv6_mapped;
1da177e4 1301 newsk->sk_backlog_rcv = tcp_v4_do_rcv;
cfb6eeb4
YH
1302#ifdef CONFIG_TCP_MD5SIG
1303 newtp->af_specific = &tcp_sock_ipv6_mapped_specific;
1304#endif
1305
1da177e4
LT
1306 newnp->pktoptions = NULL;
1307 newnp->opt = NULL;
505cbfc5 1308 newnp->mcast_oif = inet6_iif(skb);
0660e03f 1309 newnp->mcast_hops = ipv6_hdr(skb)->hop_limit;
1da177e4 1310
e6848976
ACM
1311 /*
1312 * No need to charge this sock to the relevant IPv6 refcnt debug socks count
1313 * here, tcp_create_openreq_child now does this for us, see the comment in
1314 * that function for the gory details. -acme
1da177e4 1315 */
1da177e4
LT
1316
1317 /* It is tricky place. Until this moment IPv4 tcp
8292a17a 1318 worked with IPv6 icsk.icsk_af_ops.
1da177e4
LT
1319 Sync it now.
1320 */
d83d8461 1321 tcp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie);
1da177e4
LT
1322
1323 return newsk;
1324 }
1325
78d15e82 1326 treq = inet6_rsk(req);
1da177e4
LT
1327 opt = np->opt;
1328
1329 if (sk_acceptq_is_full(sk))
1330 goto out_overflow;
1331
1da177e4
LT
1332 if (dst == NULL) {
1333 struct in6_addr *final_p = NULL, final;
1334 struct flowi fl;
1335
1336 memset(&fl, 0, sizeof(fl));
1337 fl.proto = IPPROTO_TCP;
2e6599cb 1338 ipv6_addr_copy(&fl.fl6_dst, &treq->rmt_addr);
1da177e4
LT
1339 if (opt && opt->srcrt) {
1340 struct rt0_hdr *rt0 = (struct rt0_hdr *) opt->srcrt;
1341 ipv6_addr_copy(&final, &fl.fl6_dst);
1342 ipv6_addr_copy(&fl.fl6_dst, rt0->addr);
1343 final_p = &final;
1344 }
2e6599cb 1345 ipv6_addr_copy(&fl.fl6_src, &treq->loc_addr);
1da177e4 1346 fl.oif = sk->sk_bound_dev_if;
51953d5b 1347 fl.mark = sk->sk_mark;
2e6599cb 1348 fl.fl_ip_dport = inet_rsk(req)->rmt_port;
fd507037 1349 fl.fl_ip_sport = inet_rsk(req)->loc_port;
4237c75c 1350 security_req_classify_flow(req, &fl);
1da177e4
LT
1351
1352 if (ip6_dst_lookup(sk, &dst, &fl))
1353 goto out;
1354
1355 if (final_p)
1356 ipv6_addr_copy(&fl.fl6_dst, final_p);
1357
52479b62 1358 if ((xfrm_lookup(sock_net(sk), &dst, &fl, sk, 0)) < 0)
1da177e4 1359 goto out;
1ab1457c 1360 }
1da177e4
LT
1361
1362 newsk = tcp_create_openreq_child(sk, req, skb);
1363 if (newsk == NULL)
1364 goto out;
1365
e6848976
ACM
1366 /*
1367 * No need to charge this sock to the relevant IPv6 refcnt debug socks
1368 * count here, tcp_create_openreq_child now does this for us, see the
1369 * comment in that function for the gory details. -acme
1370 */
1da177e4 1371
59eed279 1372 newsk->sk_gso_type = SKB_GSO_TCPV6;
8e1ef0a9 1373 __ip6_dst_store(newsk, dst, NULL, NULL);
1da177e4
LT
1374
1375 newtcp6sk = (struct tcp6_sock *)newsk;
1376 inet_sk(newsk)->pinet6 = &newtcp6sk->inet6;
1377
1378 newtp = tcp_sk(newsk);
1379 newinet = inet_sk(newsk);
1380 newnp = inet6_sk(newsk);
1381
1382 memcpy(newnp, np, sizeof(struct ipv6_pinfo));
1383
2e6599cb
ACM
1384 ipv6_addr_copy(&newnp->daddr, &treq->rmt_addr);
1385 ipv6_addr_copy(&newnp->saddr, &treq->loc_addr);
1386 ipv6_addr_copy(&newnp->rcv_saddr, &treq->loc_addr);
1387 newsk->sk_bound_dev_if = treq->iif;
1da177e4 1388
1ab1457c 1389 /* Now IPv6 options...
1da177e4
LT
1390
1391 First: no IPv4 options.
1392 */
1393 newinet->opt = NULL;
d35690be 1394 newnp->ipv6_fl_list = NULL;
1da177e4
LT
1395
1396 /* Clone RX bits */
1397 newnp->rxopt.all = np->rxopt.all;
1398
1399 /* Clone pktoptions received with SYN */
1400 newnp->pktoptions = NULL;
2e6599cb
ACM
1401 if (treq->pktopts != NULL) {
1402 newnp->pktoptions = skb_clone(treq->pktopts, GFP_ATOMIC);
1403 kfree_skb(treq->pktopts);
1404 treq->pktopts = NULL;
1da177e4
LT
1405 if (newnp->pktoptions)
1406 skb_set_owner_r(newnp->pktoptions, newsk);
1407 }
1408 newnp->opt = NULL;
505cbfc5 1409 newnp->mcast_oif = inet6_iif(skb);
0660e03f 1410 newnp->mcast_hops = ipv6_hdr(skb)->hop_limit;
1da177e4
LT
1411
1412 /* Clone native IPv6 options from listening socket (if any)
1413
1414 Yes, keeping reference count would be much more clever,
1415 but we make one more one thing there: reattach optmem
1416 to newsk.
1417 */
1418 if (opt) {
1419 newnp->opt = ipv6_dup_options(newsk, opt);
1420 if (opt != np->opt)
1421 sock_kfree_s(sk, opt, opt->tot_len);
1422 }
1423
d83d8461 1424 inet_csk(newsk)->icsk_ext_hdr_len = 0;
1da177e4 1425 if (newnp->opt)
d83d8461
ACM
1426 inet_csk(newsk)->icsk_ext_hdr_len = (newnp->opt->opt_nflen +
1427 newnp->opt->opt_flen);
1da177e4 1428
5d424d5a 1429 tcp_mtup_init(newsk);
1da177e4
LT
1430 tcp_sync_mss(newsk, dst_mtu(dst));
1431 newtp->advmss = dst_metric(dst, RTAX_ADVMSS);
1432 tcp_initialize_rcv_mss(newsk);
1433
1434 newinet->daddr = newinet->saddr = newinet->rcv_saddr = LOOPBACK4_IPV6;
1435
cfb6eeb4
YH
1436#ifdef CONFIG_TCP_MD5SIG
1437 /* Copy over the MD5 key from the original socket */
1438 if ((key = tcp_v6_md5_do_lookup(sk, &newnp->daddr)) != NULL) {
1439 /* We're using one, so create a matching key
1440 * on the newsk structure. If we fail to get
1441 * memory, then we end up not copying the key
1442 * across. Shucks.
1443 */
af879cc7
ACM
1444 char *newkey = kmemdup(key->key, key->keylen, GFP_ATOMIC);
1445 if (newkey != NULL)
e547bc1e 1446 tcp_v6_md5_do_add(newsk, &newnp->daddr,
cfb6eeb4 1447 newkey, key->keylen);
cfb6eeb4
YH
1448 }
1449#endif
1450
ab1e0a13 1451 __inet6_hash(newsk);
e56d8b8a 1452 __inet_inherit_port(sk, newsk);
1da177e4
LT
1453
1454 return newsk;
1455
1456out_overflow:
de0744af 1457 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS);
1da177e4 1458out:
de0744af 1459 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_LISTENDROPS);
1da177e4
LT
1460 if (opt && opt != np->opt)
1461 sock_kfree_s(sk, opt, opt->tot_len);
1462 dst_release(dst);
1463 return NULL;
1464}
1465
b51655b9 1466static __sum16 tcp_v6_checksum_init(struct sk_buff *skb)
1da177e4 1467{
84fa7933 1468 if (skb->ip_summed == CHECKSUM_COMPLETE) {
684f2176 1469 if (!tcp_v6_check(skb->len, &ipv6_hdr(skb)->saddr,
0660e03f 1470 &ipv6_hdr(skb)->daddr, skb->csum)) {
fb286bb2 1471 skb->ip_summed = CHECKSUM_UNNECESSARY;
1da177e4 1472 return 0;
fb286bb2 1473 }
1da177e4 1474 }
fb286bb2 1475
684f2176 1476 skb->csum = ~csum_unfold(tcp_v6_check(skb->len,
0660e03f
ACM
1477 &ipv6_hdr(skb)->saddr,
1478 &ipv6_hdr(skb)->daddr, 0));
fb286bb2 1479
1da177e4 1480 if (skb->len <= 76) {
fb286bb2 1481 return __skb_checksum_complete(skb);
1da177e4
LT
1482 }
1483 return 0;
1484}
1485
1486/* The socket must have it's spinlock held when we get
1487 * here.
1488 *
1489 * We have a potential double-lock case here, so even when
1490 * doing backlog processing we use the BH locking scheme.
1491 * This is because we cannot sleep with the original spinlock
1492 * held.
1493 */
1494static int tcp_v6_do_rcv(struct sock *sk, struct sk_buff *skb)
1495{
1496 struct ipv6_pinfo *np = inet6_sk(sk);
1497 struct tcp_sock *tp;
1498 struct sk_buff *opt_skb = NULL;
1499
1500 /* Imagine: socket is IPv6. IPv4 packet arrives,
1501 goes to IPv4 receive handler and backlogged.
1502 From backlog it always goes here. Kerboom...
1503 Fortunately, tcp_rcv_established and rcv_established
1504 handle them correctly, but it is not case with
1505 tcp_v6_hnd_req and tcp_v6_send_reset(). --ANK
1506 */
1507
1508 if (skb->protocol == htons(ETH_P_IP))
1509 return tcp_v4_do_rcv(sk, skb);
1510
cfb6eeb4
YH
1511#ifdef CONFIG_TCP_MD5SIG
1512 if (tcp_v6_inbound_md5_hash (sk, skb))
1513 goto discard;
1514#endif
1515
fda9ef5d 1516 if (sk_filter(sk, skb))
1da177e4
LT
1517 goto discard;
1518
1519 /*
1520 * socket locking is here for SMP purposes as backlog rcv
1521 * is currently called with bh processing disabled.
1522 */
1523
1524 /* Do Stevens' IPV6_PKTOPTIONS.
1525
1526 Yes, guys, it is the only place in our code, where we
1527 may make it not affecting IPv4.
1528 The rest of code is protocol independent,
1529 and I do not like idea to uglify IPv4.
1530
1531 Actually, all the idea behind IPV6_PKTOPTIONS
1532 looks not very well thought. For now we latch
1533 options, received in the last packet, enqueued
1534 by tcp. Feel free to propose better solution.
1ab1457c 1535 --ANK (980728)
1da177e4
LT
1536 */
1537 if (np->rxopt.all)
1538 opt_skb = skb_clone(skb, GFP_ATOMIC);
1539
1540 if (sk->sk_state == TCP_ESTABLISHED) { /* Fast path */
1541 TCP_CHECK_TIMER(sk);
aa8223c7 1542 if (tcp_rcv_established(sk, skb, tcp_hdr(skb), skb->len))
1da177e4
LT
1543 goto reset;
1544 TCP_CHECK_TIMER(sk);
1545 if (opt_skb)
1546 goto ipv6_pktoptions;
1547 return 0;
1548 }
1549
ab6a5bb6 1550 if (skb->len < tcp_hdrlen(skb) || tcp_checksum_complete(skb))
1da177e4
LT
1551 goto csum_err;
1552
1ab1457c 1553 if (sk->sk_state == TCP_LISTEN) {
1da177e4
LT
1554 struct sock *nsk = tcp_v6_hnd_req(sk, skb);
1555 if (!nsk)
1556 goto discard;
1557
1558 /*
1559 * Queue it on the new socket if the new socket is active,
1560 * otherwise we just shortcircuit this and continue with
1561 * the new socket..
1562 */
1ab1457c 1563 if(nsk != sk) {
1da177e4
LT
1564 if (tcp_child_process(sk, nsk, skb))
1565 goto reset;
1566 if (opt_skb)
1567 __kfree_skb(opt_skb);
1568 return 0;
1569 }
1570 }
1571
1572 TCP_CHECK_TIMER(sk);
aa8223c7 1573 if (tcp_rcv_state_process(sk, skb, tcp_hdr(skb), skb->len))
1da177e4
LT
1574 goto reset;
1575 TCP_CHECK_TIMER(sk);
1576 if (opt_skb)
1577 goto ipv6_pktoptions;
1578 return 0;
1579
1580reset:
cfb6eeb4 1581 tcp_v6_send_reset(sk, skb);
1da177e4
LT
1582discard:
1583 if (opt_skb)
1584 __kfree_skb(opt_skb);
1585 kfree_skb(skb);
1586 return 0;
1587csum_err:
63231bdd 1588 TCP_INC_STATS_BH(sock_net(sk), TCP_MIB_INERRS);
1da177e4
LT
1589 goto discard;
1590
1591
1592ipv6_pktoptions:
1593 /* Do you ask, what is it?
1594
1595 1. skb was enqueued by tcp.
1596 2. skb is added to tail of read queue, rather than out of order.
1597 3. socket is not in passive state.
1598 4. Finally, it really contains options, which user wants to receive.
1599 */
1600 tp = tcp_sk(sk);
1601 if (TCP_SKB_CB(opt_skb)->end_seq == tp->rcv_nxt &&
1602 !((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN))) {
333fad53 1603 if (np->rxopt.bits.rxinfo || np->rxopt.bits.rxoinfo)
505cbfc5 1604 np->mcast_oif = inet6_iif(opt_skb);
333fad53 1605 if (np->rxopt.bits.rxhlim || np->rxopt.bits.rxohlim)
0660e03f 1606 np->mcast_hops = ipv6_hdr(opt_skb)->hop_limit;
1da177e4
LT
1607 if (ipv6_opt_accepted(sk, opt_skb)) {
1608 skb_set_owner_r(opt_skb, sk);
1609 opt_skb = xchg(&np->pktoptions, opt_skb);
1610 } else {
1611 __kfree_skb(opt_skb);
1612 opt_skb = xchg(&np->pktoptions, NULL);
1613 }
1614 }
1615
800d55f1 1616 kfree_skb(opt_skb);
1da177e4
LT
1617 return 0;
1618}
1619
e5bbef20 1620static int tcp_v6_rcv(struct sk_buff *skb)
1da177e4 1621{
1ab1457c 1622 struct tcphdr *th;
1da177e4
LT
1623 struct sock *sk;
1624 int ret;
a86b1e30 1625 struct net *net = dev_net(skb->dev);
1da177e4
LT
1626
1627 if (skb->pkt_type != PACKET_HOST)
1628 goto discard_it;
1629
1630 /*
1631 * Count it even if it's bad.
1632 */
63231bdd 1633 TCP_INC_STATS_BH(net, TCP_MIB_INSEGS);
1da177e4
LT
1634
1635 if (!pskb_may_pull(skb, sizeof(struct tcphdr)))
1636 goto discard_it;
1637
aa8223c7 1638 th = tcp_hdr(skb);
1da177e4
LT
1639
1640 if (th->doff < sizeof(struct tcphdr)/4)
1641 goto bad_packet;
1642 if (!pskb_may_pull(skb, th->doff*4))
1643 goto discard_it;
1644
60476372 1645 if (!skb_csum_unnecessary(skb) && tcp_v6_checksum_init(skb))
1da177e4
LT
1646 goto bad_packet;
1647
aa8223c7 1648 th = tcp_hdr(skb);
1da177e4
LT
1649 TCP_SKB_CB(skb)->seq = ntohl(th->seq);
1650 TCP_SKB_CB(skb)->end_seq = (TCP_SKB_CB(skb)->seq + th->syn + th->fin +
1651 skb->len - th->doff*4);
1652 TCP_SKB_CB(skb)->ack_seq = ntohl(th->ack_seq);
1653 TCP_SKB_CB(skb)->when = 0;
0660e03f 1654 TCP_SKB_CB(skb)->flags = ipv6_get_dsfield(ipv6_hdr(skb));
1da177e4
LT
1655 TCP_SKB_CB(skb)->sacked = 0;
1656
9a1f27c4 1657 sk = __inet6_lookup_skb(&tcp_hashinfo, skb, th->source, th->dest);
1da177e4
LT
1658 if (!sk)
1659 goto no_tcp_socket;
1660
1661process:
1662 if (sk->sk_state == TCP_TIME_WAIT)
1663 goto do_time_wait;
1664
1665 if (!xfrm6_policy_check(sk, XFRM_POLICY_IN, skb))
1666 goto discard_and_relse;
1667
fda9ef5d 1668 if (sk_filter(sk, skb))
1da177e4
LT
1669 goto discard_and_relse;
1670
1671 skb->dev = NULL;
1672
293b9c42 1673 bh_lock_sock_nested(sk);
1da177e4
LT
1674 ret = 0;
1675 if (!sock_owned_by_user(sk)) {
1a2449a8 1676#ifdef CONFIG_NET_DMA
1ab1457c 1677 struct tcp_sock *tp = tcp_sk(sk);
b4caea8a 1678 if (!tp->ucopy.dma_chan && tp->ucopy.pinned_list)
f67b4599 1679 tp->ucopy.dma_chan = dma_find_channel(DMA_MEMCPY);
1ab1457c
YH
1680 if (tp->ucopy.dma_chan)
1681 ret = tcp_v6_do_rcv(sk, skb);
1682 else
1a2449a8
CL
1683#endif
1684 {
1685 if (!tcp_prequeue(sk, skb))
1686 ret = tcp_v6_do_rcv(sk, skb);
1687 }
1da177e4
LT
1688 } else
1689 sk_add_backlog(sk, skb);
1690 bh_unlock_sock(sk);
1691
1692 sock_put(sk);
1693 return ret ? -1 : 0;
1694
1695no_tcp_socket:
1696 if (!xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb))
1697 goto discard_it;
1698
1699 if (skb->len < (th->doff<<2) || tcp_checksum_complete(skb)) {
1700bad_packet:
63231bdd 1701 TCP_INC_STATS_BH(net, TCP_MIB_INERRS);
1da177e4 1702 } else {
cfb6eeb4 1703 tcp_v6_send_reset(NULL, skb);
1da177e4
LT
1704 }
1705
1706discard_it:
1707
1708 /*
1709 * Discard frame
1710 */
1711
1712 kfree_skb(skb);
1713 return 0;
1714
1715discard_and_relse:
1716 sock_put(sk);
1717 goto discard_it;
1718
1719do_time_wait:
1720 if (!xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb)) {
9469c7b4 1721 inet_twsk_put(inet_twsk(sk));
1da177e4
LT
1722 goto discard_it;
1723 }
1724
1725 if (skb->len < (th->doff<<2) || tcp_checksum_complete(skb)) {
63231bdd 1726 TCP_INC_STATS_BH(net, TCP_MIB_INERRS);
9469c7b4 1727 inet_twsk_put(inet_twsk(sk));
1da177e4
LT
1728 goto discard_it;
1729 }
1730
9469c7b4 1731 switch (tcp_timewait_state_process(inet_twsk(sk), skb, th)) {
1da177e4
LT
1732 case TCP_TW_SYN:
1733 {
1734 struct sock *sk2;
1735
c346dca1 1736 sk2 = inet6_lookup_listener(dev_net(skb->dev), &tcp_hashinfo,
0660e03f 1737 &ipv6_hdr(skb)->daddr,
505cbfc5 1738 ntohs(th->dest), inet6_iif(skb));
1da177e4 1739 if (sk2 != NULL) {
295ff7ed
ACM
1740 struct inet_timewait_sock *tw = inet_twsk(sk);
1741 inet_twsk_deschedule(tw, &tcp_death_row);
1742 inet_twsk_put(tw);
1da177e4
LT
1743 sk = sk2;
1744 goto process;
1745 }
1746 /* Fall through to ACK */
1747 }
1748 case TCP_TW_ACK:
1749 tcp_v6_timewait_ack(sk, skb);
1750 break;
1751 case TCP_TW_RST:
1752 goto no_tcp_socket;
1753 case TCP_TW_SUCCESS:;
1754 }
1755 goto discard_it;
1756}
1757
1da177e4
LT
1758static int tcp_v6_remember_stamp(struct sock *sk)
1759{
1760 /* Alas, not yet... */
1761 return 0;
1762}
1763
3b401a81 1764static const struct inet_connection_sock_af_ops ipv6_specific = {
543d9cfe
ACM
1765 .queue_xmit = inet6_csk_xmit,
1766 .send_check = tcp_v6_send_check,
1767 .rebuild_header = inet6_sk_rebuild_header,
1768 .conn_request = tcp_v6_conn_request,
1769 .syn_recv_sock = tcp_v6_syn_recv_sock,
1770 .remember_stamp = tcp_v6_remember_stamp,
1771 .net_header_len = sizeof(struct ipv6hdr),
1772 .setsockopt = ipv6_setsockopt,
1773 .getsockopt = ipv6_getsockopt,
1774 .addr2sockaddr = inet6_csk_addr2sockaddr,
1775 .sockaddr_len = sizeof(struct sockaddr_in6),
ab1e0a13 1776 .bind_conflict = inet6_csk_bind_conflict,
3fdadf7d 1777#ifdef CONFIG_COMPAT
543d9cfe
ACM
1778 .compat_setsockopt = compat_ipv6_setsockopt,
1779 .compat_getsockopt = compat_ipv6_getsockopt,
3fdadf7d 1780#endif
1da177e4
LT
1781};
1782
cfb6eeb4 1783#ifdef CONFIG_TCP_MD5SIG
b2e4b3de 1784static const struct tcp_sock_af_ops tcp_sock_ipv6_specific = {
cfb6eeb4 1785 .md5_lookup = tcp_v6_md5_lookup,
49a72dfb 1786 .calc_md5_hash = tcp_v6_md5_hash_skb,
cfb6eeb4
YH
1787 .md5_add = tcp_v6_md5_add_func,
1788 .md5_parse = tcp_v6_parse_md5_keys,
cfb6eeb4 1789};
a928630a 1790#endif
cfb6eeb4 1791
1da177e4
LT
1792/*
1793 * TCP over IPv4 via INET6 API
1794 */
1795
3b401a81 1796static const struct inet_connection_sock_af_ops ipv6_mapped = {
543d9cfe
ACM
1797 .queue_xmit = ip_queue_xmit,
1798 .send_check = tcp_v4_send_check,
1799 .rebuild_header = inet_sk_rebuild_header,
1800 .conn_request = tcp_v6_conn_request,
1801 .syn_recv_sock = tcp_v6_syn_recv_sock,
1802 .remember_stamp = tcp_v4_remember_stamp,
1803 .net_header_len = sizeof(struct iphdr),
1804 .setsockopt = ipv6_setsockopt,
1805 .getsockopt = ipv6_getsockopt,
1806 .addr2sockaddr = inet6_csk_addr2sockaddr,
1807 .sockaddr_len = sizeof(struct sockaddr_in6),
ab1e0a13 1808 .bind_conflict = inet6_csk_bind_conflict,
3fdadf7d 1809#ifdef CONFIG_COMPAT
543d9cfe
ACM
1810 .compat_setsockopt = compat_ipv6_setsockopt,
1811 .compat_getsockopt = compat_ipv6_getsockopt,
3fdadf7d 1812#endif
1da177e4
LT
1813};
1814
cfb6eeb4 1815#ifdef CONFIG_TCP_MD5SIG
b2e4b3de 1816static const struct tcp_sock_af_ops tcp_sock_ipv6_mapped_specific = {
cfb6eeb4 1817 .md5_lookup = tcp_v4_md5_lookup,
49a72dfb 1818 .calc_md5_hash = tcp_v4_md5_hash_skb,
cfb6eeb4
YH
1819 .md5_add = tcp_v6_md5_add_func,
1820 .md5_parse = tcp_v6_parse_md5_keys,
cfb6eeb4 1821};
a928630a 1822#endif
cfb6eeb4 1823
1da177e4
LT
1824/* NOTE: A lot of things set to zero explicitly by call to
1825 * sk_alloc() so need not be done here.
1826 */
1827static int tcp_v6_init_sock(struct sock *sk)
1828{
6687e988 1829 struct inet_connection_sock *icsk = inet_csk(sk);
1da177e4
LT
1830 struct tcp_sock *tp = tcp_sk(sk);
1831
1832 skb_queue_head_init(&tp->out_of_order_queue);
1833 tcp_init_xmit_timers(sk);
1834 tcp_prequeue_init(tp);
1835
6687e988 1836 icsk->icsk_rto = TCP_TIMEOUT_INIT;
1da177e4
LT
1837 tp->mdev = TCP_TIMEOUT_INIT;
1838
1839 /* So many TCP implementations out there (incorrectly) count the
1840 * initial SYN frame in their delayed-ACK and congestion control
1841 * algorithms that we must have the following bandaid to talk
1842 * efficiently to them. -DaveM
1843 */
1844 tp->snd_cwnd = 2;
1845
1846 /* See draft-stevens-tcpca-spec-01 for discussion of the
1847 * initialization of these values.
1848 */
0b6a05c1 1849 tp->snd_ssthresh = TCP_INFINITE_SSTHRESH;
1da177e4 1850 tp->snd_cwnd_clamp = ~0;
c1b4a7e6 1851 tp->mss_cache = 536;
1da177e4
LT
1852
1853 tp->reordering = sysctl_tcp_reordering;
1854
1855 sk->sk_state = TCP_CLOSE;
1856
8292a17a 1857 icsk->icsk_af_ops = &ipv6_specific;
6687e988 1858 icsk->icsk_ca_ops = &tcp_init_congestion_ops;
d83d8461 1859 icsk->icsk_sync_mss = tcp_sync_mss;
1da177e4
LT
1860 sk->sk_write_space = sk_stream_write_space;
1861 sock_set_flag(sk, SOCK_USE_WRITE_QUEUE);
1862
cfb6eeb4
YH
1863#ifdef CONFIG_TCP_MD5SIG
1864 tp->af_specific = &tcp_sock_ipv6_specific;
1865#endif
1866
1da177e4
LT
1867 sk->sk_sndbuf = sysctl_tcp_wmem[1];
1868 sk->sk_rcvbuf = sysctl_tcp_rmem[1];
1869
eb4dea58 1870 local_bh_disable();
1748376b 1871 percpu_counter_inc(&tcp_sockets_allocated);
eb4dea58 1872 local_bh_enable();
1da177e4
LT
1873
1874 return 0;
1875}
1876
7d06b2e0 1877static void tcp_v6_destroy_sock(struct sock *sk)
1da177e4 1878{
cfb6eeb4
YH
1879#ifdef CONFIG_TCP_MD5SIG
1880 /* Clean up the MD5 key list */
1881 if (tcp_sk(sk)->md5sig_info)
1882 tcp_v6_clear_md5_list(sk);
1883#endif
1da177e4 1884 tcp_v4_destroy_sock(sk);
7d06b2e0 1885 inet6_destroy_sock(sk);
1da177e4
LT
1886}
1887
952a10be 1888#ifdef CONFIG_PROC_FS
1da177e4 1889/* Proc filesystem TCPv6 sock list dumping. */
1ab1457c 1890static void get_openreq6(struct seq_file *seq,
60236fdd 1891 struct sock *sk, struct request_sock *req, int i, int uid)
1da177e4 1892{
1da177e4 1893 int ttd = req->expires - jiffies;
ca304b61
ACM
1894 struct in6_addr *src = &inet6_rsk(req)->loc_addr;
1895 struct in6_addr *dest = &inet6_rsk(req)->rmt_addr;
1da177e4
LT
1896
1897 if (ttd < 0)
1898 ttd = 0;
1899
1da177e4
LT
1900 seq_printf(seq,
1901 "%4d: %08X%08X%08X%08X:%04X %08X%08X%08X%08X:%04X "
1902 "%02X %08X:%08X %02X:%08lX %08X %5d %8d %d %d %p\n",
1903 i,
1904 src->s6_addr32[0], src->s6_addr32[1],
1905 src->s6_addr32[2], src->s6_addr32[3],
fd507037 1906 ntohs(inet_rsk(req)->loc_port),
1da177e4
LT
1907 dest->s6_addr32[0], dest->s6_addr32[1],
1908 dest->s6_addr32[2], dest->s6_addr32[3],
2e6599cb 1909 ntohs(inet_rsk(req)->rmt_port),
1da177e4
LT
1910 TCP_SYN_RECV,
1911 0,0, /* could print option size, but that is af dependent. */
1ab1457c
YH
1912 1, /* timers active (only the expire timer) */
1913 jiffies_to_clock_t(ttd),
1da177e4
LT
1914 req->retrans,
1915 uid,
1ab1457c 1916 0, /* non standard timer */
1da177e4
LT
1917 0, /* open_requests have no inode */
1918 0, req);
1919}
1920
1921static void get_tcp6_sock(struct seq_file *seq, struct sock *sp, int i)
1922{
1923 struct in6_addr *dest, *src;
1924 __u16 destp, srcp;
1925 int timer_active;
1926 unsigned long timer_expires;
1927 struct inet_sock *inet = inet_sk(sp);
1928 struct tcp_sock *tp = tcp_sk(sp);
463c84b9 1929 const struct inet_connection_sock *icsk = inet_csk(sp);
1da177e4
LT
1930 struct ipv6_pinfo *np = inet6_sk(sp);
1931
1932 dest = &np->daddr;
1933 src = &np->rcv_saddr;
1934 destp = ntohs(inet->dport);
1935 srcp = ntohs(inet->sport);
463c84b9
ACM
1936
1937 if (icsk->icsk_pending == ICSK_TIME_RETRANS) {
1da177e4 1938 timer_active = 1;
463c84b9
ACM
1939 timer_expires = icsk->icsk_timeout;
1940 } else if (icsk->icsk_pending == ICSK_TIME_PROBE0) {
1da177e4 1941 timer_active = 4;
463c84b9 1942 timer_expires = icsk->icsk_timeout;
1da177e4
LT
1943 } else if (timer_pending(&sp->sk_timer)) {
1944 timer_active = 2;
1945 timer_expires = sp->sk_timer.expires;
1946 } else {
1947 timer_active = 0;
1948 timer_expires = jiffies;
1949 }
1950
1951 seq_printf(seq,
1952 "%4d: %08X%08X%08X%08X:%04X %08X%08X%08X%08X:%04X "
7be87351 1953 "%02X %08X:%08X %02X:%08lX %08X %5d %8d %lu %d %p %lu %lu %u %u %d\n",
1da177e4
LT
1954 i,
1955 src->s6_addr32[0], src->s6_addr32[1],
1956 src->s6_addr32[2], src->s6_addr32[3], srcp,
1957 dest->s6_addr32[0], dest->s6_addr32[1],
1958 dest->s6_addr32[2], dest->s6_addr32[3], destp,
1ab1457c 1959 sp->sk_state,
47da8ee6
SS
1960 tp->write_seq-tp->snd_una,
1961 (sp->sk_state == TCP_LISTEN) ? sp->sk_ack_backlog : (tp->rcv_nxt - tp->copied_seq),
1da177e4
LT
1962 timer_active,
1963 jiffies_to_clock_t(timer_expires - jiffies),
463c84b9 1964 icsk->icsk_retransmits,
1da177e4 1965 sock_i_uid(sp),
6687e988 1966 icsk->icsk_probes_out,
1da177e4
LT
1967 sock_i_ino(sp),
1968 atomic_read(&sp->sk_refcnt), sp,
7be87351
SH
1969 jiffies_to_clock_t(icsk->icsk_rto),
1970 jiffies_to_clock_t(icsk->icsk_ack.ato),
463c84b9 1971 (icsk->icsk_ack.quick << 1 ) | icsk->icsk_ack.pingpong,
0b6a05c1
IJ
1972 tp->snd_cwnd,
1973 tcp_in_initial_slowstart(tp) ? -1 : tp->snd_ssthresh
1da177e4
LT
1974 );
1975}
1976
1ab1457c 1977static void get_timewait6_sock(struct seq_file *seq,
8feaf0c0 1978 struct inet_timewait_sock *tw, int i)
1da177e4
LT
1979{
1980 struct in6_addr *dest, *src;
1981 __u16 destp, srcp;
0fa1a53e 1982 struct inet6_timewait_sock *tw6 = inet6_twsk((struct sock *)tw);
1da177e4
LT
1983 int ttd = tw->tw_ttd - jiffies;
1984
1985 if (ttd < 0)
1986 ttd = 0;
1987
0fa1a53e
ACM
1988 dest = &tw6->tw_v6_daddr;
1989 src = &tw6->tw_v6_rcv_saddr;
1da177e4
LT
1990 destp = ntohs(tw->tw_dport);
1991 srcp = ntohs(tw->tw_sport);
1992
1993 seq_printf(seq,
1994 "%4d: %08X%08X%08X%08X:%04X %08X%08X%08X%08X:%04X "
1995 "%02X %08X:%08X %02X:%08lX %08X %5d %8d %d %d %p\n",
1996 i,
1997 src->s6_addr32[0], src->s6_addr32[1],
1998 src->s6_addr32[2], src->s6_addr32[3], srcp,
1999 dest->s6_addr32[0], dest->s6_addr32[1],
2000 dest->s6_addr32[2], dest->s6_addr32[3], destp,
2001 tw->tw_substate, 0, 0,
2002 3, jiffies_to_clock_t(ttd), 0, 0, 0, 0,
2003 atomic_read(&tw->tw_refcnt), tw);
2004}
2005
1da177e4
LT
2006static int tcp6_seq_show(struct seq_file *seq, void *v)
2007{
2008 struct tcp_iter_state *st;
2009
2010 if (v == SEQ_START_TOKEN) {
2011 seq_puts(seq,
2012 " sl "
2013 "local_address "
2014 "remote_address "
2015 "st tx_queue rx_queue tr tm->when retrnsmt"
2016 " uid timeout inode\n");
2017 goto out;
2018 }
2019 st = seq->private;
2020
2021 switch (st->state) {
2022 case TCP_SEQ_STATE_LISTENING:
2023 case TCP_SEQ_STATE_ESTABLISHED:
2024 get_tcp6_sock(seq, v, st->num);
2025 break;
2026 case TCP_SEQ_STATE_OPENREQ:
2027 get_openreq6(seq, st->syn_wait_sk, v, st->num, st->uid);
2028 break;
2029 case TCP_SEQ_STATE_TIME_WAIT:
2030 get_timewait6_sock(seq, v, st->num);
2031 break;
2032 }
2033out:
2034 return 0;
2035}
2036
1da177e4 2037static struct tcp_seq_afinfo tcp6_seq_afinfo = {
1da177e4
LT
2038 .name = "tcp6",
2039 .family = AF_INET6,
5f4472c5
DL
2040 .seq_fops = {
2041 .owner = THIS_MODULE,
2042 },
9427c4b3
DL
2043 .seq_ops = {
2044 .show = tcp6_seq_show,
2045 },
1da177e4
LT
2046};
2047
6f8b13bc 2048int tcp6_proc_init(struct net *net)
1da177e4 2049{
6f8b13bc 2050 return tcp_proc_register(net, &tcp6_seq_afinfo);
1da177e4
LT
2051}
2052
6f8b13bc 2053void tcp6_proc_exit(struct net *net)
1da177e4 2054{
6f8b13bc 2055 tcp_proc_unregister(net, &tcp6_seq_afinfo);
1da177e4
LT
2056}
2057#endif
2058
2059struct proto tcpv6_prot = {
2060 .name = "TCPv6",
2061 .owner = THIS_MODULE,
2062 .close = tcp_close,
2063 .connect = tcp_v6_connect,
2064 .disconnect = tcp_disconnect,
463c84b9 2065 .accept = inet_csk_accept,
1da177e4
LT
2066 .ioctl = tcp_ioctl,
2067 .init = tcp_v6_init_sock,
2068 .destroy = tcp_v6_destroy_sock,
2069 .shutdown = tcp_shutdown,
2070 .setsockopt = tcp_setsockopt,
2071 .getsockopt = tcp_getsockopt,
1da177e4
LT
2072 .recvmsg = tcp_recvmsg,
2073 .backlog_rcv = tcp_v6_do_rcv,
2074 .hash = tcp_v6_hash,
ab1e0a13
ACM
2075 .unhash = inet_unhash,
2076 .get_port = inet_csk_get_port,
1da177e4
LT
2077 .enter_memory_pressure = tcp_enter_memory_pressure,
2078 .sockets_allocated = &tcp_sockets_allocated,
2079 .memory_allocated = &tcp_memory_allocated,
2080 .memory_pressure = &tcp_memory_pressure,
0a5578cf 2081 .orphan_count = &tcp_orphan_count,
1da177e4
LT
2082 .sysctl_mem = sysctl_tcp_mem,
2083 .sysctl_wmem = sysctl_tcp_wmem,
2084 .sysctl_rmem = sysctl_tcp_rmem,
2085 .max_header = MAX_TCP_HEADER,
2086 .obj_size = sizeof(struct tcp6_sock),
3ab5aee7 2087 .slab_flags = SLAB_DESTROY_BY_RCU,
6d6ee43e 2088 .twsk_prot = &tcp6_timewait_sock_ops,
60236fdd 2089 .rsk_prot = &tcp6_request_sock_ops,
39d8cda7 2090 .h.hashinfo = &tcp_hashinfo,
543d9cfe
ACM
2091#ifdef CONFIG_COMPAT
2092 .compat_setsockopt = compat_tcp_setsockopt,
2093 .compat_getsockopt = compat_tcp_getsockopt,
2094#endif
1da177e4
LT
2095};
2096
41135cc8 2097static const struct inet6_protocol tcpv6_protocol = {
1da177e4
LT
2098 .handler = tcp_v6_rcv,
2099 .err_handler = tcp_v6_err,
a430a43d 2100 .gso_send_check = tcp_v6_gso_send_check,
adcfc7d0 2101 .gso_segment = tcp_tso_segment,
684f2176
HX
2102 .gro_receive = tcp6_gro_receive,
2103 .gro_complete = tcp6_gro_complete,
1da177e4
LT
2104 .flags = INET6_PROTO_NOPOLICY|INET6_PROTO_FINAL,
2105};
2106
1da177e4
LT
2107static struct inet_protosw tcpv6_protosw = {
2108 .type = SOCK_STREAM,
2109 .protocol = IPPROTO_TCP,
2110 .prot = &tcpv6_prot,
2111 .ops = &inet6_stream_ops,
2112 .capability = -1,
2113 .no_check = 0,
d83d8461
ACM
2114 .flags = INET_PROTOSW_PERMANENT |
2115 INET_PROTOSW_ICSK,
1da177e4
LT
2116};
2117
93ec926b
DL
2118static int tcpv6_net_init(struct net *net)
2119{
5677242f
DL
2120 return inet_ctl_sock_create(&net->ipv6.tcp_sk, PF_INET6,
2121 SOCK_RAW, IPPROTO_TCP, net);
93ec926b
DL
2122}
2123
2124static void tcpv6_net_exit(struct net *net)
2125{
5677242f 2126 inet_ctl_sock_destroy(net->ipv6.tcp_sk);
d315492b 2127 inet_twsk_purge(net, &tcp_hashinfo, &tcp_death_row, AF_INET6);
93ec926b
DL
2128}
2129
2130static struct pernet_operations tcpv6_net_ops = {
2131 .init = tcpv6_net_init,
2132 .exit = tcpv6_net_exit,
2133};
2134
7f4e4868 2135int __init tcpv6_init(void)
1da177e4 2136{
7f4e4868
DL
2137 int ret;
2138
2139 ret = inet6_add_protocol(&tcpv6_protocol, IPPROTO_TCP);
2140 if (ret)
2141 goto out;
2142
1da177e4 2143 /* register inet6 protocol */
7f4e4868
DL
2144 ret = inet6_register_protosw(&tcpv6_protosw);
2145 if (ret)
2146 goto out_tcpv6_protocol;
2147
93ec926b 2148 ret = register_pernet_subsys(&tcpv6_net_ops);
7f4e4868
DL
2149 if (ret)
2150 goto out_tcpv6_protosw;
2151out:
2152 return ret;
ae0f7d5f 2153
7f4e4868
DL
2154out_tcpv6_protocol:
2155 inet6_del_protocol(&tcpv6_protocol, IPPROTO_TCP);
2156out_tcpv6_protosw:
2157 inet6_unregister_protosw(&tcpv6_protosw);
2158 goto out;
2159}
2160
09f7709f 2161void tcpv6_exit(void)
7f4e4868 2162{
93ec926b 2163 unregister_pernet_subsys(&tcpv6_net_ops);
7f4e4868
DL
2164 inet6_unregister_protosw(&tcpv6_protosw);
2165 inet6_del_protocol(&tcpv6_protocol, IPPROTO_TCP);
1da177e4 2166}