1 // SPDX-License-Identifier: GPL-2.0-or-later
4 * Linux INET6 implementation
6 * Based on net/dccp6/ipv6.c
8 * Arnaldo Carvalho de Melo <acme@ghostprotocols.net>
11 #include <linux/module.h>
12 #include <linux/random.h>
13 #include <linux/slab.h>
14 #include <linux/xfrm.h>
15 #include <linux/string.h>
17 #include <net/addrconf.h>
18 #include <net/inet_common.h>
19 #include <net/inet_hashtables.h>
20 #include <net/inet_sock.h>
21 #include <net/inet6_connection_sock.h>
22 #include <net/inet6_hashtables.h>
23 #include <net/ip6_route.h>
25 #include <net/protocol.h>
26 #include <net/transp_v6.h>
27 #include <net/ip6_checksum.h>
29 #include <net/secure_seq.h>
30 #include <net/netns/generic.h>
37 struct dccp_v6_pernet
{
38 struct sock
*v6_ctl_sk
;
41 static unsigned int dccp_v6_pernet_id __read_mostly
;
43 /* The per-net v6_ctl_sk is used for sending RSTs and ACKs */
45 static const struct inet_connection_sock_af_ops dccp_ipv6_mapped
;
46 static const struct inet_connection_sock_af_ops dccp_ipv6_af_ops
;
48 /* add pseudo-header to DCCP checksum stored in skb->csum */
49 static inline __sum16
dccp_v6_csum_finish(struct sk_buff
*skb
,
50 const struct in6_addr
*saddr
,
51 const struct in6_addr
*daddr
)
53 return csum_ipv6_magic(saddr
, daddr
, skb
->len
, IPPROTO_DCCP
, skb
->csum
);
56 static inline void dccp_v6_send_check(struct sock
*sk
, struct sk_buff
*skb
)
58 struct ipv6_pinfo
*np
= inet6_sk(sk
);
59 struct dccp_hdr
*dh
= dccp_hdr(skb
);
61 dccp_csum_outgoing(skb
);
62 dh
->dccph_checksum
= dccp_v6_csum_finish(skb
, &np
->saddr
, &sk
->sk_v6_daddr
);
65 static inline __u64
dccp_v6_init_sequence(struct sk_buff
*skb
)
67 return secure_dccpv6_sequence_number(ipv6_hdr(skb
)->daddr
.s6_addr32
,
68 ipv6_hdr(skb
)->saddr
.s6_addr32
,
69 dccp_hdr(skb
)->dccph_dport
,
70 dccp_hdr(skb
)->dccph_sport
);
74 static int dccp_v6_err(struct sk_buff
*skb
, struct inet6_skb_parm
*opt
,
75 u8 type
, u8 code
, int offset
, __be32 info
)
77 const struct ipv6hdr
*hdr
= (const struct ipv6hdr
*)skb
->data
;
78 const struct dccp_hdr
*dh
;
80 struct ipv6_pinfo
*np
;
84 struct net
*net
= dev_net(skb
->dev
);
86 /* Only need dccph_dport & dccph_sport which are the first
87 * 4 bytes in dccp header.
88 * Our caller (icmpv6_notify()) already pulled 8 bytes for us.
90 BUILD_BUG_ON(offsetofend(struct dccp_hdr
, dccph_sport
) > 8);
91 BUILD_BUG_ON(offsetofend(struct dccp_hdr
, dccph_dport
) > 8);
92 dh
= (struct dccp_hdr
*)(skb
->data
+ offset
);
94 sk
= __inet6_lookup_established(net
, &dccp_hashinfo
,
95 &hdr
->daddr
, dh
->dccph_dport
,
96 &hdr
->saddr
, ntohs(dh
->dccph_sport
),
100 __ICMP6_INC_STATS(net
, __in6_dev_get(skb
->dev
),
105 if (sk
->sk_state
== DCCP_TIME_WAIT
) {
106 inet_twsk_put(inet_twsk(sk
));
109 seq
= dccp_hdr_seq(dh
);
110 if (sk
->sk_state
== DCCP_NEW_SYN_RECV
) {
111 dccp_req_err(sk
, seq
);
116 if (sock_owned_by_user(sk
))
117 __NET_INC_STATS(net
, LINUX_MIB_LOCKDROPPEDICMPS
);
119 if (sk
->sk_state
== DCCP_CLOSED
)
123 if ((1 << sk
->sk_state
) & ~(DCCPF_REQUESTING
| DCCPF_LISTEN
) &&
124 !between48(seq
, dp
->dccps_awl
, dp
->dccps_awh
)) {
125 __NET_INC_STATS(net
, LINUX_MIB_OUTOFWINDOWICMPS
);
131 if (type
== NDISC_REDIRECT
) {
132 if (!sock_owned_by_user(sk
)) {
133 struct dst_entry
*dst
= __sk_dst_check(sk
, np
->dst_cookie
);
136 dst
->ops
->redirect(dst
, sk
, skb
);
141 if (type
== ICMPV6_PKT_TOOBIG
) {
142 struct dst_entry
*dst
= NULL
;
144 if (!ip6_sk_accept_pmtu(sk
))
147 if (sock_owned_by_user(sk
))
149 if ((1 << sk
->sk_state
) & (DCCPF_LISTEN
| DCCPF_CLOSED
))
152 dst
= inet6_csk_update_pmtu(sk
, ntohl(info
));
156 if (inet_csk(sk
)->icsk_pmtu_cookie
> dst_mtu(dst
))
157 dccp_sync_mss(sk
, dst_mtu(dst
));
161 icmpv6_err_convert(type
, code
, &err
);
163 /* Might be for an request_sock */
164 switch (sk
->sk_state
) {
165 case DCCP_REQUESTING
:
166 case DCCP_RESPOND
: /* Cannot happen.
167 It can, it SYNs are crossed. --ANK */
168 if (!sock_owned_by_user(sk
)) {
169 __DCCP_INC_STATS(DCCP_MIB_ATTEMPTFAILS
);
172 * Wake people up to see the error
173 * (see connect in sock.c)
175 sk
->sk_error_report(sk
);
178 sk
->sk_err_soft
= err
;
182 if (!sock_owned_by_user(sk
) && np
->recverr
) {
184 sk
->sk_error_report(sk
);
186 sk
->sk_err_soft
= err
;
195 static int dccp_v6_send_response(const struct sock
*sk
, struct request_sock
*req
)
197 struct inet_request_sock
*ireq
= inet_rsk(req
);
198 struct ipv6_pinfo
*np
= inet6_sk(sk
);
200 struct in6_addr
*final_p
, final
;
203 struct dst_entry
*dst
;
205 memset(&fl6
, 0, sizeof(fl6
));
206 fl6
.flowi6_proto
= IPPROTO_DCCP
;
207 fl6
.daddr
= ireq
->ir_v6_rmt_addr
;
208 fl6
.saddr
= ireq
->ir_v6_loc_addr
;
210 fl6
.flowi6_oif
= ireq
->ir_iif
;
211 fl6
.fl6_dport
= ireq
->ir_rmt_port
;
212 fl6
.fl6_sport
= htons(ireq
->ir_num
);
213 security_req_classify_flow(req
, flowi6_to_flowi_common(&fl6
));
217 final_p
= fl6_update_dst(&fl6
, rcu_dereference(np
->opt
), &final
);
220 dst
= ip6_dst_lookup_flow(sock_net(sk
), sk
, &fl6
, final_p
);
227 skb
= dccp_make_response(sk
, dst
, req
);
229 struct dccp_hdr
*dh
= dccp_hdr(skb
);
230 struct ipv6_txoptions
*opt
;
232 dh
->dccph_checksum
= dccp_v6_csum_finish(skb
,
233 &ireq
->ir_v6_loc_addr
,
234 &ireq
->ir_v6_rmt_addr
);
235 fl6
.daddr
= ireq
->ir_v6_rmt_addr
;
237 opt
= ireq
->ipv6_opt
;
239 opt
= rcu_dereference(np
->opt
);
240 err
= ip6_xmit(sk
, skb
, &fl6
, sk
->sk_mark
, opt
, np
->tclass
,
243 err
= net_xmit_eval(err
);
251 static void dccp_v6_reqsk_destructor(struct request_sock
*req
)
253 dccp_feat_list_purge(&dccp_rsk(req
)->dreq_featneg
);
254 kfree(inet_rsk(req
)->ipv6_opt
);
255 kfree_skb(inet_rsk(req
)->pktopts
);
258 static void dccp_v6_ctl_send_reset(const struct sock
*sk
, struct sk_buff
*rxskb
)
260 const struct ipv6hdr
*rxip6h
;
263 struct net
*net
= dev_net(skb_dst(rxskb
)->dev
);
264 struct dccp_v6_pernet
*pn
;
266 struct dst_entry
*dst
;
268 if (dccp_hdr(rxskb
)->dccph_type
== DCCP_PKT_RESET
)
271 if (!ipv6_unicast_destination(rxskb
))
274 pn
= net_generic(net
, dccp_v6_pernet_id
);
275 ctl_sk
= pn
->v6_ctl_sk
;
276 skb
= dccp_ctl_make_reset(ctl_sk
, rxskb
);
280 rxip6h
= ipv6_hdr(rxskb
);
281 dccp_hdr(skb
)->dccph_checksum
= dccp_v6_csum_finish(skb
, &rxip6h
->saddr
,
284 memset(&fl6
, 0, sizeof(fl6
));
285 fl6
.daddr
= rxip6h
->saddr
;
286 fl6
.saddr
= rxip6h
->daddr
;
288 fl6
.flowi6_proto
= IPPROTO_DCCP
;
289 fl6
.flowi6_oif
= inet6_iif(rxskb
);
290 fl6
.fl6_dport
= dccp_hdr(skb
)->dccph_dport
;
291 fl6
.fl6_sport
= dccp_hdr(skb
)->dccph_sport
;
292 security_skb_classify_flow(rxskb
, flowi6_to_flowi_common(&fl6
));
294 /* sk = NULL, but it is safe for now. RST socket required. */
295 dst
= ip6_dst_lookup_flow(sock_net(ctl_sk
), ctl_sk
, &fl6
, NULL
);
297 skb_dst_set(skb
, dst
);
298 ip6_xmit(ctl_sk
, skb
, &fl6
, 0, NULL
, 0, 0);
299 DCCP_INC_STATS(DCCP_MIB_OUTSEGS
);
300 DCCP_INC_STATS(DCCP_MIB_OUTRSTS
);
307 static struct request_sock_ops dccp6_request_sock_ops
= {
309 .obj_size
= sizeof(struct dccp6_request_sock
),
310 .rtx_syn_ack
= dccp_v6_send_response
,
311 .send_ack
= dccp_reqsk_send_ack
,
312 .destructor
= dccp_v6_reqsk_destructor
,
313 .send_reset
= dccp_v6_ctl_send_reset
,
314 .syn_ack_timeout
= dccp_syn_ack_timeout
,
317 static int dccp_v6_conn_request(struct sock
*sk
, struct sk_buff
*skb
)
319 struct request_sock
*req
;
320 struct dccp_request_sock
*dreq
;
321 struct inet_request_sock
*ireq
;
322 struct ipv6_pinfo
*np
= inet6_sk(sk
);
323 const __be32 service
= dccp_hdr_request(skb
)->dccph_req_service
;
324 struct dccp_skb_cb
*dcb
= DCCP_SKB_CB(skb
);
326 if (skb
->protocol
== htons(ETH_P_IP
))
327 return dccp_v4_conn_request(sk
, skb
);
329 if (!ipv6_unicast_destination(skb
))
330 return 0; /* discard, don't send a reset here */
332 if (ipv6_addr_v4mapped(&ipv6_hdr(skb
)->saddr
)) {
333 __IP6_INC_STATS(sock_net(sk
), NULL
, IPSTATS_MIB_INHDRERRORS
);
337 if (dccp_bad_service_code(sk
, service
)) {
338 dcb
->dccpd_reset_code
= DCCP_RESET_CODE_BAD_SERVICE_CODE
;
342 * There are no SYN attacks on IPv6, yet...
344 dcb
->dccpd_reset_code
= DCCP_RESET_CODE_TOO_BUSY
;
345 if (inet_csk_reqsk_queue_is_full(sk
))
348 if (sk_acceptq_is_full(sk
))
351 req
= inet_reqsk_alloc(&dccp6_request_sock_ops
, sk
, true);
355 if (dccp_reqsk_init(req
, dccp_sk(sk
), skb
))
358 dreq
= dccp_rsk(req
);
359 if (dccp_parse_options(sk
, dreq
, skb
))
362 if (security_inet_conn_request(sk
, skb
, req
))
365 ireq
= inet_rsk(req
);
366 ireq
->ir_v6_rmt_addr
= ipv6_hdr(skb
)->saddr
;
367 ireq
->ir_v6_loc_addr
= ipv6_hdr(skb
)->daddr
;
368 ireq
->ireq_family
= AF_INET6
;
369 ireq
->ir_mark
= inet_request_mark(sk
, skb
);
371 if (ipv6_opt_accepted(sk
, skb
, IP6CB(skb
)) ||
372 np
->rxopt
.bits
.rxinfo
|| np
->rxopt
.bits
.rxoinfo
||
373 np
->rxopt
.bits
.rxhlim
|| np
->rxopt
.bits
.rxohlim
) {
374 refcount_inc(&skb
->users
);
377 ireq
->ir_iif
= sk
->sk_bound_dev_if
;
379 /* So that link locals have meaning */
380 if (!sk
->sk_bound_dev_if
&&
381 ipv6_addr_type(&ireq
->ir_v6_rmt_addr
) & IPV6_ADDR_LINKLOCAL
)
382 ireq
->ir_iif
= inet6_iif(skb
);
385 * Step 3: Process LISTEN state
387 * Set S.ISR, S.GSR, S.SWL, S.SWH from packet or Init Cookie
389 * Setting S.SWL/S.SWH to is deferred to dccp_create_openreq_child().
391 dreq
->dreq_isr
= dcb
->dccpd_seq
;
392 dreq
->dreq_gsr
= dreq
->dreq_isr
;
393 dreq
->dreq_iss
= dccp_v6_init_sequence(skb
);
394 dreq
->dreq_gss
= dreq
->dreq_iss
;
395 dreq
->dreq_service
= service
;
397 if (dccp_v6_send_response(sk
, req
))
400 inet_csk_reqsk_queue_hash_add(sk
, req
, DCCP_TIMEOUT_INIT
);
407 __DCCP_INC_STATS(DCCP_MIB_ATTEMPTFAILS
);
411 static struct sock
*dccp_v6_request_recv_sock(const struct sock
*sk
,
413 struct request_sock
*req
,
414 struct dst_entry
*dst
,
415 struct request_sock
*req_unhash
,
418 struct inet_request_sock
*ireq
= inet_rsk(req
);
419 struct ipv6_pinfo
*newnp
;
420 const struct ipv6_pinfo
*np
= inet6_sk(sk
);
421 struct ipv6_txoptions
*opt
;
422 struct inet_sock
*newinet
;
423 struct dccp6_sock
*newdp6
;
426 if (skb
->protocol
== htons(ETH_P_IP
)) {
430 newsk
= dccp_v4_request_recv_sock(sk
, skb
, req
, dst
,
431 req_unhash
, own_req
);
435 newdp6
= (struct dccp6_sock
*)newsk
;
436 newinet
= inet_sk(newsk
);
437 newinet
->pinet6
= &newdp6
->inet6
;
438 newnp
= inet6_sk(newsk
);
440 memcpy(newnp
, np
, sizeof(struct ipv6_pinfo
));
442 newnp
->saddr
= newsk
->sk_v6_rcv_saddr
;
444 inet_csk(newsk
)->icsk_af_ops
= &dccp_ipv6_mapped
;
445 newsk
->sk_backlog_rcv
= dccp_v4_do_rcv
;
446 newnp
->pktoptions
= NULL
;
448 newnp
->ipv6_mc_list
= NULL
;
449 newnp
->ipv6_ac_list
= NULL
;
450 newnp
->ipv6_fl_list
= NULL
;
451 newnp
->mcast_oif
= inet_iif(skb
);
452 newnp
->mcast_hops
= ip_hdr(skb
)->ttl
;
455 * No need to charge this sock to the relevant IPv6 refcnt debug socks count
456 * here, dccp_create_openreq_child now does this for us, see the comment in
457 * that function for the gory details. -acme
460 /* It is tricky place. Until this moment IPv4 tcp
461 worked with IPv6 icsk.icsk_af_ops.
464 dccp_sync_mss(newsk
, inet_csk(newsk
)->icsk_pmtu_cookie
);
470 if (sk_acceptq_is_full(sk
))
476 dst
= inet6_csk_route_req(sk
, &fl6
, req
, IPPROTO_DCCP
);
481 newsk
= dccp_create_openreq_child(sk
, req
, skb
);
486 * No need to charge this sock to the relevant IPv6 refcnt debug socks
487 * count here, dccp_create_openreq_child now does this for us, see the
488 * comment in that function for the gory details. -acme
491 ip6_dst_store(newsk
, dst
, NULL
, NULL
);
492 newsk
->sk_route_caps
= dst
->dev
->features
& ~(NETIF_F_IP_CSUM
|
494 newdp6
= (struct dccp6_sock
*)newsk
;
495 newinet
= inet_sk(newsk
);
496 newinet
->pinet6
= &newdp6
->inet6
;
497 newnp
= inet6_sk(newsk
);
499 memcpy(newnp
, np
, sizeof(struct ipv6_pinfo
));
501 newsk
->sk_v6_daddr
= ireq
->ir_v6_rmt_addr
;
502 newnp
->saddr
= ireq
->ir_v6_loc_addr
;
503 newsk
->sk_v6_rcv_saddr
= ireq
->ir_v6_loc_addr
;
504 newsk
->sk_bound_dev_if
= ireq
->ir_iif
;
506 /* Now IPv6 options...
508 First: no IPv4 options.
510 newinet
->inet_opt
= NULL
;
513 newnp
->rxopt
.all
= np
->rxopt
.all
;
515 newnp
->ipv6_mc_list
= NULL
;
516 newnp
->ipv6_ac_list
= NULL
;
517 newnp
->ipv6_fl_list
= NULL
;
518 newnp
->pktoptions
= NULL
;
520 newnp
->mcast_oif
= inet6_iif(skb
);
521 newnp
->mcast_hops
= ipv6_hdr(skb
)->hop_limit
;
524 * Clone native IPv6 options from listening socket (if any)
526 * Yes, keeping reference count would be much more clever, but we make
527 * one more one thing there: reattach optmem to newsk.
529 opt
= ireq
->ipv6_opt
;
531 opt
= rcu_dereference(np
->opt
);
533 opt
= ipv6_dup_options(newsk
, opt
);
534 RCU_INIT_POINTER(newnp
->opt
, opt
);
536 inet_csk(newsk
)->icsk_ext_hdr_len
= 0;
538 inet_csk(newsk
)->icsk_ext_hdr_len
= opt
->opt_nflen
+
541 dccp_sync_mss(newsk
, dst_mtu(dst
));
543 newinet
->inet_daddr
= newinet
->inet_saddr
= LOOPBACK4_IPV6
;
544 newinet
->inet_rcv_saddr
= LOOPBACK4_IPV6
;
546 if (__inet_inherit_port(sk
, newsk
) < 0) {
547 inet_csk_prepare_forced_close(newsk
);
551 *own_req
= inet_ehash_nolisten(newsk
, req_to_sk(req_unhash
), NULL
);
552 /* Clone pktoptions received with SYN, if we own the req */
553 if (*own_req
&& ireq
->pktopts
) {
554 newnp
->pktoptions
= skb_clone(ireq
->pktopts
, GFP_ATOMIC
);
555 consume_skb(ireq
->pktopts
);
556 ireq
->pktopts
= NULL
;
557 if (newnp
->pktoptions
)
558 skb_set_owner_r(newnp
->pktoptions
, newsk
);
564 __NET_INC_STATS(sock_net(sk
), LINUX_MIB_LISTENOVERFLOWS
);
568 __NET_INC_STATS(sock_net(sk
), LINUX_MIB_LISTENDROPS
);
572 /* The socket must have it's spinlock held when we get
575 * We have a potential double-lock case here, so even when
576 * doing backlog processing we use the BH locking scheme.
577 * This is because we cannot sleep with the original spinlock
580 static int dccp_v6_do_rcv(struct sock
*sk
, struct sk_buff
*skb
)
582 struct ipv6_pinfo
*np
= inet6_sk(sk
);
583 struct sk_buff
*opt_skb
= NULL
;
585 /* Imagine: socket is IPv6. IPv4 packet arrives,
586 goes to IPv4 receive handler and backlogged.
587 From backlog it always goes here. Kerboom...
588 Fortunately, dccp_rcv_established and rcv_established
589 handle them correctly, but it is not case with
590 dccp_v6_hnd_req and dccp_v6_ctl_send_reset(). --ANK
593 if (skb
->protocol
== htons(ETH_P_IP
))
594 return dccp_v4_do_rcv(sk
, skb
);
596 if (sk_filter(sk
, skb
))
600 * socket locking is here for SMP purposes as backlog rcv is currently
601 * called with bh processing disabled.
604 /* Do Stevens' IPV6_PKTOPTIONS.
606 Yes, guys, it is the only place in our code, where we
607 may make it not affecting IPv4.
608 The rest of code is protocol independent,
609 and I do not like idea to uglify IPv4.
611 Actually, all the idea behind IPV6_PKTOPTIONS
612 looks not very well thought. For now we latch
613 options, received in the last packet, enqueued
614 by tcp. Feel free to propose better solution.
618 opt_skb
= skb_clone(skb
, GFP_ATOMIC
);
620 if (sk
->sk_state
== DCCP_OPEN
) { /* Fast path */
621 if (dccp_rcv_established(sk
, skb
, dccp_hdr(skb
), skb
->len
))
624 goto ipv6_pktoptions
;
629 * Step 3: Process LISTEN state
630 * If S.state == LISTEN,
631 * If P.type == Request or P contains a valid Init Cookie option,
632 * (* Must scan the packet's options to check for Init
633 * Cookies. Only Init Cookies are processed here,
634 * however; other options are processed in Step 8. This
635 * scan need only be performed if the endpoint uses Init
637 * (* Generate a new socket and switch to that socket *)
638 * Set S := new socket for this port pair
640 * Choose S.ISS (initial seqno) or set from Init Cookies
641 * Initialize S.GAR := S.ISS
642 * Set S.ISR, S.GSR, S.SWL, S.SWH from packet or Init Cookies
643 * Continue with S.state == RESPOND
644 * (* A Response packet will be generated in Step 11 *)
646 * Generate Reset(No Connection) unless P.type == Reset
647 * Drop packet and return
649 * NOTE: the check for the packet types is done in
650 * dccp_rcv_state_process
653 if (dccp_rcv_state_process(sk
, skb
, dccp_hdr(skb
), skb
->len
))
656 goto ipv6_pktoptions
;
660 dccp_v6_ctl_send_reset(sk
, skb
);
663 __kfree_skb(opt_skb
);
667 /* Handling IPV6_PKTOPTIONS skb the similar
668 * way it's done for net/ipv6/tcp_ipv6.c
671 if (!((1 << sk
->sk_state
) & (DCCPF_CLOSED
| DCCPF_LISTEN
))) {
672 if (np
->rxopt
.bits
.rxinfo
|| np
->rxopt
.bits
.rxoinfo
)
673 np
->mcast_oif
= inet6_iif(opt_skb
);
674 if (np
->rxopt
.bits
.rxhlim
|| np
->rxopt
.bits
.rxohlim
)
675 np
->mcast_hops
= ipv6_hdr(opt_skb
)->hop_limit
;
676 if (np
->rxopt
.bits
.rxflow
|| np
->rxopt
.bits
.rxtclass
)
677 np
->rcv_flowinfo
= ip6_flowinfo(ipv6_hdr(opt_skb
));
679 np
->flow_label
= ip6_flowlabel(ipv6_hdr(opt_skb
));
680 if (ipv6_opt_accepted(sk
, opt_skb
,
681 &DCCP_SKB_CB(opt_skb
)->header
.h6
)) {
682 skb_set_owner_r(opt_skb
, sk
);
683 memmove(IP6CB(opt_skb
),
684 &DCCP_SKB_CB(opt_skb
)->header
.h6
,
685 sizeof(struct inet6_skb_parm
));
686 opt_skb
= xchg(&np
->pktoptions
, opt_skb
);
688 __kfree_skb(opt_skb
);
689 opt_skb
= xchg(&np
->pktoptions
, NULL
);
697 static int dccp_v6_rcv(struct sk_buff
*skb
)
699 const struct dccp_hdr
*dh
;
704 /* Step 1: Check header basics */
706 if (dccp_invalid_packet(skb
))
709 /* Step 1: If header checksum is incorrect, drop packet and return. */
710 if (dccp_v6_csum_finish(skb
, &ipv6_hdr(skb
)->saddr
,
711 &ipv6_hdr(skb
)->daddr
)) {
712 DCCP_WARN("dropped packet with invalid checksum\n");
718 DCCP_SKB_CB(skb
)->dccpd_seq
= dccp_hdr_seq(dh
);
719 DCCP_SKB_CB(skb
)->dccpd_type
= dh
->dccph_type
;
721 if (dccp_packet_without_ack(skb
))
722 DCCP_SKB_CB(skb
)->dccpd_ack_seq
= DCCP_PKT_WITHOUT_ACK_SEQ
;
724 DCCP_SKB_CB(skb
)->dccpd_ack_seq
= dccp_hdr_ack_seq(skb
);
727 sk
= __inet6_lookup_skb(&dccp_hashinfo
, skb
, __dccp_hdr_len(dh
),
728 dh
->dccph_sport
, dh
->dccph_dport
,
729 inet6_iif(skb
), 0, &refcounted
);
731 dccp_pr_debug("failed to look up flow ID in table and "
732 "get corresponding socket\n");
738 * ... or S.state == TIMEWAIT,
739 * Generate Reset(No Connection) unless P.type == Reset
740 * Drop packet and return
742 if (sk
->sk_state
== DCCP_TIME_WAIT
) {
743 dccp_pr_debug("sk->sk_state == DCCP_TIME_WAIT: do_time_wait\n");
744 inet_twsk_put(inet_twsk(sk
));
748 if (sk
->sk_state
== DCCP_NEW_SYN_RECV
) {
749 struct request_sock
*req
= inet_reqsk(sk
);
752 sk
= req
->rsk_listener
;
753 if (unlikely(sk
->sk_state
!= DCCP_LISTEN
)) {
754 inet_csk_reqsk_queue_drop_and_put(sk
, req
);
759 nsk
= dccp_check_req(sk
, skb
, req
);
762 goto discard_and_relse
;
766 } else if (dccp_child_process(sk
, nsk
, skb
)) {
767 dccp_v6_ctl_send_reset(sk
, skb
);
768 goto discard_and_relse
;
775 * RFC 4340, sec. 9.2.1: Minimum Checksum Coverage
776 * o if MinCsCov = 0, only packets with CsCov = 0 are accepted
777 * o if MinCsCov > 0, also accept packets with CsCov >= MinCsCov
779 min_cov
= dccp_sk(sk
)->dccps_pcrlen
;
780 if (dh
->dccph_cscov
&& (min_cov
== 0 || dh
->dccph_cscov
< min_cov
)) {
781 dccp_pr_debug("Packet CsCov %d does not satisfy MinCsCov %d\n",
782 dh
->dccph_cscov
, min_cov
);
783 /* FIXME: send Data Dropped option (see also dccp_v4_rcv) */
784 goto discard_and_relse
;
787 if (!xfrm6_policy_check(sk
, XFRM_POLICY_IN
, skb
))
788 goto discard_and_relse
;
790 return __sk_receive_skb(sk
, skb
, 1, dh
->dccph_doff
* 4,
791 refcounted
) ? -1 : 0;
794 if (!xfrm6_policy_check(NULL
, XFRM_POLICY_IN
, skb
))
799 * Generate Reset(No Connection) unless P.type == Reset
800 * Drop packet and return
802 if (dh
->dccph_type
!= DCCP_PKT_RESET
) {
803 DCCP_SKB_CB(skb
)->dccpd_reset_code
=
804 DCCP_RESET_CODE_NO_CONNECTION
;
805 dccp_v6_ctl_send_reset(sk
, skb
);
818 static int dccp_v6_connect(struct sock
*sk
, struct sockaddr
*uaddr
,
821 struct sockaddr_in6
*usin
= (struct sockaddr_in6
*)uaddr
;
822 struct inet_connection_sock
*icsk
= inet_csk(sk
);
823 struct inet_sock
*inet
= inet_sk(sk
);
824 struct ipv6_pinfo
*np
= inet6_sk(sk
);
825 struct dccp_sock
*dp
= dccp_sk(sk
);
826 struct in6_addr
*saddr
= NULL
, *final_p
, final
;
827 struct ipv6_txoptions
*opt
;
829 struct dst_entry
*dst
;
833 dp
->dccps_role
= DCCP_ROLE_CLIENT
;
835 if (addr_len
< SIN6_LEN_RFC2133
)
838 if (usin
->sin6_family
!= AF_INET6
)
839 return -EAFNOSUPPORT
;
841 memset(&fl6
, 0, sizeof(fl6
));
844 fl6
.flowlabel
= usin
->sin6_flowinfo
& IPV6_FLOWINFO_MASK
;
845 IP6_ECN_flow_init(fl6
.flowlabel
);
846 if (fl6
.flowlabel
& IPV6_FLOWLABEL_MASK
) {
847 struct ip6_flowlabel
*flowlabel
;
848 flowlabel
= fl6_sock_lookup(sk
, fl6
.flowlabel
);
849 if (IS_ERR(flowlabel
))
851 fl6_sock_release(flowlabel
);
855 * connect() to INADDR_ANY means loopback (BSD'ism).
857 if (ipv6_addr_any(&usin
->sin6_addr
))
858 usin
->sin6_addr
.s6_addr
[15] = 1;
860 addr_type
= ipv6_addr_type(&usin
->sin6_addr
);
862 if (addr_type
& IPV6_ADDR_MULTICAST
)
865 if (addr_type
& IPV6_ADDR_LINKLOCAL
) {
866 if (addr_len
>= sizeof(struct sockaddr_in6
) &&
867 usin
->sin6_scope_id
) {
868 /* If interface is set while binding, indices
871 if (sk
->sk_bound_dev_if
&&
872 sk
->sk_bound_dev_if
!= usin
->sin6_scope_id
)
875 sk
->sk_bound_dev_if
= usin
->sin6_scope_id
;
878 /* Connect to link-local address requires an interface */
879 if (!sk
->sk_bound_dev_if
)
883 sk
->sk_v6_daddr
= usin
->sin6_addr
;
884 np
->flow_label
= fl6
.flowlabel
;
889 if (addr_type
== IPV6_ADDR_MAPPED
) {
890 u32 exthdrlen
= icsk
->icsk_ext_hdr_len
;
891 struct sockaddr_in sin
;
893 SOCK_DEBUG(sk
, "connect: ipv4 mapped\n");
895 if (__ipv6_only_sock(sk
))
898 sin
.sin_family
= AF_INET
;
899 sin
.sin_port
= usin
->sin6_port
;
900 sin
.sin_addr
.s_addr
= usin
->sin6_addr
.s6_addr32
[3];
902 icsk
->icsk_af_ops
= &dccp_ipv6_mapped
;
903 sk
->sk_backlog_rcv
= dccp_v4_do_rcv
;
905 err
= dccp_v4_connect(sk
, (struct sockaddr
*)&sin
, sizeof(sin
));
907 icsk
->icsk_ext_hdr_len
= exthdrlen
;
908 icsk
->icsk_af_ops
= &dccp_ipv6_af_ops
;
909 sk
->sk_backlog_rcv
= dccp_v6_do_rcv
;
912 np
->saddr
= sk
->sk_v6_rcv_saddr
;
916 if (!ipv6_addr_any(&sk
->sk_v6_rcv_saddr
))
917 saddr
= &sk
->sk_v6_rcv_saddr
;
919 fl6
.flowi6_proto
= IPPROTO_DCCP
;
920 fl6
.daddr
= sk
->sk_v6_daddr
;
921 fl6
.saddr
= saddr
? *saddr
: np
->saddr
;
922 fl6
.flowi6_oif
= sk
->sk_bound_dev_if
;
923 fl6
.fl6_dport
= usin
->sin6_port
;
924 fl6
.fl6_sport
= inet
->inet_sport
;
925 security_sk_classify_flow(sk
, flowi6_to_flowi_common(&fl6
));
927 opt
= rcu_dereference_protected(np
->opt
, lockdep_sock_is_held(sk
));
928 final_p
= fl6_update_dst(&fl6
, opt
, &final
);
930 dst
= ip6_dst_lookup_flow(sock_net(sk
), sk
, &fl6
, final_p
);
938 sk
->sk_v6_rcv_saddr
= *saddr
;
941 /* set the source address */
943 inet
->inet_rcv_saddr
= LOOPBACK4_IPV6
;
945 ip6_dst_store(sk
, dst
, NULL
, NULL
);
947 icsk
->icsk_ext_hdr_len
= 0;
949 icsk
->icsk_ext_hdr_len
= opt
->opt_flen
+ opt
->opt_nflen
;
951 inet
->inet_dport
= usin
->sin6_port
;
953 dccp_set_state(sk
, DCCP_REQUESTING
);
954 err
= inet6_hash_connect(&dccp_death_row
, sk
);
958 dp
->dccps_iss
= secure_dccpv6_sequence_number(np
->saddr
.s6_addr32
,
959 sk
->sk_v6_daddr
.s6_addr32
,
962 err
= dccp_connect(sk
);
969 dccp_set_state(sk
, DCCP_CLOSED
);
972 inet
->inet_dport
= 0;
973 sk
->sk_route_caps
= 0;
977 static const struct inet_connection_sock_af_ops dccp_ipv6_af_ops
= {
978 .queue_xmit
= inet6_csk_xmit
,
979 .send_check
= dccp_v6_send_check
,
980 .rebuild_header
= inet6_sk_rebuild_header
,
981 .conn_request
= dccp_v6_conn_request
,
982 .syn_recv_sock
= dccp_v6_request_recv_sock
,
983 .net_header_len
= sizeof(struct ipv6hdr
),
984 .setsockopt
= ipv6_setsockopt
,
985 .getsockopt
= ipv6_getsockopt
,
986 .addr2sockaddr
= inet6_csk_addr2sockaddr
,
987 .sockaddr_len
= sizeof(struct sockaddr_in6
),
991 * DCCP over IPv4 via INET6 API
993 static const struct inet_connection_sock_af_ops dccp_ipv6_mapped
= {
994 .queue_xmit
= ip_queue_xmit
,
995 .send_check
= dccp_v4_send_check
,
996 .rebuild_header
= inet_sk_rebuild_header
,
997 .conn_request
= dccp_v6_conn_request
,
998 .syn_recv_sock
= dccp_v6_request_recv_sock
,
999 .net_header_len
= sizeof(struct iphdr
),
1000 .setsockopt
= ipv6_setsockopt
,
1001 .getsockopt
= ipv6_getsockopt
,
1002 .addr2sockaddr
= inet6_csk_addr2sockaddr
,
1003 .sockaddr_len
= sizeof(struct sockaddr_in6
),
1006 /* NOTE: A lot of things set to zero explicitly by call to
1007 * sk_alloc() so need not be done here.
1009 static int dccp_v6_init_sock(struct sock
*sk
)
1011 static __u8 dccp_v6_ctl_sock_initialized
;
1012 int err
= dccp_init_sock(sk
, dccp_v6_ctl_sock_initialized
);
1015 if (unlikely(!dccp_v6_ctl_sock_initialized
))
1016 dccp_v6_ctl_sock_initialized
= 1;
1017 inet_csk(sk
)->icsk_af_ops
= &dccp_ipv6_af_ops
;
1023 static void dccp_v6_destroy_sock(struct sock
*sk
)
1025 dccp_destroy_sock(sk
);
1026 inet6_destroy_sock(sk
);
1029 static struct timewait_sock_ops dccp6_timewait_sock_ops
= {
1030 .twsk_obj_size
= sizeof(struct dccp6_timewait_sock
),
1033 static struct proto dccp_v6_prot
= {
1035 .owner
= THIS_MODULE
,
1036 .close
= dccp_close
,
1037 .connect
= dccp_v6_connect
,
1038 .disconnect
= dccp_disconnect
,
1039 .ioctl
= dccp_ioctl
,
1040 .init
= dccp_v6_init_sock
,
1041 .setsockopt
= dccp_setsockopt
,
1042 .getsockopt
= dccp_getsockopt
,
1043 .sendmsg
= dccp_sendmsg
,
1044 .recvmsg
= dccp_recvmsg
,
1045 .backlog_rcv
= dccp_v6_do_rcv
,
1047 .unhash
= inet_unhash
,
1048 .accept
= inet_csk_accept
,
1049 .get_port
= inet_csk_get_port
,
1050 .shutdown
= dccp_shutdown
,
1051 .destroy
= dccp_v6_destroy_sock
,
1052 .orphan_count
= &dccp_orphan_count
,
1053 .max_header
= MAX_DCCP_HEADER
,
1054 .obj_size
= sizeof(struct dccp6_sock
),
1055 .slab_flags
= SLAB_TYPESAFE_BY_RCU
,
1056 .rsk_prot
= &dccp6_request_sock_ops
,
1057 .twsk_prot
= &dccp6_timewait_sock_ops
,
1058 .h
.hashinfo
= &dccp_hashinfo
,
1061 static const struct inet6_protocol dccp_v6_protocol
= {
1062 .handler
= dccp_v6_rcv
,
1063 .err_handler
= dccp_v6_err
,
1064 .flags
= INET6_PROTO_NOPOLICY
| INET6_PROTO_FINAL
,
1067 static const struct proto_ops inet6_dccp_ops
= {
1069 .owner
= THIS_MODULE
,
1070 .release
= inet6_release
,
1072 .connect
= inet_stream_connect
,
1073 .socketpair
= sock_no_socketpair
,
1074 .accept
= inet_accept
,
1075 .getname
= inet6_getname
,
1077 .ioctl
= inet6_ioctl
,
1078 .gettstamp
= sock_gettstamp
,
1079 .listen
= inet_dccp_listen
,
1080 .shutdown
= inet_shutdown
,
1081 .setsockopt
= sock_common_setsockopt
,
1082 .getsockopt
= sock_common_getsockopt
,
1083 .sendmsg
= inet_sendmsg
,
1084 .recvmsg
= sock_common_recvmsg
,
1085 .mmap
= sock_no_mmap
,
1086 .sendpage
= sock_no_sendpage
,
1087 #ifdef CONFIG_COMPAT
1088 .compat_ioctl
= inet6_compat_ioctl
,
1092 static struct inet_protosw dccp_v6_protosw
= {
1094 .protocol
= IPPROTO_DCCP
,
1095 .prot
= &dccp_v6_prot
,
1096 .ops
= &inet6_dccp_ops
,
1097 .flags
= INET_PROTOSW_ICSK
,
1100 static int __net_init
dccp_v6_init_net(struct net
*net
)
1102 struct dccp_v6_pernet
*pn
= net_generic(net
, dccp_v6_pernet_id
);
1104 if (dccp_hashinfo
.bhash
== NULL
)
1105 return -ESOCKTNOSUPPORT
;
1107 return inet_ctl_sock_create(&pn
->v6_ctl_sk
, PF_INET6
,
1108 SOCK_DCCP
, IPPROTO_DCCP
, net
);
1111 static void __net_exit
dccp_v6_exit_net(struct net
*net
)
1113 struct dccp_v6_pernet
*pn
= net_generic(net
, dccp_v6_pernet_id
);
1115 inet_ctl_sock_destroy(pn
->v6_ctl_sk
);
1118 static void __net_exit
dccp_v6_exit_batch(struct list_head
*net_exit_list
)
1120 inet_twsk_purge(&dccp_hashinfo
, AF_INET6
);
1123 static struct pernet_operations dccp_v6_ops
= {
1124 .init
= dccp_v6_init_net
,
1125 .exit
= dccp_v6_exit_net
,
1126 .exit_batch
= dccp_v6_exit_batch
,
1127 .id
= &dccp_v6_pernet_id
,
1128 .size
= sizeof(struct dccp_v6_pernet
),
1131 static int __init
dccp_v6_init(void)
1133 int err
= proto_register(&dccp_v6_prot
, 1);
1138 inet6_register_protosw(&dccp_v6_protosw
);
1140 err
= register_pernet_subsys(&dccp_v6_ops
);
1142 goto out_destroy_ctl_sock
;
1144 err
= inet6_add_protocol(&dccp_v6_protocol
, IPPROTO_DCCP
);
1146 goto out_unregister_proto
;
1150 out_unregister_proto
:
1151 unregister_pernet_subsys(&dccp_v6_ops
);
1152 out_destroy_ctl_sock
:
1153 inet6_unregister_protosw(&dccp_v6_protosw
);
1154 proto_unregister(&dccp_v6_prot
);
1158 static void __exit
dccp_v6_exit(void)
1160 inet6_del_protocol(&dccp_v6_protocol
, IPPROTO_DCCP
);
1161 unregister_pernet_subsys(&dccp_v6_ops
);
1162 inet6_unregister_protosw(&dccp_v6_protosw
);
1163 proto_unregister(&dccp_v6_prot
);
1166 module_init(dccp_v6_init
);
1167 module_exit(dccp_v6_exit
);
1170 * __stringify doesn't likes enums, so use SOCK_DCCP (6) and IPPROTO_DCCP (33)
1171 * values directly, Also cover the case where the protocol is not specified,
1172 * i.e. net-pf-PF_INET6-proto-0-type-SOCK_DCCP
1174 MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_INET6
, 33, 6);
1175 MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_INET6
, 0, 6);
1176 MODULE_LICENSE("GPL");
1177 MODULE_AUTHOR("Arnaldo Carvalho de Melo <acme@mandriva.com>");
1178 MODULE_DESCRIPTION("DCCPv6 - Datagram Congestion Controlled Protocol");