Root/
1 | /* |
2 | * DCCP over IPv6 |
3 | * Linux INET6 implementation |
4 | * |
5 | * Based on net/dccp6/ipv6.c |
6 | * |
7 | * Arnaldo Carvalho de Melo <acme@ghostprotocols.net> |
8 | * |
9 | * This program is free software; you can redistribute it and/or |
10 | * modify it under the terms of the GNU General Public License |
11 | * as published by the Free Software Foundation; either version |
12 | * 2 of the License, or (at your option) any later version. |
13 | */ |
14 | |
15 | #include <linux/module.h> |
16 | #include <linux/random.h> |
17 | #include <linux/slab.h> |
18 | #include <linux/xfrm.h> |
19 | |
20 | #include <net/addrconf.h> |
21 | #include <net/inet_common.h> |
22 | #include <net/inet_hashtables.h> |
23 | #include <net/inet_sock.h> |
24 | #include <net/inet6_connection_sock.h> |
25 | #include <net/inet6_hashtables.h> |
26 | #include <net/ip6_route.h> |
27 | #include <net/ipv6.h> |
28 | #include <net/protocol.h> |
29 | #include <net/transp_v6.h> |
30 | #include <net/ip6_checksum.h> |
31 | #include <net/xfrm.h> |
32 | |
33 | #include "dccp.h" |
34 | #include "ipv6.h" |
35 | #include "feat.h" |
36 | |
37 | /* The per-net dccp.v6_ctl_sk is used for sending RSTs and ACKs */ |
38 | |
39 | static const struct inet_connection_sock_af_ops dccp_ipv6_mapped; |
40 | static const struct inet_connection_sock_af_ops dccp_ipv6_af_ops; |
41 | |
42 | static void dccp_v6_hash(struct sock *sk) |
43 | { |
44 | if (sk->sk_state != DCCP_CLOSED) { |
45 | if (inet_csk(sk)->icsk_af_ops == &dccp_ipv6_mapped) { |
46 | inet_hash(sk); |
47 | return; |
48 | } |
49 | local_bh_disable(); |
50 | __inet6_hash(sk, NULL); |
51 | local_bh_enable(); |
52 | } |
53 | } |
54 | |
55 | /* add pseudo-header to DCCP checksum stored in skb->csum */ |
56 | static inline __sum16 dccp_v6_csum_finish(struct sk_buff *skb, |
57 | struct in6_addr *saddr, |
58 | struct in6_addr *daddr) |
59 | { |
60 | return csum_ipv6_magic(saddr, daddr, skb->len, IPPROTO_DCCP, skb->csum); |
61 | } |
62 | |
63 | static inline void dccp_v6_send_check(struct sock *sk, struct sk_buff *skb) |
64 | { |
65 | struct ipv6_pinfo *np = inet6_sk(sk); |
66 | struct dccp_hdr *dh = dccp_hdr(skb); |
67 | |
68 | dccp_csum_outgoing(skb); |
69 | dh->dccph_checksum = dccp_v6_csum_finish(skb, &np->saddr, &np->daddr); |
70 | } |
71 | |
72 | static inline __u32 secure_dccpv6_sequence_number(__be32 *saddr, __be32 *daddr, |
73 | __be16 sport, __be16 dport ) |
74 | { |
75 | return secure_tcpv6_sequence_number(saddr, daddr, sport, dport); |
76 | } |
77 | |
78 | static inline __u32 dccp_v6_init_sequence(struct sk_buff *skb) |
79 | { |
80 | return secure_dccpv6_sequence_number(ipv6_hdr(skb)->daddr.s6_addr32, |
81 | ipv6_hdr(skb)->saddr.s6_addr32, |
82 | dccp_hdr(skb)->dccph_dport, |
83 | dccp_hdr(skb)->dccph_sport ); |
84 | |
85 | } |
86 | |
87 | static void dccp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt, |
88 | u8 type, u8 code, int offset, __be32 info) |
89 | { |
90 | struct ipv6hdr *hdr = (struct ipv6hdr *)skb->data; |
91 | const struct dccp_hdr *dh = (struct dccp_hdr *)(skb->data + offset); |
92 | struct dccp_sock *dp; |
93 | struct ipv6_pinfo *np; |
94 | struct sock *sk; |
95 | int err; |
96 | __u64 seq; |
97 | struct net *net = dev_net(skb->dev); |
98 | |
99 | if (skb->len < offset + sizeof(*dh) || |
100 | skb->len < offset + __dccp_basic_hdr_len(dh)) { |
101 | ICMP6_INC_STATS_BH(net, __in6_dev_get(skb->dev), |
102 | ICMP6_MIB_INERRORS); |
103 | return; |
104 | } |
105 | |
106 | sk = inet6_lookup(net, &dccp_hashinfo, |
107 | &hdr->daddr, dh->dccph_dport, |
108 | &hdr->saddr, dh->dccph_sport, inet6_iif(skb)); |
109 | |
110 | if (sk == NULL) { |
111 | ICMP6_INC_STATS_BH(net, __in6_dev_get(skb->dev), |
112 | ICMP6_MIB_INERRORS); |
113 | return; |
114 | } |
115 | |
116 | if (sk->sk_state == DCCP_TIME_WAIT) { |
117 | inet_twsk_put(inet_twsk(sk)); |
118 | return; |
119 | } |
120 | |
121 | bh_lock_sock(sk); |
122 | if (sock_owned_by_user(sk)) |
123 | NET_INC_STATS_BH(net, LINUX_MIB_LOCKDROPPEDICMPS); |
124 | |
125 | if (sk->sk_state == DCCP_CLOSED) |
126 | goto out; |
127 | |
128 | dp = dccp_sk(sk); |
129 | seq = dccp_hdr_seq(dh); |
130 | if ((1 << sk->sk_state) & ~(DCCPF_REQUESTING | DCCPF_LISTEN) && |
131 | !between48(seq, dp->dccps_awl, dp->dccps_awh)) { |
132 | NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS); |
133 | goto out; |
134 | } |
135 | |
136 | np = inet6_sk(sk); |
137 | |
138 | if (type == ICMPV6_PKT_TOOBIG) { |
139 | struct dst_entry *dst = NULL; |
140 | |
141 | if (sock_owned_by_user(sk)) |
142 | goto out; |
143 | if ((1 << sk->sk_state) & (DCCPF_LISTEN | DCCPF_CLOSED)) |
144 | goto out; |
145 | |
146 | /* icmp should have updated the destination cache entry */ |
147 | dst = __sk_dst_check(sk, np->dst_cookie); |
148 | if (dst == NULL) { |
149 | struct inet_sock *inet = inet_sk(sk); |
150 | struct flowi fl; |
151 | |
152 | /* BUGGG_FUTURE: Again, it is not clear how |
153 | to handle rthdr case. Ignore this complexity |
154 | for now. |
155 | */ |
156 | memset(&fl, 0, sizeof(fl)); |
157 | fl.proto = IPPROTO_DCCP; |
158 | ipv6_addr_copy(&fl.fl6_dst, &np->daddr); |
159 | ipv6_addr_copy(&fl.fl6_src, &np->saddr); |
160 | fl.oif = sk->sk_bound_dev_if; |
161 | fl.fl_ip_dport = inet->inet_dport; |
162 | fl.fl_ip_sport = inet->inet_sport; |
163 | security_sk_classify_flow(sk, &fl); |
164 | |
165 | err = ip6_dst_lookup(sk, &dst, &fl); |
166 | if (err) { |
167 | sk->sk_err_soft = -err; |
168 | goto out; |
169 | } |
170 | |
171 | err = xfrm_lookup(net, &dst, &fl, sk, 0); |
172 | if (err < 0) { |
173 | sk->sk_err_soft = -err; |
174 | goto out; |
175 | } |
176 | } else |
177 | dst_hold(dst); |
178 | |
179 | if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst)) { |
180 | dccp_sync_mss(sk, dst_mtu(dst)); |
181 | } /* else let the usual retransmit timer handle it */ |
182 | dst_release(dst); |
183 | goto out; |
184 | } |
185 | |
186 | icmpv6_err_convert(type, code, &err); |
187 | |
188 | /* Might be for an request_sock */ |
189 | switch (sk->sk_state) { |
190 | struct request_sock *req, **prev; |
191 | case DCCP_LISTEN: |
192 | if (sock_owned_by_user(sk)) |
193 | goto out; |
194 | |
195 | req = inet6_csk_search_req(sk, &prev, dh->dccph_dport, |
196 | &hdr->daddr, &hdr->saddr, |
197 | inet6_iif(skb)); |
198 | if (req == NULL) |
199 | goto out; |
200 | |
201 | /* |
202 | * ICMPs are not backlogged, hence we cannot get an established |
203 | * socket here. |
204 | */ |
205 | WARN_ON(req->sk != NULL); |
206 | |
207 | if (seq != dccp_rsk(req)->dreq_iss) { |
208 | NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS); |
209 | goto out; |
210 | } |
211 | |
212 | inet_csk_reqsk_queue_drop(sk, req, prev); |
213 | goto out; |
214 | |
215 | case DCCP_REQUESTING: |
216 | case DCCP_RESPOND: /* Cannot happen. |
217 | It can, it SYNs are crossed. --ANK */ |
218 | if (!sock_owned_by_user(sk)) { |
219 | DCCP_INC_STATS_BH(DCCP_MIB_ATTEMPTFAILS); |
220 | sk->sk_err = err; |
221 | /* |
222 | * Wake people up to see the error |
223 | * (see connect in sock.c) |
224 | */ |
225 | sk->sk_error_report(sk); |
226 | dccp_done(sk); |
227 | } else |
228 | sk->sk_err_soft = err; |
229 | goto out; |
230 | } |
231 | |
232 | if (!sock_owned_by_user(sk) && np->recverr) { |
233 | sk->sk_err = err; |
234 | sk->sk_error_report(sk); |
235 | } else |
236 | sk->sk_err_soft = err; |
237 | |
238 | out: |
239 | bh_unlock_sock(sk); |
240 | sock_put(sk); |
241 | } |
242 | |
243 | |
244 | static int dccp_v6_send_response(struct sock *sk, struct request_sock *req, |
245 | struct request_values *rv_unused) |
246 | { |
247 | struct inet6_request_sock *ireq6 = inet6_rsk(req); |
248 | struct ipv6_pinfo *np = inet6_sk(sk); |
249 | struct sk_buff *skb; |
250 | struct ipv6_txoptions *opt = NULL; |
251 | struct in6_addr *final_p, final; |
252 | struct flowi fl; |
253 | int err = -1; |
254 | struct dst_entry *dst; |
255 | |
256 | memset(&fl, 0, sizeof(fl)); |
257 | fl.proto = IPPROTO_DCCP; |
258 | ipv6_addr_copy(&fl.fl6_dst, &ireq6->rmt_addr); |
259 | ipv6_addr_copy(&fl.fl6_src, &ireq6->loc_addr); |
260 | fl.fl6_flowlabel = 0; |
261 | fl.oif = ireq6->iif; |
262 | fl.fl_ip_dport = inet_rsk(req)->rmt_port; |
263 | fl.fl_ip_sport = inet_rsk(req)->loc_port; |
264 | security_req_classify_flow(req, &fl); |
265 | |
266 | opt = np->opt; |
267 | |
268 | final_p = fl6_update_dst(&fl, opt, &final); |
269 | |
270 | err = ip6_dst_lookup(sk, &dst, &fl); |
271 | if (err) |
272 | goto done; |
273 | |
274 | if (final_p) |
275 | ipv6_addr_copy(&fl.fl6_dst, final_p); |
276 | |
277 | err = xfrm_lookup(sock_net(sk), &dst, &fl, sk, 0); |
278 | if (err < 0) |
279 | goto done; |
280 | |
281 | skb = dccp_make_response(sk, dst, req); |
282 | if (skb != NULL) { |
283 | struct dccp_hdr *dh = dccp_hdr(skb); |
284 | |
285 | dh->dccph_checksum = dccp_v6_csum_finish(skb, |
286 | &ireq6->loc_addr, |
287 | &ireq6->rmt_addr); |
288 | ipv6_addr_copy(&fl.fl6_dst, &ireq6->rmt_addr); |
289 | err = ip6_xmit(sk, skb, &fl, opt); |
290 | err = net_xmit_eval(err); |
291 | } |
292 | |
293 | done: |
294 | if (opt != NULL && opt != np->opt) |
295 | sock_kfree_s(sk, opt, opt->tot_len); |
296 | dst_release(dst); |
297 | return err; |
298 | } |
299 | |
300 | static void dccp_v6_reqsk_destructor(struct request_sock *req) |
301 | { |
302 | dccp_feat_list_purge(&dccp_rsk(req)->dreq_featneg); |
303 | if (inet6_rsk(req)->pktopts != NULL) |
304 | kfree_skb(inet6_rsk(req)->pktopts); |
305 | } |
306 | |
307 | static void dccp_v6_ctl_send_reset(struct sock *sk, struct sk_buff *rxskb) |
308 | { |
309 | struct ipv6hdr *rxip6h; |
310 | struct sk_buff *skb; |
311 | struct flowi fl; |
312 | struct net *net = dev_net(skb_dst(rxskb)->dev); |
313 | struct sock *ctl_sk = net->dccp.v6_ctl_sk; |
314 | struct dst_entry *dst; |
315 | |
316 | if (dccp_hdr(rxskb)->dccph_type == DCCP_PKT_RESET) |
317 | return; |
318 | |
319 | if (!ipv6_unicast_destination(rxskb)) |
320 | return; |
321 | |
322 | skb = dccp_ctl_make_reset(ctl_sk, rxskb); |
323 | if (skb == NULL) |
324 | return; |
325 | |
326 | rxip6h = ipv6_hdr(rxskb); |
327 | dccp_hdr(skb)->dccph_checksum = dccp_v6_csum_finish(skb, &rxip6h->saddr, |
328 | &rxip6h->daddr); |
329 | |
330 | memset(&fl, 0, sizeof(fl)); |
331 | ipv6_addr_copy(&fl.fl6_dst, &rxip6h->saddr); |
332 | ipv6_addr_copy(&fl.fl6_src, &rxip6h->daddr); |
333 | |
334 | fl.proto = IPPROTO_DCCP; |
335 | fl.oif = inet6_iif(rxskb); |
336 | fl.fl_ip_dport = dccp_hdr(skb)->dccph_dport; |
337 | fl.fl_ip_sport = dccp_hdr(skb)->dccph_sport; |
338 | security_skb_classify_flow(rxskb, &fl); |
339 | |
340 | /* sk = NULL, but it is safe for now. RST socket required. */ |
341 | if (!ip6_dst_lookup(ctl_sk, &dst, &fl)) { |
342 | if (xfrm_lookup(net, &dst, &fl, NULL, 0) >= 0) { |
343 | skb_dst_set(skb, dst); |
344 | ip6_xmit(ctl_sk, skb, &fl, NULL); |
345 | DCCP_INC_STATS_BH(DCCP_MIB_OUTSEGS); |
346 | DCCP_INC_STATS_BH(DCCP_MIB_OUTRSTS); |
347 | return; |
348 | } |
349 | } |
350 | |
351 | kfree_skb(skb); |
352 | } |
353 | |
354 | static struct request_sock_ops dccp6_request_sock_ops = { |
355 | .family = AF_INET6, |
356 | .obj_size = sizeof(struct dccp6_request_sock), |
357 | .rtx_syn_ack = dccp_v6_send_response, |
358 | .send_ack = dccp_reqsk_send_ack, |
359 | .destructor = dccp_v6_reqsk_destructor, |
360 | .send_reset = dccp_v6_ctl_send_reset, |
361 | }; |
362 | |
363 | static struct sock *dccp_v6_hnd_req(struct sock *sk,struct sk_buff *skb) |
364 | { |
365 | const struct dccp_hdr *dh = dccp_hdr(skb); |
366 | const struct ipv6hdr *iph = ipv6_hdr(skb); |
367 | struct sock *nsk; |
368 | struct request_sock **prev; |
369 | /* Find possible connection requests. */ |
370 | struct request_sock *req = inet6_csk_search_req(sk, &prev, |
371 | dh->dccph_sport, |
372 | &iph->saddr, |
373 | &iph->daddr, |
374 | inet6_iif(skb)); |
375 | if (req != NULL) |
376 | return dccp_check_req(sk, skb, req, prev); |
377 | |
378 | nsk = __inet6_lookup_established(sock_net(sk), &dccp_hashinfo, |
379 | &iph->saddr, dh->dccph_sport, |
380 | &iph->daddr, ntohs(dh->dccph_dport), |
381 | inet6_iif(skb)); |
382 | if (nsk != NULL) { |
383 | if (nsk->sk_state != DCCP_TIME_WAIT) { |
384 | bh_lock_sock(nsk); |
385 | return nsk; |
386 | } |
387 | inet_twsk_put(inet_twsk(nsk)); |
388 | return NULL; |
389 | } |
390 | |
391 | return sk; |
392 | } |
393 | |
394 | static int dccp_v6_conn_request(struct sock *sk, struct sk_buff *skb) |
395 | { |
396 | struct request_sock *req; |
397 | struct dccp_request_sock *dreq; |
398 | struct inet6_request_sock *ireq6; |
399 | struct ipv6_pinfo *np = inet6_sk(sk); |
400 | const __be32 service = dccp_hdr_request(skb)->dccph_req_service; |
401 | struct dccp_skb_cb *dcb = DCCP_SKB_CB(skb); |
402 | |
403 | if (skb->protocol == htons(ETH_P_IP)) |
404 | return dccp_v4_conn_request(sk, skb); |
405 | |
406 | if (!ipv6_unicast_destination(skb)) |
407 | return 0; /* discard, don't send a reset here */ |
408 | |
409 | if (dccp_bad_service_code(sk, service)) { |
410 | dcb->dccpd_reset_code = DCCP_RESET_CODE_BAD_SERVICE_CODE; |
411 | goto drop; |
412 | } |
413 | /* |
414 | * There are no SYN attacks on IPv6, yet... |
415 | */ |
416 | dcb->dccpd_reset_code = DCCP_RESET_CODE_TOO_BUSY; |
417 | if (inet_csk_reqsk_queue_is_full(sk)) |
418 | goto drop; |
419 | |
420 | if (sk_acceptq_is_full(sk) && inet_csk_reqsk_queue_young(sk) > 1) |
421 | goto drop; |
422 | |
423 | req = inet6_reqsk_alloc(&dccp6_request_sock_ops); |
424 | if (req == NULL) |
425 | goto drop; |
426 | |
427 | if (dccp_reqsk_init(req, dccp_sk(sk), skb)) |
428 | goto drop_and_free; |
429 | |
430 | dreq = dccp_rsk(req); |
431 | if (dccp_parse_options(sk, dreq, skb)) |
432 | goto drop_and_free; |
433 | |
434 | if (security_inet_conn_request(sk, skb, req)) |
435 | goto drop_and_free; |
436 | |
437 | ireq6 = inet6_rsk(req); |
438 | ipv6_addr_copy(&ireq6->rmt_addr, &ipv6_hdr(skb)->saddr); |
439 | ipv6_addr_copy(&ireq6->loc_addr, &ipv6_hdr(skb)->daddr); |
440 | |
441 | if (ipv6_opt_accepted(sk, skb) || |
442 | np->rxopt.bits.rxinfo || np->rxopt.bits.rxoinfo || |
443 | np->rxopt.bits.rxhlim || np->rxopt.bits.rxohlim) { |
444 | atomic_inc(&skb->users); |
445 | ireq6->pktopts = skb; |
446 | } |
447 | ireq6->iif = sk->sk_bound_dev_if; |
448 | |
449 | /* So that link locals have meaning */ |
450 | if (!sk->sk_bound_dev_if && |
451 | ipv6_addr_type(&ireq6->rmt_addr) & IPV6_ADDR_LINKLOCAL) |
452 | ireq6->iif = inet6_iif(skb); |
453 | |
454 | /* |
455 | * Step 3: Process LISTEN state |
456 | * |
457 | * Set S.ISR, S.GSR, S.SWL, S.SWH from packet or Init Cookie |
458 | * |
459 | * In fact we defer setting S.GSR, S.SWL, S.SWH to |
460 | * dccp_create_openreq_child. |
461 | */ |
462 | dreq->dreq_isr = dcb->dccpd_seq; |
463 | dreq->dreq_iss = dccp_v6_init_sequence(skb); |
464 | dreq->dreq_service = service; |
465 | |
466 | if (dccp_v6_send_response(sk, req, NULL)) |
467 | goto drop_and_free; |
468 | |
469 | inet6_csk_reqsk_queue_hash_add(sk, req, DCCP_TIMEOUT_INIT); |
470 | return 0; |
471 | |
472 | drop_and_free: |
473 | reqsk_free(req); |
474 | drop: |
475 | DCCP_INC_STATS_BH(DCCP_MIB_ATTEMPTFAILS); |
476 | return -1; |
477 | } |
478 | |
479 | static struct sock *dccp_v6_request_recv_sock(struct sock *sk, |
480 | struct sk_buff *skb, |
481 | struct request_sock *req, |
482 | struct dst_entry *dst) |
483 | { |
484 | struct inet6_request_sock *ireq6 = inet6_rsk(req); |
485 | struct ipv6_pinfo *newnp, *np = inet6_sk(sk); |
486 | struct inet_sock *newinet; |
487 | struct dccp_sock *newdp; |
488 | struct dccp6_sock *newdp6; |
489 | struct sock *newsk; |
490 | struct ipv6_txoptions *opt; |
491 | |
492 | if (skb->protocol == htons(ETH_P_IP)) { |
493 | /* |
494 | * v6 mapped |
495 | */ |
496 | newsk = dccp_v4_request_recv_sock(sk, skb, req, dst); |
497 | if (newsk == NULL) |
498 | return NULL; |
499 | |
500 | newdp6 = (struct dccp6_sock *)newsk; |
501 | newdp = dccp_sk(newsk); |
502 | newinet = inet_sk(newsk); |
503 | newinet->pinet6 = &newdp6->inet6; |
504 | newnp = inet6_sk(newsk); |
505 | |
506 | memcpy(newnp, np, sizeof(struct ipv6_pinfo)); |
507 | |
508 | ipv6_addr_set_v4mapped(newinet->inet_daddr, &newnp->daddr); |
509 | |
510 | ipv6_addr_set_v4mapped(newinet->inet_saddr, &newnp->saddr); |
511 | |
512 | ipv6_addr_copy(&newnp->rcv_saddr, &newnp->saddr); |
513 | |
514 | inet_csk(newsk)->icsk_af_ops = &dccp_ipv6_mapped; |
515 | newsk->sk_backlog_rcv = dccp_v4_do_rcv; |
516 | newnp->pktoptions = NULL; |
517 | newnp->opt = NULL; |
518 | newnp->mcast_oif = inet6_iif(skb); |
519 | newnp->mcast_hops = ipv6_hdr(skb)->hop_limit; |
520 | |
521 | /* |
522 | * No need to charge this sock to the relevant IPv6 refcnt debug socks count |
523 | * here, dccp_create_openreq_child now does this for us, see the comment in |
524 | * that function for the gory details. -acme |
525 | */ |
526 | |
527 | /* It is tricky place. Until this moment IPv4 tcp |
528 | worked with IPv6 icsk.icsk_af_ops. |
529 | Sync it now. |
530 | */ |
531 | dccp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie); |
532 | |
533 | return newsk; |
534 | } |
535 | |
536 | opt = np->opt; |
537 | |
538 | if (sk_acceptq_is_full(sk)) |
539 | goto out_overflow; |
540 | |
541 | if (dst == NULL) { |
542 | struct in6_addr *final_p, final; |
543 | struct flowi fl; |
544 | |
545 | memset(&fl, 0, sizeof(fl)); |
546 | fl.proto = IPPROTO_DCCP; |
547 | ipv6_addr_copy(&fl.fl6_dst, &ireq6->rmt_addr); |
548 | final_p = fl6_update_dst(&fl, opt, &final); |
549 | ipv6_addr_copy(&fl.fl6_src, &ireq6->loc_addr); |
550 | fl.oif = sk->sk_bound_dev_if; |
551 | fl.fl_ip_dport = inet_rsk(req)->rmt_port; |
552 | fl.fl_ip_sport = inet_rsk(req)->loc_port; |
553 | security_sk_classify_flow(sk, &fl); |
554 | |
555 | if (ip6_dst_lookup(sk, &dst, &fl)) |
556 | goto out; |
557 | |
558 | if (final_p) |
559 | ipv6_addr_copy(&fl.fl6_dst, final_p); |
560 | |
561 | if ((xfrm_lookup(sock_net(sk), &dst, &fl, sk, 0)) < 0) |
562 | goto out; |
563 | } |
564 | |
565 | newsk = dccp_create_openreq_child(sk, req, skb); |
566 | if (newsk == NULL) |
567 | goto out; |
568 | |
569 | /* |
570 | * No need to charge this sock to the relevant IPv6 refcnt debug socks |
571 | * count here, dccp_create_openreq_child now does this for us, see the |
572 | * comment in that function for the gory details. -acme |
573 | */ |
574 | |
575 | __ip6_dst_store(newsk, dst, NULL, NULL); |
576 | newsk->sk_route_caps = dst->dev->features & ~(NETIF_F_IP_CSUM | |
577 | NETIF_F_TSO); |
578 | newdp6 = (struct dccp6_sock *)newsk; |
579 | newinet = inet_sk(newsk); |
580 | newinet->pinet6 = &newdp6->inet6; |
581 | newdp = dccp_sk(newsk); |
582 | newnp = inet6_sk(newsk); |
583 | |
584 | memcpy(newnp, np, sizeof(struct ipv6_pinfo)); |
585 | |
586 | ipv6_addr_copy(&newnp->daddr, &ireq6->rmt_addr); |
587 | ipv6_addr_copy(&newnp->saddr, &ireq6->loc_addr); |
588 | ipv6_addr_copy(&newnp->rcv_saddr, &ireq6->loc_addr); |
589 | newsk->sk_bound_dev_if = ireq6->iif; |
590 | |
591 | /* Now IPv6 options... |
592 | |
593 | First: no IPv4 options. |
594 | */ |
595 | newinet->opt = NULL; |
596 | |
597 | /* Clone RX bits */ |
598 | newnp->rxopt.all = np->rxopt.all; |
599 | |
600 | /* Clone pktoptions received with SYN */ |
601 | newnp->pktoptions = NULL; |
602 | if (ireq6->pktopts != NULL) { |
603 | newnp->pktoptions = skb_clone(ireq6->pktopts, GFP_ATOMIC); |
604 | kfree_skb(ireq6->pktopts); |
605 | ireq6->pktopts = NULL; |
606 | if (newnp->pktoptions) |
607 | skb_set_owner_r(newnp->pktoptions, newsk); |
608 | } |
609 | newnp->opt = NULL; |
610 | newnp->mcast_oif = inet6_iif(skb); |
611 | newnp->mcast_hops = ipv6_hdr(skb)->hop_limit; |
612 | |
613 | /* |
614 | * Clone native IPv6 options from listening socket (if any) |
615 | * |
616 | * Yes, keeping reference count would be much more clever, but we make |
617 | * one more one thing there: reattach optmem to newsk. |
618 | */ |
619 | if (opt != NULL) { |
620 | newnp->opt = ipv6_dup_options(newsk, opt); |
621 | if (opt != np->opt) |
622 | sock_kfree_s(sk, opt, opt->tot_len); |
623 | } |
624 | |
625 | inet_csk(newsk)->icsk_ext_hdr_len = 0; |
626 | if (newnp->opt != NULL) |
627 | inet_csk(newsk)->icsk_ext_hdr_len = (newnp->opt->opt_nflen + |
628 | newnp->opt->opt_flen); |
629 | |
630 | dccp_sync_mss(newsk, dst_mtu(dst)); |
631 | |
632 | newinet->inet_daddr = newinet->inet_saddr = LOOPBACK4_IPV6; |
633 | newinet->inet_rcv_saddr = LOOPBACK4_IPV6; |
634 | |
635 | __inet6_hash(newsk, NULL); |
636 | __inet_inherit_port(sk, newsk); |
637 | |
638 | return newsk; |
639 | |
640 | out_overflow: |
641 | NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS); |
642 | out: |
643 | NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_LISTENDROPS); |
644 | if (opt != NULL && opt != np->opt) |
645 | sock_kfree_s(sk, opt, opt->tot_len); |
646 | dst_release(dst); |
647 | return NULL; |
648 | } |
649 | |
650 | /* The socket must have it's spinlock held when we get |
651 | * here. |
652 | * |
653 | * We have a potential double-lock case here, so even when |
654 | * doing backlog processing we use the BH locking scheme. |
655 | * This is because we cannot sleep with the original spinlock |
656 | * held. |
657 | */ |
658 | static int dccp_v6_do_rcv(struct sock *sk, struct sk_buff *skb) |
659 | { |
660 | struct ipv6_pinfo *np = inet6_sk(sk); |
661 | struct sk_buff *opt_skb = NULL; |
662 | |
663 | /* Imagine: socket is IPv6. IPv4 packet arrives, |
664 | goes to IPv4 receive handler and backlogged. |
665 | From backlog it always goes here. Kerboom... |
666 | Fortunately, dccp_rcv_established and rcv_established |
667 | handle them correctly, but it is not case with |
668 | dccp_v6_hnd_req and dccp_v6_ctl_send_reset(). --ANK |
669 | */ |
670 | |
671 | if (skb->protocol == htons(ETH_P_IP)) |
672 | return dccp_v4_do_rcv(sk, skb); |
673 | |
674 | if (sk_filter(sk, skb)) |
675 | goto discard; |
676 | |
677 | /* |
678 | * socket locking is here for SMP purposes as backlog rcv is currently |
679 | * called with bh processing disabled. |
680 | */ |
681 | |
682 | /* Do Stevens' IPV6_PKTOPTIONS. |
683 | |
684 | Yes, guys, it is the only place in our code, where we |
685 | may make it not affecting IPv4. |
686 | The rest of code is protocol independent, |
687 | and I do not like idea to uglify IPv4. |
688 | |
689 | Actually, all the idea behind IPV6_PKTOPTIONS |
690 | looks not very well thought. For now we latch |
691 | options, received in the last packet, enqueued |
692 | by tcp. Feel free to propose better solution. |
693 | --ANK (980728) |
694 | */ |
695 | if (np->rxopt.all) |
696 | /* |
697 | * FIXME: Add handling of IPV6_PKTOPTIONS skb. See the comments below |
698 | * (wrt ipv6_pktopions) and net/ipv6/tcp_ipv6.c for an example. |
699 | */ |
700 | opt_skb = skb_clone(skb, GFP_ATOMIC); |
701 | |
702 | if (sk->sk_state == DCCP_OPEN) { /* Fast path */ |
703 | if (dccp_rcv_established(sk, skb, dccp_hdr(skb), skb->len)) |
704 | goto reset; |
705 | if (opt_skb) { |
706 | /* XXX This is where we would goto ipv6_pktoptions. */ |
707 | __kfree_skb(opt_skb); |
708 | } |
709 | return 0; |
710 | } |
711 | |
712 | /* |
713 | * Step 3: Process LISTEN state |
714 | * If S.state == LISTEN, |
715 | * If P.type == Request or P contains a valid Init Cookie option, |
716 | * (* Must scan the packet's options to check for Init |
717 | * Cookies. Only Init Cookies are processed here, |
718 | * however; other options are processed in Step 8. This |
719 | * scan need only be performed if the endpoint uses Init |
720 | * Cookies *) |
721 | * (* Generate a new socket and switch to that socket *) |
722 | * Set S := new socket for this port pair |
723 | * S.state = RESPOND |
724 | * Choose S.ISS (initial seqno) or set from Init Cookies |
725 | * Initialize S.GAR := S.ISS |
726 | * Set S.ISR, S.GSR, S.SWL, S.SWH from packet or Init Cookies |
727 | * Continue with S.state == RESPOND |
728 | * (* A Response packet will be generated in Step 11 *) |
729 | * Otherwise, |
730 | * Generate Reset(No Connection) unless P.type == Reset |
731 | * Drop packet and return |
732 | * |
733 | * NOTE: the check for the packet types is done in |
734 | * dccp_rcv_state_process |
735 | */ |
736 | if (sk->sk_state == DCCP_LISTEN) { |
737 | struct sock *nsk = dccp_v6_hnd_req(sk, skb); |
738 | |
739 | if (nsk == NULL) |
740 | goto discard; |
741 | /* |
742 | * Queue it on the new socket if the new socket is active, |
743 | * otherwise we just shortcircuit this and continue with |
744 | * the new socket.. |
745 | */ |
746 | if (nsk != sk) { |
747 | if (dccp_child_process(sk, nsk, skb)) |
748 | goto reset; |
749 | if (opt_skb != NULL) |
750 | __kfree_skb(opt_skb); |
751 | return 0; |
752 | } |
753 | } |
754 | |
755 | if (dccp_rcv_state_process(sk, skb, dccp_hdr(skb), skb->len)) |
756 | goto reset; |
757 | if (opt_skb) { |
758 | /* XXX This is where we would goto ipv6_pktoptions. */ |
759 | __kfree_skb(opt_skb); |
760 | } |
761 | return 0; |
762 | |
763 | reset: |
764 | dccp_v6_ctl_send_reset(sk, skb); |
765 | discard: |
766 | if (opt_skb != NULL) |
767 | __kfree_skb(opt_skb); |
768 | kfree_skb(skb); |
769 | return 0; |
770 | } |
771 | |
772 | static int dccp_v6_rcv(struct sk_buff *skb) |
773 | { |
774 | const struct dccp_hdr *dh; |
775 | struct sock *sk; |
776 | int min_cov; |
777 | |
778 | /* Step 1: Check header basics */ |
779 | |
780 | if (dccp_invalid_packet(skb)) |
781 | goto discard_it; |
782 | |
783 | /* Step 1: If header checksum is incorrect, drop packet and return. */ |
784 | if (dccp_v6_csum_finish(skb, &ipv6_hdr(skb)->saddr, |
785 | &ipv6_hdr(skb)->daddr)) { |
786 | DCCP_WARN("dropped packet with invalid checksum\n"); |
787 | goto discard_it; |
788 | } |
789 | |
790 | dh = dccp_hdr(skb); |
791 | |
792 | DCCP_SKB_CB(skb)->dccpd_seq = dccp_hdr_seq(dh); |
793 | DCCP_SKB_CB(skb)->dccpd_type = dh->dccph_type; |
794 | |
795 | if (dccp_packet_without_ack(skb)) |
796 | DCCP_SKB_CB(skb)->dccpd_ack_seq = DCCP_PKT_WITHOUT_ACK_SEQ; |
797 | else |
798 | DCCP_SKB_CB(skb)->dccpd_ack_seq = dccp_hdr_ack_seq(skb); |
799 | |
800 | /* Step 2: |
801 | * Look up flow ID in table and get corresponding socket */ |
802 | sk = __inet6_lookup_skb(&dccp_hashinfo, skb, |
803 | dh->dccph_sport, dh->dccph_dport); |
804 | /* |
805 | * Step 2: |
806 | * If no socket ... |
807 | */ |
808 | if (sk == NULL) { |
809 | dccp_pr_debug("failed to look up flow ID in table and " |
810 | "get corresponding socket\n"); |
811 | goto no_dccp_socket; |
812 | } |
813 | |
814 | /* |
815 | * Step 2: |
816 | * ... or S.state == TIMEWAIT, |
817 | * Generate Reset(No Connection) unless P.type == Reset |
818 | * Drop packet and return |
819 | */ |
820 | if (sk->sk_state == DCCP_TIME_WAIT) { |
821 | dccp_pr_debug("sk->sk_state == DCCP_TIME_WAIT: do_time_wait\n"); |
822 | inet_twsk_put(inet_twsk(sk)); |
823 | goto no_dccp_socket; |
824 | } |
825 | |
826 | /* |
827 | * RFC 4340, sec. 9.2.1: Minimum Checksum Coverage |
828 | * o if MinCsCov = 0, only packets with CsCov = 0 are accepted |
829 | * o if MinCsCov > 0, also accept packets with CsCov >= MinCsCov |
830 | */ |
831 | min_cov = dccp_sk(sk)->dccps_pcrlen; |
832 | if (dh->dccph_cscov && (min_cov == 0 || dh->dccph_cscov < min_cov)) { |
833 | dccp_pr_debug("Packet CsCov %d does not satisfy MinCsCov %d\n", |
834 | dh->dccph_cscov, min_cov); |
835 | /* FIXME: send Data Dropped option (see also dccp_v4_rcv) */ |
836 | goto discard_and_relse; |
837 | } |
838 | |
839 | if (!xfrm6_policy_check(sk, XFRM_POLICY_IN, skb)) |
840 | goto discard_and_relse; |
841 | |
842 | return sk_receive_skb(sk, skb, 1) ? -1 : 0; |
843 | |
844 | no_dccp_socket: |
845 | if (!xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb)) |
846 | goto discard_it; |
847 | /* |
848 | * Step 2: |
849 | * If no socket ... |
850 | * Generate Reset(No Connection) unless P.type == Reset |
851 | * Drop packet and return |
852 | */ |
853 | if (dh->dccph_type != DCCP_PKT_RESET) { |
854 | DCCP_SKB_CB(skb)->dccpd_reset_code = |
855 | DCCP_RESET_CODE_NO_CONNECTION; |
856 | dccp_v6_ctl_send_reset(sk, skb); |
857 | } |
858 | |
859 | discard_it: |
860 | kfree_skb(skb); |
861 | return 0; |
862 | |
863 | discard_and_relse: |
864 | sock_put(sk); |
865 | goto discard_it; |
866 | } |
867 | |
868 | static int dccp_v6_connect(struct sock *sk, struct sockaddr *uaddr, |
869 | int addr_len) |
870 | { |
871 | struct sockaddr_in6 *usin = (struct sockaddr_in6 *)uaddr; |
872 | struct inet_connection_sock *icsk = inet_csk(sk); |
873 | struct inet_sock *inet = inet_sk(sk); |
874 | struct ipv6_pinfo *np = inet6_sk(sk); |
875 | struct dccp_sock *dp = dccp_sk(sk); |
876 | struct in6_addr *saddr = NULL, *final_p, final; |
877 | struct flowi fl; |
878 | struct dst_entry *dst; |
879 | int addr_type; |
880 | int err; |
881 | |
882 | dp->dccps_role = DCCP_ROLE_CLIENT; |
883 | |
884 | if (addr_len < SIN6_LEN_RFC2133) |
885 | return -EINVAL; |
886 | |
887 | if (usin->sin6_family != AF_INET6) |
888 | return -EAFNOSUPPORT; |
889 | |
890 | memset(&fl, 0, sizeof(fl)); |
891 | |
892 | if (np->sndflow) { |
893 | fl.fl6_flowlabel = usin->sin6_flowinfo & IPV6_FLOWINFO_MASK; |
894 | IP6_ECN_flow_init(fl.fl6_flowlabel); |
895 | if (fl.fl6_flowlabel & IPV6_FLOWLABEL_MASK) { |
896 | struct ip6_flowlabel *flowlabel; |
897 | flowlabel = fl6_sock_lookup(sk, fl.fl6_flowlabel); |
898 | if (flowlabel == NULL) |
899 | return -EINVAL; |
900 | ipv6_addr_copy(&usin->sin6_addr, &flowlabel->dst); |
901 | fl6_sock_release(flowlabel); |
902 | } |
903 | } |
904 | /* |
905 | * connect() to INADDR_ANY means loopback (BSD'ism). |
906 | */ |
907 | if (ipv6_addr_any(&usin->sin6_addr)) |
908 | usin->sin6_addr.s6_addr[15] = 1; |
909 | |
910 | addr_type = ipv6_addr_type(&usin->sin6_addr); |
911 | |
912 | if (addr_type & IPV6_ADDR_MULTICAST) |
913 | return -ENETUNREACH; |
914 | |
915 | if (addr_type & IPV6_ADDR_LINKLOCAL) { |
916 | if (addr_len >= sizeof(struct sockaddr_in6) && |
917 | usin->sin6_scope_id) { |
918 | /* If interface is set while binding, indices |
919 | * must coincide. |
920 | */ |
921 | if (sk->sk_bound_dev_if && |
922 | sk->sk_bound_dev_if != usin->sin6_scope_id) |
923 | return -EINVAL; |
924 | |
925 | sk->sk_bound_dev_if = usin->sin6_scope_id; |
926 | } |
927 | |
928 | /* Connect to link-local address requires an interface */ |
929 | if (!sk->sk_bound_dev_if) |
930 | return -EINVAL; |
931 | } |
932 | |
933 | ipv6_addr_copy(&np->daddr, &usin->sin6_addr); |
934 | np->flow_label = fl.fl6_flowlabel; |
935 | |
936 | /* |
937 | * DCCP over IPv4 |
938 | */ |
939 | if (addr_type == IPV6_ADDR_MAPPED) { |
940 | u32 exthdrlen = icsk->icsk_ext_hdr_len; |
941 | struct sockaddr_in sin; |
942 | |
943 | SOCK_DEBUG(sk, "connect: ipv4 mapped\n"); |
944 | |
945 | if (__ipv6_only_sock(sk)) |
946 | return -ENETUNREACH; |
947 | |
948 | sin.sin_family = AF_INET; |
949 | sin.sin_port = usin->sin6_port; |
950 | sin.sin_addr.s_addr = usin->sin6_addr.s6_addr32[3]; |
951 | |
952 | icsk->icsk_af_ops = &dccp_ipv6_mapped; |
953 | sk->sk_backlog_rcv = dccp_v4_do_rcv; |
954 | |
955 | err = dccp_v4_connect(sk, (struct sockaddr *)&sin, sizeof(sin)); |
956 | if (err) { |
957 | icsk->icsk_ext_hdr_len = exthdrlen; |
958 | icsk->icsk_af_ops = &dccp_ipv6_af_ops; |
959 | sk->sk_backlog_rcv = dccp_v6_do_rcv; |
960 | goto failure; |
961 | } |
962 | ipv6_addr_set_v4mapped(inet->inet_saddr, &np->saddr); |
963 | ipv6_addr_set_v4mapped(inet->inet_rcv_saddr, &np->rcv_saddr); |
964 | |
965 | return err; |
966 | } |
967 | |
968 | if (!ipv6_addr_any(&np->rcv_saddr)) |
969 | saddr = &np->rcv_saddr; |
970 | |
971 | fl.proto = IPPROTO_DCCP; |
972 | ipv6_addr_copy(&fl.fl6_dst, &np->daddr); |
973 | ipv6_addr_copy(&fl.fl6_src, saddr ? saddr : &np->saddr); |
974 | fl.oif = sk->sk_bound_dev_if; |
975 | fl.fl_ip_dport = usin->sin6_port; |
976 | fl.fl_ip_sport = inet->inet_sport; |
977 | security_sk_classify_flow(sk, &fl); |
978 | |
979 | final_p = fl6_update_dst(&fl, np->opt, &final); |
980 | |
981 | err = ip6_dst_lookup(sk, &dst, &fl); |
982 | if (err) |
983 | goto failure; |
984 | |
985 | if (final_p) |
986 | ipv6_addr_copy(&fl.fl6_dst, final_p); |
987 | |
988 | err = __xfrm_lookup(sock_net(sk), &dst, &fl, sk, XFRM_LOOKUP_WAIT); |
989 | if (err < 0) { |
990 | if (err == -EREMOTE) |
991 | err = ip6_dst_blackhole(sk, &dst, &fl); |
992 | if (err < 0) |
993 | goto failure; |
994 | } |
995 | |
996 | if (saddr == NULL) { |
997 | saddr = &fl.fl6_src; |
998 | ipv6_addr_copy(&np->rcv_saddr, saddr); |
999 | } |
1000 | |
1001 | /* set the source address */ |
1002 | ipv6_addr_copy(&np->saddr, saddr); |
1003 | inet->inet_rcv_saddr = LOOPBACK4_IPV6; |
1004 | |
1005 | __ip6_dst_store(sk, dst, NULL, NULL); |
1006 | |
1007 | icsk->icsk_ext_hdr_len = 0; |
1008 | if (np->opt != NULL) |
1009 | icsk->icsk_ext_hdr_len = (np->opt->opt_flen + |
1010 | np->opt->opt_nflen); |
1011 | |
1012 | inet->inet_dport = usin->sin6_port; |
1013 | |
1014 | dccp_set_state(sk, DCCP_REQUESTING); |
1015 | err = inet6_hash_connect(&dccp_death_row, sk); |
1016 | if (err) |
1017 | goto late_failure; |
1018 | |
1019 | dp->dccps_iss = secure_dccpv6_sequence_number(np->saddr.s6_addr32, |
1020 | np->daddr.s6_addr32, |
1021 | inet->inet_sport, |
1022 | inet->inet_dport); |
1023 | err = dccp_connect(sk); |
1024 | if (err) |
1025 | goto late_failure; |
1026 | |
1027 | return 0; |
1028 | |
1029 | late_failure: |
1030 | dccp_set_state(sk, DCCP_CLOSED); |
1031 | __sk_dst_reset(sk); |
1032 | failure: |
1033 | inet->inet_dport = 0; |
1034 | sk->sk_route_caps = 0; |
1035 | return err; |
1036 | } |
1037 | |
1038 | static const struct inet_connection_sock_af_ops dccp_ipv6_af_ops = { |
1039 | .queue_xmit = inet6_csk_xmit, |
1040 | .send_check = dccp_v6_send_check, |
1041 | .rebuild_header = inet6_sk_rebuild_header, |
1042 | .conn_request = dccp_v6_conn_request, |
1043 | .syn_recv_sock = dccp_v6_request_recv_sock, |
1044 | .net_header_len = sizeof(struct ipv6hdr), |
1045 | .setsockopt = ipv6_setsockopt, |
1046 | .getsockopt = ipv6_getsockopt, |
1047 | .addr2sockaddr = inet6_csk_addr2sockaddr, |
1048 | .sockaddr_len = sizeof(struct sockaddr_in6), |
1049 | .bind_conflict = inet6_csk_bind_conflict, |
1050 | #ifdef CONFIG_COMPAT |
1051 | .compat_setsockopt = compat_ipv6_setsockopt, |
1052 | .compat_getsockopt = compat_ipv6_getsockopt, |
1053 | #endif |
1054 | }; |
1055 | |
1056 | /* |
1057 | * DCCP over IPv4 via INET6 API |
1058 | */ |
1059 | static const struct inet_connection_sock_af_ops dccp_ipv6_mapped = { |
1060 | .queue_xmit = ip_queue_xmit, |
1061 | .send_check = dccp_v4_send_check, |
1062 | .rebuild_header = inet_sk_rebuild_header, |
1063 | .conn_request = dccp_v6_conn_request, |
1064 | .syn_recv_sock = dccp_v6_request_recv_sock, |
1065 | .net_header_len = sizeof(struct iphdr), |
1066 | .setsockopt = ipv6_setsockopt, |
1067 | .getsockopt = ipv6_getsockopt, |
1068 | .addr2sockaddr = inet6_csk_addr2sockaddr, |
1069 | .sockaddr_len = sizeof(struct sockaddr_in6), |
1070 | #ifdef CONFIG_COMPAT |
1071 | .compat_setsockopt = compat_ipv6_setsockopt, |
1072 | .compat_getsockopt = compat_ipv6_getsockopt, |
1073 | #endif |
1074 | }; |
1075 | |
1076 | /* NOTE: A lot of things set to zero explicitly by call to |
1077 | * sk_alloc() so need not be done here. |
1078 | */ |
1079 | static int dccp_v6_init_sock(struct sock *sk) |
1080 | { |
1081 | static __u8 dccp_v6_ctl_sock_initialized; |
1082 | int err = dccp_init_sock(sk, dccp_v6_ctl_sock_initialized); |
1083 | |
1084 | if (err == 0) { |
1085 | if (unlikely(!dccp_v6_ctl_sock_initialized)) |
1086 | dccp_v6_ctl_sock_initialized = 1; |
1087 | inet_csk(sk)->icsk_af_ops = &dccp_ipv6_af_ops; |
1088 | } |
1089 | |
1090 | return err; |
1091 | } |
1092 | |
1093 | static void dccp_v6_destroy_sock(struct sock *sk) |
1094 | { |
1095 | dccp_destroy_sock(sk); |
1096 | inet6_destroy_sock(sk); |
1097 | } |
1098 | |
1099 | static struct timewait_sock_ops dccp6_timewait_sock_ops = { |
1100 | .twsk_obj_size = sizeof(struct dccp6_timewait_sock), |
1101 | }; |
1102 | |
1103 | static struct proto dccp_v6_prot = { |
1104 | .name = "DCCPv6", |
1105 | .owner = THIS_MODULE, |
1106 | .close = dccp_close, |
1107 | .connect = dccp_v6_connect, |
1108 | .disconnect = dccp_disconnect, |
1109 | .ioctl = dccp_ioctl, |
1110 | .init = dccp_v6_init_sock, |
1111 | .setsockopt = dccp_setsockopt, |
1112 | .getsockopt = dccp_getsockopt, |
1113 | .sendmsg = dccp_sendmsg, |
1114 | .recvmsg = dccp_recvmsg, |
1115 | .backlog_rcv = dccp_v6_do_rcv, |
1116 | .hash = dccp_v6_hash, |
1117 | .unhash = inet_unhash, |
1118 | .accept = inet_csk_accept, |
1119 | .get_port = inet_csk_get_port, |
1120 | .shutdown = dccp_shutdown, |
1121 | .destroy = dccp_v6_destroy_sock, |
1122 | .orphan_count = &dccp_orphan_count, |
1123 | .max_header = MAX_DCCP_HEADER, |
1124 | .obj_size = sizeof(struct dccp6_sock), |
1125 | .slab_flags = SLAB_DESTROY_BY_RCU, |
1126 | .rsk_prot = &dccp6_request_sock_ops, |
1127 | .twsk_prot = &dccp6_timewait_sock_ops, |
1128 | .h.hashinfo = &dccp_hashinfo, |
1129 | #ifdef CONFIG_COMPAT |
1130 | .compat_setsockopt = compat_dccp_setsockopt, |
1131 | .compat_getsockopt = compat_dccp_getsockopt, |
1132 | #endif |
1133 | }; |
1134 | |
1135 | static const struct inet6_protocol dccp_v6_protocol = { |
1136 | .handler = dccp_v6_rcv, |
1137 | .err_handler = dccp_v6_err, |
1138 | .flags = INET6_PROTO_NOPOLICY | INET6_PROTO_FINAL, |
1139 | }; |
1140 | |
1141 | static const struct proto_ops inet6_dccp_ops = { |
1142 | .family = PF_INET6, |
1143 | .owner = THIS_MODULE, |
1144 | .release = inet6_release, |
1145 | .bind = inet6_bind, |
1146 | .connect = inet_stream_connect, |
1147 | .socketpair = sock_no_socketpair, |
1148 | .accept = inet_accept, |
1149 | .getname = inet6_getname, |
1150 | .poll = dccp_poll, |
1151 | .ioctl = inet6_ioctl, |
1152 | .listen = inet_dccp_listen, |
1153 | .shutdown = inet_shutdown, |
1154 | .setsockopt = sock_common_setsockopt, |
1155 | .getsockopt = sock_common_getsockopt, |
1156 | .sendmsg = inet_sendmsg, |
1157 | .recvmsg = sock_common_recvmsg, |
1158 | .mmap = sock_no_mmap, |
1159 | .sendpage = sock_no_sendpage, |
1160 | #ifdef CONFIG_COMPAT |
1161 | .compat_setsockopt = compat_sock_common_setsockopt, |
1162 | .compat_getsockopt = compat_sock_common_getsockopt, |
1163 | #endif |
1164 | }; |
1165 | |
1166 | static struct inet_protosw dccp_v6_protosw = { |
1167 | .type = SOCK_DCCP, |
1168 | .protocol = IPPROTO_DCCP, |
1169 | .prot = &dccp_v6_prot, |
1170 | .ops = &inet6_dccp_ops, |
1171 | .flags = INET_PROTOSW_ICSK, |
1172 | }; |
1173 | |
1174 | static int __net_init dccp_v6_init_net(struct net *net) |
1175 | { |
1176 | if (dccp_hashinfo.bhash == NULL) |
1177 | return -ESOCKTNOSUPPORT; |
1178 | |
1179 | return inet_ctl_sock_create(&net->dccp.v6_ctl_sk, PF_INET6, |
1180 | SOCK_DCCP, IPPROTO_DCCP, net); |
1181 | } |
1182 | |
1183 | static void __net_exit dccp_v6_exit_net(struct net *net) |
1184 | { |
1185 | inet_ctl_sock_destroy(net->dccp.v6_ctl_sk); |
1186 | } |
1187 | |
1188 | static struct pernet_operations dccp_v6_ops = { |
1189 | .init = dccp_v6_init_net, |
1190 | .exit = dccp_v6_exit_net, |
1191 | }; |
1192 | |
1193 | static int __init dccp_v6_init(void) |
1194 | { |
1195 | int err = proto_register(&dccp_v6_prot, 1); |
1196 | |
1197 | if (err != 0) |
1198 | goto out; |
1199 | |
1200 | err = inet6_add_protocol(&dccp_v6_protocol, IPPROTO_DCCP); |
1201 | if (err != 0) |
1202 | goto out_unregister_proto; |
1203 | |
1204 | inet6_register_protosw(&dccp_v6_protosw); |
1205 | |
1206 | err = register_pernet_subsys(&dccp_v6_ops); |
1207 | if (err != 0) |
1208 | goto out_destroy_ctl_sock; |
1209 | out: |
1210 | return err; |
1211 | |
1212 | out_destroy_ctl_sock: |
1213 | inet6_del_protocol(&dccp_v6_protocol, IPPROTO_DCCP); |
1214 | inet6_unregister_protosw(&dccp_v6_protosw); |
1215 | out_unregister_proto: |
1216 | proto_unregister(&dccp_v6_prot); |
1217 | goto out; |
1218 | } |
1219 | |
1220 | static void __exit dccp_v6_exit(void) |
1221 | { |
1222 | unregister_pernet_subsys(&dccp_v6_ops); |
1223 | inet6_del_protocol(&dccp_v6_protocol, IPPROTO_DCCP); |
1224 | inet6_unregister_protosw(&dccp_v6_protosw); |
1225 | proto_unregister(&dccp_v6_prot); |
1226 | } |
1227 | |
1228 | module_init(dccp_v6_init); |
1229 | module_exit(dccp_v6_exit); |
1230 | |
1231 | /* |
1232 | * __stringify doesn't likes enums, so use SOCK_DCCP (6) and IPPROTO_DCCP (33) |
1233 | * values directly, Also cover the case where the protocol is not specified, |
1234 | * i.e. net-pf-PF_INET6-proto-0-type-SOCK_DCCP |
1235 | */ |
1236 | MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_INET6, 33, 6); |
1237 | MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_INET6, 0, 6); |
1238 | MODULE_LICENSE("GPL"); |
1239 | MODULE_AUTHOR("Arnaldo Carvalho de Melo <acme@mandriva.com>"); |
1240 | MODULE_DESCRIPTION("DCCPv6 - Datagram Congestion Controlled Protocol"); |
1241 |
Branches:
ben-wpan
ben-wpan-stefan
javiroman/ks7010
jz-2.6.34
jz-2.6.34-rc5
jz-2.6.34-rc6
jz-2.6.34-rc7
jz-2.6.35
jz-2.6.36
jz-2.6.37
jz-2.6.38
jz-2.6.39
jz-3.0
jz-3.1
jz-3.11
jz-3.12
jz-3.13
jz-3.15
jz-3.16
jz-3.18-dt
jz-3.2
jz-3.3
jz-3.4
jz-3.5
jz-3.6
jz-3.6-rc2-pwm
jz-3.9
jz-3.9-clk
jz-3.9-rc8
jz47xx
jz47xx-2.6.38
master
Tags:
od-2011-09-04
od-2011-09-18
v2.6.34-rc5
v2.6.34-rc6
v2.6.34-rc7
v3.9