Root/
1 | /* |
2 | * X.25 Packet Layer release 002 |
3 | * |
4 | * This is ALPHA test software. This code may break your machine, |
5 | * randomly fail to work with new releases, misbehave and/or generally |
6 | * screw up. It might even work. |
7 | * |
8 | * This code REQUIRES 2.1.15 or higher |
9 | * |
10 | * This module: |
11 | * This module is free software; you can redistribute it and/or |
12 | * modify it under the terms of the GNU General Public License |
13 | * as published by the Free Software Foundation; either version |
14 | * 2 of the License, or (at your option) any later version. |
15 | * |
16 | * History |
17 | * X.25 001 Jonathan Naylor Started coding. |
18 | * X.25 002 Jonathan Naylor Centralised disconnect handling. |
19 | * New timer architecture. |
20 | * 2000-03-11 Henner Eisen MSG_EOR handling more POSIX compliant. |
21 | * 2000-03-22 Daniela Squassoni Allowed disabling/enabling of |
22 | * facilities negotiation and increased |
23 | * the throughput upper limit. |
24 | * 2000-08-27 Arnaldo C. Melo s/suser/capable/ + micro cleanups |
25 | * 2000-09-04 Henner Eisen Set sock->state in x25_accept(). |
26 | * Fixed x25_output() related skb leakage. |
27 | * 2000-10-02 Henner Eisen Made x25_kick() single threaded per socket. |
28 | * 2000-10-27 Henner Eisen MSG_DONTWAIT for fragment allocation. |
29 | * 2000-11-14 Henner Eisen Closing datalink from NETDEV_GOING_DOWN |
30 | * 2002-10-06 Arnaldo C. Melo Get rid of cli/sti, move proc stuff to |
31 | * x25_proc.c, using seq_file |
32 | * 2005-04-02 Shaun Pereira Selective sub address matching |
33 | * with call user data |
34 | * 2005-04-15 Shaun Pereira Fast select with no restriction on |
35 | * response |
36 | */ |
37 | |
38 | #include <linux/module.h> |
39 | #include <linux/capability.h> |
40 | #include <linux/errno.h> |
41 | #include <linux/kernel.h> |
42 | #include <linux/sched.h> |
43 | #include <linux/smp_lock.h> |
44 | #include <linux/timer.h> |
45 | #include <linux/string.h> |
46 | #include <linux/net.h> |
47 | #include <linux/netdevice.h> |
48 | #include <linux/if_arp.h> |
49 | #include <linux/skbuff.h> |
50 | #include <linux/slab.h> |
51 | #include <net/sock.h> |
52 | #include <net/tcp_states.h> |
53 | #include <asm/uaccess.h> |
54 | #include <linux/fcntl.h> |
55 | #include <linux/termios.h> /* For TIOCINQ/OUTQ */ |
56 | #include <linux/notifier.h> |
57 | #include <linux/init.h> |
58 | #include <linux/compat.h> |
59 | #include <linux/ctype.h> |
60 | |
61 | #include <net/x25.h> |
62 | #include <net/compat.h> |
63 | |
64 | int sysctl_x25_restart_request_timeout = X25_DEFAULT_T20; |
65 | int sysctl_x25_call_request_timeout = X25_DEFAULT_T21; |
66 | int sysctl_x25_reset_request_timeout = X25_DEFAULT_T22; |
67 | int sysctl_x25_clear_request_timeout = X25_DEFAULT_T23; |
68 | int sysctl_x25_ack_holdback_timeout = X25_DEFAULT_T2; |
69 | int sysctl_x25_forward = 0; |
70 | |
71 | HLIST_HEAD(x25_list); |
72 | DEFINE_RWLOCK(x25_list_lock); |
73 | |
74 | static const struct proto_ops x25_proto_ops; |
75 | |
76 | static struct x25_address null_x25_address = {" "}; |
77 | |
78 | #ifdef CONFIG_COMPAT |
79 | struct compat_x25_subscrip_struct { |
80 | char device[200-sizeof(compat_ulong_t)]; |
81 | compat_ulong_t global_facil_mask; |
82 | compat_uint_t extended; |
83 | }; |
84 | #endif |
85 | |
86 | |
87 | int x25_parse_address_block(struct sk_buff *skb, |
88 | struct x25_address *called_addr, |
89 | struct x25_address *calling_addr) |
90 | { |
91 | unsigned char len; |
92 | int needed; |
93 | int rc; |
94 | |
95 | if (skb->len < 1) { |
96 | /* packet has no address block */ |
97 | rc = 0; |
98 | goto empty; |
99 | } |
100 | |
101 | len = *skb->data; |
102 | needed = 1 + (len >> 4) + (len & 0x0f); |
103 | |
104 | if (skb->len < needed) { |
105 | /* packet is too short to hold the addresses it claims |
106 | to hold */ |
107 | rc = -1; |
108 | goto empty; |
109 | } |
110 | |
111 | return x25_addr_ntoa(skb->data, called_addr, calling_addr); |
112 | |
113 | empty: |
114 | *called_addr->x25_addr = 0; |
115 | *calling_addr->x25_addr = 0; |
116 | |
117 | return rc; |
118 | } |
119 | |
120 | |
121 | int x25_addr_ntoa(unsigned char *p, struct x25_address *called_addr, |
122 | struct x25_address *calling_addr) |
123 | { |
124 | unsigned int called_len, calling_len; |
125 | char *called, *calling; |
126 | unsigned int i; |
127 | |
128 | called_len = (*p >> 0) & 0x0F; |
129 | calling_len = (*p >> 4) & 0x0F; |
130 | |
131 | called = called_addr->x25_addr; |
132 | calling = calling_addr->x25_addr; |
133 | p++; |
134 | |
135 | for (i = 0; i < (called_len + calling_len); i++) { |
136 | if (i < called_len) { |
137 | if (i % 2 != 0) { |
138 | *called++ = ((*p >> 0) & 0x0F) + '0'; |
139 | p++; |
140 | } else { |
141 | *called++ = ((*p >> 4) & 0x0F) + '0'; |
142 | } |
143 | } else { |
144 | if (i % 2 != 0) { |
145 | *calling++ = ((*p >> 0) & 0x0F) + '0'; |
146 | p++; |
147 | } else { |
148 | *calling++ = ((*p >> 4) & 0x0F) + '0'; |
149 | } |
150 | } |
151 | } |
152 | |
153 | *called = *calling = '\0'; |
154 | |
155 | return 1 + (called_len + calling_len + 1) / 2; |
156 | } |
157 | |
158 | int x25_addr_aton(unsigned char *p, struct x25_address *called_addr, |
159 | struct x25_address *calling_addr) |
160 | { |
161 | unsigned int called_len, calling_len; |
162 | char *called, *calling; |
163 | int i; |
164 | |
165 | called = called_addr->x25_addr; |
166 | calling = calling_addr->x25_addr; |
167 | |
168 | called_len = strlen(called); |
169 | calling_len = strlen(calling); |
170 | |
171 | *p++ = (calling_len << 4) | (called_len << 0); |
172 | |
173 | for (i = 0; i < (called_len + calling_len); i++) { |
174 | if (i < called_len) { |
175 | if (i % 2 != 0) { |
176 | *p |= (*called++ - '0') << 0; |
177 | p++; |
178 | } else { |
179 | *p = 0x00; |
180 | *p |= (*called++ - '0') << 4; |
181 | } |
182 | } else { |
183 | if (i % 2 != 0) { |
184 | *p |= (*calling++ - '0') << 0; |
185 | p++; |
186 | } else { |
187 | *p = 0x00; |
188 | *p |= (*calling++ - '0') << 4; |
189 | } |
190 | } |
191 | } |
192 | |
193 | return 1 + (called_len + calling_len + 1) / 2; |
194 | } |
195 | |
196 | /* |
197 | * Socket removal during an interrupt is now safe. |
198 | */ |
199 | static void x25_remove_socket(struct sock *sk) |
200 | { |
201 | write_lock_bh(&x25_list_lock); |
202 | sk_del_node_init(sk); |
203 | write_unlock_bh(&x25_list_lock); |
204 | } |
205 | |
206 | /* |
207 | * Kill all bound sockets on a dropped device. |
208 | */ |
209 | static void x25_kill_by_device(struct net_device *dev) |
210 | { |
211 | struct sock *s; |
212 | struct hlist_node *node; |
213 | |
214 | write_lock_bh(&x25_list_lock); |
215 | |
216 | sk_for_each(s, node, &x25_list) |
217 | if (x25_sk(s)->neighbour && x25_sk(s)->neighbour->dev == dev) |
218 | x25_disconnect(s, ENETUNREACH, 0, 0); |
219 | |
220 | write_unlock_bh(&x25_list_lock); |
221 | } |
222 | |
223 | /* |
224 | * Handle device status changes. |
225 | */ |
226 | static int x25_device_event(struct notifier_block *this, unsigned long event, |
227 | void *ptr) |
228 | { |
229 | struct net_device *dev = ptr; |
230 | struct x25_neigh *nb; |
231 | |
232 | if (!net_eq(dev_net(dev), &init_net)) |
233 | return NOTIFY_DONE; |
234 | |
235 | if (dev->type == ARPHRD_X25 |
236 | #if defined(CONFIG_LLC) || defined(CONFIG_LLC_MODULE) |
237 | || dev->type == ARPHRD_ETHER |
238 | #endif |
239 | ) { |
240 | switch (event) { |
241 | case NETDEV_UP: |
242 | x25_link_device_up(dev); |
243 | break; |
244 | case NETDEV_GOING_DOWN: |
245 | nb = x25_get_neigh(dev); |
246 | if (nb) { |
247 | x25_terminate_link(nb); |
248 | x25_neigh_put(nb); |
249 | } |
250 | break; |
251 | case NETDEV_DOWN: |
252 | x25_kill_by_device(dev); |
253 | x25_route_device_down(dev); |
254 | x25_link_device_down(dev); |
255 | break; |
256 | } |
257 | } |
258 | |
259 | return NOTIFY_DONE; |
260 | } |
261 | |
262 | /* |
263 | * Add a socket to the bound sockets list. |
264 | */ |
265 | static void x25_insert_socket(struct sock *sk) |
266 | { |
267 | write_lock_bh(&x25_list_lock); |
268 | sk_add_node(sk, &x25_list); |
269 | write_unlock_bh(&x25_list_lock); |
270 | } |
271 | |
272 | /* |
273 | * Find a socket that wants to accept the Call Request we just |
274 | * received. Check the full list for an address/cud match. |
275 | * If no cuds match return the next_best thing, an address match. |
276 | * Note: if a listening socket has cud set it must only get calls |
277 | * with matching cud. |
278 | */ |
279 | static struct sock *x25_find_listener(struct x25_address *addr, |
280 | struct sk_buff *skb) |
281 | { |
282 | struct sock *s; |
283 | struct sock *next_best; |
284 | struct hlist_node *node; |
285 | |
286 | read_lock_bh(&x25_list_lock); |
287 | next_best = NULL; |
288 | |
289 | sk_for_each(s, node, &x25_list) |
290 | if ((!strcmp(addr->x25_addr, |
291 | x25_sk(s)->source_addr.x25_addr) || |
292 | !strcmp(addr->x25_addr, |
293 | null_x25_address.x25_addr)) && |
294 | s->sk_state == TCP_LISTEN) { |
295 | /* |
296 | * Found a listening socket, now check the incoming |
297 | * call user data vs this sockets call user data |
298 | */ |
299 | if(skb->len > 0 && x25_sk(s)->cudmatchlength > 0) { |
300 | if((memcmp(x25_sk(s)->calluserdata.cuddata, |
301 | skb->data, |
302 | x25_sk(s)->cudmatchlength)) == 0) { |
303 | sock_hold(s); |
304 | goto found; |
305 | } |
306 | } else |
307 | next_best = s; |
308 | } |
309 | if (next_best) { |
310 | s = next_best; |
311 | sock_hold(s); |
312 | goto found; |
313 | } |
314 | s = NULL; |
315 | found: |
316 | read_unlock_bh(&x25_list_lock); |
317 | return s; |
318 | } |
319 | |
320 | /* |
321 | * Find a connected X.25 socket given my LCI and neighbour. |
322 | */ |
323 | static struct sock *__x25_find_socket(unsigned int lci, struct x25_neigh *nb) |
324 | { |
325 | struct sock *s; |
326 | struct hlist_node *node; |
327 | |
328 | sk_for_each(s, node, &x25_list) |
329 | if (x25_sk(s)->lci == lci && x25_sk(s)->neighbour == nb) { |
330 | sock_hold(s); |
331 | goto found; |
332 | } |
333 | s = NULL; |
334 | found: |
335 | return s; |
336 | } |
337 | |
338 | struct sock *x25_find_socket(unsigned int lci, struct x25_neigh *nb) |
339 | { |
340 | struct sock *s; |
341 | |
342 | read_lock_bh(&x25_list_lock); |
343 | s = __x25_find_socket(lci, nb); |
344 | read_unlock_bh(&x25_list_lock); |
345 | return s; |
346 | } |
347 | |
348 | /* |
349 | * Find a unique LCI for a given device. |
350 | */ |
351 | static unsigned int x25_new_lci(struct x25_neigh *nb) |
352 | { |
353 | unsigned int lci = 1; |
354 | struct sock *sk; |
355 | |
356 | read_lock_bh(&x25_list_lock); |
357 | |
358 | while ((sk = __x25_find_socket(lci, nb)) != NULL) { |
359 | sock_put(sk); |
360 | if (++lci == 4096) { |
361 | lci = 0; |
362 | break; |
363 | } |
364 | } |
365 | |
366 | read_unlock_bh(&x25_list_lock); |
367 | return lci; |
368 | } |
369 | |
370 | /* |
371 | * Deferred destroy. |
372 | */ |
373 | static void __x25_destroy_socket(struct sock *); |
374 | |
375 | /* |
376 | * handler for deferred kills. |
377 | */ |
378 | static void x25_destroy_timer(unsigned long data) |
379 | { |
380 | x25_destroy_socket_from_timer((struct sock *)data); |
381 | } |
382 | |
383 | /* |
384 | * This is called from user mode and the timers. Thus it protects itself |
385 | * against interrupt users but doesn't worry about being called during |
386 | * work. Once it is removed from the queue no interrupt or bottom half |
387 | * will touch it and we are (fairly 8-) ) safe. |
388 | * Not static as it's used by the timer |
389 | */ |
390 | static void __x25_destroy_socket(struct sock *sk) |
391 | { |
392 | struct sk_buff *skb; |
393 | |
394 | x25_stop_heartbeat(sk); |
395 | x25_stop_timer(sk); |
396 | |
397 | x25_remove_socket(sk); |
398 | x25_clear_queues(sk); /* Flush the queues */ |
399 | |
400 | while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { |
401 | if (skb->sk != sk) { /* A pending connection */ |
402 | /* |
403 | * Queue the unaccepted socket for death |
404 | */ |
405 | skb->sk->sk_state = TCP_LISTEN; |
406 | sock_set_flag(skb->sk, SOCK_DEAD); |
407 | x25_start_heartbeat(skb->sk); |
408 | x25_sk(skb->sk)->state = X25_STATE_0; |
409 | } |
410 | |
411 | kfree_skb(skb); |
412 | } |
413 | |
414 | if (sk_has_allocations(sk)) { |
415 | /* Defer: outstanding buffers */ |
416 | sk->sk_timer.expires = jiffies + 10 * HZ; |
417 | sk->sk_timer.function = x25_destroy_timer; |
418 | sk->sk_timer.data = (unsigned long)sk; |
419 | add_timer(&sk->sk_timer); |
420 | } else { |
421 | /* drop last reference so sock_put will free */ |
422 | __sock_put(sk); |
423 | } |
424 | } |
425 | |
426 | void x25_destroy_socket_from_timer(struct sock *sk) |
427 | { |
428 | sock_hold(sk); |
429 | bh_lock_sock(sk); |
430 | __x25_destroy_socket(sk); |
431 | bh_unlock_sock(sk); |
432 | sock_put(sk); |
433 | } |
434 | |
435 | static void x25_destroy_socket(struct sock *sk) |
436 | { |
437 | sock_hold(sk); |
438 | lock_sock(sk); |
439 | __x25_destroy_socket(sk); |
440 | release_sock(sk); |
441 | sock_put(sk); |
442 | } |
443 | |
444 | /* |
445 | * Handling for system calls applied via the various interfaces to a |
446 | * X.25 socket object. |
447 | */ |
448 | |
449 | static int x25_setsockopt(struct socket *sock, int level, int optname, |
450 | char __user *optval, unsigned int optlen) |
451 | { |
452 | int opt; |
453 | struct sock *sk = sock->sk; |
454 | int rc = -ENOPROTOOPT; |
455 | |
456 | lock_kernel(); |
457 | if (level != SOL_X25 || optname != X25_QBITINCL) |
458 | goto out; |
459 | |
460 | rc = -EINVAL; |
461 | if (optlen < sizeof(int)) |
462 | goto out; |
463 | |
464 | rc = -EFAULT; |
465 | if (get_user(opt, (int __user *)optval)) |
466 | goto out; |
467 | |
468 | x25_sk(sk)->qbitincl = !!opt; |
469 | rc = 0; |
470 | out: |
471 | unlock_kernel(); |
472 | return rc; |
473 | } |
474 | |
475 | static int x25_getsockopt(struct socket *sock, int level, int optname, |
476 | char __user *optval, int __user *optlen) |
477 | { |
478 | struct sock *sk = sock->sk; |
479 | int val, len, rc = -ENOPROTOOPT; |
480 | |
481 | lock_kernel(); |
482 | if (level != SOL_X25 || optname != X25_QBITINCL) |
483 | goto out; |
484 | |
485 | rc = -EFAULT; |
486 | if (get_user(len, optlen)) |
487 | goto out; |
488 | |
489 | len = min_t(unsigned int, len, sizeof(int)); |
490 | |
491 | rc = -EINVAL; |
492 | if (len < 0) |
493 | goto out; |
494 | |
495 | rc = -EFAULT; |
496 | if (put_user(len, optlen)) |
497 | goto out; |
498 | |
499 | val = x25_sk(sk)->qbitincl; |
500 | rc = copy_to_user(optval, &val, len) ? -EFAULT : 0; |
501 | out: |
502 | unlock_kernel(); |
503 | return rc; |
504 | } |
505 | |
506 | static int x25_listen(struct socket *sock, int backlog) |
507 | { |
508 | struct sock *sk = sock->sk; |
509 | int rc = -EOPNOTSUPP; |
510 | |
511 | lock_kernel(); |
512 | if (sk->sk_state != TCP_LISTEN) { |
513 | memset(&x25_sk(sk)->dest_addr, 0, X25_ADDR_LEN); |
514 | sk->sk_max_ack_backlog = backlog; |
515 | sk->sk_state = TCP_LISTEN; |
516 | rc = 0; |
517 | } |
518 | unlock_kernel(); |
519 | |
520 | return rc; |
521 | } |
522 | |
523 | static struct proto x25_proto = { |
524 | .name = "X25", |
525 | .owner = THIS_MODULE, |
526 | .obj_size = sizeof(struct x25_sock), |
527 | }; |
528 | |
529 | static struct sock *x25_alloc_socket(struct net *net) |
530 | { |
531 | struct x25_sock *x25; |
532 | struct sock *sk = sk_alloc(net, AF_X25, GFP_ATOMIC, &x25_proto); |
533 | |
534 | if (!sk) |
535 | goto out; |
536 | |
537 | sock_init_data(NULL, sk); |
538 | |
539 | x25 = x25_sk(sk); |
540 | skb_queue_head_init(&x25->ack_queue); |
541 | skb_queue_head_init(&x25->fragment_queue); |
542 | skb_queue_head_init(&x25->interrupt_in_queue); |
543 | skb_queue_head_init(&x25->interrupt_out_queue); |
544 | out: |
545 | return sk; |
546 | } |
547 | |
548 | static int x25_create(struct net *net, struct socket *sock, int protocol, |
549 | int kern) |
550 | { |
551 | struct sock *sk; |
552 | struct x25_sock *x25; |
553 | int rc = -EAFNOSUPPORT; |
554 | |
555 | if (!net_eq(net, &init_net)) |
556 | goto out; |
557 | |
558 | rc = -ESOCKTNOSUPPORT; |
559 | if (sock->type != SOCK_SEQPACKET) |
560 | goto out; |
561 | |
562 | rc = -EINVAL; |
563 | if (protocol) |
564 | goto out; |
565 | |
566 | rc = -ENOBUFS; |
567 | if ((sk = x25_alloc_socket(net)) == NULL) |
568 | goto out; |
569 | |
570 | x25 = x25_sk(sk); |
571 | |
572 | sock_init_data(sock, sk); |
573 | |
574 | x25_init_timers(sk); |
575 | |
576 | sock->ops = &x25_proto_ops; |
577 | sk->sk_protocol = protocol; |
578 | sk->sk_backlog_rcv = x25_backlog_rcv; |
579 | |
580 | x25->t21 = sysctl_x25_call_request_timeout; |
581 | x25->t22 = sysctl_x25_reset_request_timeout; |
582 | x25->t23 = sysctl_x25_clear_request_timeout; |
583 | x25->t2 = sysctl_x25_ack_holdback_timeout; |
584 | x25->state = X25_STATE_0; |
585 | x25->cudmatchlength = 0; |
586 | x25->accptapprv = X25_DENY_ACCPT_APPRV; /* normally no cud */ |
587 | /* on call accept */ |
588 | |
589 | x25->facilities.winsize_in = X25_DEFAULT_WINDOW_SIZE; |
590 | x25->facilities.winsize_out = X25_DEFAULT_WINDOW_SIZE; |
591 | x25->facilities.pacsize_in = X25_DEFAULT_PACKET_SIZE; |
592 | x25->facilities.pacsize_out = X25_DEFAULT_PACKET_SIZE; |
593 | x25->facilities.throughput = 0; /* by default don't negotiate |
594 | throughput */ |
595 | x25->facilities.reverse = X25_DEFAULT_REVERSE; |
596 | x25->dte_facilities.calling_len = 0; |
597 | x25->dte_facilities.called_len = 0; |
598 | memset(x25->dte_facilities.called_ae, '\0', |
599 | sizeof(x25->dte_facilities.called_ae)); |
600 | memset(x25->dte_facilities.calling_ae, '\0', |
601 | sizeof(x25->dte_facilities.calling_ae)); |
602 | |
603 | rc = 0; |
604 | out: |
605 | return rc; |
606 | } |
607 | |
608 | static struct sock *x25_make_new(struct sock *osk) |
609 | { |
610 | struct sock *sk = NULL; |
611 | struct x25_sock *x25, *ox25; |
612 | |
613 | if (osk->sk_type != SOCK_SEQPACKET) |
614 | goto out; |
615 | |
616 | if ((sk = x25_alloc_socket(sock_net(osk))) == NULL) |
617 | goto out; |
618 | |
619 | x25 = x25_sk(sk); |
620 | |
621 | sk->sk_type = osk->sk_type; |
622 | sk->sk_priority = osk->sk_priority; |
623 | sk->sk_protocol = osk->sk_protocol; |
624 | sk->sk_rcvbuf = osk->sk_rcvbuf; |
625 | sk->sk_sndbuf = osk->sk_sndbuf; |
626 | sk->sk_state = TCP_ESTABLISHED; |
627 | sk->sk_backlog_rcv = osk->sk_backlog_rcv; |
628 | sock_copy_flags(sk, osk); |
629 | |
630 | ox25 = x25_sk(osk); |
631 | x25->t21 = ox25->t21; |
632 | x25->t22 = ox25->t22; |
633 | x25->t23 = ox25->t23; |
634 | x25->t2 = ox25->t2; |
635 | x25->facilities = ox25->facilities; |
636 | x25->qbitincl = ox25->qbitincl; |
637 | x25->dte_facilities = ox25->dte_facilities; |
638 | x25->cudmatchlength = ox25->cudmatchlength; |
639 | x25->accptapprv = ox25->accptapprv; |
640 | |
641 | x25_init_timers(sk); |
642 | out: |
643 | return sk; |
644 | } |
645 | |
646 | static int x25_release(struct socket *sock) |
647 | { |
648 | struct sock *sk = sock->sk; |
649 | struct x25_sock *x25; |
650 | |
651 | lock_kernel(); |
652 | if (!sk) |
653 | goto out; |
654 | |
655 | x25 = x25_sk(sk); |
656 | |
657 | switch (x25->state) { |
658 | |
659 | case X25_STATE_0: |
660 | case X25_STATE_2: |
661 | x25_disconnect(sk, 0, 0, 0); |
662 | x25_destroy_socket(sk); |
663 | goto out; |
664 | |
665 | case X25_STATE_1: |
666 | case X25_STATE_3: |
667 | case X25_STATE_4: |
668 | x25_clear_queues(sk); |
669 | x25_write_internal(sk, X25_CLEAR_REQUEST); |
670 | x25_start_t23timer(sk); |
671 | x25->state = X25_STATE_2; |
672 | sk->sk_state = TCP_CLOSE; |
673 | sk->sk_shutdown |= SEND_SHUTDOWN; |
674 | sk->sk_state_change(sk); |
675 | sock_set_flag(sk, SOCK_DEAD); |
676 | sock_set_flag(sk, SOCK_DESTROY); |
677 | break; |
678 | } |
679 | |
680 | sock_orphan(sk); |
681 | out: |
682 | unlock_kernel(); |
683 | return 0; |
684 | } |
685 | |
686 | static int x25_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) |
687 | { |
688 | struct sock *sk = sock->sk; |
689 | struct sockaddr_x25 *addr = (struct sockaddr_x25 *)uaddr; |
690 | int len, i, rc = 0; |
691 | |
692 | lock_kernel(); |
693 | if (!sock_flag(sk, SOCK_ZAPPED) || |
694 | addr_len != sizeof(struct sockaddr_x25) || |
695 | addr->sx25_family != AF_X25) { |
696 | rc = -EINVAL; |
697 | goto out; |
698 | } |
699 | |
700 | len = strlen(addr->sx25_addr.x25_addr); |
701 | for (i = 0; i < len; i++) { |
702 | if (!isdigit(addr->sx25_addr.x25_addr[i])) { |
703 | rc = -EINVAL; |
704 | goto out; |
705 | } |
706 | } |
707 | |
708 | x25_sk(sk)->source_addr = addr->sx25_addr; |
709 | x25_insert_socket(sk); |
710 | sock_reset_flag(sk, SOCK_ZAPPED); |
711 | SOCK_DEBUG(sk, "x25_bind: socket is bound\n"); |
712 | out: |
713 | unlock_kernel(); |
714 | return rc; |
715 | } |
716 | |
717 | static int x25_wait_for_connection_establishment(struct sock *sk) |
718 | { |
719 | DECLARE_WAITQUEUE(wait, current); |
720 | int rc; |
721 | |
722 | add_wait_queue_exclusive(sk->sk_sleep, &wait); |
723 | for (;;) { |
724 | __set_current_state(TASK_INTERRUPTIBLE); |
725 | rc = -ERESTARTSYS; |
726 | if (signal_pending(current)) |
727 | break; |
728 | rc = sock_error(sk); |
729 | if (rc) { |
730 | sk->sk_socket->state = SS_UNCONNECTED; |
731 | break; |
732 | } |
733 | rc = 0; |
734 | if (sk->sk_state != TCP_ESTABLISHED) { |
735 | release_sock(sk); |
736 | schedule(); |
737 | lock_sock(sk); |
738 | } else |
739 | break; |
740 | } |
741 | __set_current_state(TASK_RUNNING); |
742 | remove_wait_queue(sk->sk_sleep, &wait); |
743 | return rc; |
744 | } |
745 | |
746 | static int x25_connect(struct socket *sock, struct sockaddr *uaddr, |
747 | int addr_len, int flags) |
748 | { |
749 | struct sock *sk = sock->sk; |
750 | struct x25_sock *x25 = x25_sk(sk); |
751 | struct sockaddr_x25 *addr = (struct sockaddr_x25 *)uaddr; |
752 | struct x25_route *rt; |
753 | int rc = 0; |
754 | |
755 | lock_kernel(); |
756 | lock_sock(sk); |
757 | if (sk->sk_state == TCP_ESTABLISHED && sock->state == SS_CONNECTING) { |
758 | sock->state = SS_CONNECTED; |
759 | goto out; /* Connect completed during a ERESTARTSYS event */ |
760 | } |
761 | |
762 | rc = -ECONNREFUSED; |
763 | if (sk->sk_state == TCP_CLOSE && sock->state == SS_CONNECTING) { |
764 | sock->state = SS_UNCONNECTED; |
765 | goto out; |
766 | } |
767 | |
768 | rc = -EISCONN; /* No reconnect on a seqpacket socket */ |
769 | if (sk->sk_state == TCP_ESTABLISHED) |
770 | goto out; |
771 | |
772 | sk->sk_state = TCP_CLOSE; |
773 | sock->state = SS_UNCONNECTED; |
774 | |
775 | rc = -EINVAL; |
776 | if (addr_len != sizeof(struct sockaddr_x25) || |
777 | addr->sx25_family != AF_X25) |
778 | goto out; |
779 | |
780 | rc = -ENETUNREACH; |
781 | rt = x25_get_route(&addr->sx25_addr); |
782 | if (!rt) |
783 | goto out; |
784 | |
785 | x25->neighbour = x25_get_neigh(rt->dev); |
786 | if (!x25->neighbour) |
787 | goto out_put_route; |
788 | |
789 | x25_limit_facilities(&x25->facilities, x25->neighbour); |
790 | |
791 | x25->lci = x25_new_lci(x25->neighbour); |
792 | if (!x25->lci) |
793 | goto out_put_neigh; |
794 | |
795 | rc = -EINVAL; |
796 | if (sock_flag(sk, SOCK_ZAPPED)) /* Must bind first - autobinding does not work */ |
797 | goto out_put_neigh; |
798 | |
799 | if (!strcmp(x25->source_addr.x25_addr, null_x25_address.x25_addr)) |
800 | memset(&x25->source_addr, '\0', X25_ADDR_LEN); |
801 | |
802 | x25->dest_addr = addr->sx25_addr; |
803 | |
804 | /* Move to connecting socket, start sending Connect Requests */ |
805 | sock->state = SS_CONNECTING; |
806 | sk->sk_state = TCP_SYN_SENT; |
807 | |
808 | x25->state = X25_STATE_1; |
809 | |
810 | x25_write_internal(sk, X25_CALL_REQUEST); |
811 | |
812 | x25_start_heartbeat(sk); |
813 | x25_start_t21timer(sk); |
814 | |
815 | /* Now the loop */ |
816 | rc = -EINPROGRESS; |
817 | if (sk->sk_state != TCP_ESTABLISHED && (flags & O_NONBLOCK)) |
818 | goto out_put_neigh; |
819 | |
820 | rc = x25_wait_for_connection_establishment(sk); |
821 | if (rc) |
822 | goto out_put_neigh; |
823 | |
824 | sock->state = SS_CONNECTED; |
825 | rc = 0; |
826 | out_put_neigh: |
827 | if (rc) |
828 | x25_neigh_put(x25->neighbour); |
829 | out_put_route: |
830 | x25_route_put(rt); |
831 | out: |
832 | release_sock(sk); |
833 | unlock_kernel(); |
834 | return rc; |
835 | } |
836 | |
837 | static int x25_wait_for_data(struct sock *sk, long timeout) |
838 | { |
839 | DECLARE_WAITQUEUE(wait, current); |
840 | int rc = 0; |
841 | |
842 | add_wait_queue_exclusive(sk->sk_sleep, &wait); |
843 | for (;;) { |
844 | __set_current_state(TASK_INTERRUPTIBLE); |
845 | if (sk->sk_shutdown & RCV_SHUTDOWN) |
846 | break; |
847 | rc = -ERESTARTSYS; |
848 | if (signal_pending(current)) |
849 | break; |
850 | rc = -EAGAIN; |
851 | if (!timeout) |
852 | break; |
853 | rc = 0; |
854 | if (skb_queue_empty(&sk->sk_receive_queue)) { |
855 | release_sock(sk); |
856 | timeout = schedule_timeout(timeout); |
857 | lock_sock(sk); |
858 | } else |
859 | break; |
860 | } |
861 | __set_current_state(TASK_RUNNING); |
862 | remove_wait_queue(sk->sk_sleep, &wait); |
863 | return rc; |
864 | } |
865 | |
866 | static int x25_accept(struct socket *sock, struct socket *newsock, int flags) |
867 | { |
868 | struct sock *sk = sock->sk; |
869 | struct sock *newsk; |
870 | struct sk_buff *skb; |
871 | int rc = -EINVAL; |
872 | |
873 | lock_kernel(); |
874 | if (!sk || sk->sk_state != TCP_LISTEN) |
875 | goto out; |
876 | |
877 | rc = -EOPNOTSUPP; |
878 | if (sk->sk_type != SOCK_SEQPACKET) |
879 | goto out; |
880 | |
881 | lock_sock(sk); |
882 | rc = x25_wait_for_data(sk, sk->sk_rcvtimeo); |
883 | if (rc) |
884 | goto out2; |
885 | skb = skb_dequeue(&sk->sk_receive_queue); |
886 | rc = -EINVAL; |
887 | if (!skb->sk) |
888 | goto out2; |
889 | newsk = skb->sk; |
890 | sock_graft(newsk, newsock); |
891 | |
892 | /* Now attach up the new socket */ |
893 | skb->sk = NULL; |
894 | kfree_skb(skb); |
895 | sk->sk_ack_backlog--; |
896 | newsock->state = SS_CONNECTED; |
897 | rc = 0; |
898 | out2: |
899 | release_sock(sk); |
900 | out: |
901 | unlock_kernel(); |
902 | return rc; |
903 | } |
904 | |
905 | static int x25_getname(struct socket *sock, struct sockaddr *uaddr, |
906 | int *uaddr_len, int peer) |
907 | { |
908 | struct sockaddr_x25 *sx25 = (struct sockaddr_x25 *)uaddr; |
909 | struct sock *sk = sock->sk; |
910 | struct x25_sock *x25 = x25_sk(sk); |
911 | int rc = 0; |
912 | |
913 | lock_kernel(); |
914 | if (peer) { |
915 | if (sk->sk_state != TCP_ESTABLISHED) { |
916 | rc = -ENOTCONN; |
917 | goto out; |
918 | } |
919 | sx25->sx25_addr = x25->dest_addr; |
920 | } else |
921 | sx25->sx25_addr = x25->source_addr; |
922 | |
923 | sx25->sx25_family = AF_X25; |
924 | *uaddr_len = sizeof(*sx25); |
925 | |
926 | out: |
927 | unlock_kernel(); |
928 | return rc; |
929 | } |
930 | |
931 | static unsigned int x25_datagram_poll(struct file *file, struct socket *sock, |
932 | poll_table *wait) |
933 | { |
934 | int rc; |
935 | |
936 | lock_kernel(); |
937 | rc = datagram_poll(file, sock, wait); |
938 | unlock_kernel(); |
939 | |
940 | return rc; |
941 | } |
942 | |
943 | int x25_rx_call_request(struct sk_buff *skb, struct x25_neigh *nb, |
944 | unsigned int lci) |
945 | { |
946 | struct sock *sk; |
947 | struct sock *make; |
948 | struct x25_sock *makex25; |
949 | struct x25_address source_addr, dest_addr; |
950 | struct x25_facilities facilities; |
951 | struct x25_dte_facilities dte_facilities; |
952 | int len, addr_len, rc; |
953 | |
954 | /* |
955 | * Remove the LCI and frame type. |
956 | */ |
957 | skb_pull(skb, X25_STD_MIN_LEN); |
958 | |
959 | /* |
960 | * Extract the X.25 addresses and convert them to ASCII strings, |
961 | * and remove them. |
962 | * |
963 | * Address block is mandatory in call request packets |
964 | */ |
965 | addr_len = x25_parse_address_block(skb, &source_addr, &dest_addr); |
966 | if (addr_len <= 0) |
967 | goto out_clear_request; |
968 | skb_pull(skb, addr_len); |
969 | |
970 | /* |
971 | * Get the length of the facilities, skip past them for the moment |
972 | * get the call user data because this is needed to determine |
973 | * the correct listener |
974 | * |
975 | * Facilities length is mandatory in call request packets |
976 | */ |
977 | if (skb->len < 1) |
978 | goto out_clear_request; |
979 | len = skb->data[0] + 1; |
980 | if (skb->len < len) |
981 | goto out_clear_request; |
982 | skb_pull(skb,len); |
983 | |
984 | /* |
985 | * Find a listener for the particular address/cud pair. |
986 | */ |
987 | sk = x25_find_listener(&source_addr,skb); |
988 | skb_push(skb,len); |
989 | |
990 | if (sk != NULL && sk_acceptq_is_full(sk)) { |
991 | goto out_sock_put; |
992 | } |
993 | |
994 | /* |
995 | * We dont have any listeners for this incoming call. |
996 | * Try forwarding it. |
997 | */ |
998 | if (sk == NULL) { |
999 | skb_push(skb, addr_len + X25_STD_MIN_LEN); |
1000 | if (sysctl_x25_forward && |
1001 | x25_forward_call(&dest_addr, nb, skb, lci) > 0) |
1002 | { |
1003 | /* Call was forwarded, dont process it any more */ |
1004 | kfree_skb(skb); |
1005 | rc = 1; |
1006 | goto out; |
1007 | } else { |
1008 | /* No listeners, can't forward, clear the call */ |
1009 | goto out_clear_request; |
1010 | } |
1011 | } |
1012 | |
1013 | /* |
1014 | * Try to reach a compromise on the requested facilities. |
1015 | */ |
1016 | len = x25_negotiate_facilities(skb, sk, &facilities, &dte_facilities); |
1017 | if (len == -1) |
1018 | goto out_sock_put; |
1019 | |
1020 | /* |
1021 | * current neighbour/link might impose additional limits |
1022 | * on certain facilties |
1023 | */ |
1024 | |
1025 | x25_limit_facilities(&facilities, nb); |
1026 | |
1027 | /* |
1028 | * Try to create a new socket. |
1029 | */ |
1030 | make = x25_make_new(sk); |
1031 | if (!make) |
1032 | goto out_sock_put; |
1033 | |
1034 | /* |
1035 | * Remove the facilities |
1036 | */ |
1037 | skb_pull(skb, len); |
1038 | |
1039 | skb->sk = make; |
1040 | make->sk_state = TCP_ESTABLISHED; |
1041 | |
1042 | makex25 = x25_sk(make); |
1043 | makex25->lci = lci; |
1044 | makex25->dest_addr = dest_addr; |
1045 | makex25->source_addr = source_addr; |
1046 | makex25->neighbour = nb; |
1047 | makex25->facilities = facilities; |
1048 | makex25->dte_facilities= dte_facilities; |
1049 | makex25->vc_facil_mask = x25_sk(sk)->vc_facil_mask; |
1050 | /* ensure no reverse facil on accept */ |
1051 | makex25->vc_facil_mask &= ~X25_MASK_REVERSE; |
1052 | /* ensure no calling address extension on accept */ |
1053 | makex25->vc_facil_mask &= ~X25_MASK_CALLING_AE; |
1054 | makex25->cudmatchlength = x25_sk(sk)->cudmatchlength; |
1055 | |
1056 | /* Normally all calls are accepted immediatly */ |
1057 | if(makex25->accptapprv & X25_DENY_ACCPT_APPRV) { |
1058 | x25_write_internal(make, X25_CALL_ACCEPTED); |
1059 | makex25->state = X25_STATE_3; |
1060 | } |
1061 | |
1062 | /* |
1063 | * Incoming Call User Data. |
1064 | */ |
1065 | skb_copy_from_linear_data(skb, makex25->calluserdata.cuddata, skb->len); |
1066 | makex25->calluserdata.cudlength = skb->len; |
1067 | |
1068 | sk->sk_ack_backlog++; |
1069 | |
1070 | x25_insert_socket(make); |
1071 | |
1072 | skb_queue_head(&sk->sk_receive_queue, skb); |
1073 | |
1074 | x25_start_heartbeat(make); |
1075 | |
1076 | if (!sock_flag(sk, SOCK_DEAD)) |
1077 | sk->sk_data_ready(sk, skb->len); |
1078 | rc = 1; |
1079 | sock_put(sk); |
1080 | out: |
1081 | return rc; |
1082 | out_sock_put: |
1083 | sock_put(sk); |
1084 | out_clear_request: |
1085 | rc = 0; |
1086 | x25_transmit_clear_request(nb, lci, 0x01); |
1087 | goto out; |
1088 | } |
1089 | |
1090 | static int x25_sendmsg(struct kiocb *iocb, struct socket *sock, |
1091 | struct msghdr *msg, size_t len) |
1092 | { |
1093 | struct sock *sk = sock->sk; |
1094 | struct x25_sock *x25 = x25_sk(sk); |
1095 | struct sockaddr_x25 *usx25 = (struct sockaddr_x25 *)msg->msg_name; |
1096 | struct sockaddr_x25 sx25; |
1097 | struct sk_buff *skb; |
1098 | unsigned char *asmptr; |
1099 | int noblock = msg->msg_flags & MSG_DONTWAIT; |
1100 | size_t size; |
1101 | int qbit = 0, rc = -EINVAL; |
1102 | |
1103 | lock_kernel(); |
1104 | if (msg->msg_flags & ~(MSG_DONTWAIT|MSG_OOB|MSG_EOR|MSG_CMSG_COMPAT)) |
1105 | goto out; |
1106 | |
1107 | /* we currently don't support segmented records at the user interface */ |
1108 | if (!(msg->msg_flags & (MSG_EOR|MSG_OOB))) |
1109 | goto out; |
1110 | |
1111 | rc = -EADDRNOTAVAIL; |
1112 | if (sock_flag(sk, SOCK_ZAPPED)) |
1113 | goto out; |
1114 | |
1115 | rc = -EPIPE; |
1116 | if (sk->sk_shutdown & SEND_SHUTDOWN) { |
1117 | send_sig(SIGPIPE, current, 0); |
1118 | goto out; |
1119 | } |
1120 | |
1121 | rc = -ENETUNREACH; |
1122 | if (!x25->neighbour) |
1123 | goto out; |
1124 | |
1125 | if (usx25) { |
1126 | rc = -EINVAL; |
1127 | if (msg->msg_namelen < sizeof(sx25)) |
1128 | goto out; |
1129 | memcpy(&sx25, usx25, sizeof(sx25)); |
1130 | rc = -EISCONN; |
1131 | if (strcmp(x25->dest_addr.x25_addr, sx25.sx25_addr.x25_addr)) |
1132 | goto out; |
1133 | rc = -EINVAL; |
1134 | if (sx25.sx25_family != AF_X25) |
1135 | goto out; |
1136 | } else { |
1137 | /* |
1138 | * FIXME 1003.1g - if the socket is like this because |
1139 | * it has become closed (not started closed) we ought |
1140 | * to SIGPIPE, EPIPE; |
1141 | */ |
1142 | rc = -ENOTCONN; |
1143 | if (sk->sk_state != TCP_ESTABLISHED) |
1144 | goto out; |
1145 | |
1146 | sx25.sx25_family = AF_X25; |
1147 | sx25.sx25_addr = x25->dest_addr; |
1148 | } |
1149 | |
1150 | /* Sanity check the packet size */ |
1151 | if (len > 65535) { |
1152 | rc = -EMSGSIZE; |
1153 | goto out; |
1154 | } |
1155 | |
1156 | SOCK_DEBUG(sk, "x25_sendmsg: sendto: Addresses built.\n"); |
1157 | |
1158 | /* Build a packet */ |
1159 | SOCK_DEBUG(sk, "x25_sendmsg: sendto: building packet.\n"); |
1160 | |
1161 | if ((msg->msg_flags & MSG_OOB) && len > 32) |
1162 | len = 32; |
1163 | |
1164 | size = len + X25_MAX_L2_LEN + X25_EXT_MIN_LEN; |
1165 | |
1166 | skb = sock_alloc_send_skb(sk, size, noblock, &rc); |
1167 | if (!skb) |
1168 | goto out; |
1169 | X25_SKB_CB(skb)->flags = msg->msg_flags; |
1170 | |
1171 | skb_reserve(skb, X25_MAX_L2_LEN + X25_EXT_MIN_LEN); |
1172 | |
1173 | /* |
1174 | * Put the data on the end |
1175 | */ |
1176 | SOCK_DEBUG(sk, "x25_sendmsg: Copying user data\n"); |
1177 | |
1178 | skb_reset_transport_header(skb); |
1179 | skb_put(skb, len); |
1180 | |
1181 | rc = memcpy_fromiovec(skb_transport_header(skb), msg->msg_iov, len); |
1182 | if (rc) |
1183 | goto out_kfree_skb; |
1184 | |
1185 | /* |
1186 | * If the Q BIT Include socket option is in force, the first |
1187 | * byte of the user data is the logical value of the Q Bit. |
1188 | */ |
1189 | if (x25->qbitincl) { |
1190 | qbit = skb->data[0]; |
1191 | skb_pull(skb, 1); |
1192 | } |
1193 | |
1194 | /* |
1195 | * Push down the X.25 header |
1196 | */ |
1197 | SOCK_DEBUG(sk, "x25_sendmsg: Building X.25 Header.\n"); |
1198 | |
1199 | if (msg->msg_flags & MSG_OOB) { |
1200 | if (x25->neighbour->extended) { |
1201 | asmptr = skb_push(skb, X25_STD_MIN_LEN); |
1202 | *asmptr++ = ((x25->lci >> 8) & 0x0F) | X25_GFI_EXTSEQ; |
1203 | *asmptr++ = (x25->lci >> 0) & 0xFF; |
1204 | *asmptr++ = X25_INTERRUPT; |
1205 | } else { |
1206 | asmptr = skb_push(skb, X25_STD_MIN_LEN); |
1207 | *asmptr++ = ((x25->lci >> 8) & 0x0F) | X25_GFI_STDSEQ; |
1208 | *asmptr++ = (x25->lci >> 0) & 0xFF; |
1209 | *asmptr++ = X25_INTERRUPT; |
1210 | } |
1211 | } else { |
1212 | if (x25->neighbour->extended) { |
1213 | /* Build an Extended X.25 header */ |
1214 | asmptr = skb_push(skb, X25_EXT_MIN_LEN); |
1215 | *asmptr++ = ((x25->lci >> 8) & 0x0F) | X25_GFI_EXTSEQ; |
1216 | *asmptr++ = (x25->lci >> 0) & 0xFF; |
1217 | *asmptr++ = X25_DATA; |
1218 | *asmptr++ = X25_DATA; |
1219 | } else { |
1220 | /* Build an Standard X.25 header */ |
1221 | asmptr = skb_push(skb, X25_STD_MIN_LEN); |
1222 | *asmptr++ = ((x25->lci >> 8) & 0x0F) | X25_GFI_STDSEQ; |
1223 | *asmptr++ = (x25->lci >> 0) & 0xFF; |
1224 | *asmptr++ = X25_DATA; |
1225 | } |
1226 | |
1227 | if (qbit) |
1228 | skb->data[0] |= X25_Q_BIT; |
1229 | } |
1230 | |
1231 | SOCK_DEBUG(sk, "x25_sendmsg: Built header.\n"); |
1232 | SOCK_DEBUG(sk, "x25_sendmsg: Transmitting buffer\n"); |
1233 | |
1234 | rc = -ENOTCONN; |
1235 | if (sk->sk_state != TCP_ESTABLISHED) |
1236 | goto out_kfree_skb; |
1237 | |
1238 | if (msg->msg_flags & MSG_OOB) |
1239 | skb_queue_tail(&x25->interrupt_out_queue, skb); |
1240 | else { |
1241 | rc = x25_output(sk, skb); |
1242 | len = rc; |
1243 | if (rc < 0) |
1244 | kfree_skb(skb); |
1245 | else if (x25->qbitincl) |
1246 | len++; |
1247 | } |
1248 | |
1249 | /* |
1250 | * lock_sock() is currently only used to serialize this x25_kick() |
1251 | * against input-driven x25_kick() calls. It currently only blocks |
1252 | * incoming packets for this socket and does not protect against |
1253 | * any other socket state changes and is not called from anywhere |
1254 | * else. As x25_kick() cannot block and as long as all socket |
1255 | * operations are BKL-wrapped, we don't need take to care about |
1256 | * purging the backlog queue in x25_release(). |
1257 | * |
1258 | * Using lock_sock() to protect all socket operations entirely |
1259 | * (and making the whole x25 stack SMP aware) unfortunately would |
1260 | * require major changes to {send,recv}msg and skb allocation methods. |
1261 | * -> 2.5 ;) |
1262 | */ |
1263 | lock_sock(sk); |
1264 | x25_kick(sk); |
1265 | release_sock(sk); |
1266 | rc = len; |
1267 | out: |
1268 | unlock_kernel(); |
1269 | return rc; |
1270 | out_kfree_skb: |
1271 | kfree_skb(skb); |
1272 | goto out; |
1273 | } |
1274 | |
1275 | |
1276 | static int x25_recvmsg(struct kiocb *iocb, struct socket *sock, |
1277 | struct msghdr *msg, size_t size, |
1278 | int flags) |
1279 | { |
1280 | struct sock *sk = sock->sk; |
1281 | struct x25_sock *x25 = x25_sk(sk); |
1282 | struct sockaddr_x25 *sx25 = (struct sockaddr_x25 *)msg->msg_name; |
1283 | size_t copied; |
1284 | int qbit; |
1285 | struct sk_buff *skb; |
1286 | unsigned char *asmptr; |
1287 | int rc = -ENOTCONN; |
1288 | |
1289 | lock_kernel(); |
1290 | /* |
1291 | * This works for seqpacket too. The receiver has ordered the queue for |
1292 | * us! We do one quick check first though |
1293 | */ |
1294 | if (sk->sk_state != TCP_ESTABLISHED) |
1295 | goto out; |
1296 | |
1297 | if (flags & MSG_OOB) { |
1298 | rc = -EINVAL; |
1299 | if (sock_flag(sk, SOCK_URGINLINE) || |
1300 | !skb_peek(&x25->interrupt_in_queue)) |
1301 | goto out; |
1302 | |
1303 | skb = skb_dequeue(&x25->interrupt_in_queue); |
1304 | |
1305 | skb_pull(skb, X25_STD_MIN_LEN); |
1306 | |
1307 | /* |
1308 | * No Q bit information on Interrupt data. |
1309 | */ |
1310 | if (x25->qbitincl) { |
1311 | asmptr = skb_push(skb, 1); |
1312 | *asmptr = 0x00; |
1313 | } |
1314 | |
1315 | msg->msg_flags |= MSG_OOB; |
1316 | } else { |
1317 | /* Now we can treat all alike */ |
1318 | skb = skb_recv_datagram(sk, flags & ~MSG_DONTWAIT, |
1319 | flags & MSG_DONTWAIT, &rc); |
1320 | if (!skb) |
1321 | goto out; |
1322 | |
1323 | qbit = (skb->data[0] & X25_Q_BIT) == X25_Q_BIT; |
1324 | |
1325 | skb_pull(skb, x25->neighbour->extended ? |
1326 | X25_EXT_MIN_LEN : X25_STD_MIN_LEN); |
1327 | |
1328 | if (x25->qbitincl) { |
1329 | asmptr = skb_push(skb, 1); |
1330 | *asmptr = qbit; |
1331 | } |
1332 | } |
1333 | |
1334 | skb_reset_transport_header(skb); |
1335 | copied = skb->len; |
1336 | |
1337 | if (copied > size) { |
1338 | copied = size; |
1339 | msg->msg_flags |= MSG_TRUNC; |
1340 | } |
1341 | |
1342 | /* Currently, each datagram always contains a complete record */ |
1343 | msg->msg_flags |= MSG_EOR; |
1344 | |
1345 | rc = skb_copy_datagram_iovec(skb, 0, msg->msg_iov, copied); |
1346 | if (rc) |
1347 | goto out_free_dgram; |
1348 | |
1349 | if (sx25) { |
1350 | sx25->sx25_family = AF_X25; |
1351 | sx25->sx25_addr = x25->dest_addr; |
1352 | } |
1353 | |
1354 | msg->msg_namelen = sizeof(struct sockaddr_x25); |
1355 | |
1356 | lock_sock(sk); |
1357 | x25_check_rbuf(sk); |
1358 | release_sock(sk); |
1359 | rc = copied; |
1360 | out_free_dgram: |
1361 | skb_free_datagram(sk, skb); |
1362 | out: |
1363 | unlock_kernel(); |
1364 | return rc; |
1365 | } |
1366 | |
1367 | |
1368 | static int x25_ioctl(struct socket *sock, unsigned int cmd, unsigned long arg) |
1369 | { |
1370 | struct sock *sk = sock->sk; |
1371 | struct x25_sock *x25 = x25_sk(sk); |
1372 | void __user *argp = (void __user *)arg; |
1373 | int rc; |
1374 | |
1375 | lock_kernel(); |
1376 | switch (cmd) { |
1377 | case TIOCOUTQ: { |
1378 | int amount = sk->sk_sndbuf - sk_wmem_alloc_get(sk); |
1379 | |
1380 | if (amount < 0) |
1381 | amount = 0; |
1382 | rc = put_user(amount, (unsigned int __user *)argp); |
1383 | break; |
1384 | } |
1385 | |
1386 | case TIOCINQ: { |
1387 | struct sk_buff *skb; |
1388 | int amount = 0; |
1389 | /* |
1390 | * These two are safe on a single CPU system as |
1391 | * only user tasks fiddle here |
1392 | */ |
1393 | if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL) |
1394 | amount = skb->len; |
1395 | rc = put_user(amount, (unsigned int __user *)argp); |
1396 | break; |
1397 | } |
1398 | |
1399 | case SIOCGSTAMP: |
1400 | rc = -EINVAL; |
1401 | if (sk) |
1402 | rc = sock_get_timestamp(sk, |
1403 | (struct timeval __user *)argp); |
1404 | break; |
1405 | case SIOCGSTAMPNS: |
1406 | rc = -EINVAL; |
1407 | if (sk) |
1408 | rc = sock_get_timestampns(sk, |
1409 | (struct timespec __user *)argp); |
1410 | break; |
1411 | case SIOCGIFADDR: |
1412 | case SIOCSIFADDR: |
1413 | case SIOCGIFDSTADDR: |
1414 | case SIOCSIFDSTADDR: |
1415 | case SIOCGIFBRDADDR: |
1416 | case SIOCSIFBRDADDR: |
1417 | case SIOCGIFNETMASK: |
1418 | case SIOCSIFNETMASK: |
1419 | case SIOCGIFMETRIC: |
1420 | case SIOCSIFMETRIC: |
1421 | rc = -EINVAL; |
1422 | break; |
1423 | case SIOCADDRT: |
1424 | case SIOCDELRT: |
1425 | rc = -EPERM; |
1426 | if (!capable(CAP_NET_ADMIN)) |
1427 | break; |
1428 | rc = x25_route_ioctl(cmd, argp); |
1429 | break; |
1430 | case SIOCX25GSUBSCRIP: |
1431 | rc = x25_subscr_ioctl(cmd, argp); |
1432 | break; |
1433 | case SIOCX25SSUBSCRIP: |
1434 | rc = -EPERM; |
1435 | if (!capable(CAP_NET_ADMIN)) |
1436 | break; |
1437 | rc = x25_subscr_ioctl(cmd, argp); |
1438 | break; |
1439 | case SIOCX25GFACILITIES: { |
1440 | struct x25_facilities fac = x25->facilities; |
1441 | rc = copy_to_user(argp, &fac, |
1442 | sizeof(fac)) ? -EFAULT : 0; |
1443 | break; |
1444 | } |
1445 | |
1446 | case SIOCX25SFACILITIES: { |
1447 | struct x25_facilities facilities; |
1448 | rc = -EFAULT; |
1449 | if (copy_from_user(&facilities, argp, |
1450 | sizeof(facilities))) |
1451 | break; |
1452 | rc = -EINVAL; |
1453 | if (sk->sk_state != TCP_LISTEN && |
1454 | sk->sk_state != TCP_CLOSE) |
1455 | break; |
1456 | if (facilities.pacsize_in < X25_PS16 || |
1457 | facilities.pacsize_in > X25_PS4096) |
1458 | break; |
1459 | if (facilities.pacsize_out < X25_PS16 || |
1460 | facilities.pacsize_out > X25_PS4096) |
1461 | break; |
1462 | if (facilities.winsize_in < 1 || |
1463 | facilities.winsize_in > 127) |
1464 | break; |
1465 | if (facilities.throughput) { |
1466 | int out = facilities.throughput & 0xf0; |
1467 | int in = facilities.throughput & 0x0f; |
1468 | if (!out) |
1469 | facilities.throughput |= |
1470 | X25_DEFAULT_THROUGHPUT << 4; |
1471 | else if (out < 0x30 || out > 0xD0) |
1472 | break; |
1473 | if (!in) |
1474 | facilities.throughput |= |
1475 | X25_DEFAULT_THROUGHPUT; |
1476 | else if (in < 0x03 || in > 0x0D) |
1477 | break; |
1478 | } |
1479 | if (facilities.reverse && |
1480 | (facilities.reverse & 0x81) != 0x81) |
1481 | break; |
1482 | x25->facilities = facilities; |
1483 | rc = 0; |
1484 | break; |
1485 | } |
1486 | |
1487 | case SIOCX25GDTEFACILITIES: { |
1488 | rc = copy_to_user(argp, &x25->dte_facilities, |
1489 | sizeof(x25->dte_facilities)); |
1490 | if (rc) |
1491 | rc = -EFAULT; |
1492 | break; |
1493 | } |
1494 | |
1495 | case SIOCX25SDTEFACILITIES: { |
1496 | struct x25_dte_facilities dtefacs; |
1497 | rc = -EFAULT; |
1498 | if (copy_from_user(&dtefacs, argp, sizeof(dtefacs))) |
1499 | break; |
1500 | rc = -EINVAL; |
1501 | if (sk->sk_state != TCP_LISTEN && |
1502 | sk->sk_state != TCP_CLOSE) |
1503 | break; |
1504 | if (dtefacs.calling_len > X25_MAX_AE_LEN) |
1505 | break; |
1506 | if (dtefacs.calling_ae == NULL) |
1507 | break; |
1508 | if (dtefacs.called_len > X25_MAX_AE_LEN) |
1509 | break; |
1510 | if (dtefacs.called_ae == NULL) |
1511 | break; |
1512 | x25->dte_facilities = dtefacs; |
1513 | rc = 0; |
1514 | break; |
1515 | } |
1516 | |
1517 | case SIOCX25GCALLUSERDATA: { |
1518 | struct x25_calluserdata cud = x25->calluserdata; |
1519 | rc = copy_to_user(argp, &cud, |
1520 | sizeof(cud)) ? -EFAULT : 0; |
1521 | break; |
1522 | } |
1523 | |
1524 | case SIOCX25SCALLUSERDATA: { |
1525 | struct x25_calluserdata calluserdata; |
1526 | |
1527 | rc = -EFAULT; |
1528 | if (copy_from_user(&calluserdata, argp, |
1529 | sizeof(calluserdata))) |
1530 | break; |
1531 | rc = -EINVAL; |
1532 | if (calluserdata.cudlength > X25_MAX_CUD_LEN) |
1533 | break; |
1534 | x25->calluserdata = calluserdata; |
1535 | rc = 0; |
1536 | break; |
1537 | } |
1538 | |
1539 | case SIOCX25GCAUSEDIAG: { |
1540 | struct x25_causediag causediag; |
1541 | causediag = x25->causediag; |
1542 | rc = copy_to_user(argp, &causediag, |
1543 | sizeof(causediag)) ? -EFAULT : 0; |
1544 | break; |
1545 | } |
1546 | |
1547 | case SIOCX25SCAUSEDIAG: { |
1548 | struct x25_causediag causediag; |
1549 | rc = -EFAULT; |
1550 | if (copy_from_user(&causediag, argp, sizeof(causediag))) |
1551 | break; |
1552 | x25->causediag = causediag; |
1553 | rc = 0; |
1554 | break; |
1555 | |
1556 | } |
1557 | |
1558 | case SIOCX25SCUDMATCHLEN: { |
1559 | struct x25_subaddr sub_addr; |
1560 | rc = -EINVAL; |
1561 | if(sk->sk_state != TCP_CLOSE) |
1562 | break; |
1563 | rc = -EFAULT; |
1564 | if (copy_from_user(&sub_addr, argp, |
1565 | sizeof(sub_addr))) |
1566 | break; |
1567 | rc = -EINVAL; |
1568 | if(sub_addr.cudmatchlength > X25_MAX_CUD_LEN) |
1569 | break; |
1570 | x25->cudmatchlength = sub_addr.cudmatchlength; |
1571 | rc = 0; |
1572 | break; |
1573 | } |
1574 | |
1575 | case SIOCX25CALLACCPTAPPRV: { |
1576 | rc = -EINVAL; |
1577 | if (sk->sk_state != TCP_CLOSE) |
1578 | break; |
1579 | x25->accptapprv = X25_ALLOW_ACCPT_APPRV; |
1580 | rc = 0; |
1581 | break; |
1582 | } |
1583 | |
1584 | case SIOCX25SENDCALLACCPT: { |
1585 | rc = -EINVAL; |
1586 | if (sk->sk_state != TCP_ESTABLISHED) |
1587 | break; |
1588 | if (x25->accptapprv) /* must call accptapprv above */ |
1589 | break; |
1590 | x25_write_internal(sk, X25_CALL_ACCEPTED); |
1591 | x25->state = X25_STATE_3; |
1592 | rc = 0; |
1593 | break; |
1594 | } |
1595 | |
1596 | default: |
1597 | rc = -ENOIOCTLCMD; |
1598 | break; |
1599 | } |
1600 | unlock_kernel(); |
1601 | |
1602 | return rc; |
1603 | } |
1604 | |
1605 | static const struct net_proto_family x25_family_ops = { |
1606 | .family = AF_X25, |
1607 | .create = x25_create, |
1608 | .owner = THIS_MODULE, |
1609 | }; |
1610 | |
1611 | #ifdef CONFIG_COMPAT |
1612 | static int compat_x25_subscr_ioctl(unsigned int cmd, |
1613 | struct compat_x25_subscrip_struct __user *x25_subscr32) |
1614 | { |
1615 | struct compat_x25_subscrip_struct x25_subscr; |
1616 | struct x25_neigh *nb; |
1617 | struct net_device *dev; |
1618 | int rc = -EINVAL; |
1619 | |
1620 | rc = -EFAULT; |
1621 | if (copy_from_user(&x25_subscr, x25_subscr32, sizeof(*x25_subscr32))) |
1622 | goto out; |
1623 | |
1624 | rc = -EINVAL; |
1625 | dev = x25_dev_get(x25_subscr.device); |
1626 | if (dev == NULL) |
1627 | goto out; |
1628 | |
1629 | nb = x25_get_neigh(dev); |
1630 | if (nb == NULL) |
1631 | goto out_dev_put; |
1632 | |
1633 | dev_put(dev); |
1634 | |
1635 | if (cmd == SIOCX25GSUBSCRIP) { |
1636 | x25_subscr.extended = nb->extended; |
1637 | x25_subscr.global_facil_mask = nb->global_facil_mask; |
1638 | rc = copy_to_user(x25_subscr32, &x25_subscr, |
1639 | sizeof(*x25_subscr32)) ? -EFAULT : 0; |
1640 | } else { |
1641 | rc = -EINVAL; |
1642 | if (x25_subscr.extended == 0 || x25_subscr.extended == 1) { |
1643 | rc = 0; |
1644 | nb->extended = x25_subscr.extended; |
1645 | nb->global_facil_mask = x25_subscr.global_facil_mask; |
1646 | } |
1647 | } |
1648 | x25_neigh_put(nb); |
1649 | out: |
1650 | return rc; |
1651 | out_dev_put: |
1652 | dev_put(dev); |
1653 | goto out; |
1654 | } |
1655 | |
1656 | static int compat_x25_ioctl(struct socket *sock, unsigned int cmd, |
1657 | unsigned long arg) |
1658 | { |
1659 | void __user *argp = compat_ptr(arg); |
1660 | struct sock *sk = sock->sk; |
1661 | |
1662 | int rc = -ENOIOCTLCMD; |
1663 | |
1664 | switch(cmd) { |
1665 | case TIOCOUTQ: |
1666 | case TIOCINQ: |
1667 | rc = x25_ioctl(sock, cmd, (unsigned long)argp); |
1668 | break; |
1669 | case SIOCGSTAMP: |
1670 | rc = -EINVAL; |
1671 | lock_kernel(); |
1672 | if (sk) |
1673 | rc = compat_sock_get_timestamp(sk, |
1674 | (struct timeval __user*)argp); |
1675 | unlock_kernel(); |
1676 | break; |
1677 | case SIOCGSTAMPNS: |
1678 | rc = -EINVAL; |
1679 | lock_kernel(); |
1680 | if (sk) |
1681 | rc = compat_sock_get_timestampns(sk, |
1682 | (struct timespec __user*)argp); |
1683 | unlock_kernel(); |
1684 | break; |
1685 | case SIOCGIFADDR: |
1686 | case SIOCSIFADDR: |
1687 | case SIOCGIFDSTADDR: |
1688 | case SIOCSIFDSTADDR: |
1689 | case SIOCGIFBRDADDR: |
1690 | case SIOCSIFBRDADDR: |
1691 | case SIOCGIFNETMASK: |
1692 | case SIOCSIFNETMASK: |
1693 | case SIOCGIFMETRIC: |
1694 | case SIOCSIFMETRIC: |
1695 | rc = -EINVAL; |
1696 | break; |
1697 | case SIOCADDRT: |
1698 | case SIOCDELRT: |
1699 | rc = -EPERM; |
1700 | if (!capable(CAP_NET_ADMIN)) |
1701 | break; |
1702 | lock_kernel(); |
1703 | rc = x25_route_ioctl(cmd, argp); |
1704 | unlock_kernel(); |
1705 | break; |
1706 | case SIOCX25GSUBSCRIP: |
1707 | lock_kernel(); |
1708 | rc = compat_x25_subscr_ioctl(cmd, argp); |
1709 | unlock_kernel(); |
1710 | break; |
1711 | case SIOCX25SSUBSCRIP: |
1712 | rc = -EPERM; |
1713 | if (!capable(CAP_NET_ADMIN)) |
1714 | break; |
1715 | lock_kernel(); |
1716 | rc = compat_x25_subscr_ioctl(cmd, argp); |
1717 | unlock_kernel(); |
1718 | break; |
1719 | case SIOCX25GFACILITIES: |
1720 | case SIOCX25SFACILITIES: |
1721 | case SIOCX25GDTEFACILITIES: |
1722 | case SIOCX25SDTEFACILITIES: |
1723 | case SIOCX25GCALLUSERDATA: |
1724 | case SIOCX25SCALLUSERDATA: |
1725 | case SIOCX25GCAUSEDIAG: |
1726 | case SIOCX25SCAUSEDIAG: |
1727 | case SIOCX25SCUDMATCHLEN: |
1728 | case SIOCX25CALLACCPTAPPRV: |
1729 | case SIOCX25SENDCALLACCPT: |
1730 | rc = x25_ioctl(sock, cmd, (unsigned long)argp); |
1731 | break; |
1732 | default: |
1733 | rc = -ENOIOCTLCMD; |
1734 | break; |
1735 | } |
1736 | return rc; |
1737 | } |
1738 | #endif |
1739 | |
1740 | static const struct proto_ops x25_proto_ops = { |
1741 | .family = AF_X25, |
1742 | .owner = THIS_MODULE, |
1743 | .release = x25_release, |
1744 | .bind = x25_bind, |
1745 | .connect = x25_connect, |
1746 | .socketpair = sock_no_socketpair, |
1747 | .accept = x25_accept, |
1748 | .getname = x25_getname, |
1749 | .poll = x25_datagram_poll, |
1750 | .ioctl = x25_ioctl, |
1751 | #ifdef CONFIG_COMPAT |
1752 | .compat_ioctl = compat_x25_ioctl, |
1753 | #endif |
1754 | .listen = x25_listen, |
1755 | .shutdown = sock_no_shutdown, |
1756 | .setsockopt = x25_setsockopt, |
1757 | .getsockopt = x25_getsockopt, |
1758 | .sendmsg = x25_sendmsg, |
1759 | .recvmsg = x25_recvmsg, |
1760 | .mmap = sock_no_mmap, |
1761 | .sendpage = sock_no_sendpage, |
1762 | }; |
1763 | |
1764 | static struct packet_type x25_packet_type __read_mostly = { |
1765 | .type = cpu_to_be16(ETH_P_X25), |
1766 | .func = x25_lapb_receive_frame, |
1767 | }; |
1768 | |
1769 | static struct notifier_block x25_dev_notifier = { |
1770 | .notifier_call = x25_device_event, |
1771 | }; |
1772 | |
1773 | void x25_kill_by_neigh(struct x25_neigh *nb) |
1774 | { |
1775 | struct sock *s; |
1776 | struct hlist_node *node; |
1777 | |
1778 | write_lock_bh(&x25_list_lock); |
1779 | |
1780 | sk_for_each(s, node, &x25_list) |
1781 | if (x25_sk(s)->neighbour == nb) |
1782 | x25_disconnect(s, ENETUNREACH, 0, 0); |
1783 | |
1784 | write_unlock_bh(&x25_list_lock); |
1785 | |
1786 | /* Remove any related forwards */ |
1787 | x25_clear_forward_by_dev(nb->dev); |
1788 | } |
1789 | |
1790 | static int __init x25_init(void) |
1791 | { |
1792 | int rc = proto_register(&x25_proto, 0); |
1793 | |
1794 | if (rc != 0) |
1795 | goto out; |
1796 | |
1797 | rc = sock_register(&x25_family_ops); |
1798 | if (rc != 0) |
1799 | goto out_proto; |
1800 | |
1801 | dev_add_pack(&x25_packet_type); |
1802 | |
1803 | rc = register_netdevice_notifier(&x25_dev_notifier); |
1804 | if (rc != 0) |
1805 | goto out_sock; |
1806 | |
1807 | printk(KERN_INFO "X.25 for Linux Version 0.2\n"); |
1808 | |
1809 | x25_register_sysctl(); |
1810 | rc = x25_proc_init(); |
1811 | if (rc != 0) |
1812 | goto out_dev; |
1813 | out: |
1814 | return rc; |
1815 | out_dev: |
1816 | unregister_netdevice_notifier(&x25_dev_notifier); |
1817 | out_sock: |
1818 | sock_unregister(AF_X25); |
1819 | out_proto: |
1820 | proto_unregister(&x25_proto); |
1821 | goto out; |
1822 | } |
1823 | module_init(x25_init); |
1824 | |
1825 | static void __exit x25_exit(void) |
1826 | { |
1827 | x25_proc_exit(); |
1828 | x25_link_free(); |
1829 | x25_route_free(); |
1830 | |
1831 | x25_unregister_sysctl(); |
1832 | |
1833 | unregister_netdevice_notifier(&x25_dev_notifier); |
1834 | |
1835 | dev_remove_pack(&x25_packet_type); |
1836 | |
1837 | sock_unregister(AF_X25); |
1838 | proto_unregister(&x25_proto); |
1839 | } |
1840 | module_exit(x25_exit); |
1841 | |
1842 | MODULE_AUTHOR("Jonathan Naylor <g4klx@g4klx.demon.co.uk>"); |
1843 | MODULE_DESCRIPTION("The X.25 Packet Layer network layer protocol"); |
1844 | MODULE_LICENSE("GPL"); |
1845 | MODULE_ALIAS_NETPROTO(PF_X25); |
1846 |
Branches:
ben-wpan
ben-wpan-stefan
javiroman/ks7010
jz-2.6.34
jz-2.6.34-rc5
jz-2.6.34-rc6
jz-2.6.34-rc7
jz-2.6.35
jz-2.6.36
jz-2.6.37
jz-2.6.38
jz-2.6.39
jz-3.0
jz-3.1
jz-3.11
jz-3.12
jz-3.13
jz-3.15
jz-3.16
jz-3.18-dt
jz-3.2
jz-3.3
jz-3.4
jz-3.5
jz-3.6
jz-3.6-rc2-pwm
jz-3.9
jz-3.9-clk
jz-3.9-rc8
jz47xx
jz47xx-2.6.38
master
Tags:
od-2011-09-04
od-2011-09-18
v2.6.34-rc5
v2.6.34-rc6
v2.6.34-rc7
v3.9