2 * This program is free software; you can redistribute it and/or modify
3 * it under the terms of the GNU General Public License as published by
4 * the Free Software Foundation; either version 2 of the License, or
5 * (at your option) any later version.
7 * Copyright Jonathan Naylor G4KLX (g4klx@g4klx.demon.co.uk)
8 * Copyright Alan Cox GW4PTS (alan@lxorguk.ukuu.org.uk)
9 * Copyright Tomi Manninen OH2BNS (oh2bns@sral.fi)
11 #include <linux/errno.h>
12 #include <linux/types.h>
13 #include <linux/socket.h>
15 #include <linux/kernel.h>
16 #include <linux/timer.h>
17 #include <linux/string.h>
18 #include <linux/sockios.h>
19 #include <linux/net.h>
21 #include <linux/inet.h>
22 #include <linux/netdevice.h>
24 #include <linux/if_arp.h>
25 #include <linux/skbuff.h>
27 #include <asm/uaccess.h>
28 #include <asm/system.h>
29 #include <linux/fcntl.h>
30 #include <linux/termios.h> /* For TIOCINQ/OUTQ */
32 #include <linux/interrupt.h>
33 #include <linux/notifier.h>
34 #include <linux/netfilter.h>
35 #include <linux/init.h>
36 #include <linux/spinlock.h>
37 #include <net/netrom.h>
38 #include <linux/seq_file.h>
40 static unsigned int nr_neigh_no = 1;
42 static HLIST_HEAD(nr_node_list);
43 static DEFINE_SPINLOCK(nr_node_list_lock);
44 static HLIST_HEAD(nr_neigh_list);
45 static DEFINE_SPINLOCK(nr_neigh_list_lock);
47 static struct nr_node *nr_node_get(ax25_address *callsign)
49 struct nr_node *found = NULL;
50 struct nr_node *nr_node;
51 struct hlist_node *node;
53 spin_lock_bh(&nr_node_list_lock);
54 nr_node_for_each(nr_node, node, &nr_node_list)
55 if (ax25cmp(callsign, &nr_node->callsign) == 0) {
56 nr_node_hold(nr_node);
60 spin_unlock_bh(&nr_node_list_lock);
64 static struct nr_neigh *nr_neigh_get_dev(ax25_address *callsign,
65 struct net_device *dev)
67 struct nr_neigh *found = NULL;
68 struct nr_neigh *nr_neigh;
69 struct hlist_node *node;
71 spin_lock_bh(&nr_neigh_list_lock);
72 nr_neigh_for_each(nr_neigh, node, &nr_neigh_list)
73 if (ax25cmp(callsign, &nr_neigh->callsign) == 0 &&
74 nr_neigh->dev == dev) {
75 nr_neigh_hold(nr_neigh);
79 spin_unlock_bh(&nr_neigh_list_lock);
83 static void nr_remove_neigh(struct nr_neigh *);
86 * Add a new route to a node, and in the process add the node and the
87 * neighbour if it is new.
89 static int __must_check nr_add_node(ax25_address *nr, const char *mnemonic,
90 ax25_address *ax25, ax25_digi *ax25_digi, struct net_device *dev,
91 int quality, int obs_count)
93 struct nr_node *nr_node;
94 struct nr_neigh *nr_neigh;
95 struct nr_route nr_route;
97 struct net_device *odev;
99 if ((odev=nr_dev_get(nr)) != NULL) { /* Can't add routes to ourself */
104 nr_node = nr_node_get(nr);
106 nr_neigh = nr_neigh_get_dev(ax25, dev);
109 * The L2 link to a neighbour has failed in the past
110 * and now a frame comes from this neighbour. We assume
111 * it was a temporary trouble with the link and reset the
112 * routes now (and not wait for a node broadcast).
114 if (nr_neigh != NULL && nr_neigh->failed != 0 && quality == 0) {
115 struct nr_node *nr_nodet;
116 struct hlist_node *node;
118 spin_lock_bh(&nr_node_list_lock);
119 nr_node_for_each(nr_nodet, node, &nr_node_list) {
120 nr_node_lock(nr_nodet);
121 for (i = 0; i < nr_nodet->count; i++)
122 if (nr_nodet->routes[i].neighbour == nr_neigh)
123 if (i < nr_nodet->which)
125 nr_node_unlock(nr_nodet);
127 spin_unlock_bh(&nr_node_list_lock);
130 if (nr_neigh != NULL)
131 nr_neigh->failed = 0;
133 if (quality == 0 && nr_neigh != NULL && nr_node != NULL) {
134 nr_neigh_put(nr_neigh);
135 nr_node_put(nr_node);
139 if (nr_neigh == NULL) {
140 if ((nr_neigh = kmalloc(sizeof(*nr_neigh), GFP_ATOMIC)) == NULL) {
142 nr_node_put(nr_node);
146 nr_neigh->callsign = *ax25;
147 nr_neigh->digipeat = NULL;
148 nr_neigh->ax25 = NULL;
150 nr_neigh->quality = sysctl_netrom_default_path_quality;
151 nr_neigh->locked = 0;
153 nr_neigh->number = nr_neigh_no++;
154 nr_neigh->failed = 0;
155 atomic_set(&nr_neigh->refcount, 1);
157 if (ax25_digi != NULL && ax25_digi->ndigi > 0) {
158 nr_neigh->digipeat = kmemdup(ax25_digi,
161 if (nr_neigh->digipeat == NULL) {
164 nr_node_put(nr_node);
169 spin_lock_bh(&nr_neigh_list_lock);
170 hlist_add_head(&nr_neigh->neigh_node, &nr_neigh_list);
171 nr_neigh_hold(nr_neigh);
172 spin_unlock_bh(&nr_neigh_list_lock);
175 if (quality != 0 && ax25cmp(nr, ax25) == 0 && !nr_neigh->locked)
176 nr_neigh->quality = quality;
178 if (nr_node == NULL) {
179 if ((nr_node = kmalloc(sizeof(*nr_node), GFP_ATOMIC)) == NULL) {
181 nr_neigh_put(nr_neigh);
185 nr_node->callsign = *nr;
186 strcpy(nr_node->mnemonic, mnemonic);
190 atomic_set(&nr_node->refcount, 1);
191 spin_lock_init(&nr_node->node_lock);
193 nr_node->routes[0].quality = quality;
194 nr_node->routes[0].obs_count = obs_count;
195 nr_node->routes[0].neighbour = nr_neigh;
197 nr_neigh_hold(nr_neigh);
200 spin_lock_bh(&nr_node_list_lock);
201 hlist_add_head(&nr_node->node_node, &nr_node_list);
202 /* refcount initialized at 1 */
203 spin_unlock_bh(&nr_node_list_lock);
207 nr_node_lock(nr_node);
210 strcpy(nr_node->mnemonic, mnemonic);
212 for (found = 0, i = 0; i < nr_node->count; i++) {
213 if (nr_node->routes[i].neighbour == nr_neigh) {
214 nr_node->routes[i].quality = quality;
215 nr_node->routes[i].obs_count = obs_count;
222 /* We have space at the bottom, slot it in */
223 if (nr_node->count < 3) {
224 nr_node->routes[2] = nr_node->routes[1];
225 nr_node->routes[1] = nr_node->routes[0];
227 nr_node->routes[0].quality = quality;
228 nr_node->routes[0].obs_count = obs_count;
229 nr_node->routes[0].neighbour = nr_neigh;
233 nr_neigh_hold(nr_neigh);
236 /* It must be better than the worst */
237 if (quality > nr_node->routes[2].quality) {
238 nr_node->routes[2].neighbour->count--;
239 nr_neigh_put(nr_node->routes[2].neighbour);
241 if (nr_node->routes[2].neighbour->count == 0 && !nr_node->routes[2].neighbour->locked)
242 nr_remove_neigh(nr_node->routes[2].neighbour);
244 nr_node->routes[2].quality = quality;
245 nr_node->routes[2].obs_count = obs_count;
246 nr_node->routes[2].neighbour = nr_neigh;
248 nr_neigh_hold(nr_neigh);
254 /* Now re-sort the routes in quality order */
255 switch (nr_node->count) {
257 if (nr_node->routes[1].quality > nr_node->routes[0].quality) {
258 switch (nr_node->which) {
259 case 0: nr_node->which = 1; break;
260 case 1: nr_node->which = 0; break;
263 nr_route = nr_node->routes[0];
264 nr_node->routes[0] = nr_node->routes[1];
265 nr_node->routes[1] = nr_route;
267 if (nr_node->routes[2].quality > nr_node->routes[1].quality) {
268 switch (nr_node->which) {
269 case 1: nr_node->which = 2;
272 case 2: nr_node->which = 1;
278 nr_route = nr_node->routes[1];
279 nr_node->routes[1] = nr_node->routes[2];
280 nr_node->routes[2] = nr_route;
283 if (nr_node->routes[1].quality > nr_node->routes[0].quality) {
284 switch (nr_node->which) {
285 case 0: nr_node->which = 1;
288 case 1: nr_node->which = 0;
293 nr_route = nr_node->routes[0];
294 nr_node->routes[0] = nr_node->routes[1];
295 nr_node->routes[1] = nr_route;
301 for (i = 0; i < nr_node->count; i++) {
302 if (nr_node->routes[i].neighbour == nr_neigh) {
303 if (i < nr_node->which)
309 nr_neigh_put(nr_neigh);
310 nr_node_unlock(nr_node);
311 nr_node_put(nr_node);
315 static inline void __nr_remove_node(struct nr_node *nr_node)
317 hlist_del_init(&nr_node->node_node);
318 nr_node_put(nr_node);
321 #define nr_remove_node_locked(__node) \
322 __nr_remove_node(__node)
324 static void nr_remove_node(struct nr_node *nr_node)
326 spin_lock_bh(&nr_node_list_lock);
327 __nr_remove_node(nr_node);
328 spin_unlock_bh(&nr_node_list_lock);
331 static inline void __nr_remove_neigh(struct nr_neigh *nr_neigh)
333 hlist_del_init(&nr_neigh->neigh_node);
334 nr_neigh_put(nr_neigh);
337 #define nr_remove_neigh_locked(__neigh) \
338 __nr_remove_neigh(__neigh)
340 static void nr_remove_neigh(struct nr_neigh *nr_neigh)
342 spin_lock_bh(&nr_neigh_list_lock);
343 __nr_remove_neigh(nr_neigh);
344 spin_unlock_bh(&nr_neigh_list_lock);
348 * "Delete" a node. Strictly speaking remove a route to a node. The node
349 * is only deleted if no routes are left to it.
351 static int nr_del_node(ax25_address *callsign, ax25_address *neighbour, struct net_device *dev)
353 struct nr_node *nr_node;
354 struct nr_neigh *nr_neigh;
357 nr_node = nr_node_get(callsign);
362 nr_neigh = nr_neigh_get_dev(neighbour, dev);
364 if (nr_neigh == NULL) {
365 nr_node_put(nr_node);
369 nr_node_lock(nr_node);
370 for (i = 0; i < nr_node->count; i++) {
371 if (nr_node->routes[i].neighbour == nr_neigh) {
373 nr_neigh_put(nr_neigh);
375 if (nr_neigh->count == 0 && !nr_neigh->locked)
376 nr_remove_neigh(nr_neigh);
377 nr_neigh_put(nr_neigh);
381 if (nr_node->count == 0) {
382 nr_remove_node(nr_node);
386 nr_node->routes[0] = nr_node->routes[1];
388 nr_node->routes[1] = nr_node->routes[2];
392 nr_node_put(nr_node);
394 nr_node_unlock(nr_node);
399 nr_neigh_put(nr_neigh);
400 nr_node_unlock(nr_node);
401 nr_node_put(nr_node);
407 * Lock a neighbour with a quality.
409 static int __must_check nr_add_neigh(ax25_address *callsign,
410 ax25_digi *ax25_digi, struct net_device *dev, unsigned int quality)
412 struct nr_neigh *nr_neigh;
414 nr_neigh = nr_neigh_get_dev(callsign, dev);
416 nr_neigh->quality = quality;
417 nr_neigh->locked = 1;
418 nr_neigh_put(nr_neigh);
422 if ((nr_neigh = kmalloc(sizeof(*nr_neigh), GFP_ATOMIC)) == NULL)
425 nr_neigh->callsign = *callsign;
426 nr_neigh->digipeat = NULL;
427 nr_neigh->ax25 = NULL;
429 nr_neigh->quality = quality;
430 nr_neigh->locked = 1;
432 nr_neigh->number = nr_neigh_no++;
433 nr_neigh->failed = 0;
434 atomic_set(&nr_neigh->refcount, 1);
436 if (ax25_digi != NULL && ax25_digi->ndigi > 0) {
437 nr_neigh->digipeat = kmemdup(ax25_digi, sizeof(*ax25_digi),
439 if (nr_neigh->digipeat == NULL) {
445 spin_lock_bh(&nr_neigh_list_lock);
446 hlist_add_head(&nr_neigh->neigh_node, &nr_neigh_list);
447 /* refcount is initialized at 1 */
448 spin_unlock_bh(&nr_neigh_list_lock);
454 * "Delete" a neighbour. The neighbour is only removed if the number
455 * of nodes that may use it is zero.
457 static int nr_del_neigh(ax25_address *callsign, struct net_device *dev, unsigned int quality)
459 struct nr_neigh *nr_neigh;
461 nr_neigh = nr_neigh_get_dev(callsign, dev);
463 if (nr_neigh == NULL) return -EINVAL;
465 nr_neigh->quality = quality;
466 nr_neigh->locked = 0;
468 if (nr_neigh->count == 0)
469 nr_remove_neigh(nr_neigh);
470 nr_neigh_put(nr_neigh);
476 * Decrement the obsolescence count by one. If a route is reduced to a
477 * count of zero, remove it. Also remove any unlocked neighbours with
478 * zero nodes routing via it.
480 static int nr_dec_obs(void)
482 struct nr_neigh *nr_neigh;
484 struct hlist_node *node, *nodet;
487 spin_lock_bh(&nr_node_list_lock);
488 nr_node_for_each_safe(s, node, nodet, &nr_node_list) {
490 for (i = 0; i < s->count; i++) {
491 switch (s->routes[i].obs_count) {
492 case 0: /* A locked entry */
495 case 1: /* From 1 -> 0 */
496 nr_neigh = s->routes[i].neighbour;
499 nr_neigh_put(nr_neigh);
501 if (nr_neigh->count == 0 && !nr_neigh->locked)
502 nr_remove_neigh(nr_neigh);
508 s->routes[0] = s->routes[1];
510 s->routes[1] = s->routes[2];
517 s->routes[i].obs_count--;
524 nr_remove_node_locked(s);
527 spin_unlock_bh(&nr_node_list_lock);
533 * A device has been removed. Remove its routes and neighbours.
535 void nr_rt_device_down(struct net_device *dev)
538 struct hlist_node *node, *nodet, *node2, *node2t;
542 spin_lock_bh(&nr_neigh_list_lock);
543 nr_neigh_for_each_safe(s, node, nodet, &nr_neigh_list) {
545 spin_lock_bh(&nr_node_list_lock);
546 nr_node_for_each_safe(t, node2, node2t, &nr_node_list) {
548 for (i = 0; i < t->count; i++) {
549 if (t->routes[i].neighbour == s) {
554 t->routes[0] = t->routes[1];
556 t->routes[1] = t->routes[2];
564 nr_remove_node_locked(t);
567 spin_unlock_bh(&nr_node_list_lock);
569 nr_remove_neigh_locked(s);
572 spin_unlock_bh(&nr_neigh_list_lock);
576 * Check that the device given is a valid AX.25 interface that is "up".
577 * Or a valid ethernet interface with an AX.25 callsign binding.
579 static struct net_device *nr_ax25_dev_get(char *devname)
581 struct net_device *dev;
583 if ((dev = dev_get_by_name(devname)) == NULL)
586 if ((dev->flags & IFF_UP) && dev->type == ARPHRD_AX25)
594 * Find the first active NET/ROM device, usually "nr0".
596 struct net_device *nr_dev_first(void)
598 struct net_device *dev, *first = NULL;
600 read_lock(&dev_base_lock);
601 for (dev = dev_base; dev != NULL; dev = dev->next) {
602 if ((dev->flags & IFF_UP) && dev->type == ARPHRD_NETROM)
603 if (first == NULL || strncmp(dev->name, first->name, 3) < 0)
608 read_unlock(&dev_base_lock);
614 * Find the NET/ROM device for the given callsign.
616 struct net_device *nr_dev_get(ax25_address *addr)
618 struct net_device *dev;
620 read_lock(&dev_base_lock);
621 for (dev = dev_base; dev != NULL; dev = dev->next) {
622 if ((dev->flags & IFF_UP) && dev->type == ARPHRD_NETROM && ax25cmp(addr, (ax25_address *)dev->dev_addr) == 0) {
628 read_unlock(&dev_base_lock);
632 static ax25_digi *nr_call_to_digi(int ndigis, ax25_address *digipeaters)
634 static ax25_digi ax25_digi;
640 for (i = 0; i < ndigis; i++) {
641 ax25_digi.calls[i] = digipeaters[i];
642 ax25_digi.repeated[i] = 0;
645 ax25_digi.ndigi = ndigis;
646 ax25_digi.lastrepeat = -1;
652 * Handle the ioctls that control the routing functions.
654 int nr_rt_ioctl(unsigned int cmd, void __user *arg)
656 struct nr_route_struct nr_route;
657 struct net_device *dev;
662 if (copy_from_user(&nr_route, arg, sizeof(struct nr_route_struct)))
664 if ((dev = nr_ax25_dev_get(nr_route.device)) == NULL)
666 if (nr_route.ndigis < 0 || nr_route.ndigis > AX25_MAX_DIGIS) {
670 switch (nr_route.type) {
672 ret = nr_add_node(&nr_route.callsign,
675 nr_call_to_digi(nr_route.ndigis, nr_route.digipeaters),
676 dev, nr_route.quality,
680 ret = nr_add_neigh(&nr_route.callsign,
681 nr_call_to_digi(nr_route.ndigis, nr_route.digipeaters),
682 dev, nr_route.quality);
691 if (copy_from_user(&nr_route, arg, sizeof(struct nr_route_struct)))
693 if ((dev = nr_ax25_dev_get(nr_route.device)) == NULL)
695 switch (nr_route.type) {
697 ret = nr_del_node(&nr_route.callsign,
698 &nr_route.neighbour, dev);
701 ret = nr_del_neigh(&nr_route.callsign,
702 dev, nr_route.quality);
721 * A level 2 link has timed out, therefore it appears to be a poor link,
722 * then don't use that neighbour until it is reset.
724 void nr_link_failed(ax25_cb *ax25, int reason)
726 struct nr_neigh *s, *nr_neigh = NULL;
727 struct hlist_node *node;
728 struct nr_node *nr_node = NULL;
730 spin_lock_bh(&nr_neigh_list_lock);
731 nr_neigh_for_each(s, node, &nr_neigh_list) {
732 if (s->ax25 == ax25) {
738 spin_unlock_bh(&nr_neigh_list_lock);
740 if (nr_neigh == NULL)
743 nr_neigh->ax25 = NULL;
746 if (++nr_neigh->failed < sysctl_netrom_link_fails_count) {
747 nr_neigh_put(nr_neigh);
750 spin_lock_bh(&nr_node_list_lock);
751 nr_node_for_each(nr_node, node, &nr_node_list) {
752 nr_node_lock(nr_node);
753 if (nr_node->which < nr_node->count &&
754 nr_node->routes[nr_node->which].neighbour == nr_neigh)
756 nr_node_unlock(nr_node);
758 spin_unlock_bh(&nr_node_list_lock);
759 nr_neigh_put(nr_neigh);
763 * Route a frame to an appropriate AX.25 connection. A NULL ax25_cb
764 * indicates an internally generated frame.
766 int nr_route_frame(struct sk_buff *skb, ax25_cb *ax25)
768 ax25_address *nr_src, *nr_dest;
769 struct nr_neigh *nr_neigh;
770 struct nr_node *nr_node;
771 struct net_device *dev;
775 struct sk_buff *skbn;
778 nr_src = (ax25_address *)(skb->data + 0);
779 nr_dest = (ax25_address *)(skb->data + 7);
782 ret = nr_add_node(nr_src, "", &ax25->dest_addr, ax25->digipeat,
783 ax25->ax25_dev->dev, 0,
784 sysctl_netrom_obsolescence_count_initialiser);
789 if ((dev = nr_dev_get(nr_dest)) != NULL) { /* Its for me */
790 if (ax25 == NULL) /* Its from me */
791 ret = nr_loopback_queue(skb);
793 ret = nr_rx_frame(skb, dev);
798 if (!sysctl_netrom_routing_control && ax25 != NULL)
801 /* Its Time-To-Live has expired */
802 if (skb->data[14] == 1) {
806 nr_node = nr_node_get(nr_dest);
809 nr_node_lock(nr_node);
811 if (nr_node->which >= nr_node->count) {
812 nr_node_unlock(nr_node);
813 nr_node_put(nr_node);
817 nr_neigh = nr_node->routes[nr_node->which].neighbour;
819 if ((dev = nr_dev_first()) == NULL) {
820 nr_node_unlock(nr_node);
821 nr_node_put(nr_node);
825 /* We are going to change the netrom headers so we should get our
826 own skb, we also did not know until now how much header space
827 we had to reserve... - RXQ */
828 if ((skbn=skb_copy_expand(skb, dev->hard_header_len, 0, GFP_ATOMIC)) == NULL) {
829 nr_node_unlock(nr_node);
830 nr_node_put(nr_node);
838 dptr = skb_push(skb, 1);
839 *dptr = AX25_P_NETROM;
841 ax25s = ax25_send_frame(skb, 256, (ax25_address *)dev->dev_addr, &nr_neigh->callsign, nr_neigh->digipeat, nr_neigh->dev);
842 if (nr_neigh->ax25 && ax25s) {
843 /* We were already holding this ax25_cb */
846 nr_neigh->ax25 = ax25s;
849 ret = (nr_neigh->ax25 != NULL);
850 nr_node_unlock(nr_node);
851 nr_node_put(nr_node);
856 #ifdef CONFIG_PROC_FS
858 static void *nr_node_start(struct seq_file *seq, loff_t *pos)
860 struct nr_node *nr_node;
861 struct hlist_node *node;
864 spin_lock_bh(&nr_node_list_lock);
866 return SEQ_START_TOKEN;
868 nr_node_for_each(nr_node, node, &nr_node_list) {
877 static void *nr_node_next(struct seq_file *seq, void *v, loff_t *pos)
879 struct hlist_node *node;
882 node = (v == SEQ_START_TOKEN)
884 : ((struct nr_node *)v)->node_node.next;
886 return hlist_entry(node, struct nr_node, node_node);
889 static void nr_node_stop(struct seq_file *seq, void *v)
891 spin_unlock_bh(&nr_node_list_lock);
894 static int nr_node_show(struct seq_file *seq, void *v)
899 if (v == SEQ_START_TOKEN)
901 "callsign mnemonic w n qual obs neigh qual obs neigh qual obs neigh\n");
903 struct nr_node *nr_node = v;
904 nr_node_lock(nr_node);
905 seq_printf(seq, "%-9s %-7s %d %d",
906 ax2asc(buf, &nr_node->callsign),
907 (nr_node->mnemonic[0] == '\0') ? "*" : nr_node->mnemonic,
911 for (i = 0; i < nr_node->count; i++) {
912 seq_printf(seq, " %3d %d %05d",
913 nr_node->routes[i].quality,
914 nr_node->routes[i].obs_count,
915 nr_node->routes[i].neighbour->number);
917 nr_node_unlock(nr_node);
924 static struct seq_operations nr_node_seqops = {
925 .start = nr_node_start,
926 .next = nr_node_next,
927 .stop = nr_node_stop,
928 .show = nr_node_show,
931 static int nr_node_info_open(struct inode *inode, struct file *file)
933 return seq_open(file, &nr_node_seqops);
936 const struct file_operations nr_nodes_fops = {
937 .owner = THIS_MODULE,
938 .open = nr_node_info_open,
941 .release = seq_release,
944 static void *nr_neigh_start(struct seq_file *seq, loff_t *pos)
946 struct nr_neigh *nr_neigh;
947 struct hlist_node *node;
950 spin_lock_bh(&nr_neigh_list_lock);
952 return SEQ_START_TOKEN;
954 nr_neigh_for_each(nr_neigh, node, &nr_neigh_list) {
961 static void *nr_neigh_next(struct seq_file *seq, void *v, loff_t *pos)
963 struct hlist_node *node;
966 node = (v == SEQ_START_TOKEN)
967 ? nr_neigh_list.first
968 : ((struct nr_neigh *)v)->neigh_node.next;
970 return hlist_entry(node, struct nr_neigh, neigh_node);
973 static void nr_neigh_stop(struct seq_file *seq, void *v)
975 spin_unlock_bh(&nr_neigh_list_lock);
978 static int nr_neigh_show(struct seq_file *seq, void *v)
983 if (v == SEQ_START_TOKEN)
984 seq_puts(seq, "addr callsign dev qual lock count failed digipeaters\n");
986 struct nr_neigh *nr_neigh = v;
988 seq_printf(seq, "%05d %-9s %-4s %3d %d %3d %3d",
990 ax2asc(buf, &nr_neigh->callsign),
991 nr_neigh->dev ? nr_neigh->dev->name : "???",
997 if (nr_neigh->digipeat != NULL) {
998 for (i = 0; i < nr_neigh->digipeat->ndigi; i++)
999 seq_printf(seq, " %s",
1000 ax2asc(buf, &nr_neigh->digipeat->calls[i]));
1003 seq_puts(seq, "\n");
1008 static struct seq_operations nr_neigh_seqops = {
1009 .start = nr_neigh_start,
1010 .next = nr_neigh_next,
1011 .stop = nr_neigh_stop,
1012 .show = nr_neigh_show,
1015 static int nr_neigh_info_open(struct inode *inode, struct file *file)
1017 return seq_open(file, &nr_neigh_seqops);
1020 const struct file_operations nr_neigh_fops = {
1021 .owner = THIS_MODULE,
1022 .open = nr_neigh_info_open,
1024 .llseek = seq_lseek,
1025 .release = seq_release,
1031 * Free all memory associated with the nodes and routes lists.
1033 void __exit nr_rt_free(void)
1035 struct nr_neigh *s = NULL;
1036 struct nr_node *t = NULL;
1037 struct hlist_node *node, *nodet;
1039 spin_lock_bh(&nr_neigh_list_lock);
1040 spin_lock_bh(&nr_node_list_lock);
1041 nr_node_for_each_safe(t, node, nodet, &nr_node_list) {
1043 nr_remove_node_locked(t);
1046 nr_neigh_for_each_safe(s, node, nodet, &nr_neigh_list) {
1051 nr_remove_neigh_locked(s);
1053 spin_unlock_bh(&nr_node_list_lock);
1054 spin_unlock_bh(&nr_neigh_list_lock);