2 * INET An implementation of the TCP/IP protocol suite for the LINUX
3 * operating system. INET is implemented using the BSD Socket
4 * interface as the means of communication with the user level.
6 * Routing netlink socket interface: protocol independent part.
8 * Authors: Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
10 * This program is free software; you can redistribute it and/or
11 * modify it under the terms of the GNU General Public License
12 * as published by the Free Software Foundation; either version
13 * 2 of the License, or (at your option) any later version.
16 * Vitaly E. Lavrov RTA_OK arithmetics was wrong.
19 #include <linux/errno.h>
20 #include <linux/module.h>
21 #include <linux/types.h>
22 #include <linux/socket.h>
23 #include <linux/kernel.h>
24 #include <linux/sched.h>
25 #include <linux/timer.h>
26 #include <linux/string.h>
27 #include <linux/sockios.h>
28 #include <linux/net.h>
29 #include <linux/fcntl.h>
31 #include <linux/slab.h>
32 #include <linux/interrupt.h>
33 #include <linux/capability.h>
34 #include <linux/skbuff.h>
35 #include <linux/init.h>
36 #include <linux/security.h>
37 #include <linux/mutex.h>
38 #include <linux/if_addr.h>
40 #include <asm/uaccess.h>
41 #include <asm/system.h>
42 #include <asm/string.h>
44 #include <linux/inet.h>
45 #include <linux/netdevice.h>
47 #include <net/protocol.h>
49 #include <net/route.h>
52 #include <net/pkt_sched.h>
53 #include <net/fib_rules.h>
54 #include <net/netlink.h>
55 #ifdef CONFIG_NET_WIRELESS_RTNETLINK
56 #include <linux/wireless.h>
57 #include <net/iw_handler.h>
58 #endif /* CONFIG_NET_WIRELESS_RTNETLINK */
60 static DEFINE_MUTEX(rtnl_mutex
);
61 static struct sock
*rtnl
;
65 mutex_lock(&rtnl_mutex
);
68 void __rtnl_unlock(void)
70 mutex_unlock(&rtnl_mutex
);
73 void rtnl_unlock(void)
75 mutex_unlock(&rtnl_mutex
);
76 if (rtnl
&& rtnl
->sk_receive_queue
.qlen
)
77 rtnl
->sk_data_ready(rtnl
, 0);
81 int rtnl_trylock(void)
83 return mutex_trylock(&rtnl_mutex
);
86 int rtattr_parse(struct rtattr
*tb
[], int maxattr
, struct rtattr
*rta
, int len
)
88 memset(tb
, 0, sizeof(struct rtattr
*)*maxattr
);
90 while (RTA_OK(rta
, len
)) {
91 unsigned flavor
= rta
->rta_type
;
92 if (flavor
&& flavor
<= maxattr
)
94 rta
= RTA_NEXT(rta
, len
);
99 struct rtnetlink_link
* rtnetlink_links
[NPROTO
];
101 static const int rtm_min
[RTM_NR_FAMILIES
] =
103 [RTM_FAM(RTM_NEWLINK
)] = NLMSG_LENGTH(sizeof(struct ifinfomsg
)),
104 [RTM_FAM(RTM_NEWADDR
)] = NLMSG_LENGTH(sizeof(struct ifaddrmsg
)),
105 [RTM_FAM(RTM_NEWROUTE
)] = NLMSG_LENGTH(sizeof(struct rtmsg
)),
106 [RTM_FAM(RTM_NEWRULE
)] = NLMSG_LENGTH(sizeof(struct fib_rule_hdr
)),
107 [RTM_FAM(RTM_NEWQDISC
)] = NLMSG_LENGTH(sizeof(struct tcmsg
)),
108 [RTM_FAM(RTM_NEWTCLASS
)] = NLMSG_LENGTH(sizeof(struct tcmsg
)),
109 [RTM_FAM(RTM_NEWTFILTER
)] = NLMSG_LENGTH(sizeof(struct tcmsg
)),
110 [RTM_FAM(RTM_NEWACTION
)] = NLMSG_LENGTH(sizeof(struct tcamsg
)),
111 [RTM_FAM(RTM_GETMULTICAST
)] = NLMSG_LENGTH(sizeof(struct rtgenmsg
)),
112 [RTM_FAM(RTM_GETANYCAST
)] = NLMSG_LENGTH(sizeof(struct rtgenmsg
)),
115 static const int rta_max
[RTM_NR_FAMILIES
] =
117 [RTM_FAM(RTM_NEWLINK
)] = IFLA_MAX
,
118 [RTM_FAM(RTM_NEWADDR
)] = IFA_MAX
,
119 [RTM_FAM(RTM_NEWROUTE
)] = RTA_MAX
,
120 [RTM_FAM(RTM_NEWRULE
)] = FRA_MAX
,
121 [RTM_FAM(RTM_NEWQDISC
)] = TCA_MAX
,
122 [RTM_FAM(RTM_NEWTCLASS
)] = TCA_MAX
,
123 [RTM_FAM(RTM_NEWTFILTER
)] = TCA_MAX
,
124 [RTM_FAM(RTM_NEWACTION
)] = TCAA_MAX
,
127 void __rta_fill(struct sk_buff
*skb
, int attrtype
, int attrlen
, const void *data
)
130 int size
= RTA_LENGTH(attrlen
);
132 rta
= (struct rtattr
*)skb_put(skb
, RTA_ALIGN(size
));
133 rta
->rta_type
= attrtype
;
135 memcpy(RTA_DATA(rta
), data
, attrlen
);
136 memset(RTA_DATA(rta
) + attrlen
, 0, RTA_ALIGN(size
) - size
);
139 size_t rtattr_strlcpy(char *dest
, const struct rtattr
*rta
, size_t size
)
141 size_t ret
= RTA_PAYLOAD(rta
);
142 char *src
= RTA_DATA(rta
);
144 if (ret
> 0 && src
[ret
- 1] == '\0')
147 size_t len
= (ret
>= size
) ? size
- 1 : ret
;
148 memset(dest
, 0, size
);
149 memcpy(dest
, src
, len
);
154 int rtnetlink_send(struct sk_buff
*skb
, u32 pid
, unsigned group
, int echo
)
158 NETLINK_CB(skb
).dst_group
= group
;
160 atomic_inc(&skb
->users
);
161 netlink_broadcast(rtnl
, skb
, pid
, group
, GFP_KERNEL
);
163 err
= netlink_unicast(rtnl
, skb
, pid
, MSG_DONTWAIT
);
167 int rtnl_unicast(struct sk_buff
*skb
, u32 pid
)
169 return nlmsg_unicast(rtnl
, skb
, pid
);
172 int rtnl_notify(struct sk_buff
*skb
, u32 pid
, u32 group
,
173 struct nlmsghdr
*nlh
, gfp_t flags
)
178 report
= nlmsg_report(nlh
);
180 return nlmsg_notify(rtnl
, skb
, pid
, group
, report
, flags
);
183 void rtnl_set_sk_err(u32 group
, int error
)
185 netlink_set_err(rtnl
, 0, group
, error
);
188 int rtnetlink_put_metrics(struct sk_buff
*skb
, u32
*metrics
)
193 mx
= nla_nest_start(skb
, RTA_METRICS
);
197 for (i
= 0; i
< RTAX_MAX
; i
++) {
200 NLA_PUT_U32(skb
, i
+1, metrics
[i
]);
205 nla_nest_cancel(skb
, mx
);
209 return nla_nest_end(skb
, mx
);
212 return nla_nest_cancel(skb
, mx
);
215 int rtnl_put_cacheinfo(struct sk_buff
*skb
, struct dst_entry
*dst
, u32 id
,
216 u32 ts
, u32 tsage
, long expires
, u32 error
)
218 struct rta_cacheinfo ci
= {
219 .rta_lastuse
= jiffies_to_clock_t(jiffies
- dst
->lastuse
),
220 .rta_used
= dst
->__use
,
221 .rta_clntref
= atomic_read(&(dst
->__refcnt
)),
229 ci
.rta_expires
= jiffies_to_clock_t(expires
);
231 return nla_put(skb
, RTA_CACHEINFO
, sizeof(ci
), &ci
);
234 EXPORT_SYMBOL_GPL(rtnl_put_cacheinfo
);
236 static void set_operstate(struct net_device
*dev
, unsigned char transition
)
238 unsigned char operstate
= dev
->operstate
;
242 if ((operstate
== IF_OPER_DORMANT
||
243 operstate
== IF_OPER_UNKNOWN
) &&
245 operstate
= IF_OPER_UP
;
248 case IF_OPER_DORMANT
:
249 if (operstate
== IF_OPER_UP
||
250 operstate
== IF_OPER_UNKNOWN
)
251 operstate
= IF_OPER_DORMANT
;
255 if (dev
->operstate
!= operstate
) {
256 write_lock_bh(&dev_base_lock
);
257 dev
->operstate
= operstate
;
258 write_unlock_bh(&dev_base_lock
);
259 netdev_state_change(dev
);
263 static void copy_rtnl_link_stats(struct rtnl_link_stats
*a
,
264 struct net_device_stats
*b
)
266 a
->rx_packets
= b
->rx_packets
;
267 a
->tx_packets
= b
->tx_packets
;
268 a
->rx_bytes
= b
->rx_bytes
;
269 a
->tx_bytes
= b
->tx_bytes
;
270 a
->rx_errors
= b
->rx_errors
;
271 a
->tx_errors
= b
->tx_errors
;
272 a
->rx_dropped
= b
->rx_dropped
;
273 a
->tx_dropped
= b
->tx_dropped
;
275 a
->multicast
= b
->multicast
;
276 a
->collisions
= b
->collisions
;
278 a
->rx_length_errors
= b
->rx_length_errors
;
279 a
->rx_over_errors
= b
->rx_over_errors
;
280 a
->rx_crc_errors
= b
->rx_crc_errors
;
281 a
->rx_frame_errors
= b
->rx_frame_errors
;
282 a
->rx_fifo_errors
= b
->rx_fifo_errors
;
283 a
->rx_missed_errors
= b
->rx_missed_errors
;
285 a
->tx_aborted_errors
= b
->tx_aborted_errors
;
286 a
->tx_carrier_errors
= b
->tx_carrier_errors
;
287 a
->tx_fifo_errors
= b
->tx_fifo_errors
;
288 a
->tx_heartbeat_errors
= b
->tx_heartbeat_errors
;
289 a
->tx_window_errors
= b
->tx_window_errors
;
291 a
->rx_compressed
= b
->rx_compressed
;
292 a
->tx_compressed
= b
->tx_compressed
;
295 static inline size_t if_nlmsg_size(int iwbuflen
)
297 return NLMSG_ALIGN(sizeof(struct ifinfomsg
))
298 + nla_total_size(IFNAMSIZ
) /* IFLA_IFNAME */
299 + nla_total_size(IFNAMSIZ
) /* IFLA_QDISC */
300 + nla_total_size(sizeof(struct rtnl_link_ifmap
))
301 + nla_total_size(sizeof(struct rtnl_link_stats
))
302 + nla_total_size(MAX_ADDR_LEN
) /* IFLA_ADDRESS */
303 + nla_total_size(MAX_ADDR_LEN
) /* IFLA_BROADCAST */
304 + nla_total_size(4) /* IFLA_TXQLEN */
305 + nla_total_size(4) /* IFLA_WEIGHT */
306 + nla_total_size(4) /* IFLA_MTU */
307 + nla_total_size(4) /* IFLA_LINK */
308 + nla_total_size(4) /* IFLA_MASTER */
309 + nla_total_size(1) /* IFLA_OPERSTATE */
310 + nla_total_size(1) /* IFLA_LINKMODE */
311 + nla_total_size(iwbuflen
);
314 static int rtnl_fill_ifinfo(struct sk_buff
*skb
, struct net_device
*dev
,
315 void *iwbuf
, int iwbuflen
, int type
, u32 pid
,
316 u32 seq
, u32 change
, unsigned int flags
)
318 struct ifinfomsg
*ifm
;
319 struct nlmsghdr
*nlh
;
321 nlh
= nlmsg_put(skb
, pid
, seq
, type
, sizeof(*ifm
), flags
);
325 ifm
= nlmsg_data(nlh
);
326 ifm
->ifi_family
= AF_UNSPEC
;
328 ifm
->ifi_type
= dev
->type
;
329 ifm
->ifi_index
= dev
->ifindex
;
330 ifm
->ifi_flags
= dev_get_flags(dev
);
331 ifm
->ifi_change
= change
;
333 NLA_PUT_STRING(skb
, IFLA_IFNAME
, dev
->name
);
334 NLA_PUT_U32(skb
, IFLA_TXQLEN
, dev
->tx_queue_len
);
335 NLA_PUT_U32(skb
, IFLA_WEIGHT
, dev
->weight
);
336 NLA_PUT_U8(skb
, IFLA_OPERSTATE
,
337 netif_running(dev
) ? dev
->operstate
: IF_OPER_DOWN
);
338 NLA_PUT_U8(skb
, IFLA_LINKMODE
, dev
->link_mode
);
339 NLA_PUT_U32(skb
, IFLA_MTU
, dev
->mtu
);
341 if (dev
->ifindex
!= dev
->iflink
)
342 NLA_PUT_U32(skb
, IFLA_LINK
, dev
->iflink
);
345 NLA_PUT_U32(skb
, IFLA_MASTER
, dev
->master
->ifindex
);
347 if (dev
->qdisc_sleeping
)
348 NLA_PUT_STRING(skb
, IFLA_QDISC
, dev
->qdisc_sleeping
->ops
->id
);
351 struct rtnl_link_ifmap map
= {
352 .mem_start
= dev
->mem_start
,
353 .mem_end
= dev
->mem_end
,
354 .base_addr
= dev
->base_addr
,
357 .port
= dev
->if_port
,
359 NLA_PUT(skb
, IFLA_MAP
, sizeof(map
), &map
);
363 NLA_PUT(skb
, IFLA_ADDRESS
, dev
->addr_len
, dev
->dev_addr
);
364 NLA_PUT(skb
, IFLA_BROADCAST
, dev
->addr_len
, dev
->broadcast
);
367 if (dev
->get_stats
) {
368 struct net_device_stats
*stats
= dev
->get_stats(dev
);
372 attr
= nla_reserve(skb
, IFLA_STATS
,
373 sizeof(struct rtnl_link_stats
));
375 goto nla_put_failure
;
377 copy_rtnl_link_stats(nla_data(attr
), stats
);
382 NLA_PUT(skb
, IFLA_WIRELESS
, iwbuflen
, iwbuf
);
384 return nlmsg_end(skb
, nlh
);
387 nlmsg_cancel(skb
, nlh
);
391 static int rtnl_dump_ifinfo(struct sk_buff
*skb
, struct netlink_callback
*cb
)
394 int s_idx
= cb
->args
[0];
395 struct net_device
*dev
;
397 read_lock(&dev_base_lock
);
398 for (dev
=dev_base
, idx
=0; dev
; dev
= dev
->next
, idx
++) {
401 if (rtnl_fill_ifinfo(skb
, dev
, NULL
, 0, RTM_NEWLINK
,
402 NETLINK_CB(cb
->skb
).pid
,
403 cb
->nlh
->nlmsg_seq
, 0, NLM_F_MULTI
) <= 0)
406 read_unlock(&dev_base_lock
);
412 static struct nla_policy ifla_policy
[IFLA_MAX
+1] __read_mostly
= {
413 [IFLA_IFNAME
] = { .type
= NLA_STRING
, .len
= IFNAMSIZ
-1 },
414 [IFLA_MAP
] = { .len
= sizeof(struct rtnl_link_ifmap
) },
415 [IFLA_MTU
] = { .type
= NLA_U32
},
416 [IFLA_TXQLEN
] = { .type
= NLA_U32
},
417 [IFLA_WEIGHT
] = { .type
= NLA_U32
},
418 [IFLA_OPERSTATE
] = { .type
= NLA_U8
},
419 [IFLA_LINKMODE
] = { .type
= NLA_U8
},
422 static int rtnl_setlink(struct sk_buff
*skb
, struct nlmsghdr
*nlh
, void *arg
)
424 struct ifinfomsg
*ifm
;
425 struct net_device
*dev
;
426 int err
, send_addr_notify
= 0, modified
= 0;
427 struct nlattr
*tb
[IFLA_MAX
+1];
428 char ifname
[IFNAMSIZ
];
430 err
= nlmsg_parse(nlh
, sizeof(*ifm
), tb
, IFLA_MAX
, ifla_policy
);
435 nla_strlcpy(ifname
, tb
[IFLA_IFNAME
], IFNAMSIZ
);
440 ifm
= nlmsg_data(nlh
);
441 if (ifm
->ifi_index
>= 0)
442 dev
= dev_get_by_index(ifm
->ifi_index
);
443 else if (tb
[IFLA_IFNAME
])
444 dev
= dev_get_by_name(ifname
);
453 if (tb
[IFLA_ADDRESS
] &&
454 nla_len(tb
[IFLA_ADDRESS
]) < dev
->addr_len
)
457 if (tb
[IFLA_BROADCAST
] &&
458 nla_len(tb
[IFLA_BROADCAST
]) < dev
->addr_len
)
462 struct rtnl_link_ifmap
*u_map
;
465 if (!dev
->set_config
) {
470 if (!netif_device_present(dev
)) {
475 u_map
= nla_data(tb
[IFLA_MAP
]);
476 k_map
.mem_start
= (unsigned long) u_map
->mem_start
;
477 k_map
.mem_end
= (unsigned long) u_map
->mem_end
;
478 k_map
.base_addr
= (unsigned short) u_map
->base_addr
;
479 k_map
.irq
= (unsigned char) u_map
->irq
;
480 k_map
.dma
= (unsigned char) u_map
->dma
;
481 k_map
.port
= (unsigned char) u_map
->port
;
483 err
= dev
->set_config(dev
, &k_map
);
490 if (tb
[IFLA_ADDRESS
]) {
494 if (!dev
->set_mac_address
) {
499 if (!netif_device_present(dev
)) {
504 len
= sizeof(sa_family_t
) + dev
->addr_len
;
505 sa
= kmalloc(len
, GFP_KERNEL
);
510 sa
->sa_family
= dev
->type
;
511 memcpy(sa
->sa_data
, nla_data(tb
[IFLA_ADDRESS
]),
513 err
= dev
->set_mac_address(dev
, sa
);
517 send_addr_notify
= 1;
522 err
= dev_set_mtu(dev
, nla_get_u32(tb
[IFLA_MTU
]));
529 * Interface selected by interface index but interface
530 * name provided implies that a name change has been
533 if (ifm
->ifi_index
>= 0 && ifname
[0]) {
534 err
= dev_change_name(dev
, ifname
);
540 #ifdef CONFIG_NET_WIRELESS_RTNETLINK
541 if (tb
[IFLA_WIRELESS
]) {
542 /* Call Wireless Extensions.
543 * Various stuff checked in there... */
544 err
= wireless_rtnetlink_set(dev
, nla_data(tb
[IFLA_WIRELESS
]),
545 nla_len(tb
[IFLA_WIRELESS
]));
549 #endif /* CONFIG_NET_WIRELESS_RTNETLINK */
551 if (tb
[IFLA_BROADCAST
]) {
552 nla_memcpy(dev
->broadcast
, tb
[IFLA_BROADCAST
], dev
->addr_len
);
553 send_addr_notify
= 1;
558 dev_change_flags(dev
, ifm
->ifi_flags
);
561 dev
->tx_queue_len
= nla_get_u32(tb
[IFLA_TXQLEN
]);
564 dev
->weight
= nla_get_u32(tb
[IFLA_WEIGHT
]);
566 if (tb
[IFLA_OPERSTATE
])
567 set_operstate(dev
, nla_get_u8(tb
[IFLA_OPERSTATE
]));
569 if (tb
[IFLA_LINKMODE
]) {
570 write_lock_bh(&dev_base_lock
);
571 dev
->link_mode
= nla_get_u8(tb
[IFLA_LINKMODE
]);
572 write_unlock_bh(&dev_base_lock
);
578 if (err
< 0 && modified
&& net_ratelimit())
579 printk(KERN_WARNING
"A link change request failed with "
580 "some changes comitted already. Interface %s may "
581 "have been left with an inconsistent configuration, "
582 "please check.\n", dev
->name
);
584 if (send_addr_notify
)
585 call_netdevice_notifiers(NETDEV_CHANGEADDR
, dev
);
592 static int rtnl_getlink(struct sk_buff
*skb
, struct nlmsghdr
* nlh
, void *arg
)
594 struct ifinfomsg
*ifm
;
595 struct nlattr
*tb
[IFLA_MAX
+1];
596 struct net_device
*dev
= NULL
;
597 struct sk_buff
*nskb
;
598 char *iw_buf
= NULL
, *iw
= NULL
;
602 err
= nlmsg_parse(nlh
, sizeof(*ifm
), tb
, IFLA_MAX
, ifla_policy
);
606 ifm
= nlmsg_data(nlh
);
607 if (ifm
->ifi_index
>= 0) {
608 dev
= dev_get_by_index(ifm
->ifi_index
);
615 #ifdef CONFIG_NET_WIRELESS_RTNETLINK
616 if (tb
[IFLA_WIRELESS
]) {
617 /* Call Wireless Extensions. We need to know the size before
618 * we can alloc. Various stuff checked in there... */
619 err
= wireless_rtnetlink_get(dev
, nla_data(tb
[IFLA_WIRELESS
]),
620 nla_len(tb
[IFLA_WIRELESS
]),
621 &iw_buf
, &iw_buf_len
);
625 iw
+= IW_EV_POINT_OFF
;
627 #endif /* CONFIG_NET_WIRELESS_RTNETLINK */
629 nskb
= nlmsg_new(if_nlmsg_size(iw_buf_len
), GFP_KERNEL
);
635 err
= rtnl_fill_ifinfo(nskb
, dev
, iw
, iw_buf_len
, RTM_NEWLINK
,
636 NETLINK_CB(skb
).pid
, nlh
->nlmsg_seq
, 0, 0);
638 /* -EMSGSIZE implies BUG in if_nlmsg_size */
639 WARN_ON(err
== -EMSGSIZE
);
643 err
= rtnl_unicast(nskb
, NETLINK_CB(skb
).pid
);
651 static int rtnl_dump_all(struct sk_buff
*skb
, struct netlink_callback
*cb
)
654 int s_idx
= cb
->family
;
658 for (idx
=1; idx
<NPROTO
; idx
++) {
659 int type
= cb
->nlh
->nlmsg_type
-RTM_BASE
;
660 if (idx
< s_idx
|| idx
== PF_PACKET
)
662 if (rtnetlink_links
[idx
] == NULL
||
663 rtnetlink_links
[idx
][type
].dumpit
== NULL
)
666 memset(&cb
->args
[0], 0, sizeof(cb
->args
));
667 if (rtnetlink_links
[idx
][type
].dumpit(skb
, cb
))
675 void rtmsg_ifinfo(int type
, struct net_device
*dev
, unsigned change
)
680 skb
= nlmsg_new(if_nlmsg_size(0), GFP_KERNEL
);
684 err
= rtnl_fill_ifinfo(skb
, dev
, NULL
, 0, type
, 0, 0, change
, 0);
686 /* -EMSGSIZE implies BUG in if_nlmsg_size() */
687 WARN_ON(err
== -EMSGSIZE
);
691 err
= rtnl_notify(skb
, 0, RTNLGRP_LINK
, NULL
, GFP_KERNEL
);
694 rtnl_set_sk_err(RTNLGRP_LINK
, err
);
697 /* Protected by RTNL sempahore. */
698 static struct rtattr
**rta_buf
;
699 static int rtattr_max
;
701 /* Process one rtnetlink message. */
703 static __inline__
int
704 rtnetlink_rcv_msg(struct sk_buff
*skb
, struct nlmsghdr
*nlh
, int *errp
)
706 struct rtnetlink_link
*link
;
707 struct rtnetlink_link
*link_tab
;
714 /* Only requests are handled by kernel now */
715 if (!(nlh
->nlmsg_flags
&NLM_F_REQUEST
))
718 type
= nlh
->nlmsg_type
;
720 /* A control message: ignore them */
724 /* Unknown message: reply with EINVAL */
730 /* All the messages must have at least 1 byte length */
731 if (nlh
->nlmsg_len
< NLMSG_LENGTH(sizeof(struct rtgenmsg
)))
734 family
= ((struct rtgenmsg
*)NLMSG_DATA(nlh
))->rtgen_family
;
735 if (family
>= NPROTO
) {
736 *errp
= -EAFNOSUPPORT
;
740 link_tab
= rtnetlink_links
[family
];
741 if (link_tab
== NULL
)
742 link_tab
= rtnetlink_links
[PF_UNSPEC
];
743 link
= &link_tab
[type
];
748 if (kind
!= 2 && security_netlink_recv(skb
, CAP_NET_ADMIN
)) {
753 if (kind
== 2 && nlh
->nlmsg_flags
&NLM_F_DUMP
) {
754 if (link
->dumpit
== NULL
)
755 link
= &(rtnetlink_links
[PF_UNSPEC
][type
]);
757 if (link
->dumpit
== NULL
)
760 if ((*errp
= netlink_dump_start(rtnl
, skb
, nlh
,
761 link
->dumpit
, NULL
)) != 0) {
765 netlink_queue_skip(nlh
, skb
);
769 memset(rta_buf
, 0, (rtattr_max
* sizeof(struct rtattr
*)));
771 min_len
= rtm_min
[sz_idx
];
772 if (nlh
->nlmsg_len
< min_len
)
775 if (nlh
->nlmsg_len
> min_len
) {
776 int attrlen
= nlh
->nlmsg_len
- NLMSG_ALIGN(min_len
);
777 struct rtattr
*attr
= (void*)nlh
+ NLMSG_ALIGN(min_len
);
779 while (RTA_OK(attr
, attrlen
)) {
780 unsigned flavor
= attr
->rta_type
;
782 if (flavor
> rta_max
[sz_idx
])
784 rta_buf
[flavor
-1] = attr
;
786 attr
= RTA_NEXT(attr
, attrlen
);
790 if (link
->doit
== NULL
)
791 link
= &(rtnetlink_links
[PF_UNSPEC
][type
]);
792 if (link
->doit
== NULL
)
794 err
= link
->doit(skb
, nlh
, (void *)&rta_buf
[0]);
804 static void rtnetlink_rcv(struct sock
*sk
, int len
)
806 unsigned int qlen
= 0;
809 mutex_lock(&rtnl_mutex
);
810 netlink_run_queue(sk
, &qlen
, &rtnetlink_rcv_msg
);
811 mutex_unlock(&rtnl_mutex
);
817 static struct rtnetlink_link link_rtnetlink_table
[RTM_NR_MSGTYPES
] =
819 [RTM_GETLINK
- RTM_BASE
] = { .doit
= rtnl_getlink
,
820 .dumpit
= rtnl_dump_ifinfo
},
821 [RTM_SETLINK
- RTM_BASE
] = { .doit
= rtnl_setlink
},
822 [RTM_GETADDR
- RTM_BASE
] = { .dumpit
= rtnl_dump_all
},
823 [RTM_GETROUTE
- RTM_BASE
] = { .dumpit
= rtnl_dump_all
},
824 [RTM_NEWNEIGH
- RTM_BASE
] = { .doit
= neigh_add
},
825 [RTM_DELNEIGH
- RTM_BASE
] = { .doit
= neigh_delete
},
826 [RTM_GETNEIGH
- RTM_BASE
] = { .dumpit
= neigh_dump_info
},
827 #ifdef CONFIG_FIB_RULES
828 [RTM_NEWRULE
- RTM_BASE
] = { .doit
= fib_nl_newrule
},
829 [RTM_DELRULE
- RTM_BASE
] = { .doit
= fib_nl_delrule
},
831 [RTM_GETRULE
- RTM_BASE
] = { .dumpit
= rtnl_dump_all
},
832 [RTM_GETNEIGHTBL
- RTM_BASE
] = { .dumpit
= neightbl_dump_info
},
833 [RTM_SETNEIGHTBL
- RTM_BASE
] = { .doit
= neightbl_set
},
836 static int rtnetlink_event(struct notifier_block
*this, unsigned long event
, void *ptr
)
838 struct net_device
*dev
= ptr
;
840 case NETDEV_UNREGISTER
:
841 rtmsg_ifinfo(RTM_DELLINK
, dev
, ~0U);
843 case NETDEV_REGISTER
:
844 rtmsg_ifinfo(RTM_NEWLINK
, dev
, ~0U);
848 rtmsg_ifinfo(RTM_NEWLINK
, dev
, IFF_UP
|IFF_RUNNING
);
851 case NETDEV_GOING_DOWN
:
854 rtmsg_ifinfo(RTM_NEWLINK
, dev
, 0);
860 static struct notifier_block rtnetlink_dev_notifier
= {
861 .notifier_call
= rtnetlink_event
,
864 void __init
rtnetlink_init(void)
869 for (i
= 0; i
< ARRAY_SIZE(rta_max
); i
++)
870 if (rta_max
[i
] > rtattr_max
)
871 rtattr_max
= rta_max
[i
];
872 rta_buf
= kmalloc(rtattr_max
* sizeof(struct rtattr
*), GFP_KERNEL
);
874 panic("rtnetlink_init: cannot allocate rta_buf\n");
876 rtnl
= netlink_kernel_create(NETLINK_ROUTE
, RTNLGRP_MAX
, rtnetlink_rcv
,
879 panic("rtnetlink_init: cannot initialize rtnetlink\n");
880 netlink_set_nonroot(NETLINK_ROUTE
, NL_NONROOT_RECV
);
881 register_netdevice_notifier(&rtnetlink_dev_notifier
);
882 rtnetlink_links
[PF_UNSPEC
] = link_rtnetlink_table
;
883 rtnetlink_links
[PF_PACKET
] = link_rtnetlink_table
;
886 EXPORT_SYMBOL(__rta_fill
);
887 EXPORT_SYMBOL(rtattr_strlcpy
);
888 EXPORT_SYMBOL(rtattr_parse
);
889 EXPORT_SYMBOL(rtnetlink_links
);
890 EXPORT_SYMBOL(rtnetlink_put_metrics
);
891 EXPORT_SYMBOL(rtnl_lock
);
892 EXPORT_SYMBOL(rtnl_trylock
);
893 EXPORT_SYMBOL(rtnl_unlock
);
894 EXPORT_SYMBOL(rtnl_unicast
);
895 EXPORT_SYMBOL(rtnl_notify
);
896 EXPORT_SYMBOL(rtnl_set_sk_err
);