2 * Copyright (c) 2015-2016
3 * Alexander V. Chernikov <melifaro@FreeBSD.org>
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
13 * 3. Neither the name of the University nor the names of its contributors
14 * may be used to endorse or promote products derived from this software
15 * without specific prior written permission.
17 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
21 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
32 #ifndef _NET_ROUTE_VAR_H_
33 #define _NET_ROUTE_VAR_H_
36 #include <net/radix.h>
39 #include <sys/epoch.h>
40 #include <netinet/in.h> /* struct sockaddr_in */
41 #include <sys/counter.h>
44 typedef int rnh_preadd_entry_f_t(u_int fibnum, const struct sockaddr *addr,
45 const struct sockaddr *mask, struct nhop_object *nh);
48 struct radix_head head;
49 rn_matchaddr_f_t *rnh_matchaddr; /* longest match for sockaddr */
50 rn_addaddr_f_t *rnh_addaddr; /* add based on sockaddr*/
51 rn_deladdr_f_t *rnh_deladdr; /* remove based on sockaddr */
52 rn_lookup_f_t *rnh_lookup; /* exact match for sockaddr */
53 rn_walktree_t *rnh_walktree; /* traverse tree */
54 rn_walktree_from_t *rnh_walktree_from; /* traverse tree below a */
55 rnh_preadd_entry_f_t *rnh_preadd; /* hook to alter record prior to insertion */
56 rt_gen_t rnh_gen; /* generation counter */
57 int rnh_multipath; /* multipath capable ? */
58 struct radix_node rnh_nodes[3]; /* empty tree for common case */
59 struct rmlock rib_lock; /* config/data path lock */
60 struct radix_mask_head rmhead; /* masks radix head */
61 struct vnet *rib_vnet; /* vnet pointer */
62 int rib_family; /* AF of the rtable */
63 u_int rib_fibnum; /* fib number */
64 struct callout expire_callout; /* Callout for expiring dynamic routes */
65 time_t next_expire; /* Next expire run ts */
66 struct nh_control *nh_control; /* nexthop subsystem data */
67 CK_STAILQ_HEAD(, rib_subscription) rnh_subscribers;/* notification subscribers */
70 #define RIB_RLOCK_TRACKER struct rm_priotracker _rib_tracker
71 #define RIB_LOCK_INIT(rh) rm_init(&(rh)->rib_lock, "rib head lock")
72 #define RIB_LOCK_DESTROY(rh) rm_destroy(&(rh)->rib_lock)
73 #define RIB_RLOCK(rh) rm_rlock(&(rh)->rib_lock, &_rib_tracker)
74 #define RIB_RUNLOCK(rh) rm_runlock(&(rh)->rib_lock, &_rib_tracker)
75 #define RIB_WLOCK(rh) rm_wlock(&(rh)->rib_lock)
76 #define RIB_WUNLOCK(rh) rm_wunlock(&(rh)->rib_lock)
77 #define RIB_LOCK_ASSERT(rh) rm_assert(&(rh)->rib_lock, RA_LOCKED)
78 #define RIB_WLOCK_ASSERT(rh) rm_assert(&(rh)->rib_lock, RA_WLOCKED)
81 #define RIB_MAX_RETRIES 3
83 /* Macro for verifying fields in af-specific 'struct route' structures */
84 #define CHK_STRUCT_FIELD_GENERIC(_s1, _f1, _s2, _f2) \
85 _Static_assert(sizeof(((_s1 *)0)->_f1) == sizeof(((_s2 *)0)->_f2), \
86 "Fields " #_f1 " and " #_f2 " size differs"); \
87 _Static_assert(__offsetof(_s1, _f1) == __offsetof(_s2, _f2), \
88 "Fields " #_f1 " and " #_f2 " offset differs");
90 #define _CHK_ROUTE_FIELD(_route_new, _field) \
91 CHK_STRUCT_FIELD_GENERIC(struct route, _field, _route_new, _field)
93 #define CHK_STRUCT_ROUTE_FIELDS(_route_new) \
94 _CHK_ROUTE_FIELD(_route_new, ro_nh) \
95 _CHK_ROUTE_FIELD(_route_new, ro_lle) \
96 _CHK_ROUTE_FIELD(_route_new, ro_prepend)\
97 _CHK_ROUTE_FIELD(_route_new, ro_plen) \
98 _CHK_ROUTE_FIELD(_route_new, ro_flags) \
99 _CHK_ROUTE_FIELD(_route_new, ro_mtu) \
100 _CHK_ROUTE_FIELD(_route_new, spare)
102 #define CHK_STRUCT_ROUTE_COMPAT(_ro_new, _dst_new) \
103 CHK_STRUCT_ROUTE_FIELDS(_ro_new); \
104 _Static_assert(__offsetof(struct route, ro_dst) == __offsetof(_ro_new, _dst_new),\
105 "ro_dst and " #_dst_new " are at different offset")
107 struct rib_head *rt_tables_get_rnh(int fib, int family);
108 void rt_mpath_init_rnh(struct rib_head *rnh);
109 int rt_getifa_fib(struct rt_addrinfo *info, u_int fibnum);
110 void rt_setmetrics(const struct rt_addrinfo *info, struct rtentry *rt);
112 struct radix_node *rt_mpath_unlink(struct rib_head *rnh,
113 struct rt_addrinfo *info, struct rtentry *rto, int *perror);
116 int add_route(struct rib_head *rnh, struct rt_addrinfo *info,
117 struct rib_cmd_info *rc);
118 int del_route(struct rib_head *rnh, struct rt_addrinfo *info,
119 struct rib_cmd_info *rc);
120 int change_route(struct rib_head *, struct rt_addrinfo *,
121 struct rib_cmd_info *rc);
123 VNET_PCPUSTAT_DECLARE(struct rtstat, rtstat);
124 #define RTSTAT_ADD(name, val) \
125 VNET_PCPUSTAT_ADD(struct rtstat, rtstat, name, (val))
126 #define RTSTAT_INC(name) RTSTAT_ADD(name, 1)
130 * Convert a 'struct radix_node *' to a 'struct rtentry *'.
131 * The operation can be done safely (in this code) because a
132 * 'struct rtentry' starts with two 'struct radix_node''s, the first
133 * one representing leaf nodes in the routing tree, which is
134 * what the code in radix.c passes us as a 'struct radix_node'.
136 * But because there are a lot of assumptions in this conversion,
137 * do not cast explicitly, but always use the macro below.
139 #define RNTORT(p) ((struct rtentry *)(p))
142 struct radix_node rt_nodes[2]; /* tree glue, and other values */
144 * XXX struct rtentry must begin with a struct radix_node (or two!)
145 * because the code does some casts of a 'struct radix_node *'
146 * to a 'struct rtentry *'
148 #define rt_key(r) (*((struct sockaddr **)(&(r)->rt_nodes->rn_key)))
149 #define rt_mask(r) (*((struct sockaddr **)(&(r)->rt_nodes->rn_mask)))
150 #define rt_key_const(r) (*((const struct sockaddr * const *)(&(r)->rt_nodes->rn_key)))
151 #define rt_mask_const(r) (*((const struct sockaddr * const *)(&(r)->rt_nodes->rn_mask)))
154 * 2 radix_node structurs above consists of 2x6 pointers, leaving
155 * 4 pointers (32 bytes) of the second cache line on amd64.
158 struct nhop_object *rt_nhop; /* nexthop data */
161 * Destination address storage.
162 * sizeof(struct sockaddr_in6) == 28, however
163 * the dataplane-relevant part (e.g. address) lies
164 * at offset 8..24, making the address not crossing
165 * cacheline boundary.
167 struct sockaddr_in rt_dst4;
168 struct sockaddr_in6 rt_dst6;
169 struct sockaddr rt_dst;
173 int rt_flags; /* up/down?, host/net */
174 u_long rt_weight; /* absolute weight */
175 u_long rt_expire; /* lifetime for route, e.g. redirect */
176 #define rt_endzero rt_mtx
177 struct mtx rt_mtx; /* mutex for routing entry */
178 struct rtentry *rt_chain; /* pointer to next rtentry to delete */
179 struct epoch_context rt_epoch_ctx; /* net epoch tracker */
182 #define RT_LOCK_INIT(_rt) \
183 mtx_init(&(_rt)->rt_mtx, "rtentry", NULL, MTX_DEF | MTX_DUPOK | MTX_NEW)
184 #define RT_LOCK(_rt) mtx_lock(&(_rt)->rt_mtx)
185 #define RT_UNLOCK(_rt) mtx_unlock(&(_rt)->rt_mtx)
186 #define RT_LOCK_DESTROY(_rt) mtx_destroy(&(_rt)->rt_mtx)
187 #define RT_LOCK_ASSERT(_rt) mtx_assert(&(_rt)->rt_mtx, MA_OWNED)
188 #define RT_UNLOCK_COND(_rt) do { \
189 if (mtx_owned(&(_rt)->rt_mtx)) \
190 mtx_unlock(&(_rt)->rt_mtx); \
194 * With the split between the routing entry and the nexthop,
195 * rt_flags has to be split between these 2 entries. As rtentry
196 * mostly contains prefix data and is thought to be generic enough
197 * so one can transparently change the nexthop pointer w/o requiring
198 * any other rtentry changes, most of rt_flags shifts to the particular nexthop.
201 * RTF_UP: rtentry, as an indication that it is linked.
202 * RTF_HOST: rtentry, nhop. The latter indication is needed for the datapath
203 * RTF_DYNAMIC: nhop, to make rtentry generic.
204 * RTF_MODIFIED: nhop, to make rtentry generic. (legacy)
205 * -- "native" path (nhop) properties:
206 * RTF_GATEWAY, RTF_STATIC, RTF_PROTO1, RTF_PROTO2, RTF_PROTO3, RTF_FIXEDMTU,
207 * RTF_PINNED, RTF_REJECT, RTF_BLACKHOLE, RTF_BROADCAST
210 /* Nexthop rt flags mask */
211 #define NHOP_RT_FLAG_MASK (RTF_GATEWAY | RTF_HOST | RTF_REJECT | RTF_DYNAMIC | \
212 RTF_MODIFIED | RTF_STATIC | RTF_BLACKHOLE | RTF_PROTO1 | RTF_PROTO2 | \
213 RTF_PROTO3 | RTF_FIXEDMTU | RTF_PINNED | RTF_BROADCAST)
215 /* rtentry rt flag mask */
216 #define RTE_RT_FLAG_MASK (RTF_UP | RTF_HOST)
218 /* Nexthop selection */
219 #define _NH2MP(_nh) ((struct nhgrp_object *)(_nh))
220 #define _SELECT_NHOP(_nh, _flowid) \
221 (_NH2MP(_nh))->nhops[(_flowid) % (_NH2MP(_nh))->mp_size]
222 #define _RT_SELECT_NHOP(_nh, _flowid) \
223 ((!NH_IS_MULTIPATH(_nh)) ? (_nh) : _SELECT_NHOP(_nh, _flowid))
224 #define RT_SELECT_NHOP(_rt, _flowid) _RT_SELECT_NHOP((_rt)->rt_nhop, _flowid)
226 /* rte<>nhop translation */
227 static inline uint16_t
228 fib_rte_to_nh_flags(int rt_flags)
232 res = (rt_flags & RTF_REJECT) ? NHF_REJECT : 0;
233 res |= (rt_flags & RTF_HOST) ? NHF_HOST : 0;
234 res |= (rt_flags & RTF_BLACKHOLE) ? NHF_BLACKHOLE : 0;
235 res |= (rt_flags & (RTF_DYNAMIC|RTF_MODIFIED)) ? NHF_REDIRECT : 0;
236 res |= (rt_flags & RTF_BROADCAST) ? NHF_BROADCAST : 0;
237 res |= (rt_flags & RTF_GATEWAY) ? NHF_GATEWAY : 0;
242 void tmproutes_update(struct rib_head *rnh, struct rtentry *rt);
243 void tmproutes_init(struct rib_head *rh);
244 void tmproutes_destroy(struct rib_head *rh);