blob: 22682e4eee4f869facd310bbc8e66fb5b0e52d1c [file] [log] [blame]
Neale Ranns948e00f2016-10-20 13:39:34 +01001/*
2 * Copyright (c) 2016 Cisco and/or its affiliates.
3 * Licensed under the Apache License, Version 2.0 (the "License");
4 * you may not use this file except in compliance with the License.
5 * You may obtain a copy of the License at:
6 *
7 * http://www.apache.org/licenses/LICENSE-2.0
8 *
9 * Unless required by applicable law or agreed to in writing, software
10 * distributed under the License is distributed on an "AS IS" BASIS,
11 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 * See the License for the specific language governing permissions and
13 * limitations under the License.
14 */
15/**
16 * @brief
17 * The data-path object representing dropping the packet
18 */
19
20#include <vnet/dpo/ip_null_dpo.h>
21#include <vnet/ip/ip.h>
22
23/**
24 * @brief A representation of the IP_NULL DPO
25 */
26typedef struct ip_null_dpo_t_
27{
28 /**
29 * @brief The action to take on a packet
30 */
31 ip_null_dpo_action_t ind_action;
32 /**
33 * @brief The next VLIB node
34 */
35 u32 ind_next_index;
36 /**
37 * rate limits
38 */
39} ip_null_dpo_t;
40
41/**
42 * @brief the IP_NULL dpos are shared by all routes, hence they are global.
43 * As the neame implies this is only for IP, hence 2.
44 */
45static ip_null_dpo_t ip_null_dpos[2 * IP_NULL_DPO_ACTION_NUM] = {
46 [0] = {
47 /* proto ip4, no action */
48 .ind_action = IP_NULL_ACTION_NONE,
49 },
50 [1] = {
51 /* proto ip4, action send unreach */
52 .ind_action = IP_NULL_ACTION_SEND_ICMP_UNREACH,
53 },
54 [2] = {
55 /* proto ip4, action send unreach */
56 .ind_action = IP_NULL_ACTION_SEND_ICMP_PROHIBIT,
57 },
58 [3] = {
59 /* proto ip6, no action */
60 .ind_action = IP_NULL_ACTION_NONE,
61 },
62 [4] = {
63 /* proto ip6, action send unreach */
64 .ind_action = IP_NULL_ACTION_SEND_ICMP_UNREACH,
65 },
66 [5] = {
67 /* proto ip6, action send unreach */
68 .ind_action = IP_NULL_ACTION_SEND_ICMP_PROHIBIT,
69 },
70};
71
72/**
73 * @brief Action strings
74 */
75const char *ip_null_action_strings[] = IP_NULL_ACTIONS;
76
77void
78ip_null_dpo_add_and_lock (dpo_proto_t proto,
79 ip_null_dpo_action_t action,
80 dpo_id_t *dpo)
81{
82 int i;
83
84 ASSERT((proto == DPO_PROTO_IP4) ||
85 (proto == DPO_PROTO_IP6));
86 ASSERT(action < IP_NULL_DPO_ACTION_NUM);
87
88 i = (proto == DPO_PROTO_IP4 ? 0 : 1);
89
90 dpo_set(dpo, DPO_IP_NULL, proto, (i*IP_NULL_DPO_ACTION_NUM) + action);
91}
92
93always_inline const ip_null_dpo_t*
94ip_null_dpo_get (index_t indi)
95{
96 return (&ip_null_dpos[indi]);
97}
98
99static void
100ip_null_dpo_lock (dpo_id_t *dpo)
101{
102 /*
103 * not maintaining a lock count on the ip_null, they are const global and
104 * never die.
105 */
106}
107static void
108ip_null_dpo_unlock (dpo_id_t *dpo)
109{
110}
111
112static u8*
113format_ip_null_dpo (u8 *s, va_list *ap)
114{
115 index_t index = va_arg(*ap, index_t);
116 CLIB_UNUSED(u32 indent) = va_arg(*ap, u32);
117 const ip_null_dpo_t *ind;
118 dpo_proto_t proto;
119
120 ind = ip_null_dpo_get(index);
121 proto = (index < IP_NULL_DPO_ACTION_NUM ? DPO_PROTO_IP4 : DPO_PROTO_IP6);
122
123 return (format(s, "%U-null action:%s",
124 format_dpo_proto, proto,
125 ip_null_action_strings[ind->ind_action]));
126}
127
128const static dpo_vft_t ip_null_vft = {
129 .dv_lock = ip_null_dpo_lock,
130 .dv_unlock = ip_null_dpo_unlock,
131 .dv_format = format_ip_null_dpo,
132};
133
134/**
135 * @brief The per-protocol VLIB graph nodes that are assigned to a ip_null
136 * object.
137 *
138 * this means that these graph nodes are ones from which a ip_null is the
139 * parent object in the DPO-graph.
140 */
141const static char* const ip4_null_nodes[] =
142{
143 "ip4-null",
144 NULL,
145};
146const static char* const ip6_null_nodes[] =
147{
148 "ip6-null",
149 NULL,
150};
151
152const static char* const * const ip_null_nodes[DPO_PROTO_NUM] =
153{
154 [DPO_PROTO_IP4] = ip4_null_nodes,
155 [DPO_PROTO_IP6] = ip6_null_nodes,
156};
157
158typedef struct ip_null_dpo_trace_t_
159{
160 index_t ind_index;
161} ip_null_dpo_trace_t;
162
163/**
164 * @brief Exit nodes from a IP_NULL
165 */
166typedef enum ip_null_next_t_
167{
168 IP_NULL_NEXT_DROP,
169 IP_NULL_NEXT_ICMP,
170 IP_NULL_NEXT_NUM,
171} ip_null_next_t;
172
173always_inline uword
174ip_null_dpo_switch (vlib_main_t * vm,
175 vlib_node_runtime_t * node,
176 vlib_frame_t * frame,
177 u8 is_ip4)
178{
179 u32 n_left_from, next_index, *from, *to_next;
180 static f64 time_last_seed_change = -1e100;
181 static u32 hash_seeds[3];
182 static uword hash_bitmap[256 / BITS (uword)];
183 f64 time_now;
184
185 from = vlib_frame_vector_args (frame);
186 n_left_from = frame->n_vectors;
187
188 time_now = vlib_time_now (vm);
189 if (time_now - time_last_seed_change > 1e-1)
190 {
191 uword i;
192 u32 * r = clib_random_buffer_get_data (&vm->random_buffer,
193 sizeof (hash_seeds));
194 for (i = 0; i < ARRAY_LEN (hash_seeds); i++)
195 hash_seeds[i] = r[i];
196
197 /* Mark all hash keys as been not-seen before. */
198 for (i = 0; i < ARRAY_LEN (hash_bitmap); i++)
199 hash_bitmap[i] = 0;
200
201 time_last_seed_change = time_now;
202 }
203
204 next_index = node->cached_next_index;
205
206 while (n_left_from > 0)
207 {
208 u32 n_left_to_next;
209
210 vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
211
212 while (n_left_from > 0 && n_left_to_next > 0)
213 {
214 u32 a0, b0, c0, m0, drop0;
215 vlib_buffer_t *p0;
216 u32 bi0, indi0, next0;
217 const ip_null_dpo_t *ind0;
218 uword bm0;
219
220 bi0 = from[0];
221 to_next[0] = bi0;
222 from += 1;
223 to_next += 1;
224 n_left_from -= 1;
225 n_left_to_next -= 1;
226
227 p0 = vlib_get_buffer (vm, bi0);
228
229 /* lookup dst + src mac */
230 indi0 = vnet_buffer (p0)->ip.adj_index[VLIB_TX];
231 ind0 = ip_null_dpo_get(indi0);
232 next0 = IP_NULL_NEXT_DROP;
233
234 /*
235 * rate limit - don't DoS the sender.
236 */
237 a0 = hash_seeds[0];
238 b0 = hash_seeds[1];
239 c0 = hash_seeds[2];
240
241 if (is_ip4)
242 {
243 ip4_header_t *ip0 = vlib_buffer_get_current (p0);
244
245 a0 ^= ip0->dst_address.data_u32;
246 b0 ^= ip0->src_address.data_u32;
247
248 hash_v3_finalize32 (a0, b0, c0);
249 }
250 else
251 {
252 ip6_header_t *ip0 = vlib_buffer_get_current (p0);
253
254 a0 ^= ip0->dst_address.as_u32[0];
255 b0 ^= ip0->src_address.as_u32[0];
256 c0 ^= ip0->src_address.as_u32[1];
257
258 hash_v3_mix32 (a0, b0, c0);
259
260 a0 ^= ip0->dst_address.as_u32[1];
261 b0 ^= ip0->src_address.as_u32[2];
262 c0 ^= ip0->src_address.as_u32[3];
263
264 hash_v3_finalize32 (a0, b0, c0);
265 }
266
267 c0 &= BITS (hash_bitmap) - 1;
268 c0 = c0 / BITS (uword);
269 m0 = (uword) 1 << (c0 % BITS (uword));
270
271 bm0 = hash_bitmap[c0];
272 drop0 = (bm0 & m0) != 0;
273
274 /* Mark it as seen. */
275 hash_bitmap[c0] = bm0 | m0;
276
277 if (PREDICT_FALSE(!drop0))
278 {
279 if (is_ip4)
280 {
281 /*
282 * There's a trade-off here. This conditinal statement
283 * versus a graph node per-condition. Given the number
284 * expect number of packets to reach a null route is 0
285 * we favour the run-time cost over the graph complexity
286 */
287 if (IP_NULL_ACTION_SEND_ICMP_UNREACH == ind0->ind_action)
288 {
289 next0 = IP_NULL_NEXT_ICMP;
290 icmp4_error_set_vnet_buffer(
291 p0,
292 ICMP4_destination_unreachable,
293 ICMP4_destination_unreachable_destination_unreachable_host,
294 0);
295 }
296 else if (IP_NULL_ACTION_SEND_ICMP_PROHIBIT == ind0->ind_action)
297 {
298 next0 = IP_NULL_NEXT_ICMP;
299 icmp4_error_set_vnet_buffer(
300 p0,
301 ICMP4_destination_unreachable,
302 ICMP4_destination_unreachable_host_administratively_prohibited,
303 0);
304 }
305 }
306 else
307 {
308 if (IP_NULL_ACTION_SEND_ICMP_UNREACH == ind0->ind_action)
309 {
310 next0 = IP_NULL_NEXT_ICMP;
311 icmp6_error_set_vnet_buffer(
312 p0,
313 ICMP6_destination_unreachable,
314 ICMP6_destination_unreachable_no_route_to_destination,
315 0);
316 }
317 else if (IP_NULL_ACTION_SEND_ICMP_PROHIBIT == ind0->ind_action)
318 {
319 next0 = IP_NULL_NEXT_ICMP;
320 icmp6_error_set_vnet_buffer(
321 p0,
322 ICMP6_destination_unreachable,
323 ICMP6_destination_unreachable_destination_administratively_prohibited,
324 0);
325 }
326 }
327 }
328
329 if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED))
330 {
331 ip_null_dpo_trace_t *tr = vlib_add_trace (vm, node, p0,
332 sizeof (*tr));
333 tr->ind_index = indi0;
334 }
335 vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
336 n_left_to_next, bi0, next0);
337 }
338
339 vlib_put_next_frame (vm, node, next_index, n_left_to_next);
340 }
341
342 return frame->n_vectors;
343}
344
345static u8 *
346format_ip_null_dpo_trace (u8 * s, va_list * args)
347{
348 CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
349 CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
350 ip_null_dpo_trace_t *t = va_arg (*args, ip_null_dpo_trace_t *);
351
352 s = format (s, "%U", format_ip_null_dpo, t->ind_index, 0);
353 return s;
354}
355
356static uword
357ip4_null_dpo_switch (vlib_main_t * vm,
358 vlib_node_runtime_t * node,
359 vlib_frame_t * frame)
360{
361 return (ip_null_dpo_switch(vm, node, frame, 1));
362}
363
364/**
365 * @brief
366 */
367VLIB_REGISTER_NODE (ip4_null_dpo_node) = {
368 .function = ip4_null_dpo_switch,
369 .name = "ip4-null",
370 .vector_size = sizeof (u32),
371
372 .format_trace = format_ip_null_dpo_trace,
373 .n_next_nodes = IP_NULL_NEXT_NUM,
374 .next_nodes = {
375 [IP_NULL_NEXT_DROP] = "ip4-drop",
376 [IP_NULL_NEXT_ICMP] = "ip4-icmp-error",
377 },
378};
379
380static uword
381ip6_null_dpo_switch (vlib_main_t * vm,
382 vlib_node_runtime_t * node,
383 vlib_frame_t * frame)
384{
385 return (ip_null_dpo_switch(vm, node, frame, 0));
386}
387
388/**
389 * @brief
390 */
391VLIB_REGISTER_NODE (ip6_null_dpo_node) = {
392 .function = ip6_null_dpo_switch,
393 .name = "ip6-null",
394 .vector_size = sizeof (u32),
395
396 .format_trace = format_ip_null_dpo_trace,
397 .n_next_nodes = IP_NULL_NEXT_NUM,
398 .next_nodes = {
399 [IP_NULL_NEXT_DROP] = "ip6-drop",
400 [IP_NULL_NEXT_ICMP] = "ip6-icmp-error",
401 },
402};
403
404void
405ip_null_dpo_module_init (void)
406{
407 dpo_register(DPO_IP_NULL, &ip_null_vft, ip_null_nodes);
408}