Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 1 | /* |
| 2 | * replication.c : packet replication |
| 3 | * |
| 4 | * Copyright (c) 2013 Cisco and/or its affiliates. |
| 5 | * Licensed under the Apache License, Version 2.0 (the "License"); |
| 6 | * you may not use this file except in compliance with the License. |
| 7 | * You may obtain a copy of the License at: |
| 8 | * |
| 9 | * http://www.apache.org/licenses/LICENSE-2.0 |
| 10 | * |
| 11 | * Unless required by applicable law or agreed to in writing, software |
| 12 | * distributed under the License is distributed on an "AS IS" BASIS, |
| 13 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| 14 | * See the License for the specific language governing permissions and |
| 15 | * limitations under the License. |
| 16 | */ |
| 17 | |
| 18 | #include <vlib/vlib.h> |
| 19 | #include <vnet/vnet.h> |
| 20 | #include <vppinfra/error.h> |
| 21 | #include <vnet/ip/ip4_packet.h> |
| 22 | #include <vnet/replication.h> |
| 23 | |
| 24 | |
| 25 | replication_main_t replication_main; |
| 26 | |
| 27 | |
| 28 | replication_context_t * |
| 29 | replication_prep (vlib_main_t * vm, |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 30 | vlib_buffer_t * b0, u32 recycle_node_index, u32 l2_packet) |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 31 | { |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 32 | replication_main_t *rm = &replication_main; |
| 33 | replication_context_t *ctx; |
Damjan Marion | 586afd7 | 2017-04-05 19:18:20 +0200 | [diff] [blame] | 34 | uword thread_index = vm->thread_index; |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 35 | ip4_header_t *ip; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 36 | u32 ctx_id; |
| 37 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 38 | /* Allocate a context, reserve context 0 */ |
Damjan Marion | 586afd7 | 2017-04-05 19:18:20 +0200 | [diff] [blame] | 39 | if (PREDICT_FALSE (rm->contexts[thread_index] == 0)) |
| 40 | pool_get_aligned (rm->contexts[thread_index], ctx, CLIB_CACHE_LINE_BYTES); |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 41 | |
Damjan Marion | 586afd7 | 2017-04-05 19:18:20 +0200 | [diff] [blame] | 42 | pool_get_aligned (rm->contexts[thread_index], ctx, CLIB_CACHE_LINE_BYTES); |
| 43 | ctx_id = ctx - rm->contexts[thread_index]; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 44 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 45 | /* Save state from vlib buffer */ |
Damjan Marion | 072401e | 2017-07-13 18:53:27 +0200 | [diff] [blame] | 46 | ctx->saved_free_list_index = vlib_buffer_get_free_list_index (b0); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 47 | ctx->current_data = b0->current_data; |
Steve Shin | 49806fe | 2017-11-13 15:46:35 -0800 | [diff] [blame] | 48 | ctx->flags = b0->flags & VNET_BUFFER_FLAGS_VLAN_BITS; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 49 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 50 | /* Set up vlib buffer hooks */ |
Dave Barach | b5adaea | 2016-06-17 14:09:56 -0400 | [diff] [blame] | 51 | b0->recycle_count = ctx_id; |
Damjan Marion | 072401e | 2017-07-13 18:53:27 +0200 | [diff] [blame] | 52 | vlib_buffer_set_free_list_index (b0, rm->recycle_list_index); |
Dave Barach | b5adaea | 2016-06-17 14:09:56 -0400 | [diff] [blame] | 53 | b0->flags |= VLIB_BUFFER_RECYCLE; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 54 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 55 | /* Save feature state */ |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 56 | ctx->recycle_node_index = recycle_node_index; |
| 57 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 58 | /* Save vnet state */ |
| 59 | clib_memcpy (ctx->vnet_buffer, vnet_buffer (b0), |
| 60 | sizeof (vnet_buffer_opaque_t)); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 61 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 62 | /* Save packet contents */ |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 63 | ctx->l2_packet = l2_packet; |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 64 | ip = (ip4_header_t *) vlib_buffer_get_current (b0); |
| 65 | if (l2_packet) |
| 66 | { |
| 67 | /* Save ethernet header */ |
| 68 | ctx->l2_header[0] = ((u64 *) ip)[0]; |
| 69 | ctx->l2_header[1] = ((u64 *) ip)[1]; |
| 70 | ctx->l2_header[2] = ((u64 *) ip)[2]; |
| 71 | /* set ip to the true ip header */ |
| 72 | ip = (ip4_header_t *) (((u8 *) ip) + vnet_buffer (b0)->l2.l2_len); |
| 73 | } |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 74 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 75 | /* |
| 76 | * Copy L3 fields. |
| 77 | * We need to save TOS for ip4 and ip6 packets. |
| 78 | * Fortunately the TOS field is |
| 79 | * in the first two bytes of both the ip4 and ip6 headers. |
| 80 | */ |
| 81 | ctx->ip_tos = *((u16 *) (ip)); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 82 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 83 | /* |
| 84 | * Save the ip4 checksum as well. We just blindly save the corresponding two |
| 85 | * bytes even for ip6 packets. |
| 86 | */ |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 87 | ctx->ip4_checksum = ip->checksum; |
| 88 | |
| 89 | return ctx; |
| 90 | } |
| 91 | |
| 92 | |
| 93 | replication_context_t * |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 94 | replication_recycle (vlib_main_t * vm, vlib_buffer_t * b0, u32 is_last) |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 95 | { |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 96 | replication_main_t *rm = &replication_main; |
| 97 | replication_context_t *ctx; |
Damjan Marion | 586afd7 | 2017-04-05 19:18:20 +0200 | [diff] [blame] | 98 | uword thread_index = vm->thread_index; |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 99 | ip4_header_t *ip; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 100 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 101 | /* Get access to the replication context */ |
Damjan Marion | 586afd7 | 2017-04-05 19:18:20 +0200 | [diff] [blame] | 102 | ctx = pool_elt_at_index (rm->contexts[thread_index], b0->recycle_count); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 103 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 104 | /* Restore vnet buffer state */ |
| 105 | clib_memcpy (vnet_buffer (b0), ctx->vnet_buffer, |
| 106 | sizeof (vnet_buffer_opaque_t)); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 107 | |
Steve Shin | 49806fe | 2017-11-13 15:46:35 -0800 | [diff] [blame] | 108 | /* Restore the vlan flags */ |
| 109 | b0->flags &= ~VNET_BUFFER_FLAGS_VLAN_BITS; |
| 110 | b0->flags |= ctx->flags; |
| 111 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 112 | /* Restore the packet start (current_data) and length */ |
| 113 | vlib_buffer_advance (b0, ctx->current_data - b0->current_data); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 114 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 115 | /* Restore packet contents */ |
| 116 | ip = (ip4_header_t *) vlib_buffer_get_current (b0); |
| 117 | if (ctx->l2_packet) |
| 118 | { |
| 119 | /* Restore ethernet header */ |
| 120 | ((u64 *) ip)[0] = ctx->l2_header[0]; |
| 121 | ((u64 *) ip)[1] = ctx->l2_header[1]; |
| 122 | ((u64 *) ip)[2] = ctx->l2_header[2]; |
| 123 | /* set ip to the true ip header */ |
| 124 | ip = (ip4_header_t *) (((u8 *) ip) + vnet_buffer (b0)->l2.l2_len); |
| 125 | } |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 126 | |
| 127 | // Restore L3 fields |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 128 | *((u16 *) (ip)) = ctx->ip_tos; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 129 | ip->checksum = ctx->ip4_checksum; |
| 130 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 131 | if (is_last) |
| 132 | { |
| 133 | /* |
| 134 | * This is the last replication in the list. |
| 135 | * Restore original buffer free functionality. |
| 136 | */ |
Damjan Marion | 072401e | 2017-07-13 18:53:27 +0200 | [diff] [blame] | 137 | vlib_buffer_set_free_list_index (b0, ctx->saved_free_list_index); |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 138 | b0->flags &= ~VLIB_BUFFER_RECYCLE; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 139 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 140 | /* Free context back to its pool */ |
Damjan Marion | 586afd7 | 2017-04-05 19:18:20 +0200 | [diff] [blame] | 141 | pool_put (rm->contexts[thread_index], ctx); |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 142 | } |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 143 | |
| 144 | return ctx; |
| 145 | } |
| 146 | |
| 147 | |
| 148 | |
| 149 | /* |
| 150 | * fish pkts back from the recycle queue/freelist |
| 151 | * un-flatten the context chains |
| 152 | */ |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 153 | static void |
| 154 | replication_recycle_callback (vlib_main_t * vm, vlib_buffer_free_list_t * fl) |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 155 | { |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 156 | vlib_frame_t *f = 0; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 157 | u32 n_left_from; |
| 158 | u32 n_left_to_next = 0; |
| 159 | u32 n_this_frame = 0; |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 160 | u32 *from; |
| 161 | u32 *to_next = 0; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 162 | u32 bi0, pi0; |
| 163 | vlib_buffer_t *b0; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 164 | int i; |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 165 | replication_main_t *rm = &replication_main; |
| 166 | replication_context_t *ctx; |
| 167 | u32 feature_node_index = 0; |
Damjan Marion | 586afd7 | 2017-04-05 19:18:20 +0200 | [diff] [blame] | 168 | uword thread_index = vm->thread_index; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 169 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 170 | /* |
| 171 | * All buffers in the list are destined to the same recycle node. |
| 172 | * Pull the recycle node index from the first buffer. |
| 173 | * Note: this could be sped up if the node index were stuffed into |
| 174 | * the freelist itself. |
| 175 | */ |
Damjan Marion | bd69a5f | 2017-02-05 23:44:42 +0100 | [diff] [blame] | 176 | if (vec_len (fl->buffers) > 0) |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 177 | { |
Damjan Marion | bd69a5f | 2017-02-05 23:44:42 +0100 | [diff] [blame] | 178 | bi0 = fl->buffers[0]; |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 179 | b0 = vlib_get_buffer (vm, bi0); |
Damjan Marion | 586afd7 | 2017-04-05 19:18:20 +0200 | [diff] [blame] | 180 | ctx = pool_elt_at_index (rm->contexts[thread_index], b0->recycle_count); |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 181 | feature_node_index = ctx->recycle_node_index; |
| 182 | } |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 183 | |
Damjan Marion | bd69a5f | 2017-02-05 23:44:42 +0100 | [diff] [blame] | 184 | /* buffers */ |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 185 | for (i = 0; i < 2; i++) |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 186 | { |
| 187 | if (i == 0) |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 188 | { |
Damjan Marion | bd69a5f | 2017-02-05 23:44:42 +0100 | [diff] [blame] | 189 | from = fl->buffers; |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 190 | n_left_from = vec_len (from); |
| 191 | } |
| 192 | |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 193 | while (n_left_from > 0) |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 194 | { |
| 195 | if (PREDICT_FALSE (n_left_to_next == 0)) |
| 196 | { |
| 197 | if (f) |
| 198 | { |
| 199 | f->n_vectors = n_this_frame; |
| 200 | vlib_put_frame_to_node (vm, feature_node_index, f); |
| 201 | } |
| 202 | |
| 203 | f = vlib_get_frame_to_node (vm, feature_node_index); |
| 204 | to_next = vlib_frame_vector_args (f); |
| 205 | n_left_to_next = VLIB_FRAME_SIZE; |
| 206 | n_this_frame = 0; |
| 207 | } |
| 208 | |
| 209 | bi0 = from[0]; |
| 210 | if (PREDICT_TRUE (n_left_from > 1)) |
| 211 | { |
| 212 | pi0 = from[1]; |
| 213 | vlib_prefetch_buffer_with_index (vm, pi0, LOAD); |
| 214 | } |
John Lo | 17f17cb | 2016-05-31 16:25:13 -0400 | [diff] [blame] | 215 | |
| 216 | b0 = vlib_get_buffer (vm, bi0); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 217 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 218 | /* Mark that this buffer was just recycled */ |
| 219 | b0->flags |= VLIB_BUFFER_IS_RECYCLED; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 220 | |
Neale Ranns | b80c536 | 2016-10-08 13:03:40 +0100 | [diff] [blame] | 221 | #if (CLIB_DEBUG > 0) |
Damjan Marion | 04a7f05 | 2017-07-10 15:06:17 +0200 | [diff] [blame] | 222 | if (vm->buffer_main->callbacks_registered == 0) |
Steven | 899a84b | 2018-01-29 20:09:09 -0800 | [diff] [blame] | 223 | vlib_buffer_set_known_state (bi0, VLIB_BUFFER_KNOWN_ALLOCATED); |
Neale Ranns | b80c536 | 2016-10-08 13:03:40 +0100 | [diff] [blame] | 224 | #endif |
| 225 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 226 | /* If buffer is traced, mark frame as traced */ |
| 227 | if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) |
| 228 | f->flags |= VLIB_FRAME_TRACE; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 229 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 230 | to_next[0] = bi0; |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 231 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 232 | from++; |
| 233 | to_next++; |
| 234 | n_this_frame++; |
| 235 | n_left_to_next--; |
| 236 | n_left_from--; |
| 237 | } |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 238 | } |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 239 | |
Damjan Marion | bd69a5f | 2017-02-05 23:44:42 +0100 | [diff] [blame] | 240 | vec_reset_length (fl->buffers); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 241 | |
| 242 | if (f) |
| 243 | { |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 244 | ASSERT (n_this_frame); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 245 | f->n_vectors = n_this_frame; |
| 246 | vlib_put_frame_to_node (vm, feature_node_index, f); |
| 247 | } |
| 248 | } |
| 249 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 250 | clib_error_t * |
| 251 | replication_init (vlib_main_t * vm) |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 252 | { |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 253 | replication_main_t *rm = &replication_main; |
| 254 | vlib_buffer_main_t *bm = vm->buffer_main; |
| 255 | vlib_buffer_free_list_t *fl; |
| 256 | __attribute__ ((unused)) replication_context_t *ctx; |
| 257 | vlib_thread_main_t *tm = vlib_get_thread_main (); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 258 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 259 | rm->vlib_main = vm; |
| 260 | rm->vnet_main = vnet_get_main (); |
| 261 | rm->recycle_list_index = |
| 262 | vlib_buffer_create_free_list (vm, 1024 /* fictional */ , |
| 263 | "replication-recycle"); |
| 264 | |
| 265 | fl = pool_elt_at_index (bm->buffer_free_list_pool, rm->recycle_list_index); |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 266 | |
| 267 | fl->buffers_added_to_freelist_function = replication_recycle_callback; |
| 268 | |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 269 | /* Verify the replication context is the expected size */ |
| 270 | ASSERT (sizeof (replication_context_t) == 128); /* 2 cache lines */ |
Ed Warnicke | cb9cada | 2015-12-08 15:45:58 -0700 | [diff] [blame] | 271 | |
| 272 | vec_validate (rm->contexts, tm->n_vlib_mains - 1); |
| 273 | return 0; |
| 274 | } |
| 275 | |
| 276 | VLIB_INIT_FUNCTION (replication_init); |
Dave Barach | ba868bb | 2016-08-08 09:51:21 -0400 | [diff] [blame] | 277 | |
| 278 | /* |
| 279 | * fd.io coding-style-patch-verification: ON |
| 280 | * |
| 281 | * Local Variables: |
| 282 | * eval: (c-set-style "gnu") |
| 283 | * End: |
| 284 | */ |