blob: 6e1e1e3f3c8fd20a398369198a1e09179f19f291 [file] [log] [blame]
/*
* Copyright (c) 2018 Cisco and/or its affiliates.
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at:
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/*
* ethernet_node.c: ethernet packet processing
*
* Copyright (c) 2008 Eliot Dresselhaus
*
* Permission is hereby granted, free of charge, to any person obtaining
* a copy of this software and associated documentation files (the
* "Software"), to deal in the Software without restriction, including
* without limitation the rights to use, copy, modify, merge, publish,
* distribute, sublicense, and/or sell copies of the Software, and to
* permit persons to whom the Software is furnished to do so, subject to
* the following conditions:
*
* The above copyright notice and this permission notice shall be
* included in all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*/
#include <vlib/vlib.h>
#include <vnet/pg/pg.h>
#include <vnet/ethernet/ethernet.h>
#include <vnet/ethernet/p2p_ethernet.h>
#include <vnet/devices/pipe/pipe.h>
#include <vppinfra/sparse_vec.h>
#include <vnet/l2/l2_bvi.h>
#define foreach_ethernet_input_next \
_ (PUNT, "error-punt") \
_ (DROP, "error-drop") \
_ (LLC, "llc-input") \
_ (IP4_INPUT, "ip4-input") \
_ (IP4_INPUT_NCS, "ip4-input-no-checksum")
typedef enum
{
#define _(s,n) ETHERNET_INPUT_NEXT_##s,
foreach_ethernet_input_next
#undef _
ETHERNET_INPUT_N_NEXT,
} ethernet_input_next_t;
typedef struct
{
u8 packet_data[32];
u16 frame_flags;
ethernet_input_frame_t frame_data;
} ethernet_input_trace_t;
static u8 *
format_ethernet_input_trace (u8 * s, va_list * va)
{
CLIB_UNUSED (vlib_main_t * vm) = va_arg (*va, vlib_main_t *);
CLIB_UNUSED (vlib_node_t * node) = va_arg (*va, vlib_node_t *);
ethernet_input_trace_t *t = va_arg (*va, ethernet_input_trace_t *);
u32 indent = format_get_indent (s);
if (t->frame_flags)
{
s = format (s, "frame: flags 0x%x", t->frame_flags);
if (t->frame_flags & ETH_INPUT_FRAME_F_SINGLE_SW_IF_IDX)
s = format (s, ", hw-if-index %u, sw-if-index %u",
t->frame_data.hw_if_index, t->frame_data.sw_if_index);
s = format (s, "\n%U", format_white_space, indent);
}
s = format (s, "%U", format_ethernet_header, t->packet_data);
return s;
}
extern vlib_node_registration_t ethernet_input_node;
typedef enum
{
ETHERNET_INPUT_VARIANT_ETHERNET,
ETHERNET_INPUT_VARIANT_ETHERNET_TYPE,
ETHERNET_INPUT_VARIANT_NOT_L2,
} ethernet_input_variant_t;
// Parse the ethernet header to extract vlan tags and innermost ethertype
static_always_inline void
parse_header (ethernet_input_variant_t variant,
vlib_buffer_t * b0,
u16 * type,
u16 * orig_type,
u16 * outer_id, u16 * inner_id, u32 * match_flags)
{
u8 vlan_count;
if (variant == ETHERNET_INPUT_VARIANT_ETHERNET
|| variant == ETHERNET_INPUT_VARIANT_NOT_L2)
{
ethernet_header_t *e0;
e0 = (void *) (b0->data + b0->current_data);
vnet_buffer (b0)->l2_hdr_offset = b0->current_data;
b0->flags |= VNET_BUFFER_F_L2_HDR_OFFSET_VALID;
vlib_buffer_advance (b0, sizeof (e0[0]));
*type = clib_net_to_host_u16 (e0->type);
}
else if (variant == ETHERNET_INPUT_VARIANT_ETHERNET_TYPE)
{
// here when prior node was LLC/SNAP processing
u16 *e0;
e0 = (void *) (b0->data + b0->current_data);
vlib_buffer_advance (b0, sizeof (e0[0]));
*type = clib_net_to_host_u16 (e0[0]);
}
// save for distinguishing between dot1q and dot1ad later
*orig_type = *type;
// default the tags to 0 (used if there is no corresponding tag)
*outer_id = 0;
*inner_id = 0;
*match_flags = SUBINT_CONFIG_VALID | SUBINT_CONFIG_MATCH_0_TAG;
vlan_count = 0;
// check for vlan encaps
if (ethernet_frame_is_tagged (*type))
{
ethernet_vlan_header_t *h0;
u16 tag;
*match_flags = SUBINT_CONFIG_VALID | SUBINT_CONFIG_MATCH_1_TAG;
h0 = (void *) (b0->data + b0->current_data);
tag = clib_net_to_host_u16 (h0->priority_cfi_and_id);
*outer_id = tag & 0xfff;
if (0 == *outer_id)
*match_flags &= ~SUBINT_CONFIG_MATCH_1_TAG;
*type = clib_net_to_host_u16 (h0->type);
vlib_buffer_advance (b0, sizeof (h0[0]));
vlan_count = 1;
if (*type == ETHERNET_TYPE_VLAN)
{
// Double tagged packet
*match_flags = SUBINT_CONFIG_VALID | SUBINT_CONFIG_MATCH_2_TAG;
h0 = (void *) (b0->data + b0->current_data);
tag = clib_net_to_host_u16 (h0->priority_cfi_and_id);
*inner_id = tag & 0xfff;
*type = clib_net_to_host_u16 (h0->type);
vlib_buffer_advance (b0, sizeof (h0[0]));
vlan_count = 2;
if (*type == ETHERNET_TYPE_VLAN)
{
// More than double tagged packet
*match_flags = SUBINT_CONFIG_VALID | SUBINT_CONFIG_MATCH_3_TAG;
vlib_buffer_advance (b0, sizeof (h0[0]));
vlan_count = 3; // "unknown" number, aka, 3-or-more
}
}
}
ethernet_buffer_set_vlan_count (b0, vlan_count);
}
// Determine the subinterface for this packet, given the result of the
// vlan table lookups and vlan header parsing. Check the most specific
// matches first.
static_always_inline void
identify_subint (vnet_hw_interface_t * hi,
vlib_buffer_t * b0,
u32 match_flags,
main_intf_t * main_intf,
vlan_intf_t * vlan_intf,
qinq_intf_t * qinq_intf,
u32 * new_sw_if_index, u8 * error0, u32 * is_l2)
{
u32 matched;
matched = eth_identify_subint (hi, b0, match_flags,
main_intf, vlan_intf, qinq_intf,
new_sw_if_index, error0, is_l2);
if (matched)
{
// Perform L3 my-mac filter
// A unicast packet arriving on an L3 interface must have a dmac matching the interface mac.
// This is required for promiscuous mode, else we will forward packets we aren't supposed to.
if (!(*is_l2))
{
ethernet_header_t *e0;
e0 = (void *) (b0->data + vnet_buffer (b0)->l2_hdr_offset);
if (!(ethernet_address_cast (e0->dst_address)))
{
if (!eth_mac_equal ((u8 *) e0, hi->hw_address))
{
*error0 = ETHERNET_ERROR_L3_MAC_MISMATCH;
}
}
}
// Check for down subinterface
*error0 = (*new_sw_if_index) != ~0 ? (*error0) : ETHERNET_ERROR_DOWN;
}
}
static_always_inline void
determine_next_node (ethernet_main_t * em,
ethernet_input_variant_t variant,
u32 is_l20,
u32 type0, vlib_buffer_t * b0, u8 * error0, u8 * next0)
{
vnet_buffer (b0)->l3_hdr_offset = b0->current_data;
b0->flags |= VNET_BUFFER_F_L3_HDR_OFFSET_VALID;
if (PREDICT_FALSE (*error0 != ETHERNET_ERROR_NONE))
{
// some error occurred
*next0 = ETHERNET_INPUT_NEXT_DROP;
}
else if (is_l20)
{
// record the L2 len and reset the buffer so the L2 header is preserved
u32 eth_start = vnet_buffer (b0)->l2_hdr_offset;
vnet_buffer (b0)->l2.l2_len = b0->current_data - eth_start;
*next0 = em->l2_next;
ASSERT (vnet_buffer (b0)->l2.l2_len ==
ethernet_buffer_header_size (b0));
vlib_buffer_advance (b0, -(vnet_buffer (b0)->l2.l2_len));
// check for common IP/MPLS ethertypes
}
else if (type0 == ETHERNET_TYPE_IP4)
{
*next0 = em->l3_next.input_next_ip4;
}
else if (type0 == ETHERNET_TYPE_IP6)
{
*next0 = em->l3_next.input_next_ip6;
}
else if (type0 == ETHERNET_TYPE_MPLS)
{
*next0 = em->l3_next.input_next_mpls;
}
else if (em->redirect_l3)
{
// L3 Redirect is on, the cached common next nodes will be
// pointing to the redirect node, catch the uncommon types here
*next0 = em->redirect_l3_next;
}
else
{
// uncommon ethertype, check table
u32 i0;
i0 = sparse_vec_index (em->l3_next.input_next_by_type, type0);
*next0 = vec_elt (em->l3_next.input_next_by_type, i0);
*error0 =
i0 ==
SPARSE_VEC_INVALID_INDEX ? ETHERNET_ERROR_UNKNOWN_TYPE : *error0;
// The table is not populated with LLC values, so check that now.
// If variant is variant_ethernet then we came from LLC processing. Don't
// go back there; drop instead using by keeping the drop/bad table result.
if ((type0 < 0x600) && (variant == ETHERNET_INPUT_VARIANT_ETHERNET))
{
*next0 = ETHERNET_INPUT_NEXT_LLC;
}
}
}
typedef enum
{
ETYPE_ID_UNKNOWN = 0,
ETYPE_ID_IP4,
ETYPE_ID_IP6,
ETYPE_ID_MPLS,
ETYPE_N_IDS,
} etype_id_t;
static_always_inline void
eth_input_advance_and_flags (vlib_main_t * vm, u32 * from, u32 n_left,
i16 advance, u32 and_flags, u32 or_flags)
{
vlib_buffer_t *b[8];
while (n_left >= 8)
{
vlib_get_buffers (vm, from, b, 8);
vlib_buffer_advance (b[0], advance);
vlib_buffer_advance (b[1], advance);
vlib_buffer_advance (b[2], advance);
vlib_buffer_advance (b[3], advance);
vlib_buffer_advance (b[4], advance);
vlib_buffer_advance (b[5], advance);
vlib_buffer_advance (b[6], advance);
vlib_buffer_advance (b[7], advance);
b[0]->flags = (b[0]->flags & and_flags) | or_flags;
b[1]->flags = (b[1]->flags & and_flags) | or_flags;
b[2]->flags = (b[2]->flags & and_flags) | or_flags;
b[3]->flags = (b[3]->flags & and_flags) | or_flags;
b[4]->flags = (b[4]->flags & and_flags) | or_flags;
b[5]->flags = (b[5]->flags & and_flags) | or_flags;
b[6]->flags = (b[6]->flags & and_flags) | or_flags;
b[7]->flags = (b[7]->flags & and_flags) | or_flags;
n_left -= 8;
from += 8;
}
while (n_left)
{
vlib_get_buffers (vm, from, b, 1);
vlib_buffer_advance (b[0], advance);
b[0]->flags = (b[0]->flags & and_flags) | or_flags;
n_left -= 1;
from += 1;
}
}
typedef struct
{
u16 etypes[VLIB_FRAME_SIZE];
u32 bufs_by_etype[ETYPE_N_IDS][VLIB_FRAME_SIZE];
u16 n_bufs_by_etype[ETYPE_N_IDS];
} eth_input_data_t;
/* following vector code relies on following assumptions */
STATIC_ASSERT_OFFSET_OF (vlib_buffer_t, current_data, 0);
STATIC_ASSERT_OFFSET_OF (vlib_buffer_t, current_length, 2);
STATIC_ASSERT_OFFSET_OF (vlib_buffer_t, flags, 4);
STATIC_ASSERT (STRUCT_OFFSET_OF (vnet_buffer_opaque_t, l2_hdr_offset) ==
STRUCT_OFFSET_OF (vnet_buffer_opaque_t, l3_hdr_offset) - 2,
"l3_hdr_offset must follow l2_hdr_offset");
static_always_inline void
eth_input_adv_and_flags_x4 (vlib_buffer_t ** b, i16 adv, u32 flags, int is_l3)
{
#ifdef CLIB_HAVE_VEC256
/* to reduce number of small loads/stores we are loading first 64 bits
of each buffer metadata into 256-bit register so we can advance
current_data, current_length and flags.
Observed saving of this code is ~2 clocks per packet */
u64x4 r, radv;
/* vector if signed 16 bit integers used in signed vector add operation
to advnce current_data and current_length */
u32x8 flags4 = { 0, flags, 0, flags, 0, flags, 0, flags };
i16x16 adv4 = {
adv, -adv, 0, 0, adv, -adv, 0, 0,
adv, -adv, 0, 0, adv, -adv, 0, 0
};
/* load 4 x 64 bits */
r = u64x4_gather (b[0], b[1], b[2], b[3]);
/* set flags */
r |= (u64x4) flags4;
/* advance buffer */
radv = (u64x4) ((i16x16) r + adv4);
/* write 4 x 64 bits */
u64x4_scatter (is_l3 ? radv : r, b[0], b[1], b[2], b[3]);
/* use old current_data as l2_hdr_offset and new current_data as
l3_hdr_offset */
r = (u64x4) u16x16_blend (r, radv << 16, 0xaa);
/* store both l2_hdr_offset and l3_hdr_offset in single store operation */
u32x8_scatter_one ((u32x8) r, 0, &vnet_buffer (b[0])->l2_hdr_offset);
u32x8_scatter_one ((u32x8) r, 2, &vnet_buffer (b[1])->l2_hdr_offset);
u32x8_scatter_one ((u32x8) r, 4, &vnet_buffer (b[2])->l2_hdr_offset);
u32x8_scatter_one ((u32x8) r, 6, &vnet_buffer (b[3])->l2_hdr_offset);
if (is_l3)
{
ASSERT (b[0]->current_data == vnet_buffer (b[0])->l3_hdr_offset);
ASSERT (b[1]->current_data == vnet_buffer (b[1])->l3_hdr_offset);
ASSERT (b[2]->current_data == vnet_buffer (b[2])->l3_hdr_offset);
ASSERT (b[3]->current_data == vnet_buffer (b[3])->l3_hdr_offset);
ASSERT (b[0]->current_data - vnet_buffer (b[0])->l2_hdr_offset == adv);
ASSERT (b[1]->current_data - vnet_buffer (b[1])->l2_hdr_offset == adv);
ASSERT (b[2]->current_data - vnet_buffer (b[2])->l2_hdr_offset == adv);
ASSERT (b[3]->current_data - vnet_buffer (b[3])->l2_hdr_offset == adv);
}
else
{
ASSERT (b[0]->current_data == vnet_buffer (b[0])->l2_hdr_offset);
ASSERT (b[1]->current_data == vnet_buffer (b[1])->l2_hdr_offset);
ASSERT (b[2]->current_data == vnet_buffer (b[2])->l2_hdr_offset);
ASSERT (b[3]->current_data == vnet_buffer (b[3])->l2_hdr_offset);
ASSERT (b[0]->current_data - vnet_buffer (b[0])->l3_hdr_offset == -adv);
ASSERT (b[1]->current_data - vnet_buffer (b[1])->l3_hdr_offset == -adv);
ASSERT (b[2]->current_data - vnet_buffer (b[2])->l3_hdr_offset == -adv);
ASSERT (b[3]->current_data - vnet_buffer (b[3])->l3_hdr_offset == -adv);
}
#else
vnet_buffer (b[0])->l2_hdr_offset = b[0]->current_data;
vnet_buffer (b[1])->l2_hdr_offset = b[1]->current_data;
vnet_buffer (b[2])->l2_hdr_offset = b[2]->current_data;
vnet_buffer (b[3])->l2_hdr_offset = b[3]->current_data;
vnet_buffer (b[0])->l3_hdr_offset = b[0]->current_data + adv;
vnet_buffer (b[1])->l3_hdr_offset = b[1]->current_data + adv;
vnet_buffer (b[2])->l3_hdr_offset = b[2]->current_data + adv;
vnet_buffer (b[3])->l3_hdr_offset = b[3]->current_data + adv;
if (is_l3)
{
vlib_buffer_advance (b[0], adv);
vlib_buffer_advance (b[1], adv);
vlib_buffer_advance (b[2], adv);
vlib_buffer_advance (b[3], adv);
}
b[0]->flags |= flags;
b[1]->flags |= flags;
b[2]->flags |= flags;
b[3]->flags |= flags;
#endif
if (!is_l3)
{
vnet_buffer (b[0])->l2.l2_len = adv;
vnet_buffer (b[1])->l2.l2_len = adv;
vnet_buffer (b[2])->l2.l2_len = adv;
vnet_buffer (b[3])->l2.l2_len = adv;
}
}
static_always_inline void
eth_input_adv_and_flags_x1 (vlib_buffer_t ** b, i16 adv, u32 flags, int is_l3)
{
vnet_buffer (b[0])->l2_hdr_offset = b[0]->current_data;
vnet_buffer (b[0])->l3_hdr_offset = b[0]->current_data + adv;
if (is_l3)
vlib_buffer_advance (b[0], adv);
b[0]->flags |= flags;
if (!is_l3)
vnet_buffer (b[0])->l2.l2_len = adv;
}
static_always_inline void
eth_input_process_frame (vlib_main_t * vm, u32 * from, u16 * etype,
u32 n_left, int is_l3)
{
vlib_buffer_t *b[16];
ethernet_header_t *e;
int adv = sizeof (ethernet_header_t);
u32 flags = VNET_BUFFER_F_L2_HDR_OFFSET_VALID |
VNET_BUFFER_F_L3_HDR_OFFSET_VALID;
while (n_left >= 16)
{
vlib_buffer_t **ph = b + 12, **pd = b + 8;
vlib_get_buffers (vm, from, b, 4);
vlib_get_buffers (vm, from + 8, b + 8, 8);
vlib_prefetch_buffer_header (ph[0], LOAD);
vlib_prefetch_buffer_data (pd[0], LOAD);
e = vlib_buffer_get_current (b[0]);
etype[0] = e->type;
vlib_prefetch_buffer_header (ph[1], LOAD);
vlib_prefetch_buffer_data (pd[1], LOAD);
e = vlib_buffer_get_current (b[1]);
etype[1] = e->type;
vlib_prefetch_buffer_header (ph[2], LOAD);
vlib_prefetch_buffer_data (pd[2], LOAD);
e = vlib_buffer_get_current (b[2]);
etype[2] = e->type;
vlib_prefetch_buffer_header (ph[3], LOAD);
vlib_prefetch_buffer_data (pd[3], LOAD);
e = vlib_buffer_get_current (b[3]);
etype[3] = e->type;
eth_input_adv_and_flags_x4 (b, adv, flags, is_l3);
/* next */
n_left -= 4;
etype += 4;
from += 4;
}
while (n_left >= 4)
{
vlib_get_buffers (vm, from, b, 4);
e = vlib_buffer_get_current (b[0]);
etype[0] = e->type;
e = vlib_buffer_get_current (b[1]);
etype[1] = e->type;
e = vlib_buffer_get_current (b[2]);
etype[2] = e->type;
e = vlib_buffer_get_current (b[3]);
etype[3] = e->type;
eth_input_adv_and_flags_x4 (b, adv, flags, is_l3);
/* next */
n_left -= 4;
etype += 4;
from += 4;
}
while (n_left)
{
vlib_get_buffers (vm, from, b, 1);
e = vlib_buffer_get_current (b[0]);
etype[0] = e->type;
eth_input_adv_and_flags_x1 (b, adv, flags, is_l3);
/* next */
n_left -= 1;
etype += 1;
from += 1;
}
}
static_always_inline void
eth_input_sort (vlib_main_t * vm, u32 * from, u32 n_packets,
eth_input_data_t * d)
{
u16 *etype = d->etypes;
i32 n_left = n_packets;
#if defined (CLIB_HAVE_VEC256)
u16x16 e16;
u16x16 et16_ip4 = u16x16_splat (clib_host_to_net_u16 (ETHERNET_TYPE_IP4));
u16x16 et16_ip6 = u16x16_splat (clib_host_to_net_u16 (ETHERNET_TYPE_IP6));
u16x16 et16_mpls = u16x16_splat (clib_host_to_net_u16 (ETHERNET_TYPE_MPLS));
u16x16 id16_ip4 = u16x16_splat (ETYPE_ID_IP4);
u16x16 id16_ip6 = u16x16_splat (ETYPE_ID_IP6);
u16x16 id16_mpls = u16x16_splat (ETYPE_ID_MPLS);
while (n_left > 0)
{
u16x16 r = { 0 };
e16 = u16x16_load_unaligned (etype);
r += (e16 == et16_ip4) & id16_ip4;
r += (e16 == et16_ip6) & id16_ip6;
r += (e16 == et16_mpls) & id16_mpls;
u16x16_store_unaligned (r, etype);
etype += 16;
n_left -= 16;
}
#elif defined (CLIB_HAVE_VEC128)
u16x8 e8;
u16x8 et8_ip4 = u16x8_splat (clib_host_to_net_u16 (ETHERNET_TYPE_IP4));
u16x8 et8_ip6 = u16x8_splat (clib_host_to_net_u16 (ETHERNET_TYPE_IP6));
u16x8 et8_mpls = u16x8_splat (clib_host_to_net_u16 (ETHERNET_TYPE_MPLS));
u16x8 id8_ip4 = u16x8_splat (ETYPE_ID_IP4);
u16x8 id8_ip6 = u16x8_splat (ETYPE_ID_IP6);
u16x8 id8_mpls = u16x8_splat (ETYPE_ID_MPLS);
while (n_left > 0)
{
u16x8 r = { 0 };
e8 = u16x8_load_unaligned (etype);
r += (e8 == et8_ip4) & id8_ip4;
r += (e8 == et8_ip6) & id8_ip6;
r += (e8 == et8_mpls) & id8_mpls;
u16x8_store_unaligned (r, etype);
etype += 8;
n_left -= 8;
}
#else
while (n_left)
{
if (etype[0] == ETHERNET_TYPE_IP4)
etype[0] = ETYPE_ID_IP4;
else if (etype[0] == ETHERNET_TYPE_IP6)
etype[0] = ETYPE_ID_IP6;
else if (etype[0] == ETHERNET_TYPE_MPLS)
etype[0] = ETYPE_ID_MPLS;
else
etype[0] = ETYPE_ID_UNKNOWN;
etype += 1;
n_left -= 1;
}
#endif
etype = d->etypes;
n_left = n_packets;
clib_memset_u16 (d->n_bufs_by_etype, 0, ETYPE_N_IDS);
while (n_left)
{
u16 x, y;
x = etype[0];
y = d->n_bufs_by_etype[x];
#ifdef CLIB_HAVE_VEC256
if (n_left >= 16 && u16x16_is_all_equal (u16x16_load_unaligned (etype),
etype[0]))
{
clib_memcpy_fast (&d->bufs_by_etype[x][y], from, 16 * sizeof (u32));
d->n_bufs_by_etype[x] += 16;
/* next */
n_left -= 16;
etype += 16;
from += 16;
continue;
}
#endif
#ifdef CLIB_HAVE_VEC128
if (n_left >= 8 && u16x8_is_all_equal (u16x8_load_unaligned (etype),
etype[0]))
{
clib_memcpy_fast (&d->bufs_by_etype[x][y], from, 8 * sizeof (u32));
d->n_bufs_by_etype[x] += 8;
/* next */
n_left -= 8;
etype += 8;
from += 8;
continue;
}
#endif
d->bufs_by_etype[x][y] = from[0];
d->n_bufs_by_etype[x]++;
/* next */
n_left -= 1;
etype += 1;
from += 1;
}
}
static_always_inline void
ethernet_input_trace (vlib_main_t * vm, vlib_node_runtime_t * node,
vlib_frame_t * from_frame)
{
u32 *from, n_left;
if ((node->flags & VLIB_NODE_FLAG_TRACE) == 0)
return;
from = vlib_frame_vector_args (from_frame);
n_left = from_frame->n_vectors;
while (n_left)
{
ethernet_input_trace_t *t0;
vlib_buffer_t *b0 = vlib_get_buffer (vm, from[0]);
if (b0->flags & VLIB_BUFFER_IS_TRACED)
{
t0 = vlib_add_trace (vm, node, b0, sizeof (ethernet_input_trace_t));
clib_memcpy_fast (t0->packet_data, b0->data + b0->current_data,
sizeof (t0->packet_data));
t0->frame_flags = from_frame->flags;
clib_memcpy_fast (&t0->frame_data,
vlib_frame_scalar_args (from_frame),
sizeof (ethernet_input_frame_t));
}
from += 1;
n_left -= 1;
}
}
static_always_inline void
ethernet_input_inline (vlib_main_t * vm,
vlib_node_runtime_t * node,
u32 * from, u32 n_packets,
ethernet_input_variant_t variant)
{
vnet_main_t *vnm = vnet_get_main ();
ethernet_main_t *em = &ethernet_main;
vlib_node_runtime_t *error_node;
u32 n_left_from, next_index, *to_next;
u32 stats_sw_if_index, stats_n_packets, stats_n_bytes;
u32 thread_index = vm->thread_index;
u32 cached_sw_if_index = ~0;
u32 cached_is_l2 = 0; /* shut up gcc */
vnet_hw_interface_t *hi = NULL; /* used for main interface only */
if (variant != ETHERNET_INPUT_VARIANT_ETHERNET)
error_node = vlib_node_get_runtime (vm, ethernet_input_node.index);
else
error_node = node;
n_left_from = n_packets;
next_index = node->cached_next_index;
stats_sw_if_index = node->runtime_data[0];
stats_n_packets = stats_n_bytes = 0;
while (n_left_from > 0)
{
u32 n_left_to_next;
vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
while (n_left_from >= 4 && n_left_to_next >= 2)
{
u32 bi0, bi1;
vlib_buffer_t *b0, *b1;
u8 next0, next1, error0, error1;
u16 type0, orig_type0, type1, orig_type1;
u16 outer_id0, inner_id0, outer_id1, inner_id1;
u32 match_flags0, match_flags1;
u32 old_sw_if_index0, new_sw_if_index0, len0, old_sw_if_index1,
new_sw_if_index1, len1;
vnet_hw_interface_t *hi0, *hi1;
main_intf_t *main_intf0, *main_intf1;
vlan_intf_t *vlan_intf0, *vlan_intf1;
qinq_intf_t *qinq_intf0, *qinq_intf1;
u32 is_l20, is_l21;
ethernet_header_t *e0, *e1;
/* Prefetch next iteration. */
{
vlib_buffer_t *b2, *b3;
b2 = vlib_get_buffer (vm, from[2]);
b3 = vlib_get_buffer (vm, from[3]);
vlib_prefetch_buffer_header (b2, STORE);
vlib_prefetch_buffer_header (b3, STORE);
CLIB_PREFETCH (b2->data, sizeof (ethernet_header_t), LOAD);
CLIB_PREFETCH (b3->data, sizeof (ethernet_header_t), LOAD);
}
bi0 = from[0];
bi1 = from[1];
to_next[0] = bi0;
to_next[1] = bi1;
from += 2;
to_next += 2;
n_left_to_next -= 2;
n_left_from -= 2;
b0 = vlib_get_buffer (vm, bi0);
b1 = vlib_get_buffer (vm, bi1);
error0 = error1 = ETHERNET_ERROR_NONE;
e0 = vlib_buffer_get_current (b0);
type0 = clib_net_to_host_u16 (e0->type);
e1 = vlib_buffer_get_current (b1);
type1 = clib_net_to_host_u16 (e1->type);
/* Set the L2 header offset for all packets */
vnet_buffer (b0)->l2_hdr_offset = b0->current_data;
vnet_buffer (b1)->l2_hdr_offset = b1->current_data;
b0->flags |= VNET_BUFFER_F_L2_HDR_OFFSET_VALID;
b1->flags |= VNET_BUFFER_F_L2_HDR_OFFSET_VALID;
/* Speed-path for the untagged case */
if (PREDICT_TRUE (variant == ETHERNET_INPUT_VARIANT_ETHERNET
&& !ethernet_frame_is_any_tagged_x2 (type0,
type1)))
{
main_intf_t *intf0;
subint_config_t *subint0;
u32 sw_if_index0, sw_if_index1;
sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX];
is_l20 = cached_is_l2;
/* This is probably wholly unnecessary */
if (PREDICT_FALSE (sw_if_index0 != sw_if_index1))
goto slowpath;
/* Now sw_if_index0 == sw_if_index1 */
if (PREDICT_FALSE (cached_sw_if_index != sw_if_index0))
{
cached_sw_if_index = sw_if_index0;
hi = vnet_get_sup_hw_interface (vnm, sw_if_index0);
intf0 = vec_elt_at_index (em->main_intfs, hi->hw_if_index);
subint0 = &intf0->untagged_subint;
cached_is_l2 = is_l20 = subint0->flags & SUBINT_CONFIG_L2;
}
if (PREDICT_TRUE (is_l20 != 0))
{
vnet_buffer (b0)->l3_hdr_offset =
vnet_buffer (b0)->l2_hdr_offset +
sizeof (ethernet_header_t);
vnet_buffer (b1)->l3_hdr_offset =
vnet_buffer (b1)->l2_hdr_offset +
sizeof (ethernet_header_t);
b0->flags |= VNET_BUFFER_F_L3_HDR_OFFSET_VALID;
b1->flags |= VNET_BUFFER_F_L3_HDR_OFFSET_VALID;
next0 = em->l2_next;
vnet_buffer (b0)->l2.l2_len = sizeof (ethernet_header_t);
next1 = em->l2_next;
vnet_buffer (b1)->l2.l2_len = sizeof (ethernet_header_t);
}
else
{
if (!ethernet_address_cast (e0->dst_address) &&
(hi->hw_address != 0) &&
!eth_mac_equal ((u8 *) e0, hi->hw_address))
error0 = ETHERNET_ERROR_L3_MAC_MISMATCH;
if (!ethernet_address_cast (e1->dst_address) &&
(hi->hw_address != 0) &&
!eth_mac_equal ((u8 *) e1, hi->hw_address))
error1 = ETHERNET_ERROR_L3_MAC_MISMATCH;
vlib_buffer_advance (b0, sizeof (ethernet_header_t));
determine_next_node (em, variant, 0, type0, b0,
&error0, &next0);
vlib_buffer_advance (b1, sizeof (ethernet_header_t));
determine_next_node (em, variant, 0, type1, b1,
&error1, &next1);
}
goto ship_it01;
}
/* Slow-path for the tagged case */
slowpath:
parse_header (variant,
b0,
&type0,
&orig_type0, &outer_id0, &inner_id0, &match_flags0);
parse_header (variant,
b1,
&type1,
&orig_type1, &outer_id1, &inner_id1, &match_flags1);
old_sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
old_sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX];
eth_vlan_table_lookups (em,
vnm,
old_sw_if_index0,
orig_type0,
outer_id0,
inner_id0,
&hi0,
&main_intf0, &vlan_intf0, &qinq_intf0);
eth_vlan_table_lookups (em,
vnm,
old_sw_if_index1,
orig_type1,
outer_id1,
inner_id1,
&hi1,
&main_intf1, &vlan_intf1, &qinq_intf1);
identify_subint (hi0,
b0,
match_flags0,
main_intf0,
vlan_intf0,
qinq_intf0, &new_sw_if_index0, &error0, &is_l20);
identify_subint (hi1,
b1,
match_flags1,
main_intf1,
vlan_intf1,
qinq_intf1, &new_sw_if_index1, &error1, &is_l21);
// Save RX sw_if_index for later nodes
vnet_buffer (b0)->sw_if_index[VLIB_RX] =
error0 !=
ETHERNET_ERROR_NONE ? old_sw_if_index0 : new_sw_if_index0;
vnet_buffer (b1)->sw_if_index[VLIB_RX] =
error1 !=
ETHERNET_ERROR_NONE ? old_sw_if_index1 : new_sw_if_index1;
// Check if there is a stat to take (valid and non-main sw_if_index for pkt 0 or pkt 1)
if (((new_sw_if_index0 != ~0)
&& (new_sw_if_index0 != old_sw_if_index0))
|| ((new_sw_if_index1 != ~0)
&& (new_sw_if_index1 != old_sw_if_index1)))
{
len0 = vlib_buffer_length_in_chain (vm, b0) + b0->current_data
- vnet_buffer (b0)->l2_hdr_offset;
len1 = vlib_buffer_length_in_chain (vm, b1) + b1->current_data
- vnet_buffer (b1)->l2_hdr_offset;
stats_n_packets += 2;
stats_n_bytes += len0 + len1;
if (PREDICT_FALSE
(!(new_sw_if_index0 == stats_sw_if_index
&& new_sw_if_index1 == stats_sw_if_index)))
{
stats_n_packets -= 2;
stats_n_bytes -= len0 + len1;
if (new_sw_if_index0 != old_sw_if_index0
&& new_sw_if_index0 != ~0)
vlib_increment_combined_counter (vnm->
interface_main.combined_sw_if_counters
+
VNET_INTERFACE_COUNTER_RX,
thread_index,
new_sw_if_index0, 1,
len0);
if (new_sw_if_index1 != old_sw_if_index1
&& new_sw_if_index1 != ~0)
vlib_increment_combined_counter (vnm->
interface_main.combined_sw_if_counters
+
VNET_INTERFACE_COUNTER_RX,
thread_index,
new_sw_if_index1, 1,
len1);
if (new_sw_if_index0 == new_sw_if_index1)
{
if (stats_n_packets > 0)
{
vlib_increment_combined_counter
(vnm->interface_main.combined_sw_if_counters
+ VNET_INTERFACE_COUNTER_RX,
thread_index,
stats_sw_if_index,
stats_n_packets, stats_n_bytes);
stats_n_packets = stats_n_bytes = 0;
}
stats_sw_if_index = new_sw_if_index0;
}
}
}
if (variant == ETHERNET_INPUT_VARIANT_NOT_L2)
is_l20 = is_l21 = 0;
determine_next_node (em, variant, is_l20, type0, b0, &error0,
&next0);
determine_next_node (em, variant, is_l21, type1, b1, &error1,
&next1);
ship_it01:
b0->error = error_node->errors[error0];
b1->error = error_node->errors[error1];
// verify speculative enqueue
vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next,
n_left_to_next, bi0, bi1, next0,
next1);
}
while (n_left_from > 0 && n_left_to_next > 0)
{
u32 bi0;
vlib_buffer_t *b0;
u8 error0, next0;
u16 type0, orig_type0;
u16 outer_id0, inner_id0;
u32 match_flags0;
u32 old_sw_if_index0, new_sw_if_index0, len0;
vnet_hw_interface_t *hi0;
main_intf_t *main_intf0;
vlan_intf_t *vlan_intf0;
qinq_intf_t *qinq_intf0;
ethernet_header_t *e0;
u32 is_l20;
// Prefetch next iteration
if (n_left_from > 1)
{
vlib_buffer_t *p2;
p2 = vlib_get_buffer (vm, from[1]);
vlib_prefetch_buffer_header (p2, STORE);
CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, LOAD);
}
bi0 = from[0];
to_next[0] = bi0;
from += 1;
to_next += 1;
n_left_from -= 1;
n_left_to_next -= 1;
b0 = vlib_get_buffer (vm, bi0);
error0 = ETHERNET_ERROR_NONE;
e0 = vlib_buffer_get_current (b0);
type0 = clib_net_to_host_u16 (e0->type);
/* Set the L2 header offset for all packets */
vnet_buffer (b0)->l2_hdr_offset = b0->current_data;
b0->flags |= VNET_BUFFER_F_L2_HDR_OFFSET_VALID;
/* Speed-path for the untagged case */
if (PREDICT_TRUE (variant == ETHERNET_INPUT_VARIANT_ETHERNET
&& !ethernet_frame_is_tagged (type0)))
{
main_intf_t *intf0;
subint_config_t *subint0;
u32 sw_if_index0;
sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
is_l20 = cached_is_l2;
if (PREDICT_FALSE (cached_sw_if_index != sw_if_index0))
{
cached_sw_if_index = sw_if_index0;
hi = vnet_get_sup_hw_interface (vnm, sw_if_index0);
intf0 = vec_elt_at_index (em->main_intfs, hi->hw_if_index);
subint0 = &intf0->untagged_subint;
cached_is_l2 = is_l20 = subint0->flags & SUBINT_CONFIG_L2;
}
if (PREDICT_TRUE (is_l20 != 0))
{
vnet_buffer (b0)->l3_hdr_offset =
vnet_buffer (b0)->l2_hdr_offset +
sizeof (ethernet_header_t);
b0->flags |= VNET_BUFFER_F_L3_HDR_OFFSET_VALID;
next0 = em->l2_next;
vnet_buffer (b0)->l2.l2_len = sizeof (ethernet_header_t);
}
else
{
if (!ethernet_address_cast (e0->dst_address) &&
(hi->hw_address != 0) &&
!eth_mac_equal ((u8 *) e0, hi->hw_address))
error0 = ETHERNET_ERROR_L3_MAC_MISMATCH;
vlib_buffer_advance (b0, sizeof (ethernet_header_t));
determine_next_node (em, variant, 0, type0, b0,
&error0, &next0);
}
goto ship_it0;
}
/* Slow-path for the tagged case */
parse_header (variant,
b0,
&type0,
&orig_type0, &outer_id0, &inner_id0, &match_flags0);
old_sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
eth_vlan_table_lookups (em,
vnm,
old_sw_if_index0,
orig_type0,
outer_id0,
inner_id0,
&hi0,
&main_intf0, &vlan_intf0, &qinq_intf0);
identify_subint (hi0,
b0,
match_flags0,
main_intf0,
vlan_intf0,
qinq_intf0, &new_sw_if_index0, &error0, &is_l20);
// Save RX sw_if_index for later nodes
vnet_buffer (b0)->sw_if_index[VLIB_RX] =
error0 !=
ETHERNET_ERROR_NONE ? old_sw_if_index0 : new_sw_if_index0;
// Increment subinterface stats
// Note that interface-level counters have already been incremented
// prior to calling this function. Thus only subinterface counters
// are incremented here.
//
// Interface level counters include packets received on the main
// interface and all subinterfaces. Subinterface level counters
// include only those packets received on that subinterface
// Increment stats if the subint is valid and it is not the main intf
if ((new_sw_if_index0 != ~0)
&& (new_sw_if_index0 != old_sw_if_index0))
{
len0 = vlib_buffer_length_in_chain (vm, b0) + b0->current_data
- vnet_buffer (b0)->l2_hdr_offset;
stats_n_packets += 1;
stats_n_bytes += len0;
// Batch stat increments from the same subinterface so counters
// don't need to be incremented for every packet.
if (PREDICT_FALSE (new_sw_if_index0 != stats_sw_if_index))
{
stats_n_packets -= 1;
stats_n_bytes -= len0;
if (new_sw_if_index0 != ~0)
vlib_increment_combined_counter
(vnm->interface_main.combined_sw_if_counters
+ VNET_INTERFACE_COUNTER_RX,
thread_index, new_sw_if_index0, 1, len0);
if (stats_n_packets > 0)
{
vlib_increment_combined_counter
(vnm->interface_main.combined_sw_if_counters
+ VNET_INTERFACE_COUNTER_RX,
thread_index,
stats_sw_if_index, stats_n_packets, stats_n_bytes);
stats_n_packets = stats_n_bytes = 0;
}
stats_sw_if_index = new_sw_if_index0;
}
}
if (variant == ETHERNET_INPUT_VARIANT_NOT_L2)
is_l20 = 0;
determine_next_node (em, variant, is_l20, type0, b0, &error0,
&next0);
ship_it0:
b0->error = error_node->errors[error0];
// verify speculative enqueue
vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
to_next, n_left_to_next,
bi0, next0);
}
vlib_put_next_frame (vm, node, next_index, n_left_to_next);
}
// Increment any remaining batched stats
if (stats_n_packets > 0)
{
vlib_increment_combined_counter
(vnm->interface_main.combined_sw_if_counters
+ VNET_INTERFACE_COUNTER_RX,
thread_index, stats_sw_if_index, stats_n_packets, stats_n_bytes);
node->runtime_data[0] = stats_sw_if_index;
}
}
static_always_inline void
eth_input_enqueue_untagged (vlib_main_t * vm, vlib_node_runtime_t * node,
eth_input_data_t * d, int ip4_cksum_ok, int is_l3)
{
ethernet_main_t *em = &ethernet_main;
etype_id_t id;
u32 next_index;
id = ETYPE_ID_IP4;
if (d->n_bufs_by_etype[id])
{
if (is_l3)
{
next_index = em->l3_next.input_next_ip4;
if (next_index == ETHERNET_INPUT_NEXT_IP4_INPUT && ip4_cksum_ok)
next_index = ETHERNET_INPUT_NEXT_IP4_INPUT_NCS;
}
else
next_index = em->l2_next;
vlib_buffer_enqueue_to_single_next (vm, node, d->bufs_by_etype[id],
next_index, d->n_bufs_by_etype[id]);
}
id = ETYPE_ID_IP6;
if (d->n_bufs_by_etype[id])
{
next_index = is_l3 ? em->l3_next.input_next_ip6 : em->l2_next;
vlib_buffer_enqueue_to_single_next (vm, node, d->bufs_by_etype[id],
next_index, d->n_bufs_by_etype[id]);
}
id = ETYPE_ID_MPLS;
if (d->n_bufs_by_etype[id])
{
next_index = is_l3 ? em->l3_next.input_next_mpls : em->l2_next;
vlib_buffer_enqueue_to_single_next (vm, node, d->bufs_by_etype[id],
next_index, d->n_bufs_by_etype[id]);
}
id = ETYPE_ID_UNKNOWN;
if (d->n_bufs_by_etype[id])
{
/* in case of l3 interfaces, we already advanced buffer so we need to
roll back */
if (is_l3)
eth_input_advance_and_flags (vm, d->bufs_by_etype[id],
d->n_bufs_by_etype[id],
-(i16) sizeof (ethernet_header_t),
~VNET_BUFFER_F_L3_HDR_OFFSET_VALID, 0);
ethernet_input_inline (vm, node, d->bufs_by_etype[id],
d->n_bufs_by_etype[id],
ETHERNET_INPUT_VARIANT_ETHERNET);
}
}
VLIB_NODE_FN (ethernet_input_node) (vlib_main_t * vm,
vlib_node_runtime_t * node,
vlib_frame_t * frame)
{
vnet_main_t *vnm = vnet_get_main ();
ethernet_main_t *em = &ethernet_main;
u32 *from = vlib_frame_vector_args (frame);
u32 n_packets = frame->n_vectors;
ethernet_input_trace (vm, node, frame);
if (frame->flags & ETH_INPUT_FRAME_F_SINGLE_SW_IF_IDX)
{
eth_input_data_t data, *d = &data;
ethernet_input_frame_t *ef = vlib_frame_scalar_args (frame);
vnet_hw_interface_t *hi = vnet_get_hw_interface (vnm, ef->hw_if_index);
main_intf_t *intf0 = vec_elt_at_index (em->main_intfs, hi->hw_if_index);
subint_config_t *subint0 = &intf0->untagged_subint;
int ip4_cksum_ok = (frame->flags & ETH_INPUT_FRAME_F_IP4_CKSUM_OK) != 0;
if (subint0->flags & SUBINT_CONFIG_L2)
{
/* untagged packets are treated as L2 */
eth_input_process_frame (vm, from, d->etypes, n_packets, 0);
eth_input_sort (vm, from, n_packets, d);
eth_input_enqueue_untagged (vm, node, d, ip4_cksum_ok, 0);
}
else
{
ethernet_interface_t *ei;
ei = pool_elt_at_index (em->interfaces, hi->hw_instance);
/* currently only slowpath deals with dmac check */
if (ei->flags & ETHERNET_INTERFACE_FLAG_ACCEPT_ALL)
goto slowpath;
/* untagged packets are treated as L3 */
eth_input_process_frame (vm, from, d->etypes, n_packets, 1);
eth_input_sort (vm, from, n_packets, d);
eth_input_enqueue_untagged (vm, node, d, ip4_cksum_ok, 1);
}
return n_packets;
}
slowpath:
ethernet_input_inline (vm, node, from, n_packets,
ETHERNET_INPUT_VARIANT_ETHERNET);
return n_packets;
}
VLIB_NODE_FN (ethernet_input_type_node) (vlib_main_t * vm,
vlib_node_runtime_t * node,
vlib_frame_t * from_frame)
{
u32 *from = vlib_frame_vector_args (from_frame);
u32 n_packets = from_frame->n_vectors;
ethernet_input_trace (vm, node, from_frame);
ethernet_input_inline (vm, node, from, n_packets,
ETHERNET_INPUT_VARIANT_ETHERNET_TYPE);
return n_packets;
}
VLIB_NODE_FN (ethernet_input_not_l2_node) (vlib_main_t * vm,
vlib_node_runtime_t * node,
vlib_frame_t * from_frame)
{
u32 *from = vlib_frame_vector_args (from_frame);
u32 n_packets = from_frame->n_vectors;
ethernet_input_trace (vm, node, from_frame);
ethernet_input_inline (vm, node, from, n_packets,
ETHERNET_INPUT_VARIANT_NOT_L2);
return n_packets;
}
// Return the subinterface config struct for the given sw_if_index
// Also return via parameter the appropriate match flags for the
// configured number of tags.
// On error (unsupported or not ethernet) return 0.
static subint_config_t *
ethernet_sw_interface_get_config (vnet_main_t * vnm,
u32 sw_if_index,
u32 * flags, u32 * unsupported)
{
ethernet_main_t *em = &ethernet_main;
vnet_hw_interface_t *hi;
vnet_sw_interface_t *si;
main_intf_t *main_intf;
vlan_table_t *vlan_table;
qinq_table_t *qinq_table;
subint_config_t *subint = 0;
hi = vnet_get_sup_hw_interface (vnm, sw_if_index);
if (!hi || (hi->hw_class_index != ethernet_hw_interface_class.index))
{
*unsupported = 0;
goto done; // non-ethernet interface
}
// ensure there's an entry for the main intf (shouldn't really be necessary)
vec_validate (em->main_intfs, hi->hw_if_index);
main_intf = vec_elt_at_index (em->main_intfs, hi->hw_if_index);
// Locate the subint for the given ethernet config
si = vnet_get_sw_interface (vnm, sw_if_index);
if (si->type == VNET_SW_INTERFACE_TYPE_P2P)
{
p2p_ethernet_main_t *p2pm = &p2p_main;
u32 p2pe_sw_if_index =
p2p_ethernet_lookup (hi->hw_if_index, si->p2p.client_mac);
if (p2pe_sw_if_index == ~0)
{
pool_get (p2pm->p2p_subif_pool, subint);
si->p2p.pool_index = subint - p2pm->p2p_subif_pool;
}
else
subint = vec_elt_at_index (p2pm->p2p_subif_pool, si->p2p.pool_index);
*flags = SUBINT_CONFIG_P2P;
}
else if (si->type == VNET_SW_INTERFACE_TYPE_PIPE)
{
pipe_t *pipe;
pipe = pipe_get (sw_if_index);
subint = &pipe->subint;
*flags = SUBINT_CONFIG_P2P;
}
else if (si->sub.eth.flags.default_sub)
{
subint = &main_intf->default_subint;
*flags = SUBINT_CONFIG_MATCH_1_TAG |
SUBINT_CONFIG_MATCH_2_TAG | SUBINT_CONFIG_MATCH_3_TAG;
}
else if ((si->sub.eth.flags.no_tags) || (si->sub.eth.raw_flags == 0))
{
// if no flags are set then this is a main interface
// so treat as untagged
subint = &main_intf->untagged_subint;
*flags = SUBINT_CONFIG_MATCH_0_TAG;
}
else
{
// one or two tags
// first get the vlan table
if (si->sub.eth.flags.dot1ad)
{
if (main_intf->dot1ad_vlans == 0)
{
// Allocate a vlan table from the pool
pool_get (em->vlan_pool, vlan_table);
main_intf->dot1ad_vlans = vlan_table - em->vlan_pool;
}
else
{
// Get ptr to existing vlan table
vlan_table =
vec_elt_at_index (em->vlan_pool, main_intf->dot1ad_vlans);
}
}
else
{ // dot1q
if (main_intf->dot1q_vlans == 0)
{
// Allocate a vlan table from the pool
pool_get (em->vlan_pool, vlan_table);
main_intf->dot1q_vlans = vlan_table - em->vlan_pool;
}
else
{
// Get ptr to existing vlan table
vlan_table =
vec_elt_at_index (em->vlan_pool, main_intf->dot1q_vlans);
}
}
if (si->sub.eth.flags.one_tag)
{
*flags = si->sub.eth.flags.exact_match ?
SUBINT_CONFIG_MATCH_1_TAG :
(SUBINT_CONFIG_MATCH_1_TAG |
SUBINT_CONFIG_MATCH_2_TAG | SUBINT_CONFIG_MATCH_3_TAG);
if (si->sub.eth.flags.outer_vlan_id_any)
{
// not implemented yet
*unsupported = 1;
goto done;
}
else
{
// a single vlan, a common case
subint =
&vlan_table->vlans[si->sub.eth.
outer_vlan_id].single_tag_subint;
}
}
else
{
// Two tags
*flags = si->sub.eth.flags.exact_match ?
SUBINT_CONFIG_MATCH_2_TAG :
(SUBINT_CONFIG_MATCH_2_TAG | SUBINT_CONFIG_MATCH_3_TAG);
if (si->sub.eth.flags.outer_vlan_id_any
&& si->sub.eth.flags.inner_vlan_id_any)
{
// not implemented yet
*unsupported = 1;
goto done;
}
if (si->sub.eth.flags.inner_vlan_id_any)
{
// a specific outer and "any" inner
// don't need a qinq table for this
subint =
&vlan_table->vlans[si->sub.eth.
outer_vlan_id].inner_any_subint;
if (si->sub.eth.flags.exact_match)
{
*flags = SUBINT_CONFIG_MATCH_2_TAG;
}
else
{
*flags = SUBINT_CONFIG_MATCH_2_TAG |
SUBINT_CONFIG_MATCH_3_TAG;
}
}
else
{
// a specific outer + specifc innner vlan id, a common case
// get the qinq table
if (vlan_table->vlans[si->sub.eth.outer_vlan_id].qinqs == 0)
{
// Allocate a qinq table from the pool
pool_get (em->qinq_pool, qinq_table);
vlan_table->vlans[si->sub.eth.outer_vlan_id].qinqs =
qinq_table - em->qinq_pool;
}
else
{
// Get ptr to existing qinq table
qinq_table =
vec_elt_at_index (em->qinq_pool,
vlan_table->vlans[si->sub.
eth.outer_vlan_id].
qinqs);
}
subint = &qinq_table->vlans[si->sub.eth.inner_vlan_id].subint;
}
}
}
done:
return subint;
}
static clib_error_t *
ethernet_sw_interface_up_down (vnet_main_t * vnm, u32 sw_if_index, u32 flags)
{
subint_config_t *subint;
u32 dummy_flags;
u32 dummy_unsup;
clib_error_t *error = 0;
// Find the config for this subinterface
subint =
ethernet_sw_interface_get_config (vnm, sw_if_index, &dummy_flags,
&dummy_unsup);
if (subint == 0)
{
// not implemented yet or not ethernet
goto done;
}
subint->sw_if_index =
((flags & VNET_SW_INTERFACE_FLAG_ADMIN_UP) ? sw_if_index : ~0);
done:
return error;
}
VNET_SW_INTERFACE_ADMIN_UP_DOWN_FUNCTION (ethernet_sw_interface_up_down);
#ifndef CLIB_MARCH_VARIANT
// Set the L2/L3 mode for the subinterface
void
ethernet_sw_interface_set_l2_mode (vnet_main_t * vnm, u32 sw_if_index, u32 l2)
{
subint_config_t *subint;
u32 dummy_flags;
u32 dummy_unsup;
int is_port;
vnet_sw_interface_t *sw = vnet_get_sw_interface (vnm, sw_if_index);
is_port = !(sw->type == VNET_SW_INTERFACE_TYPE_SUB);
// Find the config for this subinterface
subint =
ethernet_sw_interface_get_config (vnm, sw_if_index, &dummy_flags,
&dummy_unsup);
if (subint == 0)
{
// unimplemented or not ethernet
goto done;
}
// Double check that the config we found is for our interface (or the interface is down)
ASSERT ((subint->sw_if_index == sw_if_index) | (subint->sw_if_index == ~0));
if (l2)
{
subint->flags |= SUBINT_CONFIG_L2;
if (is_port)
subint->flags |=
SUBINT_CONFIG_MATCH_0_TAG | SUBINT_CONFIG_MATCH_1_TAG
| SUBINT_CONFIG_MATCH_2_TAG | SUBINT_CONFIG_MATCH_3_TAG;
}
else
{
subint->flags &= ~SUBINT_CONFIG_L2;
if (is_port)
subint->flags &=
~(SUBINT_CONFIG_MATCH_1_TAG | SUBINT_CONFIG_MATCH_2_TAG
| SUBINT_CONFIG_MATCH_3_TAG);
}
done:
return;
}
/*
* Set the L2/L3 mode for the subinterface regardless of port
*/
void
ethernet_sw_interface_set_l2_mode_noport (vnet_main_t * vnm,
u32 sw_if_index, u32 l2)
{
subint_config_t *subint;
u32 dummy_flags;
u32 dummy_unsup;
/* Find the config for this subinterface */
subint =
ethernet_sw_interface_get_config (vnm, sw_if_index, &dummy_flags,
&dummy_unsup);
if (subint == 0)
{
/* unimplemented or not ethernet */
goto done;
}
/*
* Double check that the config we found is for our interface (or the
* interface is down)
*/
ASSERT ((subint->sw_if_index == sw_if_index) | (subint->sw_if_index == ~0));
if (l2)
{
subint->flags |= SUBINT_CONFIG_L2;
}
else
{
subint->flags &= ~SUBINT_CONFIG_L2;
}
done:
return;
}
#endif
static clib_error_t *
ethernet_sw_interface_add_del (vnet_main_t * vnm,
u32 sw_if_index, u32 is_create)
{
clib_error_t *error = 0;
subint_config_t *subint;
u32 match_flags;
u32 unsupported = 0;
// Find the config for this subinterface
subint =
ethernet_sw_interface_get_config (vnm, sw_if_index, &match_flags,
&unsupported);
if (subint == 0)
{
// not implemented yet or not ethernet
if (unsupported)
{
// this is the NYI case
error = clib_error_return (0, "not implemented yet");
}
goto done;
}
if (!is_create)
{
subint->flags = 0;
return error;
}
// Initialize the subint
if (subint->flags & SUBINT_CONFIG_VALID)
{
// Error vlan already in use
error = clib_error_return (0, "vlan is already in use");
}
else
{
// Note that config is L3 by default
subint->flags = SUBINT_CONFIG_VALID | match_flags;
subint->sw_if_index = ~0; // because interfaces are initially down
}
done:
return error;
}
VNET_SW_INTERFACE_ADD_DEL_FUNCTION (ethernet_sw_interface_add_del);
static char *ethernet_error_strings[] = {
#define ethernet_error(n,c,s) s,
#include "error.def"
#undef ethernet_error
};
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (ethernet_input_node) = {
.name = "ethernet-input",
/* Takes a vector of packets. */
.vector_size = sizeof (u32),
.scalar_size = sizeof (ethernet_input_frame_t),
.n_errors = ETHERNET_N_ERROR,
.error_strings = ethernet_error_strings,
.n_next_nodes = ETHERNET_INPUT_N_NEXT,
.next_nodes = {
#define _(s,n) [ETHERNET_INPUT_NEXT_##s] = n,
foreach_ethernet_input_next
#undef _
},
.format_buffer = format_ethernet_header_with_length,
.format_trace = format_ethernet_input_trace,
.unformat_buffer = unformat_ethernet_header,
};
VLIB_REGISTER_NODE (ethernet_input_type_node) = {
.name = "ethernet-input-type",
/* Takes a vector of packets. */
.vector_size = sizeof (u32),
.n_next_nodes = ETHERNET_INPUT_N_NEXT,
.next_nodes = {
#define _(s,n) [ETHERNET_INPUT_NEXT_##s] = n,
foreach_ethernet_input_next
#undef _
},
};
VLIB_REGISTER_NODE (ethernet_input_not_l2_node) = {
.name = "ethernet-input-not-l2",
/* Takes a vector of packets. */
.vector_size = sizeof (u32),
.n_next_nodes = ETHERNET_INPUT_N_NEXT,
.next_nodes = {
#define _(s,n) [ETHERNET_INPUT_NEXT_##s] = n,
foreach_ethernet_input_next
#undef _
},
};
/* *INDENT-ON* */
#ifndef CLIB_MARCH_VARIANT
void
ethernet_set_rx_redirect (vnet_main_t * vnm,
vnet_hw_interface_t * hi, u32 enable)
{
// Insure all packets go to ethernet-input (i.e. untagged ipv4 packets
// don't go directly to ip4-input)
vnet_hw_interface_rx_redirect_to_node
(vnm, hi->hw_if_index, enable ? ethernet_input_node.index : ~0);
}
/*
* Initialization and registration for the next_by_ethernet structure
*/
clib_error_t *
next_by_ethertype_init (next_by_ethertype_t * l3_next)
{
l3_next->input_next_by_type = sparse_vec_new
( /* elt bytes */ sizeof (l3_next->input_next_by_type[0]),
/* bits in index */ BITS (((ethernet_header_t *) 0)->type));
vec_validate (l3_next->sparse_index_by_input_next_index,
ETHERNET_INPUT_NEXT_DROP);
vec_validate (l3_next->sparse_index_by_input_next_index,
ETHERNET_INPUT_NEXT_PUNT);
l3_next->sparse_index_by_input_next_index[ETHERNET_INPUT_NEXT_DROP] =
SPARSE_VEC_INVALID_INDEX;
l3_next->sparse_index_by_input_next_index[ETHERNET_INPUT_NEXT_PUNT] =
SPARSE_VEC_INVALID_INDEX;
/*
* Make sure we don't wipe out an ethernet registration by mistake
* Can happen if init function ordering constraints are missing.
*/
if (CLIB_DEBUG > 0)
{
ethernet_main_t *em = &ethernet_main;
ASSERT (em->next_by_ethertype_register_called == 0);
}
return 0;
}
// Add an ethertype -> next index mapping to the structure
clib_error_t *
next_by_ethertype_register (next_by_ethertype_t * l3_next,
u32 ethertype, u32 next_index)
{
u32 i;
u16 *n;
ethernet_main_t *em = &ethernet_main;
if (CLIB_DEBUG > 0)
{
ethernet_main_t *em = &ethernet_main;
em->next_by_ethertype_register_called = 1;
}
/* Setup ethernet type -> next index sparse vector mapping. */
n = sparse_vec_validate (l3_next->input_next_by_type, ethertype);
n[0] = next_index;
/* Rebuild next index -> sparse index inverse mapping when sparse vector
is updated. */
vec_validate (l3_next->sparse_index_by_input_next_index, next_index);
for (i = 1; i < vec_len (l3_next->input_next_by_type); i++)
l3_next->
sparse_index_by_input_next_index[l3_next->input_next_by_type[i]] = i;
// do not allow the cached next index's to be updated if L3
// redirect is enabled, as it will have overwritten them
if (!em->redirect_l3)
{
// Cache common ethertypes directly
if (ethertype == ETHERNET_TYPE_IP4)
{
l3_next->input_next_ip4 = next_index;
}
else if (ethertype == ETHERNET_TYPE_IP6)
{
l3_next->input_next_ip6 = next_index;
}
else if (ethertype == ETHERNET_TYPE_MPLS)
{
l3_next->input_next_mpls = next_index;
}
}
return 0;
}
static clib_error_t *
ethernet_input_init (vlib_main_t * vm)
{
ethernet_main_t *em = &ethernet_main;
__attribute__ ((unused)) vlan_table_t *invalid_vlan_table;
__attribute__ ((unused)) qinq_table_t *invalid_qinq_table;
ethernet_setup_node (vm, ethernet_input_node.index);
ethernet_setup_node (vm, ethernet_input_type_node.index);
ethernet_setup_node (vm, ethernet_input_not_l2_node.index);
next_by_ethertype_init (&em->l3_next);
// Initialize pools and vector for vlan parsing
vec_validate (em->main_intfs, 10); // 10 main interfaces
pool_alloc (em->vlan_pool, 10);
pool_alloc (em->qinq_pool, 1);
// The first vlan pool will always be reserved for an invalid table
pool_get (em->vlan_pool, invalid_vlan_table); // first id = 0
// The first qinq pool will always be reserved for an invalid table
pool_get (em->qinq_pool, invalid_qinq_table); // first id = 0
return 0;
}
VLIB_INIT_FUNCTION (ethernet_input_init);
void
ethernet_register_input_type (vlib_main_t * vm,
ethernet_type_t type, u32 node_index)
{
ethernet_main_t *em = &ethernet_main;
ethernet_type_info_t *ti;
u32 i;
{
clib_error_t *error = vlib_call_init_function (vm, ethernet_init);
if (error)
clib_error_report (error);
}
ti = ethernet_get_type_info (em, type);
ti->node_index = node_index;
ti->next_index = vlib_node_add_next (vm,
ethernet_input_node.index, node_index);
i = vlib_node_add_next (vm, ethernet_input_type_node.index, node_index);
ASSERT (i == ti->next_index);
i = vlib_node_add_next (vm, ethernet_input_not_l2_node.index, node_index);
ASSERT (i == ti->next_index);
// Add the L3 node for this ethertype to the next nodes structure
next_by_ethertype_register (&em->l3_next, type, ti->next_index);
// Call the registration functions for other nodes that want a mapping
l2bvi_register_input_type (vm, type, node_index);
}
void
ethernet_register_l2_input (vlib_main_t * vm, u32 node_index)
{
ethernet_main_t *em = &ethernet_main;
u32 i;
em->l2_next =
vlib_node_add_next (vm, ethernet_input_node.index, node_index);
/*
* Even if we never use these arcs, we have to align the next indices...
*/
i = vlib_node_add_next (vm, ethernet_input_type_node.index, node_index);
ASSERT (i == em->l2_next);
i = vlib_node_add_next (vm, ethernet_input_not_l2_node.index, node_index);
ASSERT (i == em->l2_next);
}
// Register a next node for L3 redirect, and enable L3 redirect
void
ethernet_register_l3_redirect (vlib_main_t * vm, u32 node_index)
{
ethernet_main_t *em = &ethernet_main;
u32 i;
em->redirect_l3 = 1;
em->redirect_l3_next = vlib_node_add_next (vm,
ethernet_input_node.index,
node_index);
/*
* Change the cached next nodes to the redirect node
*/
em->l3_next.input_next_ip4 = em->redirect_l3_next;
em->l3_next.input_next_ip6 = em->redirect_l3_next;
em->l3_next.input_next_mpls = em->redirect_l3_next;
/*
* Even if we never use these arcs, we have to align the next indices...
*/
i = vlib_node_add_next (vm, ethernet_input_type_node.index, node_index);
ASSERT (i == em->redirect_l3_next);
i = vlib_node_add_next (vm, ethernet_input_not_l2_node.index, node_index);
ASSERT (i == em->redirect_l3_next);
}
#endif
/*
* fd.io coding-style-patch-verification: ON
*
* Local Variables:
* eval: (c-set-style "gnu")
* End:
*/