blob: d848fac6b72265aa0fd76ba8173d816d241e5870 [file] [log] [blame]
/*
* Copyright (c) 2015 Cisco and/or its affiliates.
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at:
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <vnet/l2/l2_xcrw.h>
/**
* @file
* General L2 / L3 cross-connect, used to set up
* "L2 interface <--> your-favorite-tunnel-encap" tunnels.
*
* We set up a typical L2 cross-connect or (future) bridge
* to hook L2 interface(s) up to the L3 stack in arbitrary ways.
*
* Each l2_xcrw adjacency specifies 3 things:
*
* 1. The next graph node (presumably in the L3 stack) to
* process the (L2 -> L3) packet
*
* 2. A new value for vnet_buffer(b)->sw_if_index[VLIB_TX]
* (i.e. a lookup FIB index),
*
* 3. A rewrite string to apply.
*
* Example: to cross-connect an L2 interface or (future) bridge
* to an mpls-o-gre tunnel, set up the L2 rewrite string as shown in
* mpls_gre_rewrite, and use "mpls-post-rewrite" to fix the
* GRE IP header checksum and length fields.
*/
typedef struct
{
u32 next_index;
u32 tx_fib_index;
} l2_xcrw_trace_t;
/* packet trace format function */
static u8 *
format_l2_xcrw_trace (u8 * s, va_list * args)
{
CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
l2_xcrw_trace_t *t = va_arg (*args, l2_xcrw_trace_t *);
s = format (s, "L2_XCRW: next index %d tx_fib_index %d",
t->next_index, t->tx_fib_index);
return s;
}
extern l2_xcrw_main_t l2_xcrw_main;
#ifndef CLIB_MARCH_VARIANT
l2_xcrw_main_t l2_xcrw_main;
#endif /* CLIB_MARCH_VARIANT */
static char *l2_xcrw_error_strings[] = {
#define _(sym,string) string,
foreach_l2_xcrw_error
#undef _
};
VLIB_NODE_FN (l2_xcrw_node) (vlib_main_t * vm,
vlib_node_runtime_t * node, vlib_frame_t * frame)
{
u32 n_left_from, *from, *to_next;
l2_xcrw_next_t next_index;
l2_xcrw_main_t *xcm = &l2_xcrw_main;
vlib_node_t *n = vlib_get_node (vm, l2_xcrw_node.index);
u32 node_counter_base_index = n->error_heap_index;
vlib_error_main_t *em = &vm->error_main;
from = vlib_frame_vector_args (frame);
n_left_from = frame->n_vectors;
next_index = node->cached_next_index;
while (n_left_from > 0)
{
u32 n_left_to_next;
vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
while (n_left_from >= 4 && n_left_to_next >= 2)
{
u32 bi0, bi1;
vlib_buffer_t *b0, *b1;
u32 next0, next1;
u32 sw_if_index0, sw_if_index1;
l2_xcrw_adjacency_t *adj0, *adj1;
/* Prefetch next iteration. */
{
vlib_buffer_t *p2, *p3;
p2 = vlib_get_buffer (vm, from[2]);
p3 = vlib_get_buffer (vm, from[3]);
vlib_prefetch_buffer_header (p2, LOAD);
vlib_prefetch_buffer_header (p3, LOAD);
clib_prefetch_store (p2->data);
clib_prefetch_store (p3->data);
}
/* speculatively enqueue b0 and b1 to the current next frame */
to_next[0] = bi0 = from[0];
to_next[1] = bi1 = from[1];
from += 2;
to_next += 2;
n_left_from -= 2;
n_left_to_next -= 2;
b0 = vlib_get_buffer (vm, bi0);
b1 = vlib_get_buffer (vm, bi1);
sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX];
adj0 = vec_elt_at_index (xcm->adj_by_sw_if_index, sw_if_index0);
adj1 = vec_elt_at_index (xcm->adj_by_sw_if_index, sw_if_index1);
next0 = adj0->rewrite_header.next_index;
vnet_buffer (b0)->sw_if_index[VLIB_TX] =
adj0->rewrite_header.sw_if_index;
next1 = adj1->rewrite_header.next_index;
vnet_buffer (b1)->sw_if_index[VLIB_TX] =
adj1->rewrite_header.sw_if_index;
em->counters[node_counter_base_index + next1]++;
if (PREDICT_TRUE (next0 > 0))
{
u8 *h0 = vlib_buffer_get_current (b0);
vnet_rewrite_one_header (adj0[0], h0,
adj0->rewrite_header.data_bytes);
vlib_buffer_advance (b0, -adj0->rewrite_header.data_bytes);
em->counters[node_counter_base_index + L2_XCRW_ERROR_FWD]++;
}
if (PREDICT_TRUE (next1 > 0))
{
u8 *h1 = vlib_buffer_get_current (b1);
vnet_rewrite_one_header (adj1[0], h1,
adj1->rewrite_header.data_bytes);
vlib_buffer_advance (b1, -adj1->rewrite_header.data_bytes);
em->counters[node_counter_base_index + L2_XCRW_ERROR_FWD]++;
}
if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)))
{
if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
&& (b0->flags & VLIB_BUFFER_IS_TRACED)))
{
l2_xcrw_trace_t *t =
vlib_add_trace (vm, node, b0, sizeof (*t));
t->next_index = next0;
t->tx_fib_index = adj0->rewrite_header.sw_if_index;
}
if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
&& (b1->flags & VLIB_BUFFER_IS_TRACED)))
{
l2_xcrw_trace_t *t =
vlib_add_trace (vm, node, b1, sizeof (*t));
t->next_index = next1;
t->tx_fib_index = adj1->rewrite_header.sw_if_index;
}
}
/* verify speculative enqueues, maybe switch current next frame */
vlib_validate_buffer_enqueue_x2 (vm, node, next_index,
to_next, n_left_to_next,
bi0, bi1, next0, next1);
}
while (n_left_from > 0 && n_left_to_next > 0)
{
u32 bi0;
vlib_buffer_t *b0;
u32 next0;
u32 sw_if_index0;
l2_xcrw_adjacency_t *adj0;
/* speculatively enqueue b0 to the current next frame */
bi0 = from[0];
to_next[0] = bi0;
from += 1;
to_next += 1;
n_left_from -= 1;
n_left_to_next -= 1;
b0 = vlib_get_buffer (vm, bi0);
sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
adj0 = vec_elt_at_index (xcm->adj_by_sw_if_index, sw_if_index0);
next0 = adj0->rewrite_header.next_index;
vnet_buffer (b0)->sw_if_index[VLIB_TX] =
adj0->rewrite_header.sw_if_index;
if (PREDICT_TRUE (next0 > 0))
{
u8 *h0 = vlib_buffer_get_current (b0);
vnet_rewrite_one_header (adj0[0], h0,
adj0->rewrite_header.data_bytes);
vlib_buffer_advance (b0, -adj0->rewrite_header.data_bytes);
em->counters[node_counter_base_index + L2_XCRW_ERROR_FWD]++;
}
if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
&& (b0->flags & VLIB_BUFFER_IS_TRACED)))
{
l2_xcrw_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t));
t->next_index = next0;
t->tx_fib_index = adj0->rewrite_header.sw_if_index;
}
/* verify speculative enqueue, maybe switch current next frame */
vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
to_next, n_left_to_next,
bi0, next0);
}
vlib_put_next_frame (vm, node, next_index, n_left_to_next);
}
return frame->n_vectors;
}
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (l2_xcrw_node) = {
.name = "l2-xcrw",
.vector_size = sizeof (u32),
.format_trace = format_l2_xcrw_trace,
.type = VLIB_NODE_TYPE_INTERNAL,
.n_errors = ARRAY_LEN(l2_xcrw_error_strings),
.error_strings = l2_xcrw_error_strings,
.n_next_nodes = L2_XCRW_N_NEXT,
/* edit / add dispositions here */
.next_nodes = {
[L2_XCRW_NEXT_DROP] = "error-drop",
},
};
/* *INDENT-ON* */
#ifndef CLIB_MARCH_VARIANT
clib_error_t *
l2_xcrw_init (vlib_main_t * vm)
{
l2_xcrw_main_t *mp = &l2_xcrw_main;
mp->vlib_main = vm;
mp->vnet_main = &vnet_main;
mp->tunnel_index_by_l2_sw_if_index = hash_create (0, sizeof (uword));
return 0;
}
VLIB_INIT_FUNCTION (l2_xcrw_init);
static u8 *
format_xcrw_name (u8 * s, va_list * args)
{
u32 dev_instance = va_arg (*args, u32);
return format (s, "xcrw%d", dev_instance);
}
/* *INDENT-OFF* */
VNET_DEVICE_CLASS (xcrw_device_class,static) = {
.name = "Xcrw",
.format_device_name = format_xcrw_name,
};
/* *INDENT-ON* */
/* Create a sham tunnel interface and return its sw_if_index */
static u32
create_xcrw_interface (vlib_main_t * vm)
{
vnet_main_t *vnm = vnet_get_main ();
vnet_eth_interface_registration_t eir = {};
static u32 instance;
u8 address[6];
u32 hw_if_index;
vnet_hw_interface_t *hi;
u32 sw_if_index;
/* mac address doesn't really matter */
clib_memset (address, 0, sizeof (address));
address[2] = 0x12;
eir.dev_class_index = xcrw_device_class.index;
eir.dev_instance = instance++, eir.address = address;
hw_if_index = vnet_eth_register_interface (vnm, &eir);
hi = vnet_get_hw_interface (vnm, hw_if_index);
sw_if_index = hi->sw_if_index;
vnet_sw_interface_set_flags (vnm, sw_if_index,
VNET_SW_INTERFACE_FLAG_ADMIN_UP);
/* Output to the sham tunnel invokes the encap node */
hi->output_node_index = l2_xcrw_node.index;
return sw_if_index;
}
int
vnet_configure_l2_xcrw (vlib_main_t * vm, vnet_main_t * vnm,
u32 l2_sw_if_index, u32 tx_fib_index,
u8 * rewrite, u32 next_node_index, int is_add)
{
l2_xcrw_main_t *xcm = &l2_xcrw_main;
l2_xcrw_adjacency_t *a;
l2_xcrw_tunnel_t *t;
uword *p;
if (is_add)
{
pool_get (xcm->tunnels, t);
/* No interface allocated? Do it. Otherwise, set admin up */
if (t->tunnel_sw_if_index == 0)
t->tunnel_sw_if_index = create_xcrw_interface (vm);
else
vnet_sw_interface_set_flags (vnm, t->tunnel_sw_if_index,
VNET_SW_INTERFACE_FLAG_ADMIN_UP);
t->l2_sw_if_index = l2_sw_if_index;
vec_validate (xcm->adj_by_sw_if_index, t->l2_sw_if_index);
a = vec_elt_at_index (xcm->adj_by_sw_if_index, t->l2_sw_if_index);
clib_memset (a, 0, sizeof (*a));
a->rewrite_header.sw_if_index = tx_fib_index;
/*
* Add or find a dynamic disposition for the successor node,
* e.g. so we can ship pkts to mpls_post_rewrite...
*/
a->rewrite_header.next_index =
vlib_node_add_next (vm, l2_xcrw_node.index, next_node_index);
if (vec_len (rewrite))
vnet_rewrite_set_data (a[0], rewrite, vec_len (rewrite));
set_int_l2_mode (vm, vnm, MODE_L2_XC, t->l2_sw_if_index, 0,
L2_BD_PORT_TYPE_NORMAL, 0, t->tunnel_sw_if_index);
hash_set (xcm->tunnel_index_by_l2_sw_if_index,
t->l2_sw_if_index, t - xcm->tunnels);
return 0;
}
else
{
p = hash_get (xcm->tunnel_index_by_l2_sw_if_index, l2_sw_if_index);
if (p == 0)
return VNET_API_ERROR_INVALID_SW_IF_INDEX;
t = pool_elt_at_index (xcm->tunnels, p[0]);
a = vec_elt_at_index (xcm->adj_by_sw_if_index, t->l2_sw_if_index);
/* Reset adj to drop traffic */
clib_memset (a, 0, sizeof (*a));
set_int_l2_mode (vm, vnm, MODE_L3, t->l2_sw_if_index, 0,
L2_BD_PORT_TYPE_NORMAL, 0, 0);
vnet_sw_interface_set_flags (vnm, t->tunnel_sw_if_index, 0 /* down */ );
hash_unset (xcm->tunnel_index_by_l2_sw_if_index, l2_sw_if_index);
pool_put (xcm->tunnels, t);
}
return 0;
}
static clib_error_t *
set_l2_xcrw_command_fn (vlib_main_t * vm,
unformat_input_t * input, vlib_cli_command_t * cmd)
{
unformat_input_t _line_input, *line_input = &_line_input;
int is_add = 1;
int is_ipv6 = 0; /* for fib id -> fib index mapping */
u32 tx_fib_id = ~0;
u32 tx_fib_index = ~0;
u32 next_node_index = ~0;
u32 l2_sw_if_index;
u8 *rw = 0;
vnet_main_t *vnm = vnet_get_main ();
int rv;
clib_error_t *error = NULL;
if (!unformat_user (input, unformat_line_input, line_input))
return 0;
if (!unformat (line_input, "%U",
unformat_vnet_sw_interface, vnm, &l2_sw_if_index))
{
error = clib_error_return (0, "unknown input '%U'",
format_unformat_error, line_input);
goto done;
}
while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
{
if (unformat (line_input, "next %U",
unformat_vlib_node, vm, &next_node_index))
;
else if (unformat (line_input, "tx-fib-id %d", &tx_fib_id))
;
else if (unformat (line_input, "del"))
is_add = 0;
else if (unformat (line_input, "ipv6"))
is_ipv6 = 1;
else if (unformat (line_input, "rw %U", unformat_hex_string, &rw));
else
break;
}
if (next_node_index == ~0)
{
error = clib_error_return (0, "next node not specified");
goto done;
}
if (tx_fib_id != ~0)
{
uword *p;
if (is_ipv6)
p = hash_get (ip6_main.fib_index_by_table_id, tx_fib_id);
else
p = hash_get (ip4_main.fib_index_by_table_id, tx_fib_id);
if (p == 0)
{
error =
clib_error_return (0, "nonexistent tx_fib_id %d", tx_fib_id);
goto done;
}
tx_fib_index = p[0];
}
rv = vnet_configure_l2_xcrw (vm, vnm, l2_sw_if_index, tx_fib_index,
rw, next_node_index, is_add);
switch (rv)
{
case 0:
break;
case VNET_API_ERROR_INVALID_SW_IF_INDEX:
error = clib_error_return (0, "%U not cross-connected",
format_vnet_sw_if_index_name,
vnm, l2_sw_if_index);
goto done;
default:
error = clib_error_return (0, "vnet_configure_l2_xcrw returned %d", rv);
goto done;
}
done:
vec_free (rw);
unformat_free (line_input);
return error;
}
/*?
* Add or delete a Layer 2 to Layer 3 rewrite cross-connect. This is
* used to hook Layer 2 interface(s) up to the Layer 3 stack in
* arbitrary ways. For example, cross-connect an L2 interface or
* (future) bridge to an mpls-o-gre tunnel. Set up the L2 rewrite
* string as shown in mpls_gre_rewrite, and use \"mpls-post-rewrite\"
* to fix the GRE IP header checksum and length fields.
*
* @cliexpar
* @todo This is incomplete. This needs a detailed description and a
* practical example.
?*/
/* *INDENT-OFF* */
VLIB_CLI_COMMAND (set_l2_xcrw_command, static) = {
.path = "set interface l2 xcrw",
.short_help =
"set interface l2 xcrw <interface> next <node-name>\n"
" [del] [tx-fib-id <id>] [ipv6] rw <hex-bytes>",
.function = set_l2_xcrw_command_fn,
};
/* *INDENT-ON* */
#endif /* CLIB_MARCH_VARIANT */
static u8 *
format_l2xcrw (u8 * s, va_list * args)
{
vnet_main_t *vnm = va_arg (*args, vnet_main_t *);
l2_xcrw_tunnel_t *t = va_arg (*args, l2_xcrw_tunnel_t *);
l2_xcrw_main_t *xcm = &l2_xcrw_main;
vlib_main_t *vm = vlib_get_main ();
l2_xcrw_adjacency_t *a;
u8 *rewrite_string;
if (t == 0)
{
s = format (s, "%-25s%s", "L2 interface", "Tunnel Details");
return s;
}
s = format (s, "%-25U %U ",
format_vnet_sw_if_index_name, vnm, t->l2_sw_if_index,
format_vnet_sw_if_index_name, vnm, t->tunnel_sw_if_index);
a = vec_elt_at_index (xcm->adj_by_sw_if_index, t->l2_sw_if_index);
s = format (s, "next %U ",
format_vlib_next_node_name, vm, l2_xcrw_node.index,
a->rewrite_header.next_index);
if (a->rewrite_header.sw_if_index != ~0)
s = format (s, "tx fib index %d ", a->rewrite_header.sw_if_index);
if (a->rewrite_header.data_bytes)
{
rewrite_string = (u8 *) (a + 1);
rewrite_string -= a->rewrite_header.data_bytes;
s = format (s, "rewrite data: %U ",
format_hex_bytes, rewrite_string,
a->rewrite_header.data_bytes);
}
s = format (s, "\n");
return s;
}
static clib_error_t *
show_l2xcrw_command_fn (vlib_main_t * vm,
unformat_input_t * input, vlib_cli_command_t * cmd)
{
vnet_main_t *vnm = vnet_get_main ();
l2_xcrw_main_t *xcm = &l2_xcrw_main;
l2_xcrw_tunnel_t *t;
if (pool_elts (xcm->tunnels) == 0)
{
vlib_cli_output (vm, "No L2 / L3 rewrite cross-connects configured");
return 0;
}
vlib_cli_output (vm, "%U", format_l2xcrw, 0, 0);
/* *INDENT-OFF* */
pool_foreach (t, xcm->tunnels)
{
vlib_cli_output (vm, "%U", format_l2xcrw, vnm, t);
}
/* *INDENT-ON* */
return 0;
}
/*?
* Display a Layer 2 to Layer 3 rewrite cross-connect. This is used to
* hook Layer 2 interface(s) up to the Layer 3 stack in arbitrary ways.
*
* @todo This is incomplete. This needs a detailed description and a
* practical example.
?*/
/* *INDENT-OFF* */
VLIB_CLI_COMMAND (show_l2xcrw_command, static) = {
.path = "show l2xcrw",
.short_help = "show l2xcrw",
.function = show_l2xcrw_command_fn,
};
/* *INDENT-ON* */
/*
* fd.io coding-style-patch-verification: ON
*
* Local Variables:
* eval: (c-set-style "gnu")
* End:
*/