blob: 572d7f0d156283e349c9de8be644119b0a1e7565 [file] [log] [blame]
/*
* Copyright (c) 2016 Cisco and/or its affiliates.
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at:
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <vnet/fib/ip6_fib.h>
#include <vnet/fib/ip4_fib.h>
#include <vnet/fib/mpls_fib.h>
#include <vnet/adj/adj.h>
#include <vnet/dpo/load_balance.h>
#include <vnet/dpo/load_balance_map.h>
#include <vnet/dpo/mpls_label_dpo.h>
#include <vnet/dpo/lookup_dpo.h>
#include <vnet/dpo/drop_dpo.h>
#include <vnet/dpo/receive_dpo.h>
#include <vnet/dpo/ip_null_dpo.h>
#include <vnet/bfd/bfd_main.h>
#include <vnet/dpo/interface_rx_dpo.h>
#include <vnet/dpo/replicate_dpo.h>
#include <vnet/mpls/mpls.h>
#include <vnet/fib/fib_path_list.h>
#include <vnet/fib/fib_entry_src.h>
#include <vnet/fib/fib_walk.h>
#include <vnet/fib/fib_node_list.h>
#include <vnet/fib/fib_urpf_list.h>
/*
* Add debugs for passing tests
*/
static int fib_test_do_debug;
#define FIB_TEST_I(_cond, _comment, _args...) \
({ \
int _evald = (_cond); \
if (!(_evald)) { \
fformat(stderr, "FAIL:%d: " _comment "\n", \
__LINE__, ##_args); \
} else { \
if (fib_test_do_debug) \
fformat(stderr, "PASS:%d: " _comment "\n", \
__LINE__, ##_args); \
} \
_evald; \
})
#define FIB_TEST(_cond, _comment, _args...) \
{ \
if (!FIB_TEST_I(_cond, _comment, ##_args)) { \
return 1; \
ASSERT(!("FAIL: " _comment)); \
} \
}
/**
* A 'i'm not fussed is this is not efficient' store of test data
*/
typedef struct test_main_t_ {
/**
* HW if indicies
*/
u32 hw_if_indicies[4];
/**
* HW interfaces
*/
vnet_hw_interface_t * hw[4];
} test_main_t;
static test_main_t test_main;
/* fake ethernet device class, distinct from "fake-ethX" */
static u8 * format_test_interface_name (u8 * s, va_list * args)
{
u32 dev_instance = va_arg (*args, u32);
return format (s, "test-eth%d", dev_instance);
}
static uword dummy_interface_tx (vlib_main_t * vm,
vlib_node_runtime_t * node,
vlib_frame_t * frame)
{
clib_warning ("you shouldn't be here, leaking buffers...");
return frame->n_vectors;
}
static clib_error_t *
test_interface_admin_up_down (vnet_main_t * vnm,
u32 hw_if_index,
u32 flags)
{
u32 hw_flags = (flags & VNET_SW_INTERFACE_FLAG_ADMIN_UP) ?
VNET_HW_INTERFACE_FLAG_LINK_UP : 0;
vnet_hw_interface_set_flags (vnm, hw_if_index, hw_flags);
return 0;
}
VNET_DEVICE_CLASS (test_interface_device_class,static) = {
.name = "Test interface",
.format_device_name = format_test_interface_name,
.tx_function = dummy_interface_tx,
.admin_up_down_function = test_interface_admin_up_down,
};
static u8 *hw_address;
static int
fib_test_mk_intf (u32 ninterfaces)
{
clib_error_t * error = NULL;
test_main_t *tm = &test_main;
u8 byte;
u32 i;
ASSERT(ninterfaces <= ARRAY_LEN(tm->hw_if_indicies));
for (i=0; i<6; i++)
{
byte = 0xd0+i;
vec_add1(hw_address, byte);
}
for (i = 0; i < ninterfaces; i++)
{
hw_address[5] = i;
error = ethernet_register_interface(vnet_get_main(),
test_interface_device_class.index,
i /* instance */,
hw_address,
&tm->hw_if_indicies[i],
/* flag change */ 0);
FIB_TEST((NULL == error), "ADD interface %d", i);
error = vnet_hw_interface_set_flags(vnet_get_main(),
tm->hw_if_indicies[i],
VNET_HW_INTERFACE_FLAG_LINK_UP);
tm->hw[i] = vnet_get_hw_interface(vnet_get_main(),
tm->hw_if_indicies[i]);
vec_validate (ip4_main.fib_index_by_sw_if_index,
tm->hw[i]->sw_if_index);
vec_validate (ip6_main.fib_index_by_sw_if_index,
tm->hw[i]->sw_if_index);
ip4_main.fib_index_by_sw_if_index[tm->hw[i]->sw_if_index] = 0;
ip6_main.fib_index_by_sw_if_index[tm->hw[i]->sw_if_index] = 0;
error = vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[i]->sw_if_index,
VNET_SW_INTERFACE_FLAG_ADMIN_UP);
FIB_TEST((NULL == error), "UP interface %d", i);
}
/*
* re-eval after the inevitable realloc
*/
for (i = 0; i < ninterfaces; i++)
{
tm->hw[i] = vnet_get_hw_interface(vnet_get_main(),
tm->hw_if_indicies[i]);
}
return (0);
}
#define FIB_TEST_REC_FORW(_rec_prefix, _via_prefix, _bucket) \
{ \
const dpo_id_t *_rec_dpo = fib_entry_contribute_ip_forwarding( \
fib_table_lookup_exact_match(fib_index, (_rec_prefix))); \
const dpo_id_t *_via_dpo = fib_entry_contribute_ip_forwarding( \
fib_table_lookup(fib_index, (_via_prefix))); \
FIB_TEST(!dpo_cmp(_via_dpo, \
load_balance_get_bucket(_rec_dpo->dpoi_index, \
_bucket)), \
"%U is recursive via %U", \
format_fib_prefix, (_rec_prefix), \
format_fib_prefix, _via_prefix); \
}
#define FIB_TEST_LB_BUCKET_VIA_ADJ(_prefix, _bucket, _ai) \
{ \
const dpo_id_t *_dpo = fib_entry_contribute_ip_forwarding( \
fib_table_lookup_exact_match(fib_index, (_prefix))); \
const dpo_id_t *_dpo1 = \
load_balance_get_bucket(_dpo->dpoi_index, _bucket); \
FIB_TEST(DPO_ADJACENCY == _dpo1->dpoi_type, "type is %U", \
format_dpo_type, _dpo1->dpoi_type); \
FIB_TEST((_ai == _dpo1->dpoi_index), \
"%U bucket %d resolves via %U", \
format_fib_prefix, (_prefix), \
_bucket, \
format_dpo_id, _dpo1, 0); \
}
#define FIB_TEST_RPF(_cond, _comment, _args...) \
{ \
if (!FIB_TEST_I(_cond, _comment, ##_args)) { \
return (0); \
} \
}
static int
fib_test_urpf_is_equal (fib_node_index_t fei,
fib_forward_chain_type_t fct,
u32 num, ...)
{
dpo_id_t dpo = DPO_INVALID;
fib_urpf_list_t *urpf;
index_t ui;
va_list ap;
int ii;
va_start(ap, num);
fib_entry_contribute_forwarding(fei, fct, &dpo);
ui = load_balance_get_urpf(dpo.dpoi_index);
urpf = fib_urpf_list_get(ui);
FIB_TEST_RPF(num == vec_len(urpf->furpf_itfs),
"RPF:%U len %d == %d",
format_fib_urpf_list, ui,
num, vec_len(urpf->furpf_itfs));
FIB_TEST_RPF(num == fib_urpf_check_size(ui),
"RPF:%U check-size %d == %d",
format_fib_urpf_list, ui,
num, vec_len(urpf->furpf_itfs));
for (ii = 0; ii < num; ii++)
{
adj_index_t ai = va_arg(ap, adj_index_t);
FIB_TEST_RPF(ai == urpf->furpf_itfs[ii],
"RPF:%d item:%d - %d == %d",
ui, ii, ai, urpf->furpf_itfs[ii]);
FIB_TEST_RPF(fib_urpf_check(ui, ai),
"RPF:%d %d found",
ui, ai);
}
dpo_reset(&dpo);
va_end(ap);
return (1);
}
static u8*
fib_test_build_rewrite (u8 *eth_addr)
{
u8* rewrite = NULL;
vec_validate(rewrite, 13);
memcpy(rewrite, eth_addr, 6);
memcpy(rewrite+6, eth_addr, 6);
return (rewrite);
}
typedef enum fib_test_lb_bucket_type_t_ {
FT_LB_LABEL_O_ADJ,
FT_LB_LABEL_STACK_O_ADJ,
FT_LB_LABEL_O_LB,
FT_LB_O_LB,
FT_LB_SPECIAL,
FT_LB_ADJ,
FT_LB_INTF,
} fib_test_lb_bucket_type_t;
typedef struct fib_test_lb_bucket_t_ {
fib_test_lb_bucket_type_t type;
union
{
struct
{
mpls_eos_bit_t eos;
mpls_label_t label;
u8 ttl;
adj_index_t adj;
} label_o_adj;
struct
{
mpls_eos_bit_t eos;
mpls_label_t label_stack[8];
u8 label_stack_size;
u8 ttl;
adj_index_t adj;
} label_stack_o_adj;
struct
{
mpls_eos_bit_t eos;
mpls_label_t label;
u8 ttl;
index_t lb;
} label_o_lb;
struct
{
index_t adj;
} adj;
struct
{
index_t lb;
} lb;
struct
{
index_t adj;
} special;
};
} fib_test_lb_bucket_t;
typedef enum fib_test_rep_bucket_type_t_ {
FT_REP_LABEL_O_ADJ,
FT_REP_DISP_MFIB_LOOKUP,
FT_REP_INTF,
} fib_test_rep_bucket_type_t;
typedef struct fib_test_rep_bucket_t_ {
fib_test_rep_bucket_type_t type;
union
{
struct
{
mpls_eos_bit_t eos;
mpls_label_t label;
u8 ttl;
adj_index_t adj;
} label_o_adj;
struct
{
adj_index_t adj;
} adj;
};
} fib_test_rep_bucket_t;
#define FIB_TEST_LB(_cond, _comment, _args...) \
{ \
if (!FIB_TEST_I(_cond, _comment, ##_args)) { \
return (0); \
} \
}
int
fib_test_validate_rep_v (const replicate_t *rep,
u16 n_buckets,
va_list ap)
{
const fib_test_rep_bucket_t *exp;
const dpo_id_t *dpo;
int bucket;
FIB_TEST_LB((n_buckets == rep->rep_n_buckets),
"n_buckets = %d", rep->rep_n_buckets);
for (bucket = 0; bucket < n_buckets; bucket++)
{
exp = va_arg(ap, fib_test_rep_bucket_t*);
dpo = replicate_get_bucket_i(rep, bucket);
switch (exp->type)
{
case FT_REP_LABEL_O_ADJ:
{
const mpls_label_dpo_t *mld;
mpls_label_t hdr;
FIB_TEST_LB((DPO_MPLS_LABEL == dpo->dpoi_type),
"bucket %d stacks on %U",
bucket,
format_dpo_type, dpo->dpoi_type);
mld = mpls_label_dpo_get(dpo->dpoi_index);
hdr = clib_net_to_host_u32(mld->mld_hdr[0].label_exp_s_ttl);
FIB_TEST_LB((vnet_mpls_uc_get_label(hdr) ==
exp->label_o_adj.label),
"bucket %d stacks on label %d",
bucket,
exp->label_o_adj.label);
FIB_TEST_LB((vnet_mpls_uc_get_s(hdr) ==
exp->label_o_adj.eos),
"bucket %d stacks on label %d %U",
bucket,
exp->label_o_adj.label,
format_mpls_eos_bit, exp->label_o_adj.eos);
FIB_TEST_LB((DPO_ADJACENCY_INCOMPLETE == mld->mld_dpo.dpoi_type),
"bucket %d label stacks on %U",
bucket,
format_dpo_type, mld->mld_dpo.dpoi_type);
FIB_TEST_LB((exp->label_o_adj.adj == mld->mld_dpo.dpoi_index),
"bucket %d label stacks on adj %d",
bucket,
exp->label_o_adj.adj);
}
break;
case FT_REP_INTF:
FIB_TEST_LB((DPO_INTERFACE_RX == dpo->dpoi_type),
"bucket %d stacks on %U",
bucket,
format_dpo_type, dpo->dpoi_type);
FIB_TEST_LB((exp->adj.adj == dpo->dpoi_index),
"bucket %d stacks on adj %d",
bucket,
exp->adj.adj);
break;
case FT_REP_DISP_MFIB_LOOKUP:
// ASSERT(0);
break;
}
}
return (!0);
}
int
fib_test_validate_lb_v (const load_balance_t *lb,
u16 n_buckets,
va_list ap)
{
const dpo_id_t *dpo;
int bucket;
FIB_TEST_LB((n_buckets == lb->lb_n_buckets), "n_buckets = %d", lb->lb_n_buckets);
for (bucket = 0; bucket < n_buckets; bucket++)
{
const fib_test_lb_bucket_t *exp;
exp = va_arg(ap, fib_test_lb_bucket_t*);
dpo = load_balance_get_bucket_i(lb, bucket);
switch (exp->type)
{
case FT_LB_LABEL_STACK_O_ADJ:
{
const mpls_label_dpo_t *mld;
mpls_label_t hdr;
u32 ii;
FIB_TEST_LB((DPO_MPLS_LABEL == dpo->dpoi_type),
"bucket %d stacks on %U",
bucket,
format_dpo_type, dpo->dpoi_type);
mld = mpls_label_dpo_get(dpo->dpoi_index);
FIB_TEST_LB(exp->label_stack_o_adj.label_stack_size == mld->mld_n_labels,
"label stack size",
mld->mld_n_labels);
for (ii = 0; ii < mld->mld_n_labels; ii++)
{
hdr = clib_net_to_host_u32(mld->mld_hdr[ii].label_exp_s_ttl);
FIB_TEST_LB((vnet_mpls_uc_get_label(hdr) ==
exp->label_stack_o_adj.label_stack[ii]),
"bucket %d stacks on label %d",
bucket,
exp->label_stack_o_adj.label_stack[ii]);
if (ii == mld->mld_n_labels-1)
{
FIB_TEST_LB((vnet_mpls_uc_get_s(hdr) ==
exp->label_o_adj.eos),
"bucket %d stacks on label %d %U!=%U",
bucket,
exp->label_stack_o_adj.label_stack[ii],
format_mpls_eos_bit, exp->label_o_adj.eos,
format_mpls_eos_bit, vnet_mpls_uc_get_s(hdr));
}
else
{
FIB_TEST_LB((vnet_mpls_uc_get_s(hdr) == MPLS_NON_EOS),
"bucket %d stacks on label %d %U",
bucket,
exp->label_stack_o_adj.label_stack[ii],
format_mpls_eos_bit, vnet_mpls_uc_get_s(hdr));
}
}
FIB_TEST_LB((DPO_ADJACENCY_INCOMPLETE == mld->mld_dpo.dpoi_type),
"bucket %d label stacks on %U",
bucket,
format_dpo_type, mld->mld_dpo.dpoi_type);
FIB_TEST_LB((exp->label_stack_o_adj.adj == mld->mld_dpo.dpoi_index),
"bucket %d label stacks on adj %d",
bucket,
exp->label_stack_o_adj.adj);
}
break;
case FT_LB_LABEL_O_ADJ:
{
const mpls_label_dpo_t *mld;
mpls_label_t hdr;
FIB_TEST_LB((DPO_MPLS_LABEL == dpo->dpoi_type),
"bucket %d stacks on %U",
bucket,
format_dpo_type, dpo->dpoi_type);
mld = mpls_label_dpo_get(dpo->dpoi_index);
hdr = clib_net_to_host_u32(mld->mld_hdr[0].label_exp_s_ttl);
FIB_TEST_LB((vnet_mpls_uc_get_label(hdr) ==
exp->label_o_adj.label),
"bucket %d stacks on label %d",
bucket,
exp->label_o_adj.label);
FIB_TEST_LB((vnet_mpls_uc_get_s(hdr) ==
exp->label_o_adj.eos),
"bucket %d stacks on label %d %U",
bucket,
exp->label_o_adj.label,
format_mpls_eos_bit, exp->label_o_adj.eos);
FIB_TEST_LB((DPO_ADJACENCY_INCOMPLETE == mld->mld_dpo.dpoi_type),
"bucket %d label stacks on %U",
bucket,
format_dpo_type, mld->mld_dpo.dpoi_type);
FIB_TEST_LB((exp->label_o_adj.adj == mld->mld_dpo.dpoi_index),
"bucket %d label stacks on adj %d",
bucket,
exp->label_o_adj.adj);
}
break;
case FT_LB_LABEL_O_LB:
{
const mpls_label_dpo_t *mld;
mpls_label_t hdr;
FIB_TEST_LB((DPO_MPLS_LABEL == dpo->dpoi_type),
"bucket %d stacks on %U",
bucket,
format_dpo_type, dpo->dpoi_type);
mld = mpls_label_dpo_get(dpo->dpoi_index);
hdr = clib_net_to_host_u32(mld->mld_hdr[0].label_exp_s_ttl);
FIB_TEST_LB(1 == mld->mld_n_labels, "label stack size",
mld->mld_n_labels);
FIB_TEST_LB((vnet_mpls_uc_get_label(hdr) ==
exp->label_o_lb.label),
"bucket %d stacks on label %d",
bucket,
exp->label_o_lb.label);
FIB_TEST_LB((vnet_mpls_uc_get_s(hdr) ==
exp->label_o_lb.eos),
"bucket %d stacks on label %d %U",
bucket,
exp->label_o_lb.label,
format_mpls_eos_bit, exp->label_o_lb.eos);
FIB_TEST_LB((DPO_LOAD_BALANCE == mld->mld_dpo.dpoi_type),
"bucket %d label stacks on %U",
bucket,
format_dpo_type, mld->mld_dpo.dpoi_type);
FIB_TEST_LB((exp->label_o_lb.lb == mld->mld_dpo.dpoi_index),
"bucket %d label stacks on LB %d",
bucket,
exp->label_o_lb.lb);
}
break;
case FT_LB_ADJ:
FIB_TEST_I(((DPO_ADJACENCY == dpo->dpoi_type) ||
(DPO_ADJACENCY_INCOMPLETE == dpo->dpoi_type)),
"bucket %d stacks on %U",
bucket,
format_dpo_type, dpo->dpoi_type);
FIB_TEST_LB((exp->adj.adj == dpo->dpoi_index),
"bucket %d stacks on adj %d",
bucket,
exp->adj.adj);
break;
case FT_LB_INTF:
FIB_TEST_I((DPO_INTERFACE_RX == dpo->dpoi_type),
"bucket %d stacks on %U",
bucket,
format_dpo_type, dpo->dpoi_type);
FIB_TEST_LB((exp->adj.adj == dpo->dpoi_index),
"bucket %d stacks on adj %d",
bucket,
exp->adj.adj);
break;
case FT_LB_O_LB:
FIB_TEST_I((DPO_LOAD_BALANCE == dpo->dpoi_type),
"bucket %d stacks on %U",
bucket,
format_dpo_type, dpo->dpoi_type);
FIB_TEST_LB((exp->lb.lb == dpo->dpoi_index),
"bucket %d stacks on lb %d not %d",
bucket,
dpo->dpoi_index,
exp->lb.lb);
break;
case FT_LB_SPECIAL:
FIB_TEST_I((DPO_DROP == dpo->dpoi_type),
"bucket %d stacks on %U",
bucket,
format_dpo_type, dpo->dpoi_type);
FIB_TEST_LB((exp->special.adj == dpo->dpoi_index),
"bucket %d stacks on drop %d",
bucket,
exp->special.adj);
break;
}
}
return (!0);
}
int
fib_test_validate_entry (fib_node_index_t fei,
fib_forward_chain_type_t fct,
u16 n_buckets,
...)
{
dpo_id_t dpo = DPO_INVALID;
fib_prefix_t pfx;
index_t fw_lbi;
u32 fib_index;
va_list ap;
int res;
va_start(ap, n_buckets);
fib_entry_get_prefix(fei, &pfx);
fib_index = fib_entry_get_fib_index(fei);
fib_entry_contribute_forwarding(fei, fct, &dpo);
if (DPO_REPLICATE == dpo.dpoi_type)
{
const replicate_t *rep;
rep = replicate_get(dpo.dpoi_index);
res = fib_test_validate_rep_v(rep, n_buckets, ap);
}
else
{
const load_balance_t *lb;
FIB_TEST_LB((DPO_LOAD_BALANCE == dpo.dpoi_type),
"Entry links to %U",
format_dpo_type, dpo.dpoi_type);
lb = load_balance_get(dpo.dpoi_index);
res = fib_test_validate_lb_v(lb, n_buckets, ap);
/*
* ensure that the LB contributed by the entry is the
* same as the LB in the forwarding tables
*/
if (fct == fib_entry_get_default_chain_type(fib_entry_get(fei)))
{
switch (pfx.fp_proto)
{
case FIB_PROTOCOL_IP4:
fw_lbi = ip4_fib_forwarding_lookup(fib_index, &pfx.fp_addr.ip4);
break;
case FIB_PROTOCOL_IP6:
fw_lbi = ip6_fib_table_fwding_lookup(&ip6_main, fib_index, &pfx.fp_addr.ip6);
break;
case FIB_PROTOCOL_MPLS:
{
mpls_unicast_header_t hdr = {
.label_exp_s_ttl = 0,
};
vnet_mpls_uc_set_label(&hdr.label_exp_s_ttl, pfx.fp_label);
vnet_mpls_uc_set_s(&hdr.label_exp_s_ttl, pfx.fp_eos);
hdr.label_exp_s_ttl = clib_host_to_net_u32(hdr.label_exp_s_ttl);
fw_lbi = mpls_fib_table_forwarding_lookup(fib_index, &hdr);
break;
}
default:
fw_lbi = 0;
}
FIB_TEST_LB((fw_lbi == dpo.dpoi_index),
"Contributed LB = FW LB: %U\n %U",
format_load_balance, fw_lbi, 0,
format_load_balance, dpo.dpoi_index, 0);
}
}
dpo_reset(&dpo);
va_end(ap);
return (res);
}
static int
fib_test_v4 (void)
{
/*
* In the default table check for the presence and correct forwarding
* of the special entries
*/
fib_node_index_t dfrt, fei, ai, ai2, locked_ai, ai_01, ai_02, ai_03;
const dpo_id_t *dpo, *dpo1, *dpo2, *dpo_drop;
const ip_adjacency_t *adj;
const load_balance_t *lb;
test_main_t *tm;
u32 fib_index;
int lb_count;
int ii;
/* via 10.10.10.1 */
ip46_address_t nh_10_10_10_1 = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a01),
};
/* via 10.10.10.2 */
ip46_address_t nh_10_10_10_2 = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a02),
};
FIB_TEST((0 == pool_elts(load_balance_map_pool)), "LB-map pool size is %d",
pool_elts(load_balance_map_pool));
tm = &test_main;
/* record the nubmer of load-balances in use before we start */
lb_count = pool_elts(load_balance_pool);
/* Find or create FIB table 11 */
fib_index = fib_table_find_or_create_and_lock(FIB_PROTOCOL_IP4, 11,
FIB_SOURCE_API);
for (ii = 0; ii < 4; ii++)
{
ip4_main.fib_index_by_sw_if_index[tm->hw[ii]->sw_if_index] = fib_index;
}
fib_prefix_t pfx_0_0_0_0_s_0 = {
.fp_len = 0,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4 = {
{0}
},
},
};
fib_prefix_t pfx = {
.fp_len = 0,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4 = {
{0}
},
},
};
dpo_drop = drop_dpo_get(DPO_PROTO_IP4);
dfrt = fib_table_lookup(fib_index, &pfx_0_0_0_0_s_0);
FIB_TEST((FIB_NODE_INDEX_INVALID != dfrt), "default route present");
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(dfrt)),
"Default route is DROP");
pfx.fp_len = 32;
fei = fib_table_lookup(fib_index, &pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "all zeros route present");
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"all 0s route is DROP");
pfx.fp_addr.ip4.as_u32 = clib_host_to_net_u32(0xffffffff);
pfx.fp_len = 32;
fei = fib_table_lookup(fib_index, &pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "all ones route present");
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"all 1s route is DROP");
pfx.fp_addr.ip4.as_u32 = clib_host_to_net_u32(0xe0000000);
pfx.fp_len = 8;
fei = fib_table_lookup(fib_index, &pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "all-mcast route present");
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"all-mcast route is DROP");
pfx.fp_addr.ip4.as_u32 = clib_host_to_net_u32(0xf0000000);
pfx.fp_len = 8;
fei = fib_table_lookup(fib_index, &pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "class-e route present");
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"class-e route is DROP");
/*
* at this stage there are 5 entries in the test FIB (plus 5 in the default),
* all of which are special sourced and so none of which share path-lists.
* There are also 2 entries, and 2 non-shared path-lists, in the v6 default
* table, and 4 path-lists in the v6 MFIB table
*/
#define ENBR (5+5+2)
#define PNBR (5+5+6)
FIB_TEST((0 == fib_path_list_db_size()), "path list DB is empty");
FIB_TEST((PNBR == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* add interface routes.
* validate presence of /24 attached and /32 recieve.
* test for the presence of the receive address in the glean and local adj
*/
fib_prefix_t local_pfx = {
.fp_len = 24,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4 = {
.as_u32 = clib_host_to_net_u32(0x0a0a0a0a),
},
},
};
fib_table_entry_update_one_path(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_ATTACHED),
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1, // weight
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "attached interface route present");
FIB_TEST(((FIB_ENTRY_FLAG_ATTACHED | FIB_ENTRY_FLAG_CONNECTED) ==
fib_entry_get_flags(fei)),
"Flags set on attached interface");
ai = fib_entry_get_adj(fei);
FIB_TEST((FIB_NODE_INDEX_INVALID != ai),
"attached interface route adj present %d", ai);
adj = adj_get(ai);
FIB_TEST((IP_LOOKUP_NEXT_GLEAN == adj->lookup_next_index),
"attached interface adj is glean");
FIB_TEST((0 == ip46_address_cmp(&local_pfx.fp_addr,
&adj->sub_type.glean.receive_addr)),
"attached interface adj is receive ok");
local_pfx.fp_len = 32;
fib_table_entry_update_one_path(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_LOCAL),
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1, // weight
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &local_pfx);
FIB_TEST(((FIB_ENTRY_FLAG_LOCAL | FIB_ENTRY_FLAG_CONNECTED) ==
fib_entry_get_flags(fei)),
"Flags set on local interface");
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local interface route present");
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 0),
"RPF list for local length 0");
dpo = load_balance_get_bucket(dpo->dpoi_index, 0);
FIB_TEST((DPO_RECEIVE == dpo->dpoi_type),
"local interface adj is local");
receive_dpo_t *rd = receive_dpo_get(dpo->dpoi_index);
FIB_TEST((0 == ip46_address_cmp(&local_pfx.fp_addr,
&rd->rd_addr)),
"local interface adj is receive ok");
FIB_TEST((2 == fib_table_get_num_entries(fib_index,
FIB_PROTOCOL_IP4,
FIB_SOURCE_INTERFACE)),
"2 Interface Source'd prefixes");
/*
* +2 interface routes +2 non-shared path-lists
*/
FIB_TEST((0 == fib_path_list_db_size()), "path list DB is empty");
FIB_TEST((PNBR+2 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENBR+2 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Modify the default route to be via an adj not yet known.
* this sources the defalut route with the API source, which is
* a higher preference to the DEFAULT_ROUTE source
*/
pfx.fp_addr.ip4.as_u32 = 0;
pfx.fp_len = 0;
fib_table_entry_path_add(fib_index, &pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx);
FIB_TEST((FIB_ENTRY_FLAG_NONE == fib_entry_get_flags(fei)),
"Flags set on API route");
FIB_TEST((fei == dfrt), "default route same index");
ai = fib_entry_get_adj(fei);
FIB_TEST((FIB_NODE_INDEX_INVALID != ai), "default route adj present");
adj = adj_get(ai);
FIB_TEST((IP_LOOKUP_NEXT_ARP == adj->lookup_next_index),
"adj is incomplete");
FIB_TEST((0 == ip46_address_cmp(&nh_10_10_10_1, &adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
FIB_TEST((1 == fib_table_get_num_entries(fib_index,
FIB_PROTOCOL_IP4,
FIB_SOURCE_API)),
"1 API Source'd prefixes");
/*
* find the adj in the shared db
*/
locked_ai = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index);
FIB_TEST((locked_ai == ai), "ADJ NBR DB find");
adj_unlock(locked_ai);
/*
* +1 shared path-list
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+3 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENBR+2 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* remove the API source from the default route. We expected
* the route to remain, sourced by DEFAULT_ROUTE, and hence a DROP
*/
pfx.fp_addr.ip4.as_u32 = 0;
pfx.fp_len = 0;
fib_table_entry_path_remove(fib_index, &pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // non-recursive path, so no FIB index
1,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx);
FIB_TEST((fei == dfrt), "default route same index");
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"Default route is DROP");
/*
* -1 shared-path-list
*/
FIB_TEST((0 == fib_path_list_db_size()), "path list DB is empty");
FIB_TEST((PNBR+2 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENBR+2 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Add an 2 ARP entry => a complete ADJ plus adj-fib.
*/
fib_prefix_t pfx_10_10_10_1_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 10.10.10.1 */
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a01),
},
};
fib_prefix_t pfx_10_10_10_2_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 10.10.10.2 */
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a02),
},
};
fib_prefix_t pfx_11_11_11_11_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 11.11.11.11 */
.ip4.as_u32 = clib_host_to_net_u32(0x0b0b0b0b),
},
};
u8 eth_addr[] = {
0xde, 0xde, 0xde, 0xba, 0xba, 0xba,
};
ip46_address_t nh_12_12_12_12 = {
.ip4.as_u32 = clib_host_to_net_u32(0x0c0c0c0c),
};
adj_index_t ai_12_12_12_12;
/*
* Add a route via an incomplete ADJ. then complete the ADJ
* Expect the route LB is updated to use complete adj type.
*/
fei = fib_table_entry_update_one_path(fib_index,
&pfx_11_11_11_11_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&pfx_10_10_10_1_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
dpo = fib_entry_contribute_ip_forwarding(fei);
dpo1 = load_balance_get_bucket(dpo->dpoi_index, 0);
FIB_TEST(DPO_ADJACENCY_INCOMPLETE == dpo1->dpoi_type,
"11.11.11.11/32 via incomplete adj");
ai_01 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&pfx_10_10_10_1_s_32.fp_addr,
tm->hw[0]->sw_if_index);
FIB_TEST((FIB_NODE_INDEX_INVALID != ai_01), "adj created");
adj = adj_get(ai_01);
FIB_TEST((IP_LOOKUP_NEXT_ARP == adj->lookup_next_index),
"adj is incomplete");
FIB_TEST((0 == ip46_address_cmp(&pfx_10_10_10_1_s_32.fp_addr,
&adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
adj_nbr_update_rewrite(ai_01, ADJ_NBR_REWRITE_FLAG_COMPLETE,
fib_test_build_rewrite(eth_addr));
FIB_TEST((IP_LOOKUP_NEXT_REWRITE == adj->lookup_next_index),
"adj is complete");
FIB_TEST((0 == ip46_address_cmp(&pfx_10_10_10_1_s_32.fp_addr,
&adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "ADJ-FIB resolves via adj");
dpo = fib_entry_contribute_ip_forwarding(fei);
dpo1 = load_balance_get_bucket(dpo->dpoi_index, 0);
FIB_TEST(DPO_ADJACENCY == dpo1->dpoi_type,
"11.11.11.11/32 via complete adj");
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 1,
tm->hw[0]->sw_if_index),
"RPF list for adj-fib contains adj");
ai_12_12_12_12 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nh_12_12_12_12,
tm->hw[1]->sw_if_index);
FIB_TEST((FIB_NODE_INDEX_INVALID != ai_12_12_12_12), "adj created");
adj = adj_get(ai_12_12_12_12);
FIB_TEST((IP_LOOKUP_NEXT_ARP == adj->lookup_next_index),
"adj is incomplete");
FIB_TEST((0 == ip46_address_cmp(&nh_12_12_12_12,
&adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
adj_nbr_update_rewrite(ai_12_12_12_12, ADJ_NBR_REWRITE_FLAG_COMPLETE,
fib_test_build_rewrite(eth_addr));
FIB_TEST((IP_LOOKUP_NEXT_REWRITE == adj->lookup_next_index),
"adj is complete");
/*
* add the adj fib
*/
fei = fib_table_entry_path_add(fib_index,
&pfx_10_10_10_1_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&pfx_10_10_10_1_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST((FIB_ENTRY_FLAG_ATTACHED == fib_entry_get_flags(fei)),
"Flags set on adj-fib");
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "ADJ-FIB resolves via adj, %d", ai);
fib_table_entry_path_remove(fib_index,
&pfx_11_11_11_11_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_10_10_10_1_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
FIB_ROUTE_PATH_FLAG_NONE);
eth_addr[5] = 0xb2;
ai_02 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&pfx_10_10_10_2_s_32.fp_addr,
tm->hw[0]->sw_if_index);
FIB_TEST((FIB_NODE_INDEX_INVALID != ai_02), "adj created");
adj = adj_get(ai_02);
FIB_TEST((IP_LOOKUP_NEXT_ARP == adj->lookup_next_index),
"adj is incomplete");
FIB_TEST((0 == ip46_address_cmp(&pfx_10_10_10_2_s_32.fp_addr,
&adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
adj_nbr_update_rewrite(ai_02, ADJ_NBR_REWRITE_FLAG_COMPLETE,
fib_test_build_rewrite(eth_addr));
FIB_TEST((IP_LOOKUP_NEXT_REWRITE == adj->lookup_next_index),
"adj is complete");
FIB_TEST((0 == ip46_address_cmp(&pfx_10_10_10_2_s_32.fp_addr,
&adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
FIB_TEST((ai_01 != ai_02), "ADJs are different");
fib_table_entry_path_add(fib_index,
&pfx_10_10_10_2_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&pfx_10_10_10_2_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_10_10_10_2_s_32);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_02 == ai), "ADJ-FIB resolves via adj");
/*
* +2 adj-fibs, and their non-shared path-lists
*/
FIB_TEST((0 == fib_path_list_db_size()), "path list DB is empty");
FIB_TEST((PNBR+4 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+4 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Add 2 routes via the first ADJ. ensure path-list sharing
*/
fib_prefix_t pfx_1_1_1_1_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 1.1.1.1/32 */
.ip4.as_u32 = clib_host_to_net_u32(0x01010101),
},
};
fib_table_entry_path_add(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_1_1_1_1_s_32);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "1.1.1.1 resolves via 10.10.10.1");
/*
* +1 entry and a shared path-list
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB is empty");
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+5 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/* 1.1.2.0/24 */
fib_prefix_t pfx_1_1_2_0_s_24 = {
.fp_len = 24,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x01010200),
}
};
fib_table_entry_path_add(fib_index,
&pfx_1_1_2_0_s_24,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_1_1_2_0_s_24);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "1.1.2.0/24 resolves via 10.10.10.1");
/*
* +1 entry only
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB is empty");
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+6 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* modify 1.1.2.0/24 to use multipath.
*/
fib_table_entry_path_add(fib_index,
&pfx_1_1_2_0_s_24,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_1_1_2_0_s_24);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1, tm->hw[0]->sw_if_index),
"RPF list for 1.1.2.0/24 contains both adjs");
dpo1 = load_balance_get_bucket(dpo->dpoi_index, 0);
FIB_TEST(DPO_ADJACENCY == dpo1->dpoi_type, "type is %d", dpo1->dpoi_type);
FIB_TEST((ai_01 == dpo1->dpoi_index),
"1.1.2.0/24 bucket 0 resolves via 10.10.10.1 (%d=%d)",
ai_01, dpo1->dpoi_index);
dpo1 = load_balance_get_bucket(dpo->dpoi_index, 1);
FIB_TEST(DPO_ADJACENCY == dpo1->dpoi_type, "type is %d", dpo1->dpoi_type);
FIB_TEST((ai_02 == dpo1->dpoi_index),
"1.1.2.0/24 bucket 1 resolves via 10.10.10.2");
/*
* +1 shared-pathlist
*/
FIB_TEST((2 == fib_path_list_db_size()), "path list DB is empty");
FIB_TEST((PNBR+6 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+6 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* revert the modify
*/
fib_table_entry_path_remove(fib_index,
&pfx_1_1_2_0_s_24,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0,
1,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_1_1_2_0_s_24);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1, tm->hw[0]->sw_if_index),
"RPF list for 1.1.2.0/24 contains one adj");
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "1.1.2.0/24 resolves via 10.10.10.1");
/*
* +1 shared-pathlist
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB is %d",
fib_path_list_db_size());
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+6 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Add 2 recursive routes:
* 100.100.100.100/32 via 1.1.1.1/32 => the via entry is installed.
* 100.100.100.101/32 via 1.1.1.1/32 => the via entry is installed.
*/
fib_prefix_t bgp_100_pfx = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 100.100.100.100/32 */
.ip4.as_u32 = clib_host_to_net_u32(0x64646464),
},
};
/* via 1.1.1.1 */
ip46_address_t nh_1_1_1_1 = {
.ip4.as_u32 = clib_host_to_net_u32(0x01010101),
};
fei = fib_table_entry_path_add(fib_index,
&bgp_100_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_1_1_1_1,
~0, // no index provided.
fib_index, // nexthop in same fib as route
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST_REC_FORW(&bgp_100_pfx, &pfx_1_1_1_1_s_32, 0);
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 1,
tm->hw[0]->sw_if_index),
"RPF list for adj-fib contains adj");
/*
* +1 entry and +1 shared-path-list
*/
FIB_TEST((2 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+6 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+7 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
fib_prefix_t bgp_101_pfx = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 100.100.100.101/32 */
.ip4.as_u32 = clib_host_to_net_u32(0x64646465),
},
};
fib_table_entry_path_add(fib_index,
&bgp_101_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_1_1_1_1,
~0, // no index provided.
fib_index, // nexthop in same fib as route
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST_REC_FORW(&bgp_101_pfx, &pfx_1_1_1_1_s_32, 0);
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 1,
tm->hw[0]->sw_if_index),
"RPF list for adj-fib contains adj");
/*
* +1 entry, but the recursive path-list is shared.
*/
FIB_TEST((2 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+6 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+8 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* An special route; one where the user (me) provides the
* adjacency through which the route will resovle by setting the flags
*/
fib_prefix_t ex_pfx = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 4.4.4.4/32 */
.ip4.as_u32 = clib_host_to_net_u32(0x04040404),
},
};
fib_table_entry_special_add(fib_index,
&ex_pfx,
FIB_SOURCE_SPECIAL,
FIB_ENTRY_FLAG_LOCAL);
fei = fib_table_lookup_exact_match(fib_index, &ex_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
dpo = load_balance_get_bucket(dpo->dpoi_index, 0);
FIB_TEST((DPO_RECEIVE == dpo->dpoi_type),
"local interface adj is local");
fib_table_entry_special_remove(fib_index,
&ex_pfx,
FIB_SOURCE_SPECIAL);
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &ex_pfx),
"Exclusive reoute removed");
/*
* An EXCLUSIVE route; one where the user (me) provides the exclusive
* adjacency through which the route will resovle
*/
dpo_id_t ex_dpo = DPO_INVALID;
lookup_dpo_add_or_lock_w_fib_index(fib_index,
DPO_PROTO_IP4,
LOOKUP_UNICAST,
LOOKUP_INPUT_DST_ADDR,
LOOKUP_TABLE_FROM_CONFIG,
&ex_dpo);
fib_table_entry_special_dpo_add(fib_index,
&ex_pfx,
FIB_SOURCE_SPECIAL,
FIB_ENTRY_FLAG_EXCLUSIVE,
&ex_dpo);
fei = fib_table_lookup_exact_match(fib_index, &ex_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(&ex_dpo, load_balance_get_bucket(dpo->dpoi_index, 0)),
"exclusive remote uses lookup DPO");
/*
* update the exclusive to use a different DPO
*/
ip_null_dpo_add_and_lock(DPO_PROTO_IP4,
IP_NULL_ACTION_SEND_ICMP_UNREACH,
&ex_dpo);
fib_table_entry_special_dpo_update(fib_index,
&ex_pfx,
FIB_SOURCE_SPECIAL,
FIB_ENTRY_FLAG_EXCLUSIVE,
&ex_dpo);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(&ex_dpo, load_balance_get_bucket(dpo->dpoi_index, 0)),
"exclusive remote uses now uses NULL DPO");
fib_table_entry_special_remove(fib_index,
&ex_pfx,
FIB_SOURCE_SPECIAL);
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &ex_pfx),
"Exclusive reoute removed");
dpo_reset(&ex_dpo);
/*
* Add a recursive route:
* 200.200.200.200/32 via 1.1.1.2/32 => the via entry is NOT installed.
*/
fib_prefix_t bgp_200_pfx = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 200.200.200.200/32 */
.ip4.as_u32 = clib_host_to_net_u32(0xc8c8c8c8),
},
};
/* via 1.1.1.2 */
fib_prefix_t pfx_1_1_1_2_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x01010102),
},
};
fei = fib_table_entry_path_add(fib_index,
&bgp_200_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_2_s_32.fp_addr,
~0, // no index provided.
fib_index, // nexthop in same fib as route
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"Recursive via unresolved is drop");
/*
* the adj should be recursive via drop, since the route resolves via
* the default route, which is itself a DROP
*/
fei = fib_table_lookup(fib_index, &pfx_1_1_1_2_s_32);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(load_balance_is_drop(dpo1), "1.1.1.2/32 is drop");
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 0),
"RPF list for 1.1.1.2/32 contains 0 adjs");
/*
* +2 entry and +1 shared-path-list
*/
FIB_TEST((3 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+7 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+10 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Unequal Cost load-balance. 3:1 ratio. fits in a 4 bucket LB
* The paths are sort by NH first. in this case the the path with greater
* weight is first in the set. This ordering is to test the RPF sort|uniq logic
*/
fib_prefix_t pfx_1_2_3_4_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x01020304),
},
};
fib_table_entry_path_add(fib_index,
&pfx_1_2_3_4_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_entry_path_add(fib_index,
&pfx_1_2_3_4_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_12_12_12_12,
tm->hw[1]->sw_if_index,
~0,
3,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "1.2.3.4/32 presnet");
dpo = fib_entry_contribute_ip_forwarding(fei);
lb = load_balance_get(dpo->dpoi_index);
FIB_TEST((lb->lb_n_buckets == 4),
"1.2.3.4/32 LB has %d bucket",
lb->lb_n_buckets);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_4_s_32, 0, ai_12_12_12_12);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_4_s_32, 1, ai_12_12_12_12);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_4_s_32, 2, ai_12_12_12_12);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_4_s_32, 3, ai_01);
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 2,
tm->hw[0]->sw_if_index,
tm->hw[1]->sw_if_index),
"RPF list for 1.2.3.4/32 contains both adjs");
/*
* Unequal Cost load-balance. 4:1 ratio.
* fits in a 16 bucket LB with ratio 13:3
*/
fib_prefix_t pfx_1_2_3_5_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x01020305),
},
};
fib_table_entry_path_add(fib_index,
&pfx_1_2_3_5_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_12_12_12_12,
tm->hw[1]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_entry_path_add(fib_index,
&pfx_1_2_3_5_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
4,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "1.2.3.5/32 presnet");
dpo = fib_entry_contribute_ip_forwarding(fei);
lb = load_balance_get(dpo->dpoi_index);
FIB_TEST((lb->lb_n_buckets == 16),
"1.2.3.5/32 LB has %d bucket",
lb->lb_n_buckets);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 0, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 1, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 2, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 3, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 4, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 5, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 6, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 7, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 8, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 9, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 10, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 11, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 12, ai_01);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 13, ai_12_12_12_12);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 14, ai_12_12_12_12);
FIB_TEST_LB_BUCKET_VIA_ADJ(&pfx_1_2_3_5_s_32, 15, ai_12_12_12_12);
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 2,
tm->hw[0]->sw_if_index,
tm->hw[1]->sw_if_index),
"RPF list for 1.2.3.4/32 contains both adjs");
/*
* Test UCMP with a large weight skew - this produces load-balance objects with large
* numbers of buckets to accommodate the skew. By updating said load-balances we are
* laso testing the LB in placce modify code when number of buckets is large.
*/
fib_prefix_t pfx_6_6_6_6_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 1.1.1.1/32 */
.ip4.as_u32 = clib_host_to_net_u32(0x06060606),
},
};
fib_test_lb_bucket_t ip_o_10_10_10_1 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_01,
},
};
fib_test_lb_bucket_t ip_o_10_10_10_2 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_02,
},
};
fib_test_lb_bucket_t ip_6_6_6_6_o_12_12_12_12 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_12_12_12_12,
},
};
fib_table_entry_update_one_path(fib_index,
&pfx_6_6_6_6_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
0, // zero weigth
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_6_6_6_6_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_10_10_10_1),
"6.6.6.6/32 via 10.10.10.1");
fib_table_entry_path_add(fib_index,
&pfx_6_6_6_6_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
100,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_6_6_6_6_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
64,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_1),
"6.6.6.6/32 via 10.10.10.1 and 10.10.10.2 in 63:1 ratio");
fib_table_entry_path_add(fib_index,
&pfx_6_6_6_6_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_12_12_12_12,
tm->hw[1]->sw_if_index,
~0, // invalid fib index
100,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_6_6_6_6_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
128,
&ip_o_10_10_10_1,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12,
&ip_6_6_6_6_o_12_12_12_12),
"6.6.6.6/32 via 10.10.10.1 and 10.10.10.2 in 63:1 ratio");
fib_table_entry_path_remove(fib_index,
&pfx_6_6_6_6_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_12_12_12_12,
tm->hw[1]->sw_if_index,
~0, // invalid fib index
100,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_6_6_6_6_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
64,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_2,
&ip_o_10_10_10_1),
"6.6.6.6/32 via 10.10.10.1 and 10.10.10.2 in 63:1 ratio");
fib_table_entry_path_remove(fib_index,
&pfx_6_6_6_6_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
100,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_6_6_6_6_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_10_10_10_1),
"6.6.6.6/32 via 10.10.10.1");
fib_table_entry_delete(fib_index, &pfx_6_6_6_6_s_32, FIB_SOURCE_API);
/*
* A recursive via the two unequal cost entries
*/
fib_prefix_t bgp_44_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 200.200.200.201/32 */
.ip4.as_u32 = clib_host_to_net_u32(0x44444444),
},
};
fei = fib_table_entry_path_add(fib_index,
&bgp_44_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_2_3_4_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_entry_path_add(fib_index,
&bgp_44_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_2_3_5_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST_REC_FORW(&bgp_44_s_32, &pfx_1_2_3_4_s_32, 0);
FIB_TEST_REC_FORW(&bgp_44_s_32, &pfx_1_2_3_5_s_32, 1);
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 2,
tm->hw[0]->sw_if_index,
tm->hw[1]->sw_if_index),
"RPF list for 1.2.3.4/32 contains both adjs");
/*
* test the uRPF check functions
*/
dpo_id_t dpo_44 = DPO_INVALID;
index_t urpfi;
fib_entry_contribute_forwarding(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, &dpo_44);
urpfi = load_balance_get_urpf(dpo_44.dpoi_index);
FIB_TEST(fib_urpf_check(urpfi, tm->hw[0]->sw_if_index),
"uRPF check for 68.68.68.68/32 on %d OK",
tm->hw[0]->sw_if_index);
FIB_TEST(fib_urpf_check(urpfi, tm->hw[1]->sw_if_index),
"uRPF check for 68.68.68.68/32 on %d OK",
tm->hw[1]->sw_if_index);
FIB_TEST(!fib_urpf_check(urpfi, 99),
"uRPF check for 68.68.68.68/32 on 99 not-OK",
99);
dpo_reset(&dpo_44);
fib_table_entry_delete(fib_index,
&bgp_44_s_32,
FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&pfx_1_2_3_5_s_32,
FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&pfx_1_2_3_4_s_32,
FIB_SOURCE_API);
/*
* Add a recursive route:
* 200.200.200.201/32 via 1.1.1.200/32 => the via entry is NOT installed.
*/
fib_prefix_t bgp_201_pfx = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 200.200.200.201/32 */
.ip4.as_u32 = clib_host_to_net_u32(0xc8c8c8c9),
},
};
/* via 1.1.1.200 */
fib_prefix_t pfx_1_1_1_200_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x010101c8),
},
};
fei = fib_table_entry_path_add(fib_index,
&bgp_201_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_200_s_32.fp_addr,
~0, // no index provided.
fib_index, // nexthop in same fib as route
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"Recursive via unresolved is drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_200_s_32);
FIB_TEST((FIB_ENTRY_FLAG_NONE == fib_entry_get_flags(fei)),
"Flags set on RR via non-attached");
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 0),
"RPF list for BGP route empty");
/*
* +2 entry (BGP & RR) and +1 shared-path-list
*/
FIB_TEST((4 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+8 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+12 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* insert a route that covers the missing 1.1.1.2/32. we epxect
* 200.200.200.200/32 and 200.200.200.201/32 to resolve through it.
*/
fib_prefix_t pfx_1_1_1_0_s_24 = {
.fp_len = 24,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 1.1.1.0/24 */
.ip4.as_u32 = clib_host_to_net_u32(0x01010100),
},
};
fib_table_entry_path_add(fib_index,
&pfx_1_1_1_0_s_24,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_1_1_1_0_s_24);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "1.1.1.0/24 resolves via 10.10.10.1");
fei = fib_table_lookup(fib_index, &pfx_1_1_1_2_s_32);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "1.1.1.2/32 resolves via 10.10.10.1");
fei = fib_table_lookup(fib_index, &pfx_1_1_1_200_s_32);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "1.1.1.200/24 resolves via 10.10.10.1");
/*
* +1 entry. 1.1.1.1/32 already uses 10.10.10.1 so no new pah-list
*/
FIB_TEST((4 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+8 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+13 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* the recursive adj for 200.200.200.200 should be updated.
*/
FIB_TEST_REC_FORW(&bgp_201_pfx, &pfx_1_1_1_200_s_32, 0);
FIB_TEST_REC_FORW(&bgp_200_pfx, &pfx_1_1_1_2_s_32, 0);
fei = fib_table_lookup(fib_index, &bgp_200_pfx);
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 1,
tm->hw[0]->sw_if_index),
"RPF list for BGP route has itf index 0");
/*
* insert a more specific route than 1.1.1.0/24 that also covers the
* missing 1.1.1.2/32, but not 1.1.1.200/32. we epxect
* 200.200.200.200 to resolve through it.
*/
fib_prefix_t pfx_1_1_1_0_s_28 = {
.fp_len = 28,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 1.1.1.0/24 */
.ip4.as_u32 = clib_host_to_net_u32(0x01010100),
},
};
fib_table_entry_path_add(fib_index,
&pfx_1_1_1_0_s_28,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_1_1_1_0_s_28);
dpo2 = fib_entry_contribute_ip_forwarding(fei);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_02 == ai), "1.1.1.0/24 resolves via 10.10.10.2");
/*
* +1 entry. +1 shared path-list
*/
FIB_TEST((5 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+9 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+14 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* the recursive adj for 200.200.200.200 should be updated.
* 200.200.200.201 remains unchanged.
*/
FIB_TEST_REC_FORW(&bgp_201_pfx, &pfx_1_1_1_200_s_32, 0);
FIB_TEST_REC_FORW(&bgp_200_pfx, &pfx_1_1_1_2_s_32, 0);
/*
* remove this /28. 200.200.200.200/32 should revert back to via 1.1.1.0/24
*/
fib_table_entry_path_remove(fib_index,
&pfx_1_1_1_0_s_28,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0,
1,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST((fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_0_s_28) ==
FIB_NODE_INDEX_INVALID),
"1.1.1.0/28 removed");
FIB_TEST((fib_table_lookup(fib_index, &pfx_1_1_1_0_s_28) ==
fib_table_lookup(fib_index, &pfx_1_1_1_0_s_24)),
"1.1.1.0/28 lookup via /24");
FIB_TEST_REC_FORW(&bgp_201_pfx, &pfx_1_1_1_200_s_32, 0);
FIB_TEST_REC_FORW(&bgp_200_pfx, &pfx_1_1_1_2_s_32, 0);
/*
* -1 entry. -1 shared path-list
*/
FIB_TEST((4 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+8 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+13 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* remove 1.1.1.0/24. 200.200.200.200/32 should revert back to via 0.0.0.0/0
*/
fib_table_entry_path_remove(fib_index,
&pfx_1_1_1_0_s_24,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
1,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST((fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_0_s_24) ==
FIB_NODE_INDEX_INVALID),
"1.1.1.0/24 removed");
fei = fib_table_lookup(fib_index, &pfx_1_1_1_2_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"1.1.1.2/32 route is DROP");
fei = fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_200_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"1.1.1.200/32 route is DROP");
fei = fib_table_lookup_exact_match(fib_index, &bgp_201_pfx);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"201 is drop");
fei = fib_table_lookup_exact_match(fib_index, &bgp_200_pfx);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"200 is drop");
/*
* -1 entry
*/
FIB_TEST((4 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+8 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+12 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* insert the missing 1.1.1.2/32
*/
fei = fib_table_entry_path_add(fib_index,
&pfx_1_1_1_2_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai = ai_01), "1.1.1.2/32 resolves via 10.10.10.1");
fei = fib_table_lookup_exact_match(fib_index, &bgp_201_pfx);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"201 is drop");
FIB_TEST_REC_FORW(&bgp_200_pfx, &pfx_1_1_1_2_s_32, 0);
/*
* no change. 1.1.1.2/32 was already there RR sourced.
*/
FIB_TEST((4 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+8 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+12 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* give 201 a resolved path.
* it now has the unresolved 1.1.1.200 and the resolved 1.1.1.2,
* only the latter contributes forwarding.
*/
fei = fib_table_entry_path_add(fib_index,
&bgp_201_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_2_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST_REC_FORW(&bgp_201_pfx, &pfx_1_1_1_2_s_32, 0);
fib_table_entry_path_remove(fib_index,
&bgp_201_pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_1_1_1_2_s_32.fp_addr,
~0,
fib_index,
1,
FIB_ROUTE_PATH_FLAG_NONE);
/*
* remove 200.200.200.201/32 which does not have a valid via FIB
*/
fib_table_entry_path_remove(fib_index,
&bgp_201_pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_1_1_1_200_s_32.fp_addr,
~0, // no index provided.
fib_index,
1,
FIB_ROUTE_PATH_FLAG_NONE);
/*
* -2 entries (BGP and RR). -1 shared path-list;
*/
FIB_TEST((fib_table_lookup_exact_match(fib_index, &bgp_201_pfx) ==
FIB_NODE_INDEX_INVALID),
"200.200.200.201/32 removed");
FIB_TEST((fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_200_s_32) ==
FIB_NODE_INDEX_INVALID),
"1.1.1.200/32 removed");
FIB_TEST((3 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+7 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+10 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* remove 200.200.200.200/32 which does have a valid via FIB
*/
fib_table_entry_path_remove(fib_index,
&bgp_200_pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_1_1_1_2_s_32.fp_addr,
~0, // no index provided.
fib_index,
1,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST((fib_table_lookup_exact_match(fib_index, &bgp_200_pfx) ==
FIB_NODE_INDEX_INVALID),
"200.200.200.200/32 removed");
FIB_TEST((fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_2_s_32) !=
FIB_NODE_INDEX_INVALID),
"1.1.1.2/32 still present");
/*
* -1 entry (BGP, the RR source is also API sourced). -1 shared path-list;
*/
FIB_TEST((2 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+6 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+9 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* A recursive prefix that has a 2 path load-balance.
* It also shares a next-hop with other BGP prefixes and hence
* test the ref counting of RR sourced prefixes and 2 level LB.
*/
const fib_prefix_t bgp_102 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 100.100.100.101/32 */
.ip4.as_u32 = clib_host_to_net_u32(0x64646466),
},
};
fib_table_entry_path_add(fib_index,
&bgp_102,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_1_s_32.fp_addr,
~0, // no index provided.
fib_index, // same as route
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_add(fib_index,
&bgp_102,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_2_s_32.fp_addr,
~0, // no index provided.
fib_index, // same as route's FIB
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &bgp_102);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "100.100.100.102/32 presnet");
dpo = fib_entry_contribute_ip_forwarding(fei);
fei = fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_1_s_32);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
fei = fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_2_s_32);
dpo2 = fib_entry_contribute_ip_forwarding(fei);
lb = load_balance_get(dpo->dpoi_index);
FIB_TEST((lb->lb_n_buckets == 2), "Recursive LB has %d bucket", lb->lb_n_buckets);
FIB_TEST(!dpo_cmp(dpo1, load_balance_get_bucket(dpo->dpoi_index, 0)),
"First via 10.10.10.1");
FIB_TEST(!dpo_cmp(dpo2, load_balance_get_bucket(dpo->dpoi_index, 1)),
"Second via 10.10.10.1");
fib_table_entry_path_remove(fib_index,
&bgp_102,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_1_1_1_1_s_32.fp_addr,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_remove(fib_index,
&bgp_102,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_1_1_1_2_s_32.fp_addr,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &bgp_102);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "100.100.100.102/32 removed");
/*
* remove the remaining recursives
*/
fib_table_entry_path_remove(fib_index,
&bgp_100_pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_1_1_1_1_s_32.fp_addr,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_remove(fib_index,
&bgp_101_pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_1_1_1_1_s_32.fp_addr,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST((fib_table_lookup_exact_match(fib_index, &bgp_100_pfx) ==
FIB_NODE_INDEX_INVALID),
"100.100.100.100/32 removed");
FIB_TEST((fib_table_lookup_exact_match(fib_index, &bgp_101_pfx) ==
FIB_NODE_INDEX_INVALID),
"100.100.100.101/32 removed");
/*
* -2 entry (2*BGP, the RR source is also API sourced). -1 shared path-list;
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+7 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Add a recursive route via a connected cover, using an adj-fib that does exist
*/
fib_table_entry_path_add(fib_index,
&bgp_200_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
~0, // no index provided.
fib_index, // Same as route's FIB
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
/*
* +1 entry. +1 shared path-list (recursive via 10.10.10.1)
*/
FIB_TEST((2 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+6 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+8 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
fei = fib_table_lookup_exact_match(fib_index, &bgp_200_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_1_s_32);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo1, load_balance_get_bucket(dpo->dpoi_index, 0)),
"200.200.200.200/32 is recursive via adj for 10.10.10.1");
FIB_TEST((FIB_ENTRY_FLAG_ATTACHED == fib_entry_get_flags(fei)),
"Flags set on RR via existing attached");
/*
* Add a recursive route via a connected cover, using and adj-fib that does
* not exist
*/
ip46_address_t nh_10_10_10_3 = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a03),
};
fib_prefix_t pfx_10_10_10_3 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = nh_10_10_10_3,
};
fib_table_entry_path_add(fib_index,
&bgp_201_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_3,
~0, // no index provided.
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
/*
* +2 entries (BGP and RR). +1 shared path-list (recursive via 10.10.10.3) and
* one unshared non-recursive via 10.10.10.3
*/
FIB_TEST((3 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+8 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+10 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
ai_03 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nh_10_10_10_3,
tm->hw[0]->sw_if_index);
fei = fib_table_lookup_exact_match(fib_index, &bgp_201_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_3);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai == ai_03), "adj for 10.10.10.3/32 is via adj for 10.10.10.3");
FIB_TEST(((FIB_ENTRY_FLAG_ATTACHED | FIB_ENTRY_FLAG_CONNECTED) ==
fib_entry_get_flags(fei)),
"Flags set on RR via non-existing attached");
FIB_TEST(!dpo_cmp(dpo1, load_balance_get_bucket(dpo->dpoi_index, 0)),
"adj for 200.200.200.200/32 is recursive via adj for 10.10.10.3");
adj_unlock(ai_03);
/*
* remove the recursives
*/
fib_table_entry_path_remove(fib_index,
&bgp_200_pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_remove(fib_index,
&bgp_201_pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_3,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST((fib_table_lookup_exact_match(fib_index, &bgp_201_pfx) ==
FIB_NODE_INDEX_INVALID),
"200.200.200.201/32 removed");
FIB_TEST((fib_table_lookup_exact_match(fib_index, &bgp_200_pfx) ==
FIB_NODE_INDEX_INVALID),
"200.200.200.200/32 removed");
FIB_TEST((fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_3) ==
FIB_NODE_INDEX_INVALID),
"10.10.10.3/32 removed");
/*
* -3 entries (2*BGP and RR). -2 shared path-list (recursive via 10.10.10.3 &
* 10.10.10.1) and one unshared non-recursive via 10.10.10.3
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+7 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* RECURSION LOOPS
* Add 5.5.5.5/32 -> 5.5.5.6/32 -> 5.5.5.7/32 -> 5.5.5.5/32
*/
fib_prefix_t pfx_5_5_5_5_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x05050505),
},
};
fib_prefix_t pfx_5_5_5_6_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x05050506),
},
};
fib_prefix_t pfx_5_5_5_7_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x05050507),
},
};
fib_table_entry_path_add(fib_index,
&pfx_5_5_5_5_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_5_5_5_6_s_32.fp_addr,
~0, // no index provided.
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_add(fib_index,
&pfx_5_5_5_6_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_5_5_5_7_s_32.fp_addr,
~0, // no index provided.
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_add(fib_index,
&pfx_5_5_5_7_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_5_5_5_5_s_32.fp_addr,
~0, // no index provided.
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
/*
* +3 entries, +3 shared path-list
*/
FIB_TEST((4 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+8 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+10 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* All the entries have only looped paths, so they are all drop
*/
fei = fib_table_lookup(fib_index, &pfx_5_5_5_7_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.7/32 is via adj for DROP");
fei = fib_table_lookup(fib_index, &pfx_5_5_5_5_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.5/32 is via adj for DROP");
fei = fib_table_lookup(fib_index, &pfx_5_5_5_6_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.6/32 is via adj for DROP");
/*
* provide 5.5.5.6/32 with alternate path.
* this will allow only 5.5.5.6/32 to forward with this path, the others
* are still drop since the loop is still present.
*/
fib_table_entry_path_add(fib_index,
&pfx_5_5_5_6_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_5_5_5_6_s_32);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
lb = load_balance_get(dpo1->dpoi_index);
FIB_TEST((lb->lb_n_buckets == 1), "5.5.5.6 LB has %d bucket", lb->lb_n_buckets);
dpo2 = load_balance_get_bucket(dpo1->dpoi_index, 0);
FIB_TEST(DPO_ADJACENCY == dpo2->dpoi_type, "type is %d", dpo2->dpoi_type);
FIB_TEST((ai_01 == dpo2->dpoi_index),
"5.5.5.6 bucket 0 resolves via 10.10.10.2");
fei = fib_table_lookup(fib_index, &pfx_5_5_5_7_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.7/32 is via adj for DROP");
fei = fib_table_lookup(fib_index, &pfx_5_5_5_5_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.5/32 is via adj for DROP");
/*
* remove the alternate path for 5.5.5.6/32
* back to all drop
*/
fib_table_entry_path_remove(fib_index,
&pfx_5_5_5_6_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
1,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_5_5_5_7_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.7/32 is via adj for DROP");
fei = fib_table_lookup(fib_index, &pfx_5_5_5_5_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.5/32 is via adj for DROP");
fei = fib_table_lookup(fib_index, &pfx_5_5_5_6_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.6/32 is via adj for DROP");
/*
* break the loop by giving 5.5.5.5/32 a new set of paths
* expect all to forward via this new path.
*/
fib_table_entry_update_one_path(fib_index,
&pfx_5_5_5_5_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_5_5_5_5_s_32);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
lb = load_balance_get(dpo1->dpoi_index);
FIB_TEST((lb->lb_n_buckets == 1), "5.5.5.5 LB has %d bucket", lb->lb_n_buckets);
dpo2 = load_balance_get_bucket(dpo1->dpoi_index, 0);
FIB_TEST(DPO_ADJACENCY == dpo2->dpoi_type, "type is %d", dpo2->dpoi_type);
FIB_TEST((ai_01 == dpo2->dpoi_index),
"5.5.5.5 bucket 0 resolves via 10.10.10.2");
fei = fib_table_lookup_exact_match(fib_index, &pfx_5_5_5_7_s_32);
dpo2 = fib_entry_contribute_ip_forwarding(fei);
lb = load_balance_get(dpo2->dpoi_index);
FIB_TEST((lb->lb_n_buckets == 1), "Recursive LB has %d bucket", lb->lb_n_buckets);
FIB_TEST(!dpo_cmp(dpo1, load_balance_get_bucket(dpo2->dpoi_index, 0)),
"5.5.5.5.7 via 5.5.5.5");
fei = fib_table_lookup_exact_match(fib_index, &pfx_5_5_5_6_s_32);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
lb = load_balance_get(dpo1->dpoi_index);
FIB_TEST((lb->lb_n_buckets == 1), "Recursive LB has %d bucket", lb->lb_n_buckets);
FIB_TEST(!dpo_cmp(dpo2, load_balance_get_bucket(dpo1->dpoi_index, 0)),
"5.5.5.5.6 via 5.5.5.7");
/*
* revert back to the loop. so we can remove the prefixes with
* the loop intact
*/
fib_table_entry_update_one_path(fib_index,
&pfx_5_5_5_5_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_5_5_5_6_s_32.fp_addr,
~0, // no index provided.
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_5_5_5_7_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.7/32 is via adj for DROP");
fei = fib_table_lookup(fib_index, &pfx_5_5_5_5_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.5/32 is via adj for DROP");
fei = fib_table_lookup(fib_index, &pfx_5_5_5_6_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"LB for 5.5.5.6/32 is via adj for DROP");
/*
* remove all the 5.5.5.x/32 prefixes
*/
fib_table_entry_path_remove(fib_index,
&pfx_5_5_5_5_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_5_5_5_6_s_32.fp_addr,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_remove(fib_index,
&pfx_5_5_5_6_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_5_5_5_7_s_32.fp_addr,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_remove(fib_index,
&pfx_5_5_5_7_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_5_5_5_5_s_32.fp_addr,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_remove(fib_index,
&pfx_5_5_5_6_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_2,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
/*
* -3 entries, -3 shared path-list
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+7 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Single level loop 5.5.5.5/32 via 5.5.5.5/32
*/
fib_table_entry_path_add(fib_index,
&pfx_5_5_5_6_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_5_5_5_6_s_32.fp_addr,
~0, // no index provided.
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_5_5_5_6_s_32);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"1-level 5.5.5.6/32 loop is via adj for DROP");
fib_table_entry_path_remove(fib_index,
&pfx_5_5_5_6_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_5_5_5_6_s_32.fp_addr,
~0, // no index provided.
fib_index, // same as route's FIB
1,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_5_5_5_6_s_32),
"1-level 5.5.5.6/32 loop is removed");
/*
* A recursive route whose next-hop is covered by the prefix.
* This would mean the via-fib, which inherits forwarding from its
* cover, thus picks up forwarding from the prfix, which is via the
* via-fib, and we have a loop.
*/
fib_prefix_t pfx_23_23_23_0_s_24 = {
.fp_len = 24,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x17171700),
},
};
fib_prefix_t pfx_23_23_23_23_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x17171717),
},
};
fei = fib_table_entry_path_add(fib_index,
&pfx_23_23_23_0_s_24,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_23_23_23_23_s_32.fp_addr,
~0, // recursive
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(load_balance_is_drop(dpo),
"23.23.23.0/24 via covered is DROP");
fib_table_entry_delete_index(fei, FIB_SOURCE_API);
/*
* add-remove test. no change.
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+7 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Make the default route recursive via a unknown next-hop. Thus the
* next hop's cover would be the default route
*/
fei = fib_table_entry_path_add(fib_index,
&pfx_0_0_0_0_s_0,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_23_23_23_23_s_32.fp_addr,
~0, // recursive
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(load_balance_is_drop(dpo),
"0.0.0.0.0/0 via is DROP");
FIB_TEST((fib_entry_get_resolving_interface(fei) == ~0),
"no resolving interface for looped 0.0.0.0/0");
fei = fib_table_lookup_exact_match(fib_index, &pfx_23_23_23_23_s_32);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(load_balance_is_drop(dpo),
"23.23.23.23/32 via is DROP");
FIB_TEST((fib_entry_get_resolving_interface(fei) == ~0),
"no resolving interface for looped 23.23.23.23/32");
fib_table_entry_delete(fib_index, &pfx_0_0_0_0_s_0, FIB_SOURCE_API);
/*
* A recursive route with recursion constraints.
* 200.200.200.200/32 via 1.1.1.1 is recurse via host constrained
*/
fib_table_entry_path_add(fib_index,
&bgp_200_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_1_1_1_1,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_RESOLVE_VIA_HOST);
fei = fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_1_s_32);
dpo2 = fib_entry_contribute_ip_forwarding(fei);
fei = fib_table_lookup_exact_match(fib_index, &bgp_200_pfx);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo2, load_balance_get_bucket(dpo1->dpoi_index, 0)),
"adj for 200.200.200.200/32 is recursive via adj for 1.1.1.1");
/*
* save the load-balance. we expect it to be inplace modified
*/
lb = load_balance_get(dpo1->dpoi_index);
/*
* add a covering prefix for the via fib that would otherwise serve
* as the resolving route when the host is removed
*/
fib_table_entry_path_add(fib_index,
&pfx_1_1_1_0_s_28,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_0_s_28);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai == ai_01),
"adj for 1.1.1.0/28 is via adj for 1.1.1.1");
/*
* remove the host via FIB - expect the BGP prefix to be drop
*/
fib_table_entry_path_remove(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo1->dpoi_index, 0)),
"adj for 200.200.200.200/32 is recursive via adj for DROP");
/*
* add the via-entry host reoute back. expect to resolve again
*/
fib_table_entry_path_add(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(!dpo_cmp(dpo2, load_balance_get_bucket(dpo1->dpoi_index, 0)),
"adj for 200.200.200.200/32 is recursive via adj for 1.1.1.1");
/*
* add another path for the recursive. it will then have 2.
*/
fib_prefix_t pfx_1_1_1_3_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x01010103),
},
};
fib_table_entry_path_add(fib_index,
&pfx_1_1_1_3_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_add(fib_index,
&bgp_200_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_3_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_RESOLVE_VIA_HOST);
/*
* add a bunch load more entries using this path combo so that we get
* an LB-map created.
*/
#define N_P 128
fib_prefix_t bgp_78s[N_P];
for (ii = 0; ii < N_P; ii++)
{
bgp_78s[ii].fp_len = 32;
bgp_78s[ii].fp_proto = FIB_PROTOCOL_IP4;
bgp_78s[ii].fp_addr.ip4.as_u32 = clib_host_to_net_u32(0x4e000000+ii);
fib_table_entry_path_add(fib_index,
&bgp_78s[ii],
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_3_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_RESOLVE_VIA_HOST);
fib_table_entry_path_add(fib_index,
&bgp_78s[ii],
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_1_1_1_1,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_RESOLVE_VIA_HOST);
}
fei = fib_table_lookup_exact_match(fib_index, &bgp_200_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
fei = fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_1_s_32);
dpo2 = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo2, load_balance_get_bucket(dpo->dpoi_index, 0)),
"adj for 200.200.200.200/32 is recursive via adj for 1.1.1.1");
fei = fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_3_s_32);
dpo1 = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo1, load_balance_get_bucket(dpo->dpoi_index, 1)),
"adj for 200.200.200.200/32 is recursive via adj for 1.1.1.3");
/*
* expect the lb-map used by the recursive's load-balance is using both buckets
*/
load_balance_map_t *lbm;
index_t lbmi;
lb = load_balance_get(dpo->dpoi_index);
lbmi = lb->lb_map;
load_balance_map_lock(lbmi);
lbm = load_balance_map_get(lbmi);
FIB_TEST(lbm->lbm_buckets[0] == 0,
"LB maps's bucket 0 is %d",
lbm->lbm_buckets[0]);
FIB_TEST(lbm->lbm_buckets[1] == 1,
"LB maps's bucket 1 is %d",
lbm->lbm_buckets[1]);
/*
* withdraw one of the /32 via-entrys.
* that ECMP path will be unresolved and forwarding should continue on the
* other available path. this is an iBGP PIC edge failover.
* Test the forwarding changes without re-fetching the adj from the
* recursive entry. this ensures its the same one that is updated; i.e. an
* inplace-modify.
*/
fib_table_entry_path_remove(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
FIB_ROUTE_PATH_FLAG_NONE);
/* suspend so the update walk kicks int */
vlib_process_suspend(vlib_get_main(), 1e-5);
fei = fib_table_lookup_exact_match(fib_index, &bgp_200_pfx);
FIB_TEST(!dpo_cmp(dpo, fib_entry_contribute_ip_forwarding(fei)),
"post PIC 200.200.200.200/32 was inplace modified");
FIB_TEST(!dpo_cmp(dpo1, load_balance_get_bucket_i(lb, 0)),
"post PIC adj for 200.200.200.200/32 is recursive"
" via adj for 1.1.1.3");
/*
* the LB maps that was locked above should have been modified to remove
* the path that was down, and thus its bucket points to a path that is
* still up.
*/
FIB_TEST(lbm->lbm_buckets[0] == 1,
"LB maps's bucket 0 is %d",
lbm->lbm_buckets[0]);
FIB_TEST(lbm->lbm_buckets[1] == 1,
"LB maps's bucket 1 is %d",
lbm->lbm_buckets[1]);
load_balance_map_unlock(lbmi);
/*
* add it back. again
*/
fib_table_entry_path_add(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
/* suspend so the update walk kicks in */
vlib_process_suspend(vlib_get_main(), 1e-5);
FIB_TEST(!dpo_cmp(dpo2, load_balance_get_bucket_i(lb, 0)),
"post PIC recovery adj for 200.200.200.200/32 is recursive "
"via adj for 1.1.1.1");
FIB_TEST(!dpo_cmp(dpo1, load_balance_get_bucket_i(lb, 1)),
"post PIC recovery adj for 200.200.200.200/32 is recursive "
"via adj for 1.1.1.3");
fei = fib_table_lookup_exact_match(fib_index, &bgp_200_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(lb == load_balance_get(dpo->dpoi_index),
"post PIC 200.200.200.200/32 was inplace modified");
/*
* add a 3rd path. this makes the LB 16 buckets.
*/
fib_table_entry_path_add(fib_index,
&bgp_200_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_2_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_RESOLVE_VIA_HOST);
for (ii = 0; ii < N_P; ii++)
{
fib_table_entry_path_add(fib_index,
&bgp_78s[ii],
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_2_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_RESOLVE_VIA_HOST);
}
fei = fib_table_lookup_exact_match(fib_index, &bgp_200_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(lb == load_balance_get(dpo->dpoi_index),
"200.200.200.200/32 was inplace modified for 3rd path");
FIB_TEST(16 == lb->lb_n_buckets,
"200.200.200.200/32 was inplace modified for 3rd path to 16 buckets");
lbmi = lb->lb_map;
load_balance_map_lock(lbmi);
lbm = load_balance_map_get(lbmi);
for (ii = 0; ii < 16; ii++)
{
FIB_TEST(lbm->lbm_buckets[ii] == ii,
"LB Map for 200.200.200.200/32 at %d is %d",
ii, lbm->lbm_buckets[ii]);
}
/*
* trigger PIC by removing the first via-entry
* the first 6 buckets of the map should map to the next 6
*/
fib_table_entry_path_remove(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
1,
FIB_ROUTE_PATH_FLAG_NONE);
/* suspend so the update walk kicks int */
vlib_process_suspend(vlib_get_main(), 1e-5);
fei = fib_table_lookup_exact_match(fib_index, &bgp_200_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(lb == load_balance_get(dpo->dpoi_index),
"200.200.200.200/32 was inplace modified for 3rd path");
FIB_TEST(2 == lb->lb_n_buckets,
"200.200.200.200/32 was inplace modified for 3rd path remove to 2 buckets");
for (ii = 0; ii < 6; ii++)
{
FIB_TEST(lbm->lbm_buckets[ii] == ii+6,
"LB Map for 200.200.200.200/32 at %d is %d",
ii, lbm->lbm_buckets[ii]);
}
for (ii = 6; ii < 16; ii++)
{
FIB_TEST(lbm->lbm_buckets[ii] == ii,
"LB Map for 200.200.200.200/32 at %d is %d",
ii, lbm->lbm_buckets[ii]);
}
load_balance_map_unlock(lbmi);
/*
* tidy up
*/
fib_table_entry_path_add(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
for (ii = 0; ii < N_P; ii++)
{
fib_table_entry_delete(fib_index,
&bgp_78s[ii],
FIB_SOURCE_API);
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &bgp_78s[ii])),
"%U removed",
format_fib_prefix, &bgp_78s[ii]);
}
fib_table_entry_path_remove(fib_index,
&bgp_200_pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_1_1_1_2_s_32.fp_addr,
~0,
fib_index,
1,
MPLS_LABEL_INVALID);
fib_table_entry_path_remove(fib_index,
&bgp_200_pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_1_1_1_1,
~0,
fib_index,
1,
FIB_ROUTE_PATH_RESOLVE_VIA_HOST);
fib_table_entry_path_remove(fib_index,
&bgp_200_pfx,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&pfx_1_1_1_3_s_32.fp_addr,
~0,
fib_index,
1,
FIB_ROUTE_PATH_RESOLVE_VIA_HOST);
fib_table_entry_delete(fib_index,
&pfx_1_1_1_3_s_32,
FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&pfx_1_1_1_0_s_28,
FIB_SOURCE_API);
/* suspend so the update walk kicks int */
vlib_process_suspend(vlib_get_main(), 1e-5);
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_0_s_28)),
"1.1.1.1/28 removed");
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_3_s_32)),
"1.1.1.3/32 removed");
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &bgp_200_pfx)),
"200.200.200.200/32 removed");
/*
* add-remove test. no change.
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+7 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* A route whose paths are built up iteratively and then removed
* all at once
*/
fib_prefix_t pfx_4_4_4_4_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 4.4.4.4/32 */
.ip4.as_u32 = clib_host_to_net_u32(0x04040404),
},
};
fib_table_entry_path_add(fib_index,
&pfx_4_4_4_4_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_add(fib_index,
&pfx_4_4_4_4_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_add(fib_index,
&pfx_4_4_4_4_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_3,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(FIB_NODE_INDEX_INVALID !=
fib_table_lookup_exact_match(fib_index, &pfx_4_4_4_4_s_32),
"4.4.4.4/32 present");
fib_table_entry_delete(fib_index,
&pfx_4_4_4_4_s_32,
FIB_SOURCE_API);
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_4_4_4_4_s_32),
"4.4.4.4/32 removed");
/*
* add-remove test. no change.
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+7 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* A route with multiple paths at once
*/
fib_route_path_t *r_paths = NULL;
for (ii = 0; ii < 4; ii++)
{
fib_route_path_t r_path = {
.frp_proto = DPO_PROTO_IP4,
.frp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a02 + ii),
},
.frp_sw_if_index = tm->hw[0]->sw_if_index,
.frp_weight = 1,
.frp_fib_index = ~0,
};
vec_add1(r_paths, r_path);
}
fib_table_entry_update(fib_index,
&pfx_4_4_4_4_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
r_paths);
fei = fib_table_lookup_exact_match(fib_index, &pfx_4_4_4_4_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "4.4.4.4/32 present");
dpo = fib_entry_contribute_ip_forwarding(fei);
lb = load_balance_get(dpo->dpoi_index);
FIB_TEST((lb->lb_n_buckets == 4), "4.4.4.4/32 lb over %d paths", lb->lb_n_buckets);
fib_table_entry_delete(fib_index,
&pfx_4_4_4_4_s_32,
FIB_SOURCE_API);
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_4_4_4_4_s_32),
"4.4.4.4/32 removed");
vec_free(r_paths);
/*
* add-remove test. no change.
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+7 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* A route deag route
*/
fib_table_entry_path_add(fib_index,
&pfx_4_4_4_4_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&zero_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_4_4_4_4_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "4.4.4.4/32 present");
dpo = fib_entry_contribute_ip_forwarding(fei);
dpo = load_balance_get_bucket(dpo->dpoi_index, 0);
lookup_dpo_t *lkd = lookup_dpo_get(dpo->dpoi_index);
FIB_TEST((fib_index == lkd->lkd_fib_index),
"4.4.4.4/32 is deag in %d %U",
lkd->lkd_fib_index,
format_dpo_id, dpo, 0);
fib_table_entry_delete(fib_index,
&pfx_4_4_4_4_s_32,
FIB_SOURCE_API);
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_4_4_4_4_s_32),
"4.4.4.4/32 removed");
vec_free(r_paths);
/*
* add-remove test. no change.
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+7 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Duplicate paths:
* add a recursive with duplicate paths. Expect the duplicate to be ignored.
*/
fib_prefix_t pfx_34_1_1_1_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x22010101),
},
};
fib_prefix_t pfx_34_34_1_1_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x22220101),
},
};
fei = fib_table_entry_path_add(fib_index,
&pfx_34_34_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_entry_path_add(fib_index,
&pfx_34_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_34_34_1_1_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_entry_path_add(fib_index,
&pfx_34_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_34_34_1_1_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST_REC_FORW(&pfx_34_1_1_1_s_32, &pfx_34_34_1_1_s_32, 0);
fib_table_entry_delete_index(fei, FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&pfx_34_34_1_1_s_32,
FIB_SOURCE_API);
/*
* CLEANUP
* remove: 1.1.1.2/32, 1.1.2.0/24 and 1.1.1.1/32
* all of which are via 10.10.10.1, Itf1
*/
fib_table_entry_path_remove(fib_index,
&pfx_1_1_1_2_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
1,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_remove(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
1,
FIB_ROUTE_PATH_FLAG_NONE);
fib_table_entry_path_remove(fib_index,
&pfx_1_1_2_0_s_24,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0,
1,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_1_s_32),
"1.1.1.1/32 removed");
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_1_1_1_2_s_32),
"1.1.1.2/32 removed");
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_1_1_2_0_s_24),
"1.1.2.0/24 removed");
/*
* -3 entries and -1 shared path-list
*/
FIB_TEST((0 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+4 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+4 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* An attached-host route. Expect to link to the incomplete adj
*/
fib_prefix_t pfx_4_1_1_1_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 4.1.1.1/32 */
.ip4.as_u32 = clib_host_to_net_u32(0x04010101),
},
};
fib_table_entry_path_add(fib_index,
&pfx_4_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&zero_addr,
tm->hw[0]->sw_if_index,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_4_1_1_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "4.1.1.1/32 present");
ai = fib_entry_get_adj(fei);
ai2 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&pfx_4_1_1_1_s_32.fp_addr,
tm->hw[0]->sw_if_index);
FIB_TEST((ai == ai2), "Attached-host link to incomplete ADJ");
adj_unlock(ai2);
/*
* +1 entry and +1 shared path-list
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+5 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
fib_table_entry_delete(fib_index,
&pfx_4_1_1_1_s_32,
FIB_SOURCE_API);
FIB_TEST((0 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+4 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+4 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* add a v6 prefix via v4 next-hops
*/
fib_prefix_t pfx_2001_s_64 = {
.fp_len = 64,
.fp_proto = FIB_PROTOCOL_IP6,
.fp_addr = {
.ip6.as_u64[0] = clib_host_to_net_u64(0x2001000000000000),
},
};
fei = fib_table_entry_path_add(0, //default v6 table
&pfx_2001_s_64,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(0, &pfx_2001_s_64);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "2001::/64 present");
ai = fib_entry_get_adj(fei);
adj = adj_get(ai);
FIB_TEST((adj->lookup_next_index == IP_LOOKUP_NEXT_ARP),
"2001::/64 via ARP-adj");
FIB_TEST((adj->ia_link == VNET_LINK_IP6),
"2001::/64 is link type v6");
FIB_TEST((adj->ia_nh_proto == FIB_PROTOCOL_IP4),
"2001::/64 ADJ-adj is NH proto v4");
fib_table_entry_delete(0, &pfx_2001_s_64, FIB_SOURCE_API);
/*
* add a uRPF exempt prefix:
* test:
* - it's forwarding is drop
* - it's uRPF list is not empty
* - the uRPF list for the default route (it's cover) is empty
*/
fei = fib_table_entry_special_add(fib_index,
&pfx_4_1_1_1_s_32,
FIB_SOURCE_URPF_EXEMPT,
FIB_ENTRY_FLAG_DROP);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(load_balance_is_drop(dpo),
"uRPF exempt 4.1.1.1/32 DROP");
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 1, 0),
"uRPF list for exempt prefix has itf index 0");
fei = fib_table_lookup_exact_match(fib_index, &pfx_0_0_0_0_s_0);
FIB_TEST(fib_test_urpf_is_equal(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, 0),
"uRPF list for 0.0.0.0/0 empty");
fib_table_entry_delete(fib_index, &pfx_4_1_1_1_s_32, FIB_SOURCE_URPF_EXEMPT);
/*
* An adj-fib that fails the refinement criteria - no connected cover
*/
fib_prefix_t pfx_12_10_10_2_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 12.10.10.2 */
.ip4.as_u32 = clib_host_to_net_u32(0x0c0a0a02),
},
};
fib_table_entry_path_add(fib_index,
&pfx_12_10_10_2_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&pfx_12_10_10_2_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_12_10_10_2_s_32);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_id_is_valid(dpo),
"no connected cover adj-fib fails refinement");
fib_table_entry_delete(fib_index,
&pfx_12_10_10_2_s_32,
FIB_SOURCE_ADJ);
/*
* An adj-fib that fails the refinement criteria - cover is connected
* but on a different interface
*/
fib_prefix_t pfx_10_10_10_127_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 10.10.10.127 */
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a7f),
},
};
fib_table_entry_path_add(fib_index,
&pfx_10_10_10_127_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&pfx_10_10_10_127_s_32.fp_addr,
tm->hw[1]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_127_s_32);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_id_is_valid(dpo),
"wrong interface adj-fib fails refinement");
fib_table_entry_delete(fib_index,
&pfx_10_10_10_127_s_32,
FIB_SOURCE_ADJ);
/*
* add a second path to an adj-fib
* this is a sumiluation of another ARP entry created
* on an interface on which the connected prefi does not exist.
* The second path fails refinement. Expect to forward through the
* first.
*/
fib_prefix_t pfx_10_10_10_3_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 10.10.10.3 */
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a03),
},
};
ai_03 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nh_10_10_10_3,
tm->hw[0]->sw_if_index);
fib_test_lb_bucket_t ip_o_10_10_10_3 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_03,
},
};
fei = fib_table_entry_path_add(fib_index,
&pfx_10_10_10_3_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_3,
tm->hw[0]->sw_if_index,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_entry_path_add(fib_index,
&pfx_10_10_10_3_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_12_12_12_12,
tm->hw[1]->sw_if_index,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_10_10_10_3),
"10.10.10.3 via 10.10.10.3/Eth0 only");
/*
* remove the path that refines the cover, should go unresolved
*/
fib_table_entry_path_remove(fib_index,
&pfx_10_10_10_3_s_32,
FIB_SOURCE_ADJ,
DPO_PROTO_IP4,
&nh_10_10_10_3,
tm->hw[0]->sw_if_index,
fib_index,
1,
FIB_ROUTE_PATH_FLAG_NONE);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_id_is_valid(dpo),
"wrong interface adj-fib fails refinement");
/*
* add back the path that refines the cover
*/
fei = fib_table_entry_path_add(fib_index,
&pfx_10_10_10_3_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_3,
tm->hw[0]->sw_if_index,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_10_10_10_3),
"10.10.10.3 via 10.10.10.3/Eth0 only");
/*
* remove the path that does not refine the cover
*/
fib_table_entry_path_remove(fib_index,
&pfx_10_10_10_3_s_32,
FIB_SOURCE_ADJ,
DPO_PROTO_IP4,
&nh_12_12_12_12,
tm->hw[1]->sw_if_index,
fib_index,
1,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_10_10_10_3),
"10.10.10.3 via 10.10.10.3/Eth0 only");
/*
* remove the path that does refine, it's the last path, so
* the entry should be gone
*/
fib_table_entry_path_remove(fib_index,
&pfx_10_10_10_3_s_32,
FIB_SOURCE_ADJ,
DPO_PROTO_IP4,
&nh_10_10_10_3,
tm->hw[0]->sw_if_index,
fib_index,
1,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_3_s_32);
FIB_TEST((fei == FIB_NODE_INDEX_INVALID), "10.10.10.3 gone");
adj_unlock(ai_03);
/*
* change the table's flow-hash config - expect the update to propagete to
* the entries' load-balance objects
*/
flow_hash_config_t old_hash_config, new_hash_config;
old_hash_config = fib_table_get_flow_hash_config(fib_index,
FIB_PROTOCOL_IP4);
new_hash_config = (IP_FLOW_HASH_SRC_ADDR |
IP_FLOW_HASH_DST_ADDR);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_1_s_32);
dpo = fib_entry_contribute_ip_forwarding(fei);
lb = load_balance_get(dpo->dpoi_index);
FIB_TEST((lb->lb_hash_config == old_hash_config),
"Table and LB hash config match: %U",
format_ip_flow_hash_config, lb->lb_hash_config);
fib_table_set_flow_hash_config(fib_index, FIB_PROTOCOL_IP4, new_hash_config);
FIB_TEST((lb->lb_hash_config == new_hash_config),
"Table and LB newhash config match: %U",
format_ip_flow_hash_config, lb->lb_hash_config);
/*
* CLEANUP
* remove adj-fibs:
*/
fib_table_entry_delete(fib_index,
&pfx_10_10_10_1_s_32,
FIB_SOURCE_ADJ);
fib_table_entry_delete(fib_index,
&pfx_10_10_10_2_s_32,
FIB_SOURCE_ADJ);
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_1_s_32),
"10.10.10.1/32 adj-fib removed");
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_2_s_32),
"10.10.10.2/32 adj-fib removed");
/*
* -2 entries and -2 non-shared path-list
*/
FIB_TEST((0 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR+2 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR+2 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* unlock the adjacencies for which this test provided a rewrite.
* These are the last locks on these adjs. they should thus go away.
*/
adj_unlock(ai_02);
adj_unlock(ai_01);
adj_unlock(ai_12_12_12_12);
FIB_TEST((0 == adj_nbr_db_size()), "ADJ DB size is %d",
adj_nbr_db_size());
/*
* CLEANUP
* remove the interface prefixes
*/
local_pfx.fp_len = 32;
fib_table_entry_special_remove(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE);
fei = fib_table_lookup(fib_index, &local_pfx);
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &local_pfx),
"10.10.10.10/32 adj-fib removed");
local_pfx.fp_len = 24;
fib_table_entry_delete(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE);
FIB_TEST(FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &local_pfx),
"10.10.10.10/24 adj-fib removed");
/*
* -2 entries and -2 non-shared path-list
*/
FIB_TEST((0 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Last but not least, remove the VRF
*/
FIB_TEST((0 == fib_table_get_num_entries(fib_index,
FIB_PROTOCOL_IP4,
FIB_SOURCE_API)),
"NO API Source'd prefixes");
FIB_TEST((0 == fib_table_get_num_entries(fib_index,
FIB_PROTOCOL_IP4,
FIB_SOURCE_RR)),
"NO RR Source'd prefixes");
FIB_TEST((0 == fib_table_get_num_entries(fib_index,
FIB_PROTOCOL_IP4,
FIB_SOURCE_INTERFACE)),
"NO INterface Source'd prefixes");
fib_table_unlock(fib_index, FIB_PROTOCOL_IP4, FIB_SOURCE_API);
FIB_TEST((0 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNBR-5 == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENBR-5 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
FIB_TEST((ENBR-5 == pool_elts(fib_urpf_list_pool)), "uRPF pool size is %d",
pool_elts(fib_urpf_list_pool));
FIB_TEST((0 == pool_elts(load_balance_map_pool)), "LB-map pool size is %d",
pool_elts(load_balance_map_pool));
FIB_TEST((lb_count == pool_elts(load_balance_pool)), "LB pool size is %d",
pool_elts(load_balance_pool));
return 0;
}
static int
fib_test_v6 (void)
{
/*
* In the default table check for the presence and correct forwarding
* of the special entries
*/
fib_node_index_t dfrt, fei, ai, locked_ai, ai_01, ai_02;
const dpo_id_t *dpo, *dpo_drop;
const ip_adjacency_t *adj;
const receive_dpo_t *rd;
test_main_t *tm;
u32 fib_index;
int ii;
FIB_TEST((0 == adj_nbr_db_size()), "ADJ DB size is %d",
adj_nbr_db_size());
/* via 2001:0:0:1::2 */
ip46_address_t nh_2001_2 = {
.ip6 = {
.as_u64 = {
[0] = clib_host_to_net_u64(0x2001000000000001),
[1] = clib_host_to_net_u64(0x0000000000000002),
},
},
};
tm = &test_main;
dpo_drop = drop_dpo_get(DPO_PROTO_IP6);
/* Find or create FIB table 11 */
fib_index = fib_table_find_or_create_and_lock(FIB_PROTOCOL_IP6, 11,
FIB_SOURCE_API);
for (ii = 0; ii < 4; ii++)
{
ip6_main.fib_index_by_sw_if_index[tm->hw[ii]->sw_if_index] = fib_index;
}
fib_prefix_t pfx_0_0 = {
.fp_len = 0,
.fp_proto = FIB_PROTOCOL_IP6,
.fp_addr = {
.ip6 = {
{0, 0},
},
},
};
dfrt = fib_table_lookup(fib_index, &pfx_0_0);
FIB_TEST((FIB_NODE_INDEX_INVALID != dfrt), "default route present");
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(dfrt)),
"Default route is DROP");
dpo = fib_entry_contribute_ip_forwarding(dfrt);
FIB_TEST((dpo->dpoi_index == ip6_fib_table_fwding_lookup(
&ip6_main,
1,
&pfx_0_0.fp_addr.ip6)),
"default-route; fwd and non-fwd tables match");
// FIXME - check specials.
/*
* At this stage there is one v4 FIB with 5 routes and two v6 FIBs
* each with 2 entries and a v6 mfib with 4 path-lists.
* All entries are special so no path-list sharing.
*/
#define ENPS (5+4)
#define PNPS (5+4+4)
FIB_TEST((0 == fib_path_list_db_size()), "path list DB is empty");
FIB_TEST((PNPS == fib_path_list_pool_size()), "path list pool size is %d",
fib_path_list_pool_size());
FIB_TEST((ENPS == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* add interface routes.
* validate presence of /64 attached and /128 recieve.
* test for the presence of the receive address in the glean and local adj
*
* receive on 2001:0:0:1::1/128
*/
fib_prefix_t local_pfx = {
.fp_len = 64,
.fp_proto = FIB_PROTOCOL_IP6,
.fp_addr = {
.ip6 = {
.as_u64 = {
[0] = clib_host_to_net_u64(0x2001000000000001),
[1] = clib_host_to_net_u64(0x0000000000000001),
},
},
}
};
fib_table_entry_update_one_path(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_ATTACHED),
DPO_PROTO_IP6,
NULL,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "attached interface route present");
ai = fib_entry_get_adj(fei);
FIB_TEST((FIB_NODE_INDEX_INVALID != ai), "attached interface route adj present");
adj = adj_get(ai);
FIB_TEST((IP_LOOKUP_NEXT_GLEAN == adj->lookup_next_index),
"attached interface adj is glean");
FIB_TEST((0 == ip46_address_cmp(&local_pfx.fp_addr,
&adj->sub_type.glean.receive_addr)),
"attached interface adj is receive ok");
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST((dpo->dpoi_index == ip6_fib_table_fwding_lookup(
&ip6_main,
1,
&local_pfx.fp_addr.ip6)),
"attached-route; fwd and non-fwd tables match");
local_pfx.fp_len = 128;
fib_table_entry_update_one_path(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_LOCAL),
DPO_PROTO_IP6,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local interface route present");
dpo = fib_entry_contribute_ip_forwarding(fei);
dpo = load_balance_get_bucket(dpo->dpoi_index, 0);
FIB_TEST((DPO_RECEIVE == dpo->dpoi_type),
"local interface adj is local");
rd = receive_dpo_get(dpo->dpoi_index);
FIB_TEST((0 == ip46_address_cmp(&local_pfx.fp_addr,
&rd->rd_addr)),
"local interface adj is receive ok");
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST((dpo->dpoi_index == ip6_fib_table_fwding_lookup(
&ip6_main,
1,
&local_pfx.fp_addr.ip6)),
"local-route; fwd and non-fwd tables match");
/*
* +2 entries. +2 unshared path-lists
*/
FIB_TEST((0 == fib_path_list_db_size()), "path list DB is empty");
FIB_TEST((PNPS+2 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENPS+2 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Modify the default route to be via an adj not yet known.
* this sources the defalut route with the API source, which is
* a higher preference to the DEFAULT_ROUTE source
*/
fib_table_entry_path_add(fib_index, &pfx_0_0,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP6,
&nh_2001_2,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_0_0);
FIB_TEST((fei == dfrt), "default route same index");
ai = fib_entry_get_adj(fei);
FIB_TEST((FIB_NODE_INDEX_INVALID != ai), "default route adj present");
adj = adj_get(ai);
FIB_TEST((IP_LOOKUP_NEXT_ARP == adj->lookup_next_index),
"adj is incomplete");
FIB_TEST((0 == ip46_address_cmp(&nh_2001_2, &adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
/*
* find the adj in the shared db
*/
locked_ai = adj_nbr_add_or_lock(FIB_PROTOCOL_IP6,
VNET_LINK_IP6,
&nh_2001_2,
tm->hw[0]->sw_if_index);
FIB_TEST((locked_ai == ai), "ADJ NBR DB find");
adj_unlock(locked_ai);
/*
* no more entires. +1 shared path-list
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNPS+3 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENPS+2 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* remove the API source from the default route. We expected
* the route to remain, sourced by DEFAULT_ROUTE, and hence a DROP
*/
fib_table_entry_path_remove(fib_index, &pfx_0_0,
FIB_SOURCE_API,
DPO_PROTO_IP6,
&nh_2001_2,
tm->hw[0]->sw_if_index,
~0,
1,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_0_0);
FIB_TEST((fei == dfrt), "default route same index");
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(dfrt)),
"Default route is DROP");
/*
* no more entires. -1 shared path-list
*/
FIB_TEST((0 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNPS+2 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENPS+2 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Add an 2 ARP entry => a complete ADJ plus adj-fib.
*/
fib_prefix_t pfx_2001_1_2_s_128 = {
.fp_len = 128,
.fp_proto = FIB_PROTOCOL_IP6,
.fp_addr = {
.ip6 = {
.as_u64 = {
[0] = clib_host_to_net_u64(0x2001000000000001),
[1] = clib_host_to_net_u64(0x0000000000000002),
},
},
}
};
fib_prefix_t pfx_2001_1_3_s_128 = {
.fp_len = 128,
.fp_proto = FIB_PROTOCOL_IP6,
.fp_addr = {
.ip6 = {
.as_u64 = {
[0] = clib_host_to_net_u64(0x2001000000000001),
[1] = clib_host_to_net_u64(0x0000000000000003),
},
},
}
};
u8 eth_addr[] = {
0xde, 0xde, 0xde, 0xba, 0xba, 0xba,
};
ai_01 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP6,
VNET_LINK_IP6,
&pfx_2001_1_2_s_128.fp_addr,
tm->hw[0]->sw_if_index);
FIB_TEST((FIB_NODE_INDEX_INVALID != ai_01), "adj created");
adj = adj_get(ai_01);
FIB_TEST((IP_LOOKUP_NEXT_ARP == adj->lookup_next_index),
"adj is incomplete");
FIB_TEST((0 == ip46_address_cmp(&pfx_2001_1_2_s_128.fp_addr,
&adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
adj_nbr_update_rewrite(ai_01, ADJ_NBR_REWRITE_FLAG_COMPLETE,
fib_test_build_rewrite(eth_addr));
FIB_TEST((IP_LOOKUP_NEXT_REWRITE == adj->lookup_next_index),
"adj is complete");
FIB_TEST((0 == ip46_address_cmp(&pfx_2001_1_2_s_128.fp_addr,
&adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
fib_table_entry_path_add(fib_index,
&pfx_2001_1_2_s_128,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP6,
&pfx_2001_1_2_s_128.fp_addr,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_2001_1_2_s_128);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "ADJ-FIB resolves via adj");
eth_addr[5] = 0xb2;
ai_02 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP6,
VNET_LINK_IP6,
&pfx_2001_1_3_s_128.fp_addr,
tm->hw[0]->sw_if_index);
FIB_TEST((FIB_NODE_INDEX_INVALID != ai_02), "adj created");
adj = adj_get(ai_02);
FIB_TEST((IP_LOOKUP_NEXT_ARP == adj->lookup_next_index),
"adj is incomplete");
FIB_TEST((0 == ip46_address_cmp(&pfx_2001_1_3_s_128.fp_addr,
&adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
adj_nbr_update_rewrite(ai_02, ADJ_NBR_REWRITE_FLAG_COMPLETE,
fib_test_build_rewrite(eth_addr));
FIB_TEST((IP_LOOKUP_NEXT_REWRITE == adj->lookup_next_index),
"adj is complete");
FIB_TEST((0 == ip46_address_cmp(&pfx_2001_1_3_s_128.fp_addr,
&adj->sub_type.nbr.next_hop)),
"adj nbr next-hop ok");
FIB_TEST((ai_01 != ai_02), "ADJs are different");
fib_table_entry_path_add(fib_index,
&pfx_2001_1_3_s_128,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP6,
&pfx_2001_1_3_s_128.fp_addr,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_2001_1_3_s_128);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_02 == ai), "ADJ-FIB resolves via adj");
/*
* +2 entries, +2 unshread path-lists.
*/
FIB_TEST((0 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNPS+4 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENPS+4 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Add a 2 routes via the first ADJ. ensure path-list sharing
*/
fib_prefix_t pfx_2001_a_s_64 = {
.fp_len = 64,
.fp_proto = FIB_PROTOCOL_IP6,
.fp_addr = {
.ip6 = {
.as_u64 = {
[0] = clib_host_to_net_u64(0x200100000000000a),
[1] = clib_host_to_net_u64(0x0000000000000000),
},
},
}
};
fib_prefix_t pfx_2001_b_s_64 = {
.fp_len = 64,
.fp_proto = FIB_PROTOCOL_IP6,
.fp_addr = {
.ip6 = {
.as_u64 = {
[0] = clib_host_to_net_u64(0x200100000000000b),
[1] = clib_host_to_net_u64(0x0000000000000000),
},
},
}
};
fib_table_entry_path_add(fib_index,
&pfx_2001_a_s_64,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP6,
&nh_2001_2,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_2001_a_s_64);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "2001::a/64 resolves via 2001:0:0:1::1");
fib_table_entry_path_add(fib_index,
&pfx_2001_b_s_64,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP6,
&nh_2001_2,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_2001_b_s_64);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "2001::b/64 resolves via 2001:0:0:1::1");
/*
* +2 entries, +1 shared path-list.
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNPS+5 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENPS+6 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* add a v4 prefix via a v6 next-hop
*/
fib_prefix_t pfx_1_1_1_1_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = 0x01010101,
},
};
fei = fib_table_entry_path_add(0, // default table
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP6,
&nh_2001_2,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fei == fib_table_lookup_exact_match(0, &pfx_1_1_1_1_s_32),
"1.1.1.1/32 o v6 route present");
ai = fib_entry_get_adj(fei);
adj = adj_get(ai);
FIB_TEST((adj->lookup_next_index == IP_LOOKUP_NEXT_ARP),
"1.1.1.1/32 via ARP-adj");
FIB_TEST((adj->ia_link == VNET_LINK_IP4),
"1.1.1.1/32 ADJ-adj is link type v4");
FIB_TEST((adj->ia_nh_proto == FIB_PROTOCOL_IP6),
"1.1.1.1/32 ADJ-adj is NH proto v6");
fib_table_entry_delete(0, &pfx_1_1_1_1_s_32, FIB_SOURCE_API);
/*
* An attached route
*/
fib_prefix_t pfx_2001_c_s_64 = {
.fp_len = 64,
.fp_proto = FIB_PROTOCOL_IP6,
.fp_addr = {
.ip6 = {
.as_u64 = {
[0] = clib_host_to_net_u64(0x200100000000000c),
[1] = clib_host_to_net_u64(0x0000000000000000),
},
},
}
};
fib_table_entry_path_add(fib_index,
&pfx_2001_c_s_64,
FIB_SOURCE_CLI,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP6,
NULL,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_c_s_64);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "attached route present");
ai = fib_entry_get_adj(fei);
adj = adj_get(ai);
FIB_TEST((adj->lookup_next_index == IP_LOOKUP_NEXT_GLEAN),
"2001:0:0:c/64 attached resolves via glean");
fib_table_entry_path_remove(fib_index,
&pfx_2001_c_s_64,
FIB_SOURCE_CLI,
DPO_PROTO_IP6,
NULL,
tm->hw[0]->sw_if_index,
~0,
1,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_c_s_64);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "attached route removed");
/*
* Shutdown the interface on which we have a connected and through
* which the routes are reachable.
* This will result in the connected, adj-fibs, and routes linking to drop
* The local/for-us prefix continues to receive.
*/
clib_error_t * error;
error = vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[0]->sw_if_index,
~VNET_SW_INTERFACE_FLAG_ADMIN_UP);
FIB_TEST((NULL == error), "Interface shutdown OK");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_b_s_64);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001::b/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_a_s_64);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001::a/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_3_s_128);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001:0:0:1::3/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_2_s_128);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001:0:0:1::2/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001:0:0:1::1/128 not drop");
local_pfx.fp_len = 64;
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001:0:0:1/64 resolves via drop");
/*
* no change
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNPS+5 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENPS+6 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* shutdown one of the other interfaces, then add a connected.
* and swap one of the routes to it.
*/
error = vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[1]->sw_if_index,
~VNET_SW_INTERFACE_FLAG_ADMIN_UP);
FIB_TEST((NULL == error), "Interface 1 shutdown OK");
fib_prefix_t connected_pfx = {
.fp_len = 64,
.fp_proto = FIB_PROTOCOL_IP6,
.fp_addr = {
.ip6 = {
/* 2001:0:0:2::1/64 */
.as_u64 = {
[0] = clib_host_to_net_u64(0x2001000000000002),
[1] = clib_host_to_net_u64(0x0000000000000001),
},
},
}
};
fib_table_entry_update_one_path(fib_index, &connected_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_ATTACHED),
DPO_PROTO_IP6,
NULL,
tm->hw[1]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &connected_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "attached interface route present");
dpo = fib_entry_contribute_ip_forwarding(fei);
dpo = load_balance_get_bucket(dpo->dpoi_index, 0);
FIB_TEST(!dpo_cmp(dpo, dpo_drop),
"2001:0:0:2/64 not resolves via drop");
connected_pfx.fp_len = 128;
fib_table_entry_update_one_path(fib_index, &connected_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_LOCAL),
DPO_PROTO_IP6,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &connected_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local interface route present");
dpo = fib_entry_contribute_ip_forwarding(fei);
dpo = load_balance_get_bucket(dpo->dpoi_index, 0);
FIB_TEST((DPO_RECEIVE == dpo->dpoi_type),
"local interface adj is local");
rd = receive_dpo_get(dpo->dpoi_index);
FIB_TEST((0 == ip46_address_cmp(&connected_pfx.fp_addr,
&rd->rd_addr)),
"local interface adj is receive ok");
/*
* +2 entries, +2 unshared path-lists
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNPS+7 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENPS+8 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* bring the interface back up. we expected the routes to return
* to normal forwarding.
*/
error = vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[0]->sw_if_index,
VNET_SW_INTERFACE_FLAG_ADMIN_UP);
FIB_TEST((NULL == error), "Interface bring-up OK");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_a_s_64);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "2001::a/64 resolves via 2001:0:0:1::1");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_b_s_64);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "2001::b/64 resolves via 2001:0:0:1::1");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_3_s_128);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_02 == ai), "ADJ-FIB resolves via adj");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_2_s_128);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "ADJ-FIB resolves via adj");
local_pfx.fp_len = 64;
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
ai = fib_entry_get_adj(fei);
adj = adj_get(ai);
FIB_TEST((IP_LOOKUP_NEXT_GLEAN == adj->lookup_next_index),
"attached interface adj is glean");
/*
* Same test as above, but this time the HW interface goes down
*/
error = vnet_hw_interface_set_flags(vnet_get_main(),
tm->hw_if_indicies[0],
~VNET_HW_INTERFACE_FLAG_LINK_UP);
FIB_TEST((NULL == error), "Interface shutdown OK");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_b_s_64);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001::b/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_a_s_64);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001::a/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_3_s_128);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001:0:0:1::3/128 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_2_s_128);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001:0:0:1::2/128 resolves via drop");
local_pfx.fp_len = 128;
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001:0:0:1::1/128 not drop");
local_pfx.fp_len = 64;
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(!dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"2001:0:0:1/64 resolves via drop");
error = vnet_hw_interface_set_flags(vnet_get_main(),
tm->hw_if_indicies[0],
VNET_HW_INTERFACE_FLAG_LINK_UP);
FIB_TEST((NULL == error), "Interface bring-up OK");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_a_s_64);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "2001::a/64 resolves via 2001:0:0:1::1");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_b_s_64);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "2001::b/64 resolves via 2001:0:0:1::1");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_3_s_128);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_02 == ai), "ADJ-FIB resolves via adj");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_2_s_128);
ai = fib_entry_get_adj(fei);
FIB_TEST((ai_01 == ai), "ADJ-FIB resolves via adj");
local_pfx.fp_len = 64;
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
ai = fib_entry_get_adj(fei);
adj = adj_get(ai);
FIB_TEST((IP_LOOKUP_NEXT_GLEAN == adj->lookup_next_index),
"attached interface adj is glean");
/*
* Delete the interface that the routes reolve through.
* Again no routes are removed. They all point to drop.
*
* This is considered an error case. The control plane should
* not remove interfaces through which routes resolve, but
* such things can happen. ALL affected routes will drop.
*/
vnet_delete_hw_interface(vnet_get_main(), tm->hw_if_indicies[0]);
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_b_s_64);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001::b/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_a_s_64);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001::b/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_3_s_128);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001:0:0:1::3/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_2_s_128);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001:0:0:1::2/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001:0:0:1::1/128 is drop");
local_pfx.fp_len = 64;
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001:0:0:1/64 resolves via drop");
/*
* no change
*/
FIB_TEST((1 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNPS+7 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENPS+8 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* Add the interface back. routes stay unresolved.
*/
error = ethernet_register_interface(vnet_get_main(),
test_interface_device_class.index,
0 /* instance */,
hw_address,
&tm->hw_if_indicies[0],
/* flag change */ 0);
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_b_s_64);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001::b/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_a_s_64);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001::b/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_3_s_128);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001:0:0:1::3/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &pfx_2001_1_2_s_128);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001:0:0:1::2/64 resolves via drop");
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001:0:0:1::1/128 is drop");
local_pfx.fp_len = 64;
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"2001:0:0:1/64 resolves via drop");
/*
* CLEANUP ALL the routes
*/
fib_table_entry_delete(fib_index,
&pfx_2001_c_s_64,
FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&pfx_2001_a_s_64,
FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&pfx_2001_b_s_64,
FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&pfx_2001_1_3_s_128,
FIB_SOURCE_ADJ);
fib_table_entry_delete(fib_index,
&pfx_2001_1_2_s_128,
FIB_SOURCE_ADJ);
local_pfx.fp_len = 64;
fib_table_entry_delete(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE);
local_pfx.fp_len = 128;
fib_table_entry_special_remove(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE);
connected_pfx.fp_len = 64;
fib_table_entry_delete(fib_index, &connected_pfx,
FIB_SOURCE_INTERFACE);
connected_pfx.fp_len = 128;
fib_table_entry_special_remove(fib_index, &connected_pfx,
FIB_SOURCE_INTERFACE);
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_2001_a_s_64)),
"2001::a/64 removed");
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_2001_b_s_64)),
"2001::b/64 removed");
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_2001_1_3_s_128)),
"2001:0:0:1::3/128 removed");
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &pfx_2001_1_2_s_128)),
"2001:0:0:1::3/128 removed");
local_pfx.fp_len = 64;
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &local_pfx)),
"2001:0:0:1/64 removed");
local_pfx.fp_len = 128;
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &local_pfx)),
"2001:0:0:1::1/128 removed");
connected_pfx.fp_len = 64;
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &connected_pfx)),
"2001:0:0:2/64 removed");
connected_pfx.fp_len = 128;
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup_exact_match(fib_index, &connected_pfx)),
"2001:0:0:2::1/128 removed");
/*
* -8 entries. -7 path-lists (1 was shared).
*/
FIB_TEST((0 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNPS == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENPS == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
/*
* now remove the VRF
*/
fib_table_unlock(fib_index, FIB_PROTOCOL_IP6, FIB_SOURCE_API);
FIB_TEST((0 == fib_path_list_db_size()), "path list DB population:%d",
fib_path_list_db_size());
FIB_TEST((PNPS-2 == fib_path_list_pool_size()), "path list pool size is%d",
fib_path_list_pool_size());
FIB_TEST((ENPS-2 == fib_entry_pool_size()), "entry pool size is %d",
fib_entry_pool_size());
adj_unlock(ai_02);
adj_unlock(ai_01);
/*
* return the interfaces to up state
*/
error = vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[0]->sw_if_index,
VNET_SW_INTERFACE_FLAG_ADMIN_UP);
error = vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[1]->sw_if_index,
VNET_SW_INTERFACE_FLAG_ADMIN_UP);
FIB_TEST((0 == adj_nbr_db_size()), "ADJ DB size is %d",
adj_nbr_db_size());
return (0);
}
/*
* Test Attached Exports
*/
static int
fib_test_ae (void)
{
const dpo_id_t *dpo, *dpo_drop;
const u32 fib_index = 0;
fib_node_index_t fei;
test_main_t *tm;
ip4_main_t *im;
tm = &test_main;
im = &ip4_main;
FIB_TEST((0 == adj_nbr_db_size()), "ADJ DB size is %d",
adj_nbr_db_size());
/*
* add interface routes. We'll assume this works. It's more rigorously
* tested elsewhere.
*/
fib_prefix_t local_pfx = {
.fp_len = 24,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4 = {
/* 10.10.10.10 */
.as_u32 = clib_host_to_net_u32(0x0a0a0a0a),
},
},
};
vec_validate(im->fib_index_by_sw_if_index, tm->hw[0]->sw_if_index);
im->fib_index_by_sw_if_index[tm->hw[0]->sw_if_index] = fib_index;
dpo_drop = drop_dpo_get(DPO_PROTO_IP4);
fib_table_entry_update_one_path(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_ATTACHED),
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei),
"attached interface route present");
local_pfx.fp_len = 32;
fib_table_entry_update_one_path(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_LOCAL),
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei),
"local interface route present");
/*
* Add an 2 ARP entry => a complete ADJ plus adj-fib.
*/
fib_prefix_t pfx_10_10_10_1_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 10.10.10.1 */
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a01),
},
};
fib_node_index_t ai;
fib_table_entry_path_add(fib_index,
&pfx_10_10_10_1_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&pfx_10_10_10_1_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib1 created");
ai = fib_entry_get_adj(fei);
/*
* create another FIB table into which routes will be imported
*/
u32 import_fib_index1;
import_fib_index1 = fib_table_find_or_create_and_lock(FIB_PROTOCOL_IP4,
11,
FIB_SOURCE_CLI);
/*
* Add an attached route in the import FIB
*/
local_pfx.fp_len = 24;
fib_table_entry_update_one_path(import_fib_index1,
&local_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(import_fib_index1, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "attached export created");
/*
* check for the presence of the adj-fibs in the import table
*/
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib1 imported");
FIB_TEST((ai == fib_entry_get_adj(fei)),
"adj-fib1 Import uses same adj as export");
/*
* check for the presence of the local in the import table
*/
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(import_fib_index1, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local imported");
/*
* Add another adj-fin in the export table. Expect this
* to get magically exported;
*/
fib_prefix_t pfx_10_10_10_2_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 10.10.10.2 */
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a02),
},
};
fib_table_entry_path_add(fib_index,
&pfx_10_10_10_2_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&pfx_10_10_10_2_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib2 present");
ai = fib_entry_get_adj(fei);
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib2 imported");
FIB_TEST((ai == fib_entry_get_adj(fei)),
"Import uses same adj as export");
FIB_TEST((FIB_ENTRY_FLAG_ATTACHED & fib_entry_get_flags(fei)),
"ADJ-fib2 imported flags %d",
fib_entry_get_flags(fei));
/*
* create a 2nd FIB table into which routes will be imported
*/
u32 import_fib_index2;
import_fib_index2 = fib_table_find_or_create_and_lock(FIB_PROTOCOL_IP4, 12,
FIB_SOURCE_CLI);
/*
* Add an attached route in the import FIB
*/
local_pfx.fp_len = 24;
fib_table_entry_update_one_path(import_fib_index2,
&local_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(import_fib_index1, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "attached export created");
/*
* check for the presence of all the adj-fibs and local in the import table
*/
fei = fib_table_lookup_exact_match(import_fib_index2, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib1 imported");
fei = fib_table_lookup_exact_match(import_fib_index2, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib2 imported");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(import_fib_index2, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local imported");
/*
* add a 3rd adj-fib. expect it to be exported to both tables.
*/
fib_prefix_t pfx_10_10_10_3_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 10.10.10.3 */
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a03),
},
};
fib_table_entry_path_add(fib_index,
&pfx_10_10_10_3_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&pfx_10_10_10_3_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_3_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib3 present");
ai = fib_entry_get_adj(fei);
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_3_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib3 imported to FIB1");
FIB_TEST((ai == fib_entry_get_adj(fei)),
"Import uses same adj as export");
fei = fib_table_lookup_exact_match(import_fib_index2, &pfx_10_10_10_3_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib3 imported to FIB2");
FIB_TEST((ai == fib_entry_get_adj(fei)),
"Import uses same adj as export");
/*
* remove the 3rd adj fib. we expect it to be removed from both FIBs
*/
fib_table_entry_delete(fib_index,
&pfx_10_10_10_3_s_32,
FIB_SOURCE_ADJ);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_3_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "ADJ-fib3 remved");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_3_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "ADJ-fib3 removed from FIB1");
fei = fib_table_lookup_exact_match(import_fib_index2, &pfx_10_10_10_3_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "ADJ-fib3 removed from FIB2");
/*
* remove the attached route from the 2nd FIB. expect the imported
* entires to be removed
*/
local_pfx.fp_len = 24;
fib_table_entry_delete(import_fib_index2,
&local_pfx,
FIB_SOURCE_API);
fei = fib_table_lookup_exact_match(import_fib_index2, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "attached export removed");
fei = fib_table_lookup_exact_match(import_fib_index2, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "ADJ-fib1 removed from FIB2");
fei = fib_table_lookup_exact_match(import_fib_index2, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "ADJ-fib2 removed from FIB2");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(import_fib_index2, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "local removed from FIB2");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib1 still in FIB1");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib2 still in FIB1");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(import_fib_index1, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local still in FIB1");
/*
* modify the route in FIB1 so it is no longer attached. expect the imported
* entires to be removed
*/
local_pfx.fp_len = 24;
fib_table_entry_update_one_path(import_fib_index1,
&local_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_10_10_10_2_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "ADJ-fib1 removed from FIB1");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "ADJ-fib2 removed from FIB1");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(import_fib_index1, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "local removed from FIB1");
/*
* modify it back to attached. expect the adj-fibs back
*/
local_pfx.fp_len = 24;
fib_table_entry_update_one_path(import_fib_index1,
&local_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib1 imported in FIB1");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib2 imported in FIB1");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(import_fib_index1, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local imported in FIB1");
/*
* add a covering attached next-hop for the interface address, so we have
* a valid adj to find when we check the forwarding tables
*/
fib_prefix_t pfx_10_0_0_0_s_8 = {
.fp_len = 8,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
/* 10.0.0.0 */
.ip4.as_u32 = clib_host_to_net_u32(0x0a000000),
},
};
fei = fib_table_entry_update_one_path(fib_index,
&pfx_10_0_0_0_s_8,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_10_10_10_3_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
dpo = fib_entry_contribute_ip_forwarding(fei);
/*
* remove the route in the export fib. expect the adj-fibs to be removed
*/
local_pfx.fp_len = 24;
fib_table_entry_delete(fib_index,
&local_pfx,
FIB_SOURCE_INTERFACE);
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "Delete export: ADJ-fib1 removed from FIB1");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "ADJ-fib2 removed from FIB1");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(import_fib_index1, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "local removed from FIB1");
/*
* the adj-fibs in the export VRF are present in the FIB table,
* but not installed in forwarding, since they have no attached cover.
* Consequently a lookup in the MTRIE gives the adj for the covering
* route 10.0.0.0/8.
*/
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib1 in export");
index_t lbi;
lbi = ip4_fib_forwarding_lookup(fib_index, &pfx_10_10_10_1_s_32.fp_addr.ip4);
FIB_TEST(lbi == dpo->dpoi_index,
"10.10.10.1 forwards on \n%U not \n%U",
format_load_balance, lbi, 0,
format_dpo_id, dpo, 0);
lbi = ip4_fib_forwarding_lookup(fib_index, &pfx_10_10_10_2_s_32.fp_addr.ip4);
FIB_TEST(lbi == dpo->dpoi_index,
"10.10.10.2 forwards on %U", format_dpo_id, dpo, 0);
lbi = ip4_fib_forwarding_lookup(fib_index, &pfx_10_10_10_3_s_32.fp_addr.ip4);
FIB_TEST(lbi == dpo->dpoi_index,
"10.10.10.3 forwards on %U", format_dpo_id, dpo, 0);
/*
* add the export prefix back, but not as attached.
* No adj-fibs in export nor import tables
*/
local_pfx.fp_len = 24;
fei = fib_table_entry_update_one_path(fib_index,
&local_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_10_10_10_1_s_32.fp_addr,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
dpo = fib_entry_contribute_ip_forwarding(fei);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "non-attached in export: ADJ-fib1 in export");
lbi = ip4_fib_forwarding_lookup(fib_index, &pfx_10_10_10_1_s_32.fp_addr.ip4);
FIB_TEST(lbi == dpo->dpoi_index,
"10.10.10.1 forwards on %U", format_dpo_id, dpo, 0);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib1 in export");
lbi = ip4_fib_forwarding_lookup(fib_index, &pfx_10_10_10_2_s_32.fp_addr.ip4);
FIB_TEST(lbi == dpo->dpoi_index,
"10.10.10.2 forwards on %U", format_dpo_id, dpo, 0);
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "ADJ-fib1 removed from FIB1");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "ADJ-fib2 removed from FIB1");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(import_fib_index1, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID == fei), "local removed from FIB1");
/*
* modify the export prefix so it is attached. expect all covereds to return
*/
local_pfx.fp_len = 24;
fib_table_entry_update_one_path(fib_index,
&local_pfx,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib1 reinstalled in export");
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"Adj-fib1 is not drop in export");
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib2 reinstalled in export");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local reinstalled in export");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "attached in export: ADJ-fib1 imported");
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"Adj-fib1 is not drop in export");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib1 imported");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib2 imported");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(import_fib_index1, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local imported");
/*
* modify the export prefix so connected. no change.
*/
local_pfx.fp_len = 24;
fib_table_entry_update_one_path(fib_index, &local_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_ATTACHED),
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib1 reinstalled in export");
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"Adj-fib1 is not drop in export");
fei = fib_table_lookup_exact_match(fib_index, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib2 reinstalled in export");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(fib_index, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local reinstalled in export");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "attached in export: ADJ-fib1 imported");
dpo = fib_entry_contribute_ip_forwarding(fei);
FIB_TEST(dpo_cmp(dpo_drop, load_balance_get_bucket(dpo->dpoi_index, 0)),
"Adj-fib1 is not drop in export");
fei = fib_table_lookup_exact_match(import_fib_index1, &pfx_10_10_10_2_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "ADJ-fib2 imported");
local_pfx.fp_len = 32;
fei = fib_table_lookup_exact_match(import_fib_index1, &local_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "local imported");
/*
* CLEANUP
*/
fib_table_entry_delete(fib_index,
&pfx_10_0_0_0_s_8,
FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&pfx_10_10_10_1_s_32,
FIB_SOURCE_ADJ);
fib_table_entry_delete(fib_index,
&pfx_10_10_10_2_s_32,
FIB_SOURCE_ADJ);
local_pfx.fp_len = 32;
fib_table_entry_delete(fib_index,
&local_pfx,
FIB_SOURCE_INTERFACE);
local_pfx.fp_len = 24;
fib_table_entry_delete(fib_index,
&local_pfx,
FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&local_pfx,
FIB_SOURCE_INTERFACE);
local_pfx.fp_len = 24;
fib_table_entry_delete(import_fib_index1,
&local_pfx,
FIB_SOURCE_API);
fib_table_unlock(import_fib_index1, FIB_PROTOCOL_IP4, FIB_SOURCE_CLI);
fib_table_unlock(import_fib_index2, FIB_PROTOCOL_IP4, FIB_SOURCE_CLI);
FIB_TEST((0 == adj_nbr_db_size()), "ADJ DB size is %d",
adj_nbr_db_size());
return (0);
}
/*
* Test Path Preference
*/
static int
fib_test_pref (void)
{
test_main_t *tm = &test_main;
const fib_prefix_t pfx_1_1_1_1_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4 = {
.as_u32 = clib_host_to_net_u32(0x01010101),
},
},
};
/*
* 2 high, 2 medium and 2 low preference non-recursive paths
*/
fib_route_path_t nr_path_hi_1 = {
.frp_proto = DPO_PROTO_IP4,
.frp_sw_if_index = tm->hw[0]->sw_if_index,
.frp_fib_index = ~0,
.frp_weight = 1,
.frp_preference = 0,
.frp_flags = FIB_ROUTE_PATH_FLAG_NONE,
.frp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a01),
},
};
fib_route_path_t nr_path_hi_2 = {
.frp_proto = DPO_PROTO_IP4,
.frp_sw_if_index = tm->hw[0]->sw_if_index,
.frp_fib_index = ~0,
.frp_weight = 1,
.frp_preference = 0,
.frp_flags = FIB_ROUTE_PATH_FLAG_NONE,
.frp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a02),
},
};
fib_route_path_t nr_path_med_1 = {
.frp_proto = DPO_PROTO_IP4,
.frp_sw_if_index = tm->hw[1]->sw_if_index,
.frp_fib_index = ~0,
.frp_weight = 1,
.frp_preference = 1,
.frp_flags = FIB_ROUTE_PATH_FLAG_NONE,
.frp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0c01),
},
};
fib_route_path_t nr_path_med_2 = {
.frp_proto = DPO_PROTO_IP4,
.frp_sw_if_index = tm->hw[1]->sw_if_index,
.frp_fib_index = ~0,
.frp_weight = 1,
.frp_preference = 1,
.frp_flags = FIB_ROUTE_PATH_FLAG_NONE,
.frp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0c01),
},
};
fib_route_path_t nr_path_low_1 = {
.frp_proto = DPO_PROTO_IP4,
.frp_sw_if_index = tm->hw[2]->sw_if_index,
.frp_fib_index = ~0,
.frp_weight = 1,
.frp_preference = 2,
.frp_flags = FIB_ROUTE_PATH_FLAG_NONE,
.frp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0b01),
},
};
fib_route_path_t nr_path_low_2 = {
.frp_proto = DPO_PROTO_IP4,
.frp_sw_if_index = tm->hw[2]->sw_if_index,
.frp_fib_index = ~0,
.frp_weight = 1,
.frp_preference = 2,
.frp_flags = FIB_ROUTE_PATH_FLAG_NONE,
.frp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0b02),
},
};
fib_route_path_t *nr_paths = NULL;
vec_add1(nr_paths, nr_path_hi_1);
vec_add1(nr_paths, nr_path_hi_2);
vec_add1(nr_paths, nr_path_med_1);
vec_add1(nr_paths, nr_path_med_2);
vec_add1(nr_paths, nr_path_low_1);
vec_add1(nr_paths, nr_path_low_2);
adj_index_t ai_hi_1 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nr_path_hi_1.frp_addr,
nr_path_hi_1.frp_sw_if_index);
adj_index_t ai_hi_2 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nr_path_hi_2.frp_addr,
nr_path_hi_2.frp_sw_if_index);
adj_index_t ai_med_1 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nr_path_med_1.frp_addr,
nr_path_med_1.frp_sw_if_index);
adj_index_t ai_med_2 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nr_path_med_2.frp_addr,
nr_path_med_2.frp_sw_if_index);
adj_index_t ai_low_1 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nr_path_low_1.frp_addr,
nr_path_low_1.frp_sw_if_index);
adj_index_t ai_low_2 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nr_path_low_2.frp_addr,
nr_path_low_2.frp_sw_if_index);
fib_test_lb_bucket_t ip_hi_1 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_hi_1,
},
};
fib_test_lb_bucket_t ip_hi_2 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_hi_2,
},
};
fib_test_lb_bucket_t ip_med_1 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_med_1,
},
};
fib_test_lb_bucket_t ip_med_2 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_med_2,
},
};
fib_test_lb_bucket_t ip_low_1 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_low_1,
},
};
fib_test_lb_bucket_t ip_low_2 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_low_2,
},
};
fib_node_index_t fei;
fei = fib_table_entry_path_add2(0,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
nr_paths);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&ip_hi_1,
&ip_hi_2),
"1.1.1.1/32 via high preference paths");
/*
* bring down the interface on which the high preference path lie
*/
vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[0]->sw_if_index,
0);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&ip_med_1,
&ip_med_2),
"1.1.1.1/32 via medium preference paths");
/*
* bring down the interface on which the medium preference path lie
*/
vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[1]->sw_if_index,
0);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&ip_low_1,
&ip_low_2),
"1.1.1.1/32 via low preference paths");
/*
* bring up the interface on which the high preference path lie
*/
vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[0]->sw_if_index,
VNET_SW_INTERFACE_FLAG_ADMIN_UP);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&ip_hi_1,
&ip_hi_2),
"1.1.1.1/32 via high preference paths");
/*
* bring up the interface on which the medium preference path lie
*/
vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[1]->sw_if_index,
VNET_SW_INTERFACE_FLAG_ADMIN_UP);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&ip_hi_1,
&ip_hi_2),
"1.1.1.1/32 via high preference paths");
dpo_id_t ip_1_1_1_1 = DPO_INVALID;
fib_entry_contribute_forwarding(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
&ip_1_1_1_1);
/*
* 3 recursive paths of different preference
*/
const fib_prefix_t pfx_1_1_1_2_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4 = {
.as_u32 = clib_host_to_net_u32(0x01010102),
},
},
};
const fib_prefix_t pfx_1_1_1_3_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4 = {
.as_u32 = clib_host_to_net_u32(0x01010103),
},
},
};
fei = fib_table_entry_path_add2(0,
&pfx_1_1_1_2_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
nr_paths);
dpo_id_t ip_1_1_1_2 = DPO_INVALID;
fib_entry_contribute_forwarding(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
&ip_1_1_1_2);
fei = fib_table_entry_path_add2(0,
&pfx_1_1_1_3_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
nr_paths);
dpo_id_t ip_1_1_1_3 = DPO_INVALID;
fib_entry_contribute_forwarding(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
&ip_1_1_1_3);
fib_test_lb_bucket_t ip_o_1_1_1_1 = {
.type = FT_LB_O_LB,
.lb = {
.lb = ip_1_1_1_1.dpoi_index,
},
};
fib_test_lb_bucket_t ip_o_1_1_1_2 = {
.type = FT_LB_O_LB,
.lb = {
.lb = ip_1_1_1_2.dpoi_index,
},
};
fib_test_lb_bucket_t ip_o_1_1_1_3 = {
.type = FT_LB_O_LB,
.lb = {
.lb = ip_1_1_1_3.dpoi_index,
},
};
fib_route_path_t r_path_hi = {
.frp_proto = DPO_PROTO_IP4,
.frp_sw_if_index = ~0,
.frp_fib_index = 0,
.frp_weight = 1,
.frp_preference = 0,
.frp_flags = FIB_ROUTE_PATH_RESOLVE_VIA_HOST,
.frp_addr = pfx_1_1_1_1_s_32.fp_addr,
};
fib_route_path_t r_path_med = {
.frp_proto = DPO_PROTO_IP4,
.frp_sw_if_index = ~0,
.frp_fib_index = 0,
.frp_weight = 1,
.frp_preference = 10,
.frp_flags = FIB_ROUTE_PATH_FLAG_NONE,
.frp_addr = pfx_1_1_1_2_s_32.fp_addr,
};
fib_route_path_t r_path_low = {
.frp_proto = DPO_PROTO_IP4,
.frp_sw_if_index = ~0,
.frp_fib_index = 0,
.frp_weight = 1,
.frp_preference = 255,
.frp_flags = FIB_ROUTE_PATH_RESOLVE_VIA_HOST,
.frp_addr = pfx_1_1_1_3_s_32.fp_addr,
};
fib_route_path_t *r_paths = NULL;
vec_add1(r_paths, r_path_hi);
vec_add1(r_paths, r_path_low);
vec_add1(r_paths, r_path_med);
/*
* add many recursive so we get the LB MAp created
*/
#define N_PFXS 64
fib_prefix_t pfx_r[N_PFXS];
unsigned int n_pfxs;
for (n_pfxs = 0; n_pfxs < N_PFXS; n_pfxs++)
{
pfx_r[n_pfxs].fp_len = 32;
pfx_r[n_pfxs].fp_proto = FIB_PROTOCOL_IP4;
pfx_r[n_pfxs].fp_addr.ip4.as_u32 =
clib_host_to_net_u32(0x02000000 + n_pfxs);
fei = fib_table_entry_path_add2(0,
&pfx_r[n_pfxs],
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
r_paths);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_1_1_1_1),
"recursive via high preference paths");
/*
* withdraw hig pref resolving entry
*/
fib_table_entry_delete(0,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API);
/* suspend so the update walk kicks int */
vlib_process_suspend(vlib_get_main(), 1e-5);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_1_1_1_2),
"recursive via medium preference paths");
/*
* withdraw medium pref resolving entry
*/
fib_table_entry_delete(0,
&pfx_1_1_1_2_s_32,
FIB_SOURCE_API);
/* suspend so the update walk kicks int */
vlib_process_suspend(vlib_get_main(), 1e-5);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_1_1_1_3),
"recursive via low preference paths");
/*
* add back paths for next iteration
*/
fei = fib_table_entry_update(0,
&pfx_1_1_1_2_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
nr_paths);
fei = fib_table_entry_update(0,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
nr_paths);
/* suspend so the update walk kicks int */
vlib_process_suspend(vlib_get_main(), 1e-5);
fei = fib_table_lookup_exact_match(0, &pfx_r[n_pfxs]);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_1_1_1_1),
"recursive via high preference paths");
}
fib_table_entry_delete(0,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API);
/* suspend so the update walk kicks int */
vlib_process_suspend(vlib_get_main(), 1e-5);
for (n_pfxs = 0; n_pfxs < N_PFXS; n_pfxs++)
{
fei = fib_table_lookup_exact_match(0, &pfx_r[n_pfxs]);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_1_1_1_2),
"recursive via medium preference paths");
}
for (n_pfxs = 0; n_pfxs < N_PFXS; n_pfxs++)
{
fib_table_entry_delete(0,
&pfx_r[n_pfxs],
FIB_SOURCE_API);
}
/*
* Cleanup
*/
fib_table_entry_delete(0,
&pfx_1_1_1_2_s_32,
FIB_SOURCE_API);
fib_table_entry_delete(0,
&pfx_1_1_1_3_s_32,
FIB_SOURCE_API);
dpo_reset(&ip_1_1_1_1);
dpo_reset(&ip_1_1_1_2);
dpo_reset(&ip_1_1_1_3);
adj_unlock(ai_low_2);
adj_unlock(ai_low_1);
adj_unlock(ai_med_2);
adj_unlock(ai_med_1);
adj_unlock(ai_hi_2);
adj_unlock(ai_hi_1);
return (0);
}
/*
* Test the recursive route route handling for GRE tunnels
*/
static int
fib_test_label (void)
{
fib_node_index_t fei, ai_mpls_10_10_10_1, ai_v4_10_10_11_1, ai_v4_10_10_11_2, ai_mpls_10_10_11_2, ai_mpls_10_10_11_1;
const u32 fib_index = 0;
test_main_t *tm;
ip4_main_t *im;
int lb_count, ii;
lb_count = pool_elts(load_balance_pool);
tm = &test_main;
im = &ip4_main;
/*
* add interface routes. We'll assume this works. It's more rigorously
* tested elsewhere.
*/
fib_prefix_t local0_pfx = {
.fp_len = 24,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4 = {
/* 10.10.10.10 */
.as_u32 = clib_host_to_net_u32(0x0a0a0a0a),
},
},
};
FIB_TEST((0 == adj_nbr_db_size()), "ADJ DB size is %d",
adj_nbr_db_size());
vec_validate(im->fib_index_by_sw_if_index, tm->hw[0]->sw_if_index);
im->fib_index_by_sw_if_index[tm->hw[0]->sw_if_index] = fib_index;
fib_table_entry_update_one_path(fib_index, &local0_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_ATTACHED),
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &local0_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei),
"attached interface route present");
local0_pfx.fp_len = 32;
fib_table_entry_update_one_path(fib_index, &local0_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_LOCAL),
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &local0_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei),
"local interface route present");
fib_prefix_t local1_pfx = {
.fp_len = 24,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4 = {
/* 10.10.11.10 */
.as_u32 = clib_host_to_net_u32(0x0a0a0b0a),
},
},
};
vec_validate(im->fib_index_by_sw_if_index, tm->hw[1]->sw_if_index);
im->fib_index_by_sw_if_index[tm->hw[1]->sw_if_index] = fib_index;
fib_table_entry_update_one_path(fib_index, &local1_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_ATTACHED),
DPO_PROTO_IP4,
NULL,
tm->hw[1]->sw_if_index,
~0,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &local1_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei),
"attached interface route present");
local1_pfx.fp_len = 32;
fib_table_entry_update_one_path(fib_index, &local1_pfx,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_LOCAL),
DPO_PROTO_IP4,
NULL,
tm->hw[1]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup_exact_match(fib_index, &local1_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei),
"local interface route present");
ip46_address_t nh_10_10_10_1 = {
.ip4 = {
.as_u32 = clib_host_to_net_u32(0x0a0a0a01),
},
};
ip46_address_t nh_10_10_11_1 = {
.ip4 = {
.as_u32 = clib_host_to_net_u32(0x0a0a0b01),
},
};
ip46_address_t nh_10_10_11_2 = {
.ip4 = {
.as_u32 = clib_host_to_net_u32(0x0a0a0b02),
},
};
ai_v4_10_10_11_1 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nh_10_10_11_1,
tm->hw[1]->sw_if_index);
ai_v4_10_10_11_2 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nh_10_10_11_2,
tm->hw[1]->sw_if_index);
ai_mpls_10_10_10_1 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_MPLS,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index);
ai_mpls_10_10_11_2 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_MPLS,
&nh_10_10_11_2,
tm->hw[1]->sw_if_index);
ai_mpls_10_10_11_1 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_MPLS,
&nh_10_10_11_1,
tm->hw[1]->sw_if_index);
/*
* Add an etry with one path with a real out-going label
*/
fib_prefix_t pfx_1_1_1_1_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x01010101),
},
};
fib_test_lb_bucket_t l99_eos_o_10_10_10_1 = {
.type = FT_LB_LABEL_O_ADJ,
.label_o_adj = {
.adj = ai_mpls_10_10_10_1,
.label = 99,
.eos = MPLS_EOS,
},
};
fib_test_lb_bucket_t l99_neos_o_10_10_10_1 = {
.type = FT_LB_LABEL_O_ADJ,
.label_o_adj = {
.adj = ai_mpls_10_10_10_1,
.label = 99,
.eos = MPLS_NON_EOS,
},
};
mpls_label_t *l99 = NULL;
vec_add1(l99, 99);
fib_table_entry_update_one_path(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
l99,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_1_1_1_1_s_32);
FIB_TEST((FIB_NODE_INDEX_INVALID != fei), "1.1.1.1/32 created");
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&l99_eos_o_10_10_10_1),
"1.1.1.1/32 LB 1 bucket via label 99 over 10.10.10.1");
/*
* add a path with an implicit NULL label
*/
fib_test_lb_bucket_t a_o_10_10_11_1 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_v4_10_10_11_1,
},
};
fib_test_lb_bucket_t a_mpls_o_10_10_11_1 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_mpls_10_10_11_1,
},
};
mpls_label_t *l_imp_null = NULL;
vec_add1(l_imp_null, MPLS_IETF_IMPLICIT_NULL_LABEL);
fei = fib_table_entry_path_add(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_11_1,
tm->hw[1]->sw_if_index,
~0, // invalid fib index
1,
l_imp_null,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&l99_eos_o_10_10_10_1,
&a_o_10_10_11_1),
"1.1.1.1/32 LB 2 buckets via: "
"label 99 over 10.10.10.1, "
"adj over 10.10.11.1");
/*
* assign the route a local label
*/
fib_table_entry_local_label_add(fib_index,
&pfx_1_1_1_1_s_32,
24001);
fib_prefix_t pfx_24001_eos = {
.fp_proto = FIB_PROTOCOL_MPLS,
.fp_label = 24001,
.fp_eos = MPLS_EOS,
};
fib_prefix_t pfx_24001_neos = {
.fp_proto = FIB_PROTOCOL_MPLS,
.fp_label = 24001,
.fp_eos = MPLS_NON_EOS,
};
/*
* The EOS entry should link to both the paths,
* and use an ip adj for the imp-null
* The NON-EOS entry should link to both the paths,
* and use an mpls adj for the imp-null
*/
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_eos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
2,
&l99_eos_o_10_10_10_1,
&a_o_10_10_11_1),
"24001/eos LB 2 buckets via: "
"label 99 over 10.10.10.1, "
"adj over 10.10.11.1");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_neos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
2,
&l99_neos_o_10_10_10_1,
&a_mpls_o_10_10_11_1),
"24001/neos LB 1 bucket via: "
"label 99 over 10.10.10.1 ",
"mpls-adj via 10.10.11.1");
/*
* add an unlabelled path, this is excluded from the neos chains,
*/
fib_test_lb_bucket_t adj_o_10_10_11_2 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_v4_10_10_11_2,
},
};
fei = fib_table_entry_path_add(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_11_2,
tm->hw[1]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
16, // 3 choices spread over 16 buckets
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2),
"1.1.1.1/32 LB 16 buckets via: "
"label 99 over 10.10.10.1, "
"adj over 10.10.11.1",
"adj over 10.10.11.2");
/*
* get and lock a reference to the non-eos of the via entry 1.1.1.1/32
*/
dpo_id_t non_eos_1_1_1_1 = DPO_INVALID;
fib_entry_contribute_forwarding(fei,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
&non_eos_1_1_1_1);
/*
* n-eos has only the 2 labelled paths
*/
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_neos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
2,
&l99_neos_o_10_10_10_1,
&a_mpls_o_10_10_11_1),
"24001/neos LB 2 buckets via: "
"label 99 over 10.10.10.1, "
"adj-mpls over 10.10.11.2");
/*
* A labelled recursive
*/
fib_prefix_t pfx_2_2_2_2_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x02020202),
},
};
fib_test_lb_bucket_t l1600_eos_o_1_1_1_1 = {
.type = FT_LB_LABEL_O_LB,
.label_o_lb = {
.lb = non_eos_1_1_1_1.dpoi_index,
.label = 1600,
.eos = MPLS_EOS,
},
};
mpls_label_t *l1600 = NULL;
vec_add1(l1600, 1600);
fib_table_entry_update_one_path(fib_index,
&pfx_2_2_2_2_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_1_s_32.fp_addr,
~0,
fib_index,
1,
l1600,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_2_2_2_2_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&l1600_eos_o_1_1_1_1),
"2.2.2.2.2/32 LB 1 buckets via: "
"label 1600 over 1.1.1.1");
dpo_id_t dpo_44 = DPO_INVALID;
index_t urpfi;
fib_entry_contribute_forwarding(fei, FIB_FORW_CHAIN_TYPE_UNICAST_IP4, &dpo_44);
urpfi = load_balance_get_urpf(dpo_44.dpoi_index);
FIB_TEST(fib_urpf_check(urpfi, tm->hw[0]->sw_if_index),
"uRPF check for 2.2.2.2/32 on %d OK",
tm->hw[0]->sw_if_index);
FIB_TEST(fib_urpf_check(urpfi, tm->hw[1]->sw_if_index),
"uRPF check for 2.2.2.2/32 on %d OK",
tm->hw[1]->sw_if_index);
FIB_TEST(!fib_urpf_check(urpfi, 99),
"uRPF check for 2.2.2.2/32 on 99 not-OK",
99);
fib_entry_contribute_forwarding(fei, FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS, &dpo_44);
FIB_TEST(urpfi == load_balance_get_urpf(dpo_44.dpoi_index),
"Shared uRPF on IP and non-EOS chain");
dpo_reset(&dpo_44);
/*
* we are holding a lock on the non-eos LB of the via-entry.
* do a PIC-core failover by shutting the link of the via-entry.
*
* shut down the link with the valid label
*/
vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[0]->sw_if_index,
0);
fei = fib_table_lookup(fib_index, &pfx_1_1_1_1_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&a_o_10_10_11_1,
&adj_o_10_10_11_2),
"1.1.1.1/32 LB 2 buckets via: "
"adj over 10.10.11.1, ",
"adj-v4 over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_eos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
2,
&a_o_10_10_11_1,
&adj_o_10_10_11_2),
"24001/eos LB 2 buckets via: "
"adj over 10.10.11.1, ",
"adj-v4 over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_neos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
1,
&a_mpls_o_10_10_11_1),
"24001/neos LB 1 buckets via: "
"adj-mpls over 10.10.11.2");
/*
* test that the pre-failover load-balance has been in-place
* modified
*/
dpo_id_t current = DPO_INVALID;
fib_entry_contribute_forwarding(fei,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
&current);
FIB_TEST(!dpo_cmp(&non_eos_1_1_1_1,
&current),
"PIC-core LB inplace modified %U %U",
format_dpo_id, &non_eos_1_1_1_1, 0,
format_dpo_id, &current, 0);
dpo_reset(&non_eos_1_1_1_1);
dpo_reset(&current);
/*
* no-shut the link with the valid label
*/
vnet_sw_interface_set_flags(vnet_get_main(),
tm->hw[0]->sw_if_index,
VNET_SW_INTERFACE_FLAG_ADMIN_UP);
fei = fib_table_lookup(fib_index, &pfx_1_1_1_1_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
16, // 3 choices spread over 16 buckets
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2),
"1.1.1.1/32 LB 16 buckets via: "
"label 99 over 10.10.10.1, "
"adj over 10.10.11.1",
"adj-v4 over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_eos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
16, // 3 choices spread over 16 buckets
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&l99_eos_o_10_10_10_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&a_o_10_10_11_1,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2,
&adj_o_10_10_11_2),
"24001/eos LB 16 buckets via: "
"label 99 over 10.10.10.1, "
"adj over 10.10.11.1",
"adj-v4 over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_neos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
2,
&l99_neos_o_10_10_10_1,
&a_mpls_o_10_10_11_1),
"24001/neos LB 2 buckets via: "
"label 99 over 10.10.10.1, "
"adj-mpls over 10.10.11.2");
/*
* remove the first path with the valid label
*/
fib_table_entry_path_remove(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_1_1_1_1_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&a_o_10_10_11_1,
&adj_o_10_10_11_2),
"1.1.1.1/32 LB 2 buckets via: "
"adj over 10.10.11.1, "
"adj-v4 over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_eos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
2,
&a_o_10_10_11_1,
&adj_o_10_10_11_2),
"24001/eos LB 2 buckets via: "
"adj over 10.10.11.1, "
"adj-v4 over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_neos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
1,
&a_mpls_o_10_10_11_1),
"24001/neos LB 1 buckets via: "
"adj-mpls over 10.10.11.2");
/*
* remove the other path with a valid label
*/
fib_test_lb_bucket_t bucket_drop = {
.type = FT_LB_SPECIAL,
.special = {
.adj = DPO_PROTO_IP4,
},
};
fib_test_lb_bucket_t mpls_bucket_drop = {
.type = FT_LB_SPECIAL,
.special = {
.adj = DPO_PROTO_MPLS,
},
};
fib_table_entry_path_remove(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
DPO_PROTO_IP4,
&nh_10_10_11_1,
tm->hw[1]->sw_if_index,
~0, // invalid fib index
1,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_1_1_1_1_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&adj_o_10_10_11_2),
"1.1.1.1/32 LB 1 buckets via: "
"adj over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_eos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
1,
&adj_o_10_10_11_2),
"24001/eos LB 1 buckets via: "
"adj over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_neos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
1,
&mpls_bucket_drop),
"24001/neos LB 1 buckets via: DROP");
/*
* add back the path with the valid label
*/
l99 = NULL;
vec_add1(l99, 99);
fib_table_entry_path_add(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
l99,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_1_1_1_1_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&l99_eos_o_10_10_10_1,
&adj_o_10_10_11_2),
"1.1.1.1/32 LB 2 buckets via: "
"label 99 over 10.10.10.1, "
"adj over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_eos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
2,
&l99_eos_o_10_10_10_1,
&adj_o_10_10_11_2),
"24001/eos LB 2 buckets via: "
"label 99 over 10.10.10.1, "
"adj over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_24001_neos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
1,
&l99_neos_o_10_10_10_1),
"24001/neos LB 1 buckets via: "
"label 99 over 10.10.10.1");
/*
* change the local label
*/
fib_table_entry_local_label_add(fib_index,
&pfx_1_1_1_1_s_32,
25005);
fib_prefix_t pfx_25005_eos = {
.fp_proto = FIB_PROTOCOL_MPLS,
.fp_label = 25005,
.fp_eos = MPLS_EOS,
};
fib_prefix_t pfx_25005_neos = {
.fp_proto = FIB_PROTOCOL_MPLS,
.fp_label = 25005,
.fp_eos = MPLS_NON_EOS,
};
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup(fib_index, &pfx_24001_eos)),
"24001/eos removed after label change");
FIB_TEST((FIB_NODE_INDEX_INVALID ==
fib_table_lookup(fib_index, &pfx_24001_neos)),
"24001/eos removed after label change");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_25005_eos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
2,
&l99_eos_o_10_10_10_1,
&adj_o_10_10_11_2),
"25005/eos LB 2 buckets via: "
"label 99 over 10.10.10.1, "
"adj over 10.10.11.2");
fei = fib_table_lookup(MPLS_FIB_DEFAULT_TABLE_ID,
&pfx_25005_neos);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
1,
&l99_neos_o_10_10_10_1),
"25005/neos LB 1 buckets via: "
"label 99 over 10.10.10.1");
/*
* remove the local label.
* the check that the MPLS entries are gone is done by the fact the
* MPLS table is no longer present.
*/
fib_table_entry_local_label_remove(fib_index,
&pfx_1_1_1_1_s_32,
25005);
fei = fib_table_lookup(fib_index, &pfx_1_1_1_1_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&l99_eos_o_10_10_10_1,
&adj_o_10_10_11_2),
"24001/eos LB 2 buckets via: "
"label 99 over 10.10.10.1, "
"adj over 10.10.11.2");
FIB_TEST((FIB_NODE_INDEX_INVALID ==
mpls_fib_index_from_table_id(MPLS_FIB_DEFAULT_TABLE_ID)),
"No more MPLS FIB entries => table removed");
/*
* add another via-entry for the recursive
*/
fib_prefix_t pfx_1_1_1_2_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x01010102),
},
};
fib_test_lb_bucket_t l101_eos_o_10_10_10_1 = {
.type = FT_LB_LABEL_O_ADJ,
.label_o_adj = {
.adj = ai_mpls_10_10_10_1,
.label = 101,
.eos = MPLS_EOS,
},
};
mpls_label_t *l101 = NULL;
vec_add1(l101, 101);
fei = fib_table_entry_update_one_path(fib_index,
&pfx_1_1_1_2_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
l101,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&l101_eos_o_10_10_10_1),
"1.1.1.2/32 LB 1 buckets via: "
"label 101 over 10.10.10.1");
dpo_id_t non_eos_1_1_1_2 = DPO_INVALID;
fib_entry_contribute_forwarding(fib_table_lookup(fib_index,
&pfx_1_1_1_1_s_32),
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
&non_eos_1_1_1_1);
fib_entry_contribute_forwarding(fib_table_lookup(fib_index,
&pfx_1_1_1_2_s_32),
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
&non_eos_1_1_1_2);
fib_test_lb_bucket_t l1601_eos_o_1_1_1_2 = {
.type = FT_LB_LABEL_O_LB,
.label_o_lb = {
.lb = non_eos_1_1_1_2.dpoi_index,
.label = 1601,
.eos = MPLS_EOS,
},
};
mpls_label_t *l1601 = NULL;
vec_add1(l1601, 1601);
l1600_eos_o_1_1_1_1.label_o_lb.lb = non_eos_1_1_1_1.dpoi_index;
fei = fib_table_entry_path_add(fib_index,
&pfx_2_2_2_2_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_2_s_32.fp_addr,
~0,
fib_index,
1,
l1601,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&l1600_eos_o_1_1_1_1,
&l1601_eos_o_1_1_1_2),
"2.2.2.2/32 LB 2 buckets via: "
"label 1600 via 1.1,1.1, "
"label 16001 via 1.1.1.2");
/*
* update the via-entry so it no longer has an imp-null path.
* the LB for the recursive can use an imp-null
*/
l_imp_null = NULL;
vec_add1(l_imp_null, MPLS_IETF_IMPLICIT_NULL_LABEL);
fei = fib_table_entry_update_one_path(fib_index,
&pfx_1_1_1_2_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_11_1,
tm->hw[1]->sw_if_index,
~0, // invalid fib index
1,
l_imp_null,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&a_o_10_10_11_1),
"1.1.1.2/32 LB 1 buckets via: "
"adj 10.10.11.1");
fei = fib_table_lookup(fib_index, &pfx_2_2_2_2_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&l1600_eos_o_1_1_1_1,
&l1601_eos_o_1_1_1_2),
"2.2.2.2/32 LB 2 buckets via: "
"label 1600 via 1.1,1.1, "
"label 16001 via 1.1.1.2");
/*
* update the via-entry so it no longer has labelled paths.
* the LB for the recursive should exclue this via form its LB
*/
fei = fib_table_entry_update_one_path(fib_index,
&pfx_1_1_1_2_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_11_1,
tm->hw[1]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&a_o_10_10_11_1),
"1.1.1.2/32 LB 1 buckets via: "
"adj 10.10.11.1");
fei = fib_table_lookup(fib_index, &pfx_2_2_2_2_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&l1600_eos_o_1_1_1_1),
"2.2.2.2/32 LB 1 buckets via: "
"label 1600 via 1.1,1.1");
dpo_reset(&non_eos_1_1_1_1);
dpo_reset(&non_eos_1_1_1_2);
/*
* Add a recursive with no out-labels. We expect to use the IP of the via
*/
fib_prefix_t pfx_2_2_2_3_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x02020203),
},
};
dpo_id_t ip_1_1_1_1 = DPO_INVALID;
fib_table_entry_update_one_path(fib_index,
&pfx_2_2_2_3_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_1_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fib_entry_contribute_forwarding(fib_table_lookup(fib_index,
&pfx_1_1_1_1_s_32),
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
&ip_1_1_1_1);
fib_test_lb_bucket_t ip_o_1_1_1_1 = {
.type = FT_LB_O_LB,
.lb = {
.lb = ip_1_1_1_1.dpoi_index,
},
};
fei = fib_table_lookup(fib_index, &pfx_2_2_2_3_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_1_1_1_1),
"2.2.2.2.3/32 LB 1 buckets via: "
"ip 1.1.1.1");
/*
* Add a recursive with an imp-null out-label.
* We expect to use the IP of the via
*/
fib_prefix_t pfx_2_2_2_4_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x02020204),
},
};
fib_table_entry_update_one_path(fib_index,
&pfx_2_2_2_4_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&pfx_1_1_1_1_s_32.fp_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fei = fib_table_lookup(fib_index, &pfx_2_2_2_4_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_1_1_1_1),
"2.2.2.2.4/32 LB 1 buckets via: "
"ip 1.1.1.1");
dpo_reset(&ip_1_1_1_1);
/*
* Create an entry with a deep label stack
*/
fib_prefix_t pfx_2_2_5_5_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x02020505),
},
};
fib_test_lb_bucket_t ls_eos_o_10_10_10_1 = {
.type = FT_LB_LABEL_STACK_O_ADJ,
.label_stack_o_adj = {
.adj = ai_mpls_10_10_11_1,
.label_stack_size = 8,
.label_stack = {
200, 201, 202, 203, 204, 205, 206, 207
},
.eos = MPLS_EOS,
},
};
mpls_label_t *label_stack = NULL;
vec_validate(label_stack, 7);
for (ii = 0; ii < 8; ii++)
{
label_stack[ii] = ii + 200;
}
fei = fib_table_entry_update_one_path(fib_index,
&pfx_2_2_5_5_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_11_1,
tm->hw[1]->sw_if_index,
~0, // invalid fib index
1,
label_stack,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ls_eos_o_10_10_10_1),
"2.2.5.5/32 LB 1 buckets via: "
"adj 10.10.11.1");
fib_table_entry_delete_index(fei, FIB_SOURCE_API);
/*
* cleanup
*/
fib_table_entry_delete(fib_index,
&pfx_1_1_1_2_s_32,
FIB_SOURCE_API);
fei = fib_table_lookup(fib_index, &pfx_2_2_2_2_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&l1600_eos_o_1_1_1_1),
"2.2.2.2/32 LB 1 buckets via: "
"label 1600 via 1.1,1.1");
fib_table_entry_delete(fib_index,
&pfx_1_1_1_1_s_32,
FIB_SOURCE_API);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&bucket_drop),
"2.2.2.2/32 LB 1 buckets via: DROP");
fib_table_entry_delete(fib_index,
&pfx_2_2_2_2_s_32,
FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&pfx_2_2_2_3_s_32,
FIB_SOURCE_API);
fib_table_entry_delete(fib_index,
&pfx_2_2_2_4_s_32,
FIB_SOURCE_API);
adj_unlock(ai_mpls_10_10_10_1);
adj_unlock(ai_mpls_10_10_11_2);
adj_unlock(ai_v4_10_10_11_1);
adj_unlock(ai_v4_10_10_11_2);
adj_unlock(ai_mpls_10_10_11_1);
FIB_TEST((0 == adj_nbr_db_size()), "ADJ DB size is %d",
adj_nbr_db_size());
local0_pfx.fp_len = 32;
fib_table_entry_delete(fib_index,
&local0_pfx,
FIB_SOURCE_INTERFACE);
local0_pfx.fp_len = 24;
fib_table_entry_delete(fib_index,
&local0_pfx,
FIB_SOURCE_INTERFACE);
local1_pfx.fp_len = 32;
fib_table_entry_delete(fib_index,
&local1_pfx,
FIB_SOURCE_INTERFACE);
local1_pfx.fp_len = 24;
fib_table_entry_delete(fib_index,
&local1_pfx,
FIB_SOURCE_INTERFACE);
/*
* +1 for the drop LB in the MPLS tables.
*/
FIB_TEST(lb_count+1 == pool_elts(load_balance_pool),
"Load-balance resources freed %d of %d",
lb_count+1, pool_elts(load_balance_pool));
return (0);
}
#define N_TEST_CHILDREN 4
#define PARENT_INDEX 0
typedef struct fib_node_test_t_
{
fib_node_t node;
u32 sibling;
u32 index;
fib_node_back_walk_ctx_t *ctxs;
u32 destroyed;
} fib_node_test_t;
static fib_node_test_t fib_test_nodes[N_TEST_CHILDREN+1];
#define PARENT() (&fib_test_nodes[PARENT_INDEX].node)
#define FOR_EACH_TEST_CHILD(_tc) \
for (ii = 1, (_tc) = &fib_test_nodes[1]; \
ii < N_TEST_CHILDREN+1; \
ii++, (_tc) = &fib_test_nodes[ii])
static fib_node_t *
fib_test_child_get_node (fib_node_index_t index)
{
return (&fib_test_nodes[index].node);
}
static int fib_test_walk_spawns_walks;
static fib_node_back_walk_rc_t
fib_test_child_back_walk_notify (fib_node_t *node,
fib_node_back_walk_ctx_t *ctx)
{
fib_node_test_t *tc = (fib_node_test_t*) node;
vec_add1(tc->ctxs, *ctx);
if (1 == fib_test_walk_spawns_walks)
fib_walk_sync(FIB_NODE_TYPE_TEST, tc->index, ctx);
if (2 == fib_test_walk_spawns_walks)
fib_walk_async(FIB_NODE_TYPE_TEST, tc->index,
FIB_WALK_PRIORITY_HIGH, ctx);
return (FIB_NODE_BACK_WALK_CONTINUE);
}
static void
fib_test_child_last_lock_gone (fib_node_t *node)
{
fib_node_test_t *tc = (fib_node_test_t *)node;
tc->destroyed = 1;
}
/**
* The FIB walk's graph node virtual function table
*/
static const fib_node_vft_t fib_test_child_vft = {
.fnv_get = fib_test_child_get_node,
.fnv_last_lock = fib_test_child_last_lock_gone,
.fnv_back_walk = fib_test_child_back_walk_notify,
};
/*
* the function (that should have been static but isn't so I can do this)
* that processes the walk from the async queue,
*/
f64 fib_walk_process_queues(vlib_main_t * vm,
const f64 quota);
u32 fib_walk_queue_get_size(fib_walk_priority_t prio);
static int
fib_test_walk (void)
{
fib_node_back_walk_ctx_t high_ctx = {}, low_ctx = {};
fib_node_test_t *tc;
vlib_main_t *vm;
u32 ii;
vm = vlib_get_main();
fib_node_register_type(FIB_NODE_TYPE_TEST, &fib_test_child_vft);
/*
* init a fake node on which we will add children
*/
fib_node_init(&fib_test_nodes[PARENT_INDEX].node,
FIB_NODE_TYPE_TEST);
FOR_EACH_TEST_CHILD(tc)
{
fib_node_init(&tc->node, FIB_NODE_TYPE_TEST);
fib_node_lock(&tc->node);
tc->ctxs = NULL;
tc->index = ii;
tc->sibling = fib_node_child_add(FIB_NODE_TYPE_TEST,
PARENT_INDEX,
FIB_NODE_TYPE_TEST, ii);
}
/*
* enqueue a walk across the parents children.
*/
high_ctx.fnbw_reason = FIB_NODE_BW_REASON_FLAG_RESOLVE;
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &high_ctx);
FIB_TEST(N_TEST_CHILDREN+1 == fib_node_list_get_size(PARENT()->fn_children),
"Parent has %d children pre-walk",
fib_node_list_get_size(PARENT()->fn_children));
/*
* give the walk a large amount of time so it gets to the end
*/
fib_walk_process_queues(vm, 1);
FOR_EACH_TEST_CHILD(tc)
{
FIB_TEST(1 == vec_len(tc->ctxs),
"%d child visitsed %d times",
ii, vec_len(tc->ctxs));
vec_free(tc->ctxs);
}
FIB_TEST(0 == fib_walk_queue_get_size(FIB_WALK_PRIORITY_HIGH),
"Queue is empty post walk");
FIB_TEST(N_TEST_CHILDREN == fib_node_list_get_size(PARENT()->fn_children),
"Parent has %d children post walk",
fib_node_list_get_size(PARENT()->fn_children));
/*
* walk again. should be no increase in the number of visits, since
* the walk will have terminated.
*/
fib_walk_process_queues(vm, 1);
FOR_EACH_TEST_CHILD(tc)
{
FIB_TEST(0 == vec_len(tc->ctxs),
"%d child visitsed %d times",
ii, vec_len(tc->ctxs));
}
/*
* schedule a low and hig priority walk. expect the high to be performed
* before the low.
* schedule the high prio walk first so that it is further from the head
* of the dependency list. that way it won't merge with the low one.
*/
high_ctx.fnbw_reason = FIB_NODE_BW_REASON_FLAG_RESOLVE;
low_ctx.fnbw_reason = FIB_NODE_BW_REASON_FLAG_ADJ_UPDATE;
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &high_ctx);
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_LOW, &low_ctx);
fib_walk_process_queues(vm, 1);
FOR_EACH_TEST_CHILD(tc)
{
FIB_TEST(high_ctx.fnbw_reason == tc->ctxs[0].fnbw_reason,
"%d child visitsed by high prio walk", ii);
FIB_TEST(low_ctx.fnbw_reason == tc->ctxs[1].fnbw_reason,
"%d child visitsed by low prio walk", ii);
vec_free(tc->ctxs);
}
FIB_TEST(0 == fib_walk_queue_get_size(FIB_WALK_PRIORITY_HIGH),
"Queue is empty post prio walk");
FIB_TEST(N_TEST_CHILDREN == fib_node_list_get_size(PARENT()->fn_children),
"Parent has %d children post prio walk",
fib_node_list_get_size(PARENT()->fn_children));
/*
* schedule 2 walks of the same priority that can be megred.
* expect that each child is thus visited only once.
*/
high_ctx.fnbw_reason = FIB_NODE_BW_REASON_FLAG_RESOLVE;
low_ctx.fnbw_reason = FIB_NODE_BW_REASON_FLAG_RESOLVE;
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &high_ctx);
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &low_ctx);
fib_walk_process_queues(vm, 1);
FOR_EACH_TEST_CHILD(tc)
{
FIB_TEST(1 == vec_len(tc->ctxs),
"%d child visitsed %d times during merge walk",
ii, vec_len(tc->ctxs));
vec_free(tc->ctxs);
}
FIB_TEST(0 == fib_walk_queue_get_size(FIB_WALK_PRIORITY_HIGH),
"Queue is empty post merge walk");
FIB_TEST(N_TEST_CHILDREN == fib_node_list_get_size(PARENT()->fn_children),
"Parent has %d children post merge walk",
fib_node_list_get_size(PARENT()->fn_children));
/*
* schedule 2 walks of the same priority that cannot be megred.
* expect that each child is thus visited twice and in the order
* in which the walks were scheduled.
*/
high_ctx.fnbw_reason = FIB_NODE_BW_REASON_FLAG_RESOLVE;
low_ctx.fnbw_reason = FIB_NODE_BW_REASON_FLAG_ADJ_UPDATE;
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &high_ctx);
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &low_ctx);
fib_walk_process_queues(vm, 1);
FOR_EACH_TEST_CHILD(tc)
{
FIB_TEST(high_ctx.fnbw_reason == tc->ctxs[0].fnbw_reason,
"%d child visitsed by high prio walk", ii);
FIB_TEST(low_ctx.fnbw_reason == tc->ctxs[1].fnbw_reason,
"%d child visitsed by low prio walk", ii);
vec_free(tc->ctxs);
}
FIB_TEST(0 == fib_walk_queue_get_size(FIB_WALK_PRIORITY_HIGH),
"Queue is empty post no-merge walk");
FIB_TEST(N_TEST_CHILDREN == fib_node_list_get_size(PARENT()->fn_children),
"Parent has %d children post no-merge walk",
fib_node_list_get_size(PARENT()->fn_children));
/*
* schedule a walk that makes one one child progress.
* we do this by giving the queue draining process zero
* time quanta. it's a do..while loop, so it does something.
*/
high_ctx.fnbw_reason = FIB_NODE_BW_REASON_FLAG_RESOLVE;
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &high_ctx);
fib_walk_process_queues(vm, 0);
FOR_EACH_TEST_CHILD(tc)
{
if (ii == N_TEST_CHILDREN)
{
FIB_TEST(1 == vec_len(tc->ctxs),
"%d child visitsed %d times in zero quanta walk",
ii, vec_len(tc->ctxs));
}
else
{
FIB_TEST(0 == vec_len(tc->ctxs),
"%d child visitsed %d times in 0 quanta walk",
ii, vec_len(tc->ctxs));
}
}
FIB_TEST(1 == fib_walk_queue_get_size(FIB_WALK_PRIORITY_HIGH),
"Queue is not empty post zero quanta walk");
FIB_TEST(N_TEST_CHILDREN+1 == fib_node_list_get_size(PARENT()->fn_children),
"Parent has %d children post zero qunta walk",
fib_node_list_get_size(PARENT()->fn_children));
/*
* another one step
*/
fib_walk_process_queues(vm, 0);
FOR_EACH_TEST_CHILD(tc)
{
if (ii >= N_TEST_CHILDREN-1)
{
FIB_TEST(1 == vec_len(tc->ctxs),
"%d child visitsed %d times in 2nd zero quanta walk",
ii, vec_len(tc->ctxs));
}
else
{
FIB_TEST(0 == vec_len(tc->ctxs),
"%d child visitsed %d times in 2nd 0 quanta walk",
ii, vec_len(tc->ctxs));
}
}
FIB_TEST(1 == fib_walk_queue_get_size(FIB_WALK_PRIORITY_HIGH),
"Queue is not empty post zero quanta walk");
FIB_TEST(N_TEST_CHILDREN+1 == fib_node_list_get_size(PARENT()->fn_children),
"Parent has %d children post zero qunta walk",
fib_node_list_get_size(PARENT()->fn_children));
/*
* schedule another walk that will catch-up and merge.
*/
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &high_ctx);
fib_walk_process_queues(vm, 1);
FOR_EACH_TEST_CHILD(tc)
{
if (ii >= N_TEST_CHILDREN-1)
{
FIB_TEST(2 == vec_len(tc->ctxs),
"%d child visitsed %d times in 2nd zero quanta merge walk",
ii, vec_len(tc->ctxs));
vec_free(tc->ctxs);
}
else
{
FIB_TEST(1 == vec_len(tc->ctxs),
"%d child visitsed %d times in 2nd 0 quanta merge walk",
ii, vec_len(tc->ctxs));
vec_free(tc->ctxs);
}
}
FIB_TEST(0 == fib_walk_queue_get_size(FIB_WALK_PRIORITY_HIGH),
"Queue is not empty post 2nd zero quanta merge walk");
FIB_TEST(N_TEST_CHILDREN == fib_node_list_get_size(PARENT()->fn_children),
"Parent has %d children post 2nd zero qunta merge walk",
fib_node_list_get_size(PARENT()->fn_children));
/*
* park a async walk in the middle of the list, then have an sync walk catch
* it. same expectations as async catches async.
*/
high_ctx.fnbw_reason = FIB_NODE_BW_REASON_FLAG_RESOLVE;
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &high_ctx);
fib_walk_process_queues(vm, 0);
fib_walk_process_queues(vm, 0);
fib_walk_sync(FIB_NODE_TYPE_TEST, PARENT_INDEX, &high_ctx);
FOR_EACH_TEST_CHILD(tc)
{
if (ii >= N_TEST_CHILDREN-1)
{
FIB_TEST(2 == vec_len(tc->ctxs),
"%d child visitsed %d times in sync catches async walk",
ii, vec_len(tc->ctxs));
vec_free(tc->ctxs);
}
else
{
FIB_TEST(1 == vec_len(tc->ctxs),
"%d child visitsed %d times in sync catches async walk",
ii, vec_len(tc->ctxs));
vec_free(tc->ctxs);
}
}
FIB_TEST(0 == fib_walk_queue_get_size(FIB_WALK_PRIORITY_HIGH),
"Queue is not empty post 2nd zero quanta merge walk");
FIB_TEST(N_TEST_CHILDREN == fib_node_list_get_size(PARENT()->fn_children),
"Parent has %d children post 2nd zero qunta merge walk",
fib_node_list_get_size(PARENT()->fn_children));
/*
* make the parent a child of one of its children, thus inducing a routing loop.
*/
fib_test_nodes[PARENT_INDEX].sibling =
fib_node_child_add(FIB_NODE_TYPE_TEST,
1, // the first child
FIB_NODE_TYPE_TEST,
PARENT_INDEX);
/*
* execute a sync walk from the parent. each child visited spawns more sync
* walks. we expect the walk to terminate.
*/
fib_test_walk_spawns_walks = 1;
fib_walk_sync(FIB_NODE_TYPE_TEST, PARENT_INDEX, &high_ctx);
FOR_EACH_TEST_CHILD(tc)
{
/*
* child 1 - which is last in the list - has the loop.
* the other children a re thus visitsed first. the we meet
* child 1. we go round the loop again, visting the other children.
* then we meet the walk in the dep list and bail. child 1 is not visitsed
* again.
*/
if (1 == ii)
{
FIB_TEST(1 == vec_len(tc->ctxs),
"child %d visitsed %d times during looped sync walk",
ii, vec_len(tc->ctxs));
}
else
{
FIB_TEST(2 == vec_len(tc->ctxs),
"child %d visitsed %d times during looped sync walk",
ii, vec_len(tc->ctxs));
}
vec_free(tc->ctxs);
}
FIB_TEST(N_TEST_CHILDREN == fib_node_list_get_size(PARENT()->fn_children),
"Parent has %d children post sync loop walk",
fib_node_list_get_size(PARENT()->fn_children));
/*
* the walk doesn't reach the max depth because the infra knows that sync
* meets sync implies a loop and bails early.
*/
FIB_TEST(high_ctx.fnbw_depth == 9,
"Walk context depth %d post sync loop walk",
high_ctx.fnbw_depth);
/*
* execute an async walk of the graph loop, with each child spawns sync walks
*/
high_ctx.fnbw_depth = 0;
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &high_ctx);
fib_walk_process_queues(vm, 1);
FOR_EACH_TEST_CHILD(tc)
{
/*
* we don't really care how many times the children are visisted, as long as
* it is more than once.
*/
FIB_TEST(1 <= vec_len(tc->ctxs),
"child %d visitsed %d times during looped aync spawns sync walk",
ii, vec_len(tc->ctxs));
vec_free(tc->ctxs);
}
/*
* execute an async walk of the graph loop, with each child spawns async walks
*/
fib_test_walk_spawns_walks = 2;
high_ctx.fnbw_depth = 0;
fib_walk_async(FIB_NODE_TYPE_TEST, PARENT_INDEX,
FIB_WALK_PRIORITY_HIGH, &high_ctx);
fib_walk_process_queues(vm, 1);
FOR_EACH_TEST_CHILD(tc)
{
/*
* we don't really care how many times the children are visisted, as long as
* it is more than once.
*/
FIB_TEST(1 <= vec_len(tc->ctxs),
"child %d visitsed %d times during looped async spawns async walk",
ii, vec_len(tc->ctxs));
vec_free(tc->ctxs);
}
fib_node_child_remove(FIB_NODE_TYPE_TEST,
1, // the first child
fib_test_nodes[PARENT_INDEX].sibling);
/*
* cleanup
*/
FOR_EACH_TEST_CHILD(tc)
{
fib_node_child_remove(FIB_NODE_TYPE_TEST, PARENT_INDEX,
tc->sibling);
fib_node_deinit(&tc->node);
fib_node_unlock(&tc->node);
}
fib_node_deinit(PARENT());
/*
* The parent will be destroyed when the last lock on it goes.
* this test ensures all the walk objects are unlocking it.
*/
FIB_TEST((1 == fib_test_nodes[PARENT_INDEX].destroyed),
"Parent was destroyed");
return (0);
}
/*
* declaration of the otherwise static callback functions
*/
void fib_bfd_notify (bfd_listen_event_e event,
const bfd_session_t *session);
void adj_bfd_notify (bfd_listen_event_e event,
const bfd_session_t *session);
/**
* Test BFD session interaction with FIB
*/
static int
fib_test_bfd (void)
{
fib_node_index_t fei;
test_main_t *tm;
int n_feis;
/* via 10.10.10.1 */
ip46_address_t nh_10_10_10_1 = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a01),
};
/* via 10.10.10.2 */
ip46_address_t nh_10_10_10_2 = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a02),
};
/* via 10.10.10.10 */
ip46_address_t nh_10_10_10_10 = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a0a),
};
n_feis = fib_entry_pool_size();
tm = &test_main;
/*
* add interface routes. we'll assume this works. it's tested elsewhere
*/
fib_prefix_t pfx_10_10_10_10_s_24 = {
.fp_len = 24,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = nh_10_10_10_10,
};
fib_table_entry_update_one_path(0, &pfx_10_10_10_10_s_24,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_ATTACHED),
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1, // weight
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
fib_prefix_t pfx_10_10_10_10_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = nh_10_10_10_10,
};
fib_table_entry_update_one_path(0, &pfx_10_10_10_10_s_32,
FIB_SOURCE_INTERFACE,
(FIB_ENTRY_FLAG_CONNECTED |
FIB_ENTRY_FLAG_LOCAL),
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1, // weight
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
/*
* A BFD session via a neighbour we do not yet know
*/
bfd_session_t bfd_10_10_10_1 = {
.udp = {
.key = {
.fib_index = 0,
.peer_addr = nh_10_10_10_1,
},
},
.hop_type = BFD_HOP_TYPE_MULTI,
.local_state = BFD_STATE_init,
};
fib_bfd_notify (BFD_LISTEN_EVENT_CREATE, &bfd_10_10_10_1);
/*
* A new entry will be created that forwards via the adj
*/
adj_index_t ai_10_10_10_1 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index);
fib_prefix_t pfx_10_10_10_1_s_32 = {
.fp_addr = nh_10_10_10_1,
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
};
fib_test_lb_bucket_t adj_o_10_10_10_1 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_10_10_10_1,
},
};
fei = fib_table_lookup_exact_match(0, &pfx_10_10_10_1_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&adj_o_10_10_10_1),
"BFD sourced %U via %U",
format_fib_prefix, &pfx_10_10_10_1_s_32,
format_ip_adjacency, ai_10_10_10_1, FORMAT_IP_ADJACENCY_NONE);
/*
* Delete the BFD session. Expect the fib_entry to be removed
*/
fib_bfd_notify (BFD_LISTEN_EVENT_DELETE, &bfd_10_10_10_1);
fei = fib_table_lookup_exact_match(0, &pfx_10_10_10_1_s_32);
FIB_TEST(FIB_NODE_INDEX_INVALID == fei,
"BFD sourced %U removed",
format_fib_prefix, &pfx_10_10_10_1_s_32);
/*
* Add the BFD source back
*/
fib_bfd_notify (BFD_LISTEN_EVENT_CREATE, &bfd_10_10_10_1);
/*
* source the entry via the ADJ fib
*/
fei = fib_table_entry_path_add(0,
&pfx_10_10_10_1_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
/*
* Delete the BFD session. Expect the fib_entry to remain
*/
fib_bfd_notify (BFD_LISTEN_EVENT_DELETE, &bfd_10_10_10_1);
fei = fib_table_lookup_exact_match(0, &pfx_10_10_10_1_s_32);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&adj_o_10_10_10_1),
"BFD sourced %U remains via %U",
format_fib_prefix, &pfx_10_10_10_1_s_32,
format_ip_adjacency, ai_10_10_10_1, FORMAT_IP_ADJACENCY_NONE);
/*
* Add the BFD source back
*/
fib_bfd_notify (BFD_LISTEN_EVENT_CREATE, &bfd_10_10_10_1);
/*
* Create another ADJ FIB
*/
fib_prefix_t pfx_10_10_10_2_s_32 = {
.fp_addr = nh_10_10_10_2,
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
};
fib_table_entry_path_add(0,
&pfx_10_10_10_2_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
/*
* A BFD session for the new ADJ FIB
*/
bfd_session_t bfd_10_10_10_2 = {
.udp = {
.key = {
.fib_index = 0,
.peer_addr = nh_10_10_10_2,
},
},
.hop_type = BFD_HOP_TYPE_MULTI,
.local_state = BFD_STATE_init,
};
fib_bfd_notify (BFD_LISTEN_EVENT_CREATE, &bfd_10_10_10_2);
/*
* remove the adj-fib source whilst the session is present
* then add it back
*/
fib_table_entry_delete(0, &pfx_10_10_10_2_s_32, FIB_SOURCE_ADJ);
fib_table_entry_path_add(0,
&pfx_10_10_10_2_s_32,
FIB_SOURCE_ADJ,
FIB_ENTRY_FLAG_ATTACHED,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
/*
* Before adding a recursive via the BFD tracked ADJ-FIBs,
* bring one of the sessions UP, leave the other down
*/
bfd_10_10_10_1.local_state = BFD_STATE_up;
fib_bfd_notify (BFD_LISTEN_EVENT_UPDATE, &bfd_10_10_10_1);
bfd_10_10_10_2.local_state = BFD_STATE_down;
fib_bfd_notify (BFD_LISTEN_EVENT_UPDATE, &bfd_10_10_10_2);
/*
* A recursive prefix via both of the ADJ FIBs
*/
fib_prefix_t pfx_200_0_0_0_s_24 = {
.fp_proto = FIB_PROTOCOL_IP4,
.fp_len = 32,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0xc8000000),
},
};
const dpo_id_t *dpo_10_10_10_1, *dpo_10_10_10_2;
dpo_10_10_10_1 =
fib_entry_contribute_ip_forwarding(
fib_table_lookup_exact_match(0, &pfx_10_10_10_1_s_32));
dpo_10_10_10_2 =
fib_entry_contribute_ip_forwarding(
fib_table_lookup_exact_match(0, &pfx_10_10_10_2_s_32));
fib_test_lb_bucket_t lb_o_10_10_10_1 = {
.type = FT_LB_O_LB,
.lb = {
.lb = dpo_10_10_10_1->dpoi_index,
},
};
fib_test_lb_bucket_t lb_o_10_10_10_2 = {
.type = FT_LB_O_LB,
.lb = {
.lb = dpo_10_10_10_2->dpoi_index,
},
};
/*
* A prefix via the adj-fib that is BFD down => DROP
*/
fei = fib_table_entry_path_add(0,
&pfx_200_0_0_0_s_24,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_2,
~0, // recursive
0, // default fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"%U resolves via drop",
format_fib_prefix, &pfx_200_0_0_0_s_24);
/*
* add a path via the UP BFD adj-fib.
* we expect that the DOWN BFD ADJ FIB is not used.
*/
fei = fib_table_entry_path_add(0,
&pfx_200_0_0_0_s_24,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
~0, // recursive
0, // default fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&lb_o_10_10_10_1),
"Recursive %U only UP BFD adj-fibs",
format_fib_prefix, &pfx_200_0_0_0_s_24);
/*
* Send a BFD state change to UP - both sessions are now up
* the recursive prefix should LB over both
*/
bfd_10_10_10_2.local_state = BFD_STATE_up;
fib_bfd_notify (BFD_LISTEN_EVENT_UPDATE, &bfd_10_10_10_2);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&lb_o_10_10_10_1,
&lb_o_10_10_10_2),
"Recursive %U via both UP BFD adj-fibs",
format_fib_prefix, &pfx_200_0_0_0_s_24);
/*
* Send a BFD state change to DOWN
* the recursive prefix should exclude the down
*/
bfd_10_10_10_2.local_state = BFD_STATE_down;
fib_bfd_notify (BFD_LISTEN_EVENT_UPDATE, &bfd_10_10_10_2);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&lb_o_10_10_10_1),
"Recursive %U via only UP",
format_fib_prefix, &pfx_200_0_0_0_s_24);
/*
* Delete the BFD session while it is in the DOWN state.
* FIB should consider the entry's state as back up
*/
fib_bfd_notify (BFD_LISTEN_EVENT_DELETE, &bfd_10_10_10_2);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&lb_o_10_10_10_1,
&lb_o_10_10_10_2),
"Recursive %U via both UP BFD adj-fibs post down session delete",
format_fib_prefix, &pfx_200_0_0_0_s_24);
/*
* Delete the BFD other session while it is in the UP state.
*/
fib_bfd_notify (BFD_LISTEN_EVENT_DELETE, &bfd_10_10_10_1);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&lb_o_10_10_10_1,
&lb_o_10_10_10_2),
"Recursive %U via both UP BFD adj-fibs post up session delete",
format_fib_prefix, &pfx_200_0_0_0_s_24);
/*
* cleaup
*/
fib_table_entry_delete(0, &pfx_200_0_0_0_s_24, FIB_SOURCE_API);
fib_table_entry_delete(0, &pfx_10_10_10_1_s_32, FIB_SOURCE_ADJ);
fib_table_entry_delete(0, &pfx_10_10_10_2_s_32, FIB_SOURCE_ADJ);
fib_table_entry_delete(0, &pfx_10_10_10_10_s_32, FIB_SOURCE_INTERFACE);
fib_table_entry_delete(0, &pfx_10_10_10_10_s_24, FIB_SOURCE_INTERFACE);
adj_unlock(ai_10_10_10_1);
/*
* test no-one left behind
*/
FIB_TEST((n_feis == fib_entry_pool_size()), "Entries gone");
FIB_TEST(0 == adj_nbr_db_size(), "All adjacencies removed");
/*
* Single-hop BFD tests
*/
bfd_10_10_10_1.hop_type = BFD_HOP_TYPE_SINGLE;
bfd_10_10_10_1.udp.key.sw_if_index = tm->hw[0]->sw_if_index;
adj_bfd_notify(BFD_LISTEN_EVENT_CREATE, &bfd_10_10_10_1);
ai_10_10_10_1 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index);
/*
* whilst the BFD session is not signalled, the adj is up
*/
FIB_TEST(adj_is_up(ai_10_10_10_1), "Adj state up on uninit session");
/*
* bring the BFD session up
*/
bfd_10_10_10_1.local_state = BFD_STATE_up;
adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfd_10_10_10_1);
FIB_TEST(adj_is_up(ai_10_10_10_1), "Adj state up on UP session");
/*
* bring the BFD session down
*/
bfd_10_10_10_1.local_state = BFD_STATE_down;
adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfd_10_10_10_1);
FIB_TEST(!adj_is_up(ai_10_10_10_1), "Adj state down on DOWN session");
/*
* add an attached next hop FIB entry via the down adj
*/
fib_prefix_t pfx_5_5_5_5_s_32 = {
.fp_addr = {
.ip4 = {
.as_u32 = clib_host_to_net_u32(0x05050505),
},
},
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
};
fei = fib_table_entry_path_add(0,
&pfx_5_5_5_5_s_32,
FIB_SOURCE_CLI,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(load_balance_is_drop(fib_entry_contribute_ip_forwarding(fei)),
"%U resolves via drop",
format_fib_prefix, &pfx_5_5_5_5_s_32);
/*
* Add a path via an ADJ that is up
*/
adj_index_t ai_10_10_10_2 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index);
fib_test_lb_bucket_t adj_o_10_10_10_2 = {
.type = FT_LB_ADJ,
.adj = {
.adj = ai_10_10_10_2,
},
};
adj_o_10_10_10_1.adj.adj = ai_10_10_10_1;
fei = fib_table_entry_path_add(0,
&pfx_5_5_5_5_s_32,
FIB_SOURCE_CLI,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_2,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&adj_o_10_10_10_2),
"BFD sourced %U via %U",
format_fib_prefix, &pfx_5_5_5_5_s_32,
format_ip_adjacency, ai_10_10_10_2, FORMAT_IP_ADJACENCY_NONE);
/*
* Bring up the down session - should now LB
*/
bfd_10_10_10_1.local_state = BFD_STATE_up;
adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfd_10_10_10_1);
FIB_TEST(fib_test_validate_entry(fei,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
2,
&adj_o_10_10_10_1,
&adj_o_10_10_10_2),
"BFD sourced %U via noth adjs",
format_fib_prefix, &pfx_5_5_5_5_s_32);
/*
* remove the BFD session state from the adj
*/
adj_bfd_notify(BFD_LISTEN_EVENT_DELETE, &bfd_10_10_10_1);
/*
* clean-up
*/
fib_table_entry_delete(0, &pfx_5_5_5_5_s_32, FIB_SOURCE_CLI);
adj_unlock(ai_10_10_10_1);
adj_unlock(ai_10_10_10_2);
/*
* test no-one left behind
*/
FIB_TEST((n_feis == fib_entry_pool_size()), "Entries gone");
FIB_TEST(0 == adj_nbr_db_size(), "All adjacencies removed");
return (0);
}
static int
lfib_test (void)
{
const mpls_label_t deag_label = 50;
const u32 lfib_index = 0;
const u32 fib_index = 0;
dpo_id_t dpo = DPO_INVALID;
const dpo_id_t *dpo1;
fib_node_index_t lfe;
lookup_dpo_t *lkd;
test_main_t *tm;
int lb_count;
adj_index_t ai_mpls_10_10_10_1;
tm = &test_main;
lb_count = pool_elts(load_balance_pool);
FIB_TEST((0 == adj_nbr_db_size()), "ADJ DB size is %d",
adj_nbr_db_size());
/*
* MPLS enable an interface so we get the MPLS table created
*/
mpls_table_create(MPLS_FIB_DEFAULT_TABLE_ID, FIB_SOURCE_API);
mpls_sw_interface_enable_disable(&mpls_main,
tm->hw[0]->sw_if_index,
1, 1);
ip46_address_t nh_10_10_10_1 = {
.ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a01),
};
ai_mpls_10_10_10_1 = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
VNET_LINK_MPLS,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index);
/*
* Test the specials stack properly.
*/
fib_prefix_t exp_null_v6_pfx = {
.fp_proto = FIB_PROTOCOL_MPLS,
.fp_eos = MPLS_EOS,
.fp_label = MPLS_IETF_IPV6_EXPLICIT_NULL_LABEL,
.fp_payload_proto = DPO_PROTO_IP6,
};
lfe = fib_table_lookup(lfib_index, &exp_null_v6_pfx);
FIB_TEST((FIB_NODE_INDEX_INVALID != lfe),
"%U/%U present",
format_mpls_unicast_label, MPLS_IETF_IPV6_EXPLICIT_NULL_LABEL,
format_mpls_eos_bit, MPLS_EOS);
fib_entry_contribute_forwarding(lfe,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
&dpo);
dpo1 = load_balance_get_bucket(dpo.dpoi_index, 0);
lkd = lookup_dpo_get(dpo1->dpoi_index);
FIB_TEST((fib_index == lkd->lkd_fib_index),
"%U/%U is deag in %d %U",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_EOS,
lkd->lkd_fib_index,
format_dpo_id, &dpo, 0);
FIB_TEST((LOOKUP_INPUT_DST_ADDR == lkd->lkd_input),
"%U/%U is dst deag",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_EOS);
FIB_TEST((LOOKUP_TABLE_FROM_INPUT_INTERFACE == lkd->lkd_table),
"%U/%U is lookup in interface's table",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_EOS);
FIB_TEST((DPO_PROTO_IP6 == lkd->lkd_proto),
"%U/%U is %U dst deag",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_EOS,
format_dpo_proto, lkd->lkd_proto);
/*
* A route deag route for EOS
*/
fib_prefix_t pfx = {
.fp_proto = FIB_PROTOCOL_MPLS,
.fp_eos = MPLS_EOS,
.fp_label = deag_label,
.fp_payload_proto = DPO_PROTO_IP4,
};
lfe = fib_table_entry_path_add(lfib_index,
&pfx,
FIB_SOURCE_CLI,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&zero_addr,
~0,
fib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST((lfe == fib_table_lookup(lfib_index, &pfx)),
"%U/%U present",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_EOS);
fib_entry_contribute_forwarding(lfe,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
&dpo);
dpo1 = load_balance_get_bucket(dpo.dpoi_index, 0);
lkd = lookup_dpo_get(dpo1->dpoi_index);
FIB_TEST((fib_index == lkd->lkd_fib_index),
"%U/%U is deag in %d %U",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_EOS,
lkd->lkd_fib_index,
format_dpo_id, &dpo, 0);
FIB_TEST((LOOKUP_INPUT_DST_ADDR == lkd->lkd_input),
"%U/%U is dst deag",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_EOS);
FIB_TEST((DPO_PROTO_IP4 == lkd->lkd_proto),
"%U/%U is %U dst deag",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_EOS,
format_dpo_proto, lkd->lkd_proto);
fib_table_entry_delete_index(lfe, FIB_SOURCE_CLI);
FIB_TEST((FIB_NODE_INDEX_INVALID == fib_table_lookup(lfib_index,
&pfx)),
"%U/%U not present",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_EOS);
/*
* A route deag route for non-EOS
*/
pfx.fp_eos = MPLS_NON_EOS;
lfe = fib_table_entry_path_add(lfib_index,
&pfx,
FIB_SOURCE_CLI,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&zero_addr,
~0,
lfib_index,
1,
NULL,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST((lfe == fib_table_lookup(lfib_index, &pfx)),
"%U/%U present",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_NON_EOS);
fib_entry_contribute_forwarding(lfe,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
&dpo);
dpo1 = load_balance_get_bucket(dpo.dpoi_index, 0);
lkd = lookup_dpo_get(dpo1->dpoi_index);
FIB_TEST((fib_index == lkd->lkd_fib_index),
"%U/%U is deag in %d %U",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_NON_EOS,
lkd->lkd_fib_index,
format_dpo_id, &dpo, 0);
FIB_TEST((LOOKUP_INPUT_DST_ADDR == lkd->lkd_input),
"%U/%U is dst deag",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_NON_EOS);
FIB_TEST((DPO_PROTO_MPLS == lkd->lkd_proto),
"%U/%U is %U dst deag",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_NON_EOS,
format_dpo_proto, lkd->lkd_proto);
fib_table_entry_delete_index(lfe, FIB_SOURCE_CLI);
FIB_TEST((FIB_NODE_INDEX_INVALID == fib_table_lookup(lfib_index,
&pfx)),
"%U/%U not present",
format_mpls_unicast_label, deag_label,
format_mpls_eos_bit, MPLS_EOS);
dpo_reset(&dpo);
/*
* An MPLS x-connect
*/
fib_prefix_t pfx_1200 = {
.fp_len = 21,
.fp_proto = FIB_PROTOCOL_MPLS,
.fp_label = 1200,
.fp_eos = MPLS_NON_EOS,
};
fib_test_lb_bucket_t neos_o_10_10_10_1 = {
.type = FT_LB_LABEL_STACK_O_ADJ,
.label_stack_o_adj = {
.adj = ai_mpls_10_10_10_1,
.label_stack_size = 4,
.label_stack = {
200, 300, 400, 500,
},
.eos = MPLS_NON_EOS,
},
};
dpo_id_t neos_1200 = DPO_INVALID;
dpo_id_t ip_1200 = DPO_INVALID;
mpls_label_t *l200 = NULL;
vec_add1(l200, 200);
vec_add1(l200, 300);
vec_add1(l200, 400);
vec_add1(l200, 500);
lfe = fib_table_entry_update_one_path(fib_index,
&pfx_1200,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
l200,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(lfe,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
1,
&neos_o_10_10_10_1),
"1200/0 LB 1 buckets via: "
"adj 10.10.11.1");
/*
* A recursive route via the MPLS x-connect
*/
fib_prefix_t pfx_2_2_2_3_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x02020203),
},
};
fib_route_path_t *rpaths = NULL, rpath = {
.frp_proto = DPO_PROTO_MPLS,
.frp_local_label = 1200,
.frp_eos = MPLS_NON_EOS,
.frp_sw_if_index = ~0, // recurive
.frp_fib_index = 0, // Default MPLS fib
.frp_weight = 1,
.frp_flags = FIB_ROUTE_PATH_FLAG_NONE,
.frp_label_stack = NULL,
};
vec_add1(rpaths, rpath);
fib_table_entry_path_add2(fib_index,
&pfx_2_2_2_3_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
rpaths);
/*
* A labelled recursive route via the MPLS x-connect
*/
fib_prefix_t pfx_2_2_2_4_s_32 = {
.fp_len = 32,
.fp_proto = FIB_PROTOCOL_IP4,
.fp_addr = {
.ip4.as_u32 = clib_host_to_net_u32(0x02020204),
},
};
mpls_label_t *l999 = NULL;
vec_add1(l999, 999);
rpaths[0].frp_label_stack = l999,
fib_table_entry_path_add2(fib_index,
&pfx_2_2_2_4_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
rpaths);
fib_entry_contribute_forwarding(fib_table_lookup(fib_index, &pfx_1200),
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
&ip_1200);
fib_entry_contribute_forwarding(fib_table_lookup(fib_index, &pfx_1200),
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
&neos_1200);
fib_test_lb_bucket_t ip_o_1200 = {
.type = FT_LB_O_LB,
.lb = {
.lb = ip_1200.dpoi_index,
},
};
fib_test_lb_bucket_t mpls_o_1200 = {
.type = FT_LB_LABEL_O_LB,
.label_o_lb = {
.lb = neos_1200.dpoi_index,
.label = 999,
.eos = MPLS_EOS,
},
};
lfe = fib_table_lookup(fib_index, &pfx_2_2_2_3_s_32);
FIB_TEST(fib_test_validate_entry(lfe,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&ip_o_1200),
"2.2.2.2.3/32 LB 1 buckets via: label 1200 EOS");
lfe = fib_table_lookup(fib_index, &pfx_2_2_2_4_s_32);
FIB_TEST(fib_test_validate_entry(lfe,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&mpls_o_1200),
"2.2.2.2.4/32 LB 1 buckets via: label 1200 non-EOS");
fib_table_entry_delete(fib_index, &pfx_1200, FIB_SOURCE_API);
fib_table_entry_delete(fib_index, &pfx_2_2_2_3_s_32, FIB_SOURCE_API);
fib_table_entry_delete(fib_index, &pfx_2_2_2_4_s_32, FIB_SOURCE_API);
dpo_reset(&neos_1200);
dpo_reset(&ip_1200);
/*
* A recursive via a label that does not exist
*/
fib_test_lb_bucket_t bucket_drop = {
.type = FT_LB_SPECIAL,
.special = {
.adj = DPO_PROTO_IP4,
},
};
fib_test_lb_bucket_t mpls_bucket_drop = {
.type = FT_LB_SPECIAL,
.special = {
.adj = DPO_PROTO_MPLS,
},
};
rpaths[0].frp_label_stack = NULL;
lfe = fib_table_entry_path_add2(fib_index,
&pfx_2_2_2_4_s_32,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
rpaths);
fib_entry_contribute_forwarding(fib_table_lookup(fib_index, &pfx_1200),
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
&ip_1200);
ip_o_1200.lb.lb = ip_1200.dpoi_index;
FIB_TEST(fib_test_validate_entry(lfe,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&bucket_drop),
"2.2.2.2.4/32 LB 1 buckets via: drop");
lfe = fib_table_lookup(fib_index, &pfx_1200);
FIB_TEST(fib_test_validate_entry(lfe,
FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
1,
&bucket_drop),
"1200/neos LB 1 buckets via: ip4-DROP");
FIB_TEST(fib_test_validate_entry(lfe,
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
1,
&mpls_bucket_drop),
"1200/neos LB 1 buckets via: mpls-DROP");
fib_table_entry_delete(fib_index, &pfx_2_2_2_4_s_32, FIB_SOURCE_API);
dpo_reset(&ip_1200);
/*
* An rx-interface route.
* like the tail of an mcast LSP
*/
dpo_id_t idpo = DPO_INVALID;
interface_rx_dpo_add_or_lock(DPO_PROTO_IP4,
tm->hw[0]->sw_if_index,
&idpo);
fib_prefix_t pfx_2500 = {
.fp_len = 21,
.fp_proto = FIB_PROTOCOL_MPLS,
.fp_label = 2500,
.fp_eos = MPLS_EOS,
.fp_payload_proto = DPO_PROTO_IP4,
};
fib_test_lb_bucket_t rx_intf_0 = {
.type = FT_LB_INTF,
.adj = {
.adj = idpo.dpoi_index,
},
};
lfe = fib_table_entry_update_one_path(fib_index,
&pfx_2500,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_NONE,
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
0,
NULL,
FIB_ROUTE_PATH_INTF_RX);
FIB_TEST(fib_test_validate_entry(lfe,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
1,
&rx_intf_0),
"2500 rx-interface 0");
fib_table_entry_delete(fib_index, &pfx_2500, FIB_SOURCE_API);
/*
* An MPLS mulicast entry
*/
fib_prefix_t pfx_3500 = {
.fp_len = 21,
.fp_proto = FIB_PROTOCOL_MPLS,
.fp_label = 3500,
.fp_eos = MPLS_EOS,
.fp_payload_proto = DPO_PROTO_IP4,
};
fib_test_rep_bucket_t mc_0 = {
.type = FT_REP_LABEL_O_ADJ,
.label_o_adj = {
.adj = ai_mpls_10_10_10_1,
.label = 3300,
.eos = MPLS_EOS,
},
};
fib_test_rep_bucket_t mc_intf_0 = {
.type = FT_REP_INTF,
.adj = {
.adj = idpo.dpoi_index,
},
};
mpls_label_t *l3300 = NULL;
vec_add1(l3300, 3300);
lfe = fib_table_entry_update_one_path(lfib_index,
&pfx_3500,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_MULTICAST,
DPO_PROTO_IP4,
&nh_10_10_10_1,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
1,
l3300,
FIB_ROUTE_PATH_FLAG_NONE);
FIB_TEST(fib_test_validate_entry(lfe,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
1,
&mc_0),
"3500 via replicate over 10.10.10.1");
/*
* MPLS Bud-node. Add a replication via an interface-receieve path
*/
lfe = fib_table_entry_path_add(lfib_index,
&pfx_3500,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_MULTICAST,
DPO_PROTO_IP4,
NULL,
tm->hw[0]->sw_if_index,
~0, // invalid fib index
0,
NULL,
FIB_ROUTE_PATH_INTF_RX);
FIB_TEST(fib_test_validate_entry(lfe,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
2,
&mc_0,
&mc_intf_0),
"3500 via replicate over 10.10.10.1 and interface-rx");
/*
* Add a replication via an interface-free for-us path
*/
fib_test_rep_bucket_t mc_disp = {
.type = FT_REP_DISP_MFIB_LOOKUP,
.adj = {
.adj = idpo.dpoi_index,
},
};
lfe = fib_table_entry_path_add(lfib_index,
&pfx_3500,
FIB_SOURCE_API,
FIB_ENTRY_FLAG_MULTICAST,
DPO_PROTO_IP4,
NULL,
5, // rpf-id
0, // default table
0,
NULL,
FIB_ROUTE_PATH_RPF_ID);
FIB_TEST(fib_test_validate_entry(lfe,
FIB_FORW_CHAIN_TYPE_MPLS_EOS,
3,
&mc_0,
&mc_disp,
&mc_intf_0),
"3500 via replicate over 10.10.10.1 and interface-rx");
fib_table_entry_delete(fib_index, &pfx_3500, FIB_SOURCE_API);
dpo_reset(&idpo);
/*
* cleanup
*/
mpls_sw_interface_enable_disable(&mpls_main,
tm->hw[0]->sw_if_index,
0, 1);
mpls_table_delete(MPLS_FIB_DEFAULT_TABLE_ID, FIB_SOURCE_API);
FIB_TEST(lb_count == pool_elts(load_balance_pool),
"Load-balance resources freed %d of %d",
lb_count, pool_elts(load_balance_pool));
FIB_TEST(0 == pool_elts(interface_rx_dpo_pool),
"interface_rx_dpo resources freed %d of %d",
0, pool_elts(interface_rx_dpo_pool));
return (0);
}
static clib_error_t *
fib_test (vlib_main_t * vm,
unformat_input_t * input,
vlib_cli_command_t * cmd_arg)
{
int res;
res = 0;
fib_test_mk_intf(4);
if (unformat (input, "debug"))
{
fib_test_do_debug = 1;
}
if (unformat (input, "ip"))
{
res += fib_test_v4();
res += fib_test_v6();
}
else if (unformat (input, "label"))
{
res += fib_test_label();
}
else if (unformat (input, "ae"))
{
res += fib_test_ae();
}
else if (unformat (input, "pref"))
{
res += fib_test_pref();
}
else if (unformat (input, "lfib"))
{
res += lfib_test();
}
else if (unformat (input, "walk"))
{
res += fib_test_walk();
}
else if (unformat (input, "bfd"))
{
res += fib_test_bfd();
}
else
{
res += fib_test_v4();
res += fib_test_v6();
res += fib_test_ae();
res += fib_test_bfd();
res += fib_test_pref();
res += fib_test_label();
res += lfib_test();
/*
* fib-walk process must be disabled in order for the walk tests to work
*/
fib_walk_process_disable();
res += fib_test_walk();
fib_walk_process_enable();
}
if (res)
{
return clib_error_return(0, "FIB Unit Test Failed");
}
else
{
return (NULL);
}
}
VLIB_CLI_COMMAND (test_fib_command, static) = {
.path = "test fib",
.short_help = "fib unit tests - DO NOT RUN ON A LIVE SYSTEM",
.function = fib_test,
};
clib_error_t *
fib_test_init (vlib_main_t *vm)
{
return 0;
}
VLIB_INIT_FUNCTION (fib_test_init);