| /* |
| *------------------------------------------------------------------ |
| * mpls_api.c - mpls api |
| * |
| * Copyright (c) 2016 Cisco and/or its affiliates. |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at: |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| *------------------------------------------------------------------ |
| */ |
| |
| #include <vnet/vnet.h> |
| #include <vlibmemory/api.h> |
| |
| #include <vnet/interface.h> |
| #include <vnet/api_errno.h> |
| #include <vnet/mpls/mpls.h> |
| #include <vnet/mpls/mpls_tunnel.h> |
| #include <vnet/fib/fib_table.h> |
| #include <vnet/fib/fib_api.h> |
| #include <vnet/fib/mpls_fib.h> |
| #include <vnet/fib/fib_path_list.h> |
| |
| #include <vnet/vnet_msg_enum.h> |
| |
| #define vl_typedefs /* define message structures */ |
| #include <vnet/vnet_all_api_h.h> |
| #undef vl_typedefs |
| |
| #define vl_endianfun /* define message structures */ |
| #include <vnet/vnet_all_api_h.h> |
| #undef vl_endianfun |
| |
| /* instantiate all the print functions we know about */ |
| #define vl_print(handle, ...) vlib_cli_output (handle, __VA_ARGS__) |
| #define vl_printfun |
| #include <vnet/vnet_all_api_h.h> |
| #undef vl_printfun |
| |
| #include <vlibapi/api_helper_macros.h> |
| |
| #define foreach_vpe_api_msg \ |
| _(MPLS_IP_BIND_UNBIND, mpls_ip_bind_unbind) \ |
| _(MPLS_ROUTE_ADD_DEL, mpls_route_add_del) \ |
| _(MPLS_TABLE_ADD_DEL, mpls_table_add_del) \ |
| _(MPLS_TUNNEL_ADD_DEL, mpls_tunnel_add_del) \ |
| _(MPLS_TUNNEL_DUMP, mpls_tunnel_dump) \ |
| _(MPLS_FIB_DUMP, mpls_fib_dump) |
| |
| extern void stats_dslock_with_hint (int hint, int tag); |
| extern void stats_dsunlock (void); |
| |
| void |
| mpls_table_delete (u32 table_id, u8 is_api) |
| { |
| u32 fib_index; |
| |
| /* |
| * The MPLS defult table must also be explicitly created via the API. |
| * So in contrast to IP, it gets no special treatment here. |
| * |
| * The API holds only one lock on the table. |
| * i.e. it can be added many times via the API but needs to be |
| * deleted only once. |
| */ |
| fib_index = fib_table_find (FIB_PROTOCOL_MPLS, table_id); |
| |
| if (~0 != fib_index) |
| { |
| fib_table_unlock (fib_index, |
| FIB_PROTOCOL_MPLS, |
| (is_api ? FIB_SOURCE_API : FIB_SOURCE_CLI)); |
| } |
| } |
| |
| void |
| vl_api_mpls_table_add_del_t_handler (vl_api_mpls_table_add_del_t * mp) |
| { |
| vl_api_mpls_table_add_del_reply_t *rmp; |
| vnet_main_t *vnm; |
| int rv = 0; |
| |
| vnm = vnet_get_main (); |
| vnm->api_errno = 0; |
| |
| if (mp->mt_is_add) |
| mpls_table_create (ntohl (mp->mt_table_id), 1, mp->mt_name); |
| else |
| mpls_table_delete (ntohl (mp->mt_table_id), 1); |
| |
| // NB: Nothing sets rv; none of the above returns an error |
| |
| REPLY_MACRO (VL_API_MPLS_TABLE_ADD_DEL_REPLY); |
| } |
| |
| static int |
| mpls_ip_bind_unbind_handler (vnet_main_t * vnm, |
| vl_api_mpls_ip_bind_unbind_t * mp) |
| { |
| u32 mpls_fib_index, ip_fib_index; |
| |
| mpls_fib_index = |
| fib_table_find (FIB_PROTOCOL_MPLS, ntohl (mp->mb_mpls_table_id)); |
| |
| if (~0 == mpls_fib_index) |
| { |
| return VNET_API_ERROR_NO_SUCH_FIB; |
| } |
| |
| ip_fib_index = fib_table_find ((mp->mb_is_ip4 ? |
| FIB_PROTOCOL_IP4 : |
| FIB_PROTOCOL_IP6), |
| ntohl (mp->mb_ip_table_id)); |
| if (~0 == ip_fib_index) |
| return VNET_API_ERROR_NO_SUCH_FIB; |
| |
| fib_prefix_t pfx = { |
| .fp_len = mp->mb_address_length, |
| }; |
| |
| if (mp->mb_is_ip4) |
| { |
| pfx.fp_proto = FIB_PROTOCOL_IP4; |
| clib_memcpy (&pfx.fp_addr.ip4, mp->mb_address, |
| sizeof (pfx.fp_addr.ip4)); |
| } |
| else |
| { |
| pfx.fp_proto = FIB_PROTOCOL_IP6; |
| clib_memcpy (&pfx.fp_addr.ip6, mp->mb_address, |
| sizeof (pfx.fp_addr.ip6)); |
| } |
| |
| if (mp->mb_is_bind) |
| fib_table_entry_local_label_add (ip_fib_index, &pfx, |
| ntohl (mp->mb_label)); |
| else |
| fib_table_entry_local_label_remove (ip_fib_index, &pfx, |
| ntohl (mp->mb_label)); |
| |
| return (0); |
| } |
| |
| void |
| vl_api_mpls_ip_bind_unbind_t_handler (vl_api_mpls_ip_bind_unbind_t * mp) |
| { |
| vl_api_mpls_ip_bind_unbind_reply_t *rmp; |
| vnet_main_t *vnm; |
| int rv; |
| |
| vnm = vnet_get_main (); |
| vnm->api_errno = 0; |
| |
| rv = mpls_ip_bind_unbind_handler (vnm, mp); |
| rv = (rv == 0) ? vnm->api_errno : rv; |
| |
| REPLY_MACRO (VL_API_MPLS_IP_BIND_UNBIND_REPLY); |
| } |
| |
| static int |
| mpls_route_add_del_t_handler (vnet_main_t * vnm, |
| vl_api_mpls_route_add_del_t * mp) |
| { |
| u32 fib_index, next_hop_fib_index; |
| mpls_label_t *label_stack = NULL; |
| int rv, ii, n_labels;; |
| |
| fib_prefix_t pfx = { |
| .fp_len = 21, |
| .fp_proto = FIB_PROTOCOL_MPLS, |
| .fp_eos = mp->mr_eos, |
| .fp_label = ntohl (mp->mr_label), |
| }; |
| if (pfx.fp_eos) |
| { |
| pfx.fp_payload_proto = mp->mr_next_hop_proto; |
| } |
| else |
| { |
| pfx.fp_payload_proto = DPO_PROTO_MPLS; |
| } |
| |
| rv = add_del_route_check (FIB_PROTOCOL_MPLS, |
| mp->mr_table_id, |
| mp->mr_next_hop_sw_if_index, |
| pfx.fp_payload_proto, |
| mp->mr_next_hop_table_id, |
| mp->mr_is_rpf_id, |
| &fib_index, &next_hop_fib_index); |
| |
| if (0 != rv) |
| return (rv); |
| |
| ip46_address_t nh; |
| memset (&nh, 0, sizeof (nh)); |
| |
| if (DPO_PROTO_IP4 == mp->mr_next_hop_proto) |
| memcpy (&nh.ip4, mp->mr_next_hop, sizeof (nh.ip4)); |
| else if (DPO_PROTO_IP6 == mp->mr_next_hop_proto) |
| memcpy (&nh.ip6, mp->mr_next_hop, sizeof (nh.ip6)); |
| |
| n_labels = mp->mr_next_hop_n_out_labels; |
| if (n_labels == 0) |
| ; |
| else if (1 == n_labels) |
| vec_add1 (label_stack, ntohl (mp->mr_next_hop_out_label_stack[0])); |
| else |
| { |
| vec_validate (label_stack, n_labels - 1); |
| for (ii = 0; ii < n_labels; ii++) |
| label_stack[ii] = ntohl (mp->mr_next_hop_out_label_stack[ii]); |
| } |
| |
| return (add_del_route_t_handler (mp->mr_is_multipath, mp->mr_is_add, 0, // mp->is_drop, |
| 0, // mp->is_unreach, |
| 0, // mp->is_prohibit, |
| 0, // mp->is_local, |
| mp->mr_is_multicast, |
| mp->mr_is_classify, |
| mp->mr_classify_table_index, |
| mp->mr_is_resolve_host, |
| mp->mr_is_resolve_attached, |
| mp->mr_is_interface_rx, |
| mp->mr_is_rpf_id, |
| fib_index, &pfx, |
| mp->mr_next_hop_proto, |
| &nh, ntohl (mp->mr_next_hop_sw_if_index), |
| next_hop_fib_index, |
| mp->mr_next_hop_weight, |
| mp->mr_next_hop_preference, |
| ntohl (mp->mr_next_hop_via_label), |
| label_stack)); |
| } |
| |
| void |
| vl_api_mpls_route_add_del_t_handler (vl_api_mpls_route_add_del_t * mp) |
| { |
| vl_api_mpls_route_add_del_reply_t *rmp; |
| vnet_main_t *vnm; |
| int rv; |
| |
| vnm = vnet_get_main (); |
| vnm->api_errno = 0; |
| |
| rv = mpls_route_add_del_t_handler (vnm, mp); |
| |
| rv = (rv == 0) ? vnm->api_errno : rv; |
| |
| REPLY_MACRO (VL_API_MPLS_ROUTE_ADD_DEL_REPLY); |
| } |
| |
| void |
| mpls_table_create (u32 table_id, u8 is_api, const u8 * name) |
| { |
| u32 fib_index; |
| |
| /* |
| * The MPLS defult table must also be explicitly created via the API. |
| * So in contrast to IP, it gets no special treatment here. |
| */ |
| |
| /* |
| * The API holds only one lock on the table. |
| * i.e. it can be added many times via the API but needs to be |
| * deleted only once. |
| */ |
| fib_index = fib_table_find (FIB_PROTOCOL_MPLS, table_id); |
| |
| if (~0 == fib_index) |
| { |
| fib_table_find_or_create_and_lock_w_name (FIB_PROTOCOL_MPLS, |
| table_id, |
| (is_api ? |
| FIB_SOURCE_API : |
| FIB_SOURCE_CLI), name); |
| } |
| } |
| |
| static void |
| vl_api_mpls_tunnel_add_del_t_handler (vl_api_mpls_tunnel_add_del_t * mp) |
| { |
| vl_api_mpls_tunnel_add_del_reply_t *rmp; |
| int rv = 0; |
| u32 tunnel_sw_if_index; |
| int ii; |
| fib_route_path_t rpath, *rpaths = NULL; |
| |
| memset (&rpath, 0, sizeof (rpath)); |
| |
| stats_dslock_with_hint (1 /* release hint */ , 5 /* tag */ ); |
| |
| if (mp->mt_next_hop_proto_is_ip4) |
| { |
| rpath.frp_proto = DPO_PROTO_IP4; |
| clib_memcpy (&rpath.frp_addr.ip4, |
| mp->mt_next_hop, sizeof (rpath.frp_addr.ip4)); |
| } |
| else |
| { |
| rpath.frp_proto = DPO_PROTO_IP6; |
| clib_memcpy (&rpath.frp_addr.ip6, |
| mp->mt_next_hop, sizeof (rpath.frp_addr.ip6)); |
| } |
| rpath.frp_sw_if_index = ntohl (mp->mt_next_hop_sw_if_index); |
| rpath.frp_weight = 1; |
| |
| if (mp->mt_is_add) |
| { |
| for (ii = 0; ii < mp->mt_next_hop_n_out_labels; ii++) |
| vec_add1 (rpath.frp_label_stack, |
| ntohl (mp->mt_next_hop_out_label_stack[ii])); |
| } |
| |
| vec_add1 (rpaths, rpath); |
| |
| tunnel_sw_if_index = ntohl (mp->mt_sw_if_index); |
| |
| if (mp->mt_is_add) |
| { |
| if (~0 == tunnel_sw_if_index) |
| tunnel_sw_if_index = vnet_mpls_tunnel_create (mp->mt_l2_only, |
| mp->mt_is_multicast); |
| vnet_mpls_tunnel_path_add (tunnel_sw_if_index, rpaths); |
| } |
| else |
| { |
| tunnel_sw_if_index = ntohl (mp->mt_sw_if_index); |
| if (!vnet_mpls_tunnel_path_remove (tunnel_sw_if_index, rpaths)) |
| vnet_mpls_tunnel_del (tunnel_sw_if_index); |
| } |
| |
| vec_free (rpaths); |
| |
| stats_dsunlock (); |
| |
| /* *INDENT-OFF* */ |
| REPLY_MACRO2(VL_API_MPLS_TUNNEL_ADD_DEL_REPLY, |
| ({ |
| rmp->sw_if_index = ntohl(tunnel_sw_if_index); |
| })); |
| /* *INDENT-ON* */ |
| } |
| |
| typedef struct mpls_tunnel_send_walk_ctx_t_ |
| { |
| unix_shared_memory_queue_t *q; |
| u32 index; |
| u32 context; |
| } mpls_tunnel_send_walk_ctx_t; |
| |
| static void |
| send_mpls_tunnel_entry (u32 mti, void *arg) |
| { |
| fib_route_path_encode_t *api_rpaths, *api_rpath; |
| mpls_tunnel_send_walk_ctx_t *ctx; |
| vl_api_mpls_tunnel_details_t *mp; |
| const mpls_tunnel_t *mt; |
| vl_api_fib_path2_t *fp; |
| u32 n; |
| |
| ctx = arg; |
| |
| if (~0 != ctx->index && mti != ctx->index) |
| return; |
| |
| mt = mpls_tunnel_get (mti); |
| n = fib_path_list_get_n_paths (mt->mt_path_list); |
| |
| mp = vl_msg_api_alloc (sizeof (*mp) + n * sizeof (vl_api_fib_path2_t)); |
| memset (mp, 0, sizeof (*mp) + n * sizeof (vl_api_fib_path2_t)); |
| |
| mp->_vl_msg_id = ntohs (VL_API_MPLS_TUNNEL_DETAILS); |
| mp->context = ctx->context; |
| |
| mp->mt_tunnel_index = ntohl (mti); |
| mp->mt_count = ntohl (n); |
| |
| fib_path_list_walk (mt->mt_path_list, fib_path_encode, &api_rpaths); |
| |
| fp = mp->mt_paths; |
| vec_foreach (api_rpath, api_rpaths) |
| { |
| memset (fp, 0, sizeof (*fp)); |
| |
| fp->weight = api_rpath->rpath.frp_weight; |
| fp->preference = api_rpath->rpath.frp_preference; |
| fp->sw_if_index = htonl (api_rpath->rpath.frp_sw_if_index); |
| copy_fib_next_hop (api_rpath, fp); |
| fp++; |
| } |
| |
| // FIXME |
| // memcpy (mp->mt_next_hop_out_labels, |
| // mt->mt_label_stack, nlabels * sizeof (u32)); |
| |
| |
| vl_msg_api_send_shmem (ctx->q, (u8 *) & mp); |
| } |
| |
| static void |
| vl_api_mpls_tunnel_dump_t_handler (vl_api_mpls_tunnel_dump_t * mp) |
| { |
| unix_shared_memory_queue_t *q; |
| |
| q = vl_api_client_index_to_input_queue (mp->client_index); |
| if (q == 0) |
| return; |
| |
| mpls_tunnel_send_walk_ctx_t ctx = { |
| .q = q, |
| .index = ntohl (mp->tunnel_index), |
| .context = mp->context, |
| }; |
| mpls_tunnel_walk (send_mpls_tunnel_entry, &ctx); |
| } |
| |
| static void |
| send_mpls_fib_details (vpe_api_main_t * am, |
| unix_shared_memory_queue_t * q, |
| const fib_table_t * table, |
| u32 label, u32 eos, |
| fib_route_path_encode_t * api_rpaths, u32 context) |
| { |
| vl_api_mpls_fib_details_t *mp; |
| fib_route_path_encode_t *api_rpath; |
| vl_api_fib_path2_t *fp; |
| int path_count; |
| |
| path_count = vec_len (api_rpaths); |
| mp = vl_msg_api_alloc (sizeof (*mp) + path_count * sizeof (*fp)); |
| if (!mp) |
| return; |
| memset (mp, 0, sizeof (*mp)); |
| mp->_vl_msg_id = ntohs (VL_API_MPLS_FIB_DETAILS); |
| mp->context = context; |
| |
| mp->table_id = htonl (table->ft_table_id); |
| memcpy (mp->table_name, table->ft_desc, |
| clib_min (vec_len (table->ft_desc), sizeof (mp->table_name))); |
| mp->eos_bit = eos; |
| mp->label = htonl (label); |
| |
| mp->count = htonl (path_count); |
| fp = mp->path; |
| vec_foreach (api_rpath, api_rpaths) |
| { |
| memset (fp, 0, sizeof (*fp)); |
| fp->weight = api_rpath->rpath.frp_weight; |
| fp->preference = api_rpath->rpath.frp_preference; |
| fp->sw_if_index = htonl (api_rpath->rpath.frp_sw_if_index); |
| copy_fib_next_hop (api_rpath, fp); |
| fp++; |
| } |
| |
| vl_msg_api_send_shmem (q, (u8 *) & mp); |
| } |
| |
| typedef struct vl_api_mpls_fib_dump_table_walk_ctx_t_ |
| { |
| fib_node_index_t *lfeis; |
| } vl_api_mpls_fib_dump_table_walk_ctx_t; |
| |
| static int |
| vl_api_mpls_fib_dump_table_walk (fib_node_index_t fei, void *arg) |
| { |
| vl_api_mpls_fib_dump_table_walk_ctx_t *ctx = arg; |
| |
| vec_add1 (ctx->lfeis, fei); |
| |
| return (1); |
| } |
| |
| static void |
| vl_api_mpls_fib_dump_t_handler (vl_api_mpls_fib_dump_t * mp) |
| { |
| vpe_api_main_t *am = &vpe_api_main; |
| unix_shared_memory_queue_t *q; |
| mpls_main_t *mm = &mpls_main; |
| fib_table_t *fib_table; |
| mpls_fib_t *mpls_fib; |
| fib_node_index_t *lfeip = NULL; |
| fib_prefix_t pfx; |
| u32 fib_index; |
| fib_route_path_encode_t *api_rpaths; |
| vl_api_mpls_fib_dump_table_walk_ctx_t ctx = { |
| .lfeis = NULL, |
| }; |
| |
| q = vl_api_client_index_to_input_queue (mp->client_index); |
| if (q == 0) |
| return; |
| |
| /* *INDENT-OFF* */ |
| pool_foreach (mpls_fib, mm->mpls_fibs, |
| ({ |
| mpls_fib_table_walk (mpls_fib, |
| vl_api_mpls_fib_dump_table_walk, |
| &ctx); |
| })); |
| /* *INDENT-ON* */ |
| vec_sort_with_function (ctx.lfeis, fib_entry_cmp_for_sort); |
| |
| vec_foreach (lfeip, ctx.lfeis) |
| { |
| fib_entry_get_prefix (*lfeip, &pfx); |
| fib_index = fib_entry_get_fib_index (*lfeip); |
| fib_table = fib_table_get (fib_index, pfx.fp_proto); |
| api_rpaths = NULL; |
| fib_entry_encode (*lfeip, &api_rpaths); |
| send_mpls_fib_details (am, q, |
| fib_table, pfx.fp_label, |
| pfx.fp_eos, api_rpaths, mp->context); |
| vec_free (api_rpaths); |
| } |
| |
| vec_free (ctx.lfeis); |
| } |
| |
| /* |
| * mpls_api_hookup |
| * Add vpe's API message handlers to the table. |
| * vlib has alread mapped shared memory and |
| * added the client registration handlers. |
| * See .../vlib-api/vlibmemory/memclnt_vlib.c:memclnt_process() |
| */ |
| #define vl_msg_name_crc_list |
| #include <vnet/vnet_all_api_h.h> |
| #undef vl_msg_name_crc_list |
| |
| static void |
| setup_message_id_table (api_main_t * am) |
| { |
| #define _(id,n,crc) vl_msg_api_add_msg_name_crc (am, #n "_" #crc, id); |
| foreach_vl_msg_name_crc_mpls; |
| #undef _ |
| } |
| |
| static clib_error_t * |
| mpls_api_hookup (vlib_main_t * vm) |
| { |
| api_main_t *am = &api_main; |
| |
| #define _(N,n) \ |
| vl_msg_api_set_handlers(VL_API_##N, #n, \ |
| vl_api_##n##_t_handler, \ |
| vl_noop_handler, \ |
| vl_api_##n##_t_endian, \ |
| vl_api_##n##_t_print, \ |
| sizeof(vl_api_##n##_t), 1); |
| foreach_vpe_api_msg; |
| #undef _ |
| |
| /* |
| * Trace space for 8 MPLS encap labels |
| */ |
| am->api_trace_cfg[VL_API_MPLS_TUNNEL_ADD_DEL].size += 8 * sizeof (u32); |
| |
| /* |
| * Set up the (msg_name, crc, message-id) table |
| */ |
| setup_message_id_table (am); |
| |
| return 0; |
| } |
| |
| VLIB_API_INIT_FUNCTION (mpls_api_hookup); |
| |
| /* |
| * fd.io coding-style-patch-verification: ON |
| * |
| * Local Variables: |
| * eval: (c-set-style "gnu") |
| * End: |
| */ |