blob: e1186f49c0ad4006172a7da0e09e3dcd4f921159 [file] [log] [blame]
/*
* Copyright (c) 2020 Cisco and/or its affiliates.
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at:
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <stdbool.h>
#include <vlib/vlib.h>
#include <vnet/crypto/crypto.h>
typedef enum
{
#define _(sym,str) VNET_CRYPTO_ASYNC_ERROR_##sym,
foreach_crypto_op_status
#undef _
VNET_CRYPTO_ASYNC_N_ERROR,
} vnet_crypto_async_error_t;
static char *vnet_crypto_async_error_strings[] = {
#define _(sym,string) string,
foreach_crypto_op_status
#undef _
};
#define foreach_crypto_dispatch_next \
_(ERR_DROP, "error-drop")
typedef enum
{
#define _(n, s) CRYPTO_DISPATCH_NEXT_##n,
foreach_crypto_dispatch_next
#undef _
CRYPTO_DISPATCH_N_NEXT,
} crypto_dispatch_next_t;
typedef struct
{
vnet_crypto_op_status_t op_status;
vnet_crypto_async_op_id_t op;
} crypto_dispatch_trace_t;
static u8 *
format_crypto_dispatch_trace (u8 * s, va_list * args)
{
CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
crypto_dispatch_trace_t *t = va_arg (*args, crypto_dispatch_trace_t *);
s = format (s, "%U: %U", format_vnet_crypto_async_op, t->op,
format_vnet_crypto_op_status, t->op_status);
return s;
}
static void
vnet_crypto_async_add_trace (vlib_main_t * vm, vlib_node_runtime_t * node,
vlib_buffer_t * b,
vnet_crypto_async_op_id_t op_id,
vnet_crypto_op_status_t status)
{
crypto_dispatch_trace_t *tr = vlib_add_trace (vm, node, b, sizeof (*tr));
tr->op_status = status;
tr->op = op_id;
}
static_always_inline u32
crypto_dequeue_frame (vlib_main_t * vm, vlib_node_runtime_t * node,
vnet_crypto_thread_t * ct,
vnet_crypto_frame_dequeue_t * hdl, u32 n_cache,
u32 * n_total)
{
vnet_crypto_main_t *cm = &crypto_main;
u32 n_elts = 0;
u32 enqueue_thread_idx = ~0;
vnet_crypto_async_frame_t *cf = (hdl) (vm, &n_elts, &enqueue_thread_idx);
*n_total += n_elts;
while (cf || n_elts)
{
if (cf)
{
vec_validate (ct->buffer_indices, n_cache + cf->n_elts);
vec_validate (ct->nexts, n_cache + cf->n_elts);
clib_memcpy_fast (ct->buffer_indices + n_cache, cf->buffer_indices,
sizeof (u32) * cf->n_elts);
if (cf->state == VNET_CRYPTO_FRAME_STATE_SUCCESS)
{
clib_memcpy_fast (ct->nexts + n_cache, cf->next_node_index,
sizeof (u16) * cf->n_elts);
}
else
{
u32 i;
for (i = 0; i < cf->n_elts; i++)
{
if (cf->elts[i].status != VNET_CRYPTO_OP_STATUS_COMPLETED)
{
ct->nexts[i + n_cache] = CRYPTO_DISPATCH_NEXT_ERR_DROP;
vlib_node_increment_counter (vm, node->node_index,
cf->elts[i].status, 1);
}
else
ct->nexts[i + n_cache] = cf->next_node_index[i];
}
}
n_cache += cf->n_elts;
if (n_cache >= VLIB_FRAME_SIZE)
{
vlib_buffer_enqueue_to_next (vm, node, ct->buffer_indices,
ct->nexts, n_cache);
n_cache = 0;
}
if (PREDICT_FALSE (node->flags & VLIB_NODE_FLAG_TRACE))
{
u32 i;
for (i = 0; i < cf->n_elts; i++)
{
vlib_buffer_t *b = vlib_get_buffer (vm,
cf->buffer_indices[i]);
if (b->flags & VLIB_BUFFER_IS_TRACED)
vnet_crypto_async_add_trace (vm, node, b, cf->op,
cf->elts[i].status);
}
}
vnet_crypto_async_free_frame (vm, cf);
}
/* signal enqueue-thread to dequeue the processed frame (n_elts>0) */
if (cm->dispatch_mode == VNET_CRYPTO_ASYNC_DISPATCH_INTERRUPT
&& n_elts > 0)
{
vlib_node_set_interrupt_pending (vlib_mains[enqueue_thread_idx],
cm->crypto_node_index);
}
n_elts = 0;
enqueue_thread_idx = 0;
cf = (hdl) (vm, &n_elts, &enqueue_thread_idx);
*n_total += n_elts;
}
return n_cache;
}
VLIB_NODE_FN (crypto_dispatch_node) (vlib_main_t * vm,
vlib_node_runtime_t * node,
vlib_frame_t * frame)
{
vnet_crypto_main_t *cm = &crypto_main;
vnet_crypto_thread_t *ct = cm->threads + vm->thread_index;
u32 n_dispatched = 0, n_cache = 0;
u32 index;
/* *INDENT-OFF* */
clib_bitmap_foreach (index, cm->async_active_ids) {
n_cache = crypto_dequeue_frame (vm, node, ct, cm->dequeue_handlers[index],
n_cache, &n_dispatched);
}
/* *INDENT-ON* */
if (n_cache)
vlib_buffer_enqueue_to_next (vm, node, ct->buffer_indices, ct->nexts,
n_cache);
return n_dispatched;
}
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (crypto_dispatch_node) = {
.name = "crypto-dispatch",
.type = VLIB_NODE_TYPE_INPUT,
.state = VLIB_NODE_STATE_DISABLED,
.format_trace = format_crypto_dispatch_trace,
.n_errors = ARRAY_LEN(vnet_crypto_async_error_strings),
.error_strings = vnet_crypto_async_error_strings,
.n_next_nodes = CRYPTO_DISPATCH_N_NEXT,
.next_nodes = {
#define _(n, s) \
[CRYPTO_DISPATCH_NEXT_##n] = s,
foreach_crypto_dispatch_next
#undef _
},
};
/* *INDENT-ON* */
/*
* fd.io coding-style-patch-verification: ON
*
* Local Variables:
* eval: (c-set-style "gnu")
* End:
*/