blob: 0ef90673596126a8c9f9fbdab4096b9320db5202 [file] [log] [blame]
/*
* ipsec.c : IPSEC module functions
*
* Copyright (c) 2015 Cisco and/or its affiliates.
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at:
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <vnet/vnet.h>
#include <vnet/api_errno.h>
#include <vnet/ip/ip.h>
#include <vnet/interface.h>
#include <vnet/udp/udp.h>
#include <vnet/ipsec/ipsec.h>
#include <vnet/ipsec/esp.h>
#include <vnet/ipsec/ah.h>
ipsec_main_t ipsec_main;
esp_async_post_next_t esp_encrypt_async_next;
esp_async_post_next_t esp_decrypt_async_next;
static clib_error_t *
ipsec_check_ah_support (ipsec_sa_t * sa)
{
ipsec_main_t *im = &ipsec_main;
if (sa->integ_alg == IPSEC_INTEG_ALG_NONE)
return clib_error_return (0, "unsupported none integ-alg");
if (!vnet_crypto_is_set_handler (im->integ_algs[sa->integ_alg].alg))
return clib_error_return (0, "No crypto engine support for %U",
format_ipsec_integ_alg, sa->integ_alg);
return 0;
}
static clib_error_t *
ipsec_check_esp_support (ipsec_sa_t * sa)
{
ipsec_main_t *im = &ipsec_main;
if (IPSEC_INTEG_ALG_NONE != sa->integ_alg)
{
if (!vnet_crypto_is_set_handler (im->integ_algs[sa->integ_alg].alg))
return clib_error_return (0, "No crypto engine support for %U",
format_ipsec_integ_alg, sa->integ_alg);
}
if (IPSEC_CRYPTO_ALG_NONE != sa->crypto_alg)
{
if (!vnet_crypto_is_set_handler (im->crypto_algs[sa->crypto_alg].alg))
return clib_error_return (0, "No crypto engine support for %U",
format_ipsec_crypto_alg, sa->crypto_alg);
}
return (0);
}
clib_error_t *
ipsec_add_del_sa_sess_cb (ipsec_main_t * im, u32 sa_index, u8 is_add)
{
ipsec_ah_backend_t *ah =
pool_elt_at_index (im->ah_backends, im->ah_current_backend);
if (ah->add_del_sa_sess_cb)
{
clib_error_t *err = ah->add_del_sa_sess_cb (sa_index, is_add);
if (err)
return err;
}
ipsec_esp_backend_t *esp =
pool_elt_at_index (im->esp_backends, im->esp_current_backend);
if (esp->add_del_sa_sess_cb)
{
clib_error_t *err = esp->add_del_sa_sess_cb (sa_index, is_add);
if (err)
return err;
}
return 0;
}
clib_error_t *
ipsec_check_support_cb (ipsec_main_t * im, ipsec_sa_t * sa)
{
clib_error_t *error = 0;
if (PREDICT_FALSE (sa->protocol == IPSEC_PROTOCOL_AH))
{
ipsec_ah_backend_t *ah =
pool_elt_at_index (im->ah_backends, im->ah_current_backend);
ASSERT (ah->check_support_cb);
error = ah->check_support_cb (sa);
}
else
{
ipsec_esp_backend_t *esp =
pool_elt_at_index (im->esp_backends, im->esp_current_backend);
ASSERT (esp->check_support_cb);
error = esp->check_support_cb (sa);
}
return error;
}
static void
ipsec_add_node (vlib_main_t * vm, const char *node_name,
const char *prev_node_name, u32 * out_node_index,
u32 * out_next_index)
{
vlib_node_t *prev_node, *node;
prev_node = vlib_get_node_by_name (vm, (u8 *) prev_node_name);
ASSERT (prev_node);
node = vlib_get_node_by_name (vm, (u8 *) node_name);
ASSERT (node);
*out_node_index = node->index;
*out_next_index = vlib_node_add_next (vm, prev_node->index, node->index);
}
void
ipsec_add_feature (const char *arc_name,
const char *node_name, u32 * out_feature_index)
{
u8 arc;
arc = vnet_get_feature_arc_index (arc_name);
ASSERT (arc != (u8) ~ 0);
*out_feature_index = vnet_get_feature_index (arc, node_name);
}
void
ipsec_unregister_udp_port (u16 port)
{
ipsec_main_t *im = &ipsec_main;
u32 n_regs;
uword *p;
p = hash_get (im->udp_port_registrations, port);
ASSERT (p);
n_regs = p[0];
if (0 == --n_regs)
{
udp_unregister_dst_port (vlib_get_main (), port, 1);
hash_unset (im->udp_port_registrations, port);
}
else
{
hash_unset (im->udp_port_registrations, port);
hash_set (im->udp_port_registrations, port, n_regs);
}
}
void
ipsec_register_udp_port (u16 port)
{
ipsec_main_t *im = &ipsec_main;
u32 n_regs;
uword *p;
p = hash_get (im->udp_port_registrations, port);
n_regs = (p ? p[0] : 0);
if (0 == n_regs++)
udp_register_dst_port (vlib_get_main (), port,
ipsec4_tun_input_node.index, 1);
hash_unset (im->udp_port_registrations, port);
hash_set (im->udp_port_registrations, port, n_regs);
}
u32
ipsec_register_ah_backend (vlib_main_t * vm, ipsec_main_t * im,
const char *name,
const char *ah4_encrypt_node_name,
const char *ah4_decrypt_node_name,
const char *ah6_encrypt_node_name,
const char *ah6_decrypt_node_name,
check_support_cb_t ah_check_support_cb,
add_del_sa_sess_cb_t ah_add_del_sa_sess_cb)
{
ipsec_ah_backend_t *b;
pool_get (im->ah_backends, b);
b->name = format (0, "%s%c", name, 0);
ipsec_add_node (vm, ah4_encrypt_node_name, "ipsec4-output-feature",
&b->ah4_encrypt_node_index, &b->ah4_encrypt_next_index);
ipsec_add_node (vm, ah4_decrypt_node_name, "ipsec4-input-feature",
&b->ah4_decrypt_node_index, &b->ah4_decrypt_next_index);
ipsec_add_node (vm, ah6_encrypt_node_name, "ipsec6-output-feature",
&b->ah6_encrypt_node_index, &b->ah6_encrypt_next_index);
ipsec_add_node (vm, ah6_decrypt_node_name, "ipsec6-input-feature",
&b->ah6_decrypt_node_index, &b->ah6_decrypt_next_index);
b->check_support_cb = ah_check_support_cb;
b->add_del_sa_sess_cb = ah_add_del_sa_sess_cb;
return b - im->ah_backends;
}
u32
ipsec_register_esp_backend (vlib_main_t * vm, ipsec_main_t * im,
const char *name,
const char *esp4_encrypt_node_name,
const char *esp4_encrypt_node_tun_name,
const char *esp4_decrypt_node_name,
const char *esp4_decrypt_tun_node_name,
const char *esp6_encrypt_node_name,
const char *esp6_encrypt_node_tun_name,
const char *esp6_decrypt_node_name,
const char *esp6_decrypt_tun_node_name,
check_support_cb_t esp_check_support_cb,
add_del_sa_sess_cb_t esp_add_del_sa_sess_cb,
enable_disable_cb_t enable_disable_cb)
{
ipsec_esp_backend_t *b;
pool_get (im->esp_backends, b);
b->name = format (0, "%s%c", name, 0);
ipsec_add_node (vm, esp4_encrypt_node_name, "ipsec4-output-feature",
&b->esp4_encrypt_node_index, &b->esp4_encrypt_next_index);
ipsec_add_node (vm, esp4_decrypt_node_name, "ipsec4-input-feature",
&b->esp4_decrypt_node_index, &b->esp4_decrypt_next_index);
ipsec_add_node (vm, esp6_encrypt_node_name, "ipsec6-output-feature",
&b->esp6_encrypt_node_index, &b->esp6_encrypt_next_index);
ipsec_add_node (vm, esp6_decrypt_node_name, "ipsec6-input-feature",
&b->esp6_decrypt_node_index, &b->esp6_decrypt_next_index);
ipsec_add_node (vm, esp4_decrypt_tun_node_name, "ipsec4-tun-input",
&b->esp4_decrypt_tun_node_index,
&b->esp4_decrypt_tun_next_index);
ipsec_add_node (vm, esp6_decrypt_tun_node_name, "ipsec6-tun-input",
&b->esp6_decrypt_tun_node_index,
&b->esp6_decrypt_tun_next_index);
b->esp6_encrypt_tun_node_index =
vlib_get_node_by_name (vm, (u8 *) esp6_encrypt_node_tun_name)->index;
b->esp4_encrypt_tun_node_index =
vlib_get_node_by_name (vm, (u8 *) esp4_encrypt_node_tun_name)->index;
b->check_support_cb = esp_check_support_cb;
b->add_del_sa_sess_cb = esp_add_del_sa_sess_cb;
b->enable_disable_cb = enable_disable_cb;
return b - im->esp_backends;
}
clib_error_t *
ipsec_rsc_in_use (ipsec_main_t * im)
{
/* return an error is crypto resource are in use */
if (pool_elts (im->sad) > 0)
return clib_error_return (0,
"%d SA entries configured",
pool_elts (im->sad));
return (NULL);
}
int
ipsec_select_ah_backend (ipsec_main_t * im, u32 backend_idx)
{
if (ipsec_rsc_in_use (im))
return VNET_API_ERROR_RSRC_IN_USE;
if (pool_is_free_index (im->ah_backends, backend_idx))
return VNET_API_ERROR_INVALID_VALUE;
ipsec_ah_backend_t *b = pool_elt_at_index (im->ah_backends, backend_idx);
im->ah_current_backend = backend_idx;
im->ah4_encrypt_node_index = b->ah4_encrypt_node_index;
im->ah4_decrypt_node_index = b->ah4_decrypt_node_index;
im->ah4_encrypt_next_index = b->ah4_encrypt_next_index;
im->ah4_decrypt_next_index = b->ah4_decrypt_next_index;
im->ah6_encrypt_node_index = b->ah6_encrypt_node_index;
im->ah6_decrypt_node_index = b->ah6_decrypt_node_index;
im->ah6_encrypt_next_index = b->ah6_encrypt_next_index;
im->ah6_decrypt_next_index = b->ah6_decrypt_next_index;
return 0;
}
int
ipsec_select_esp_backend (ipsec_main_t * im, u32 backend_idx)
{
if (ipsec_rsc_in_use (im))
return VNET_API_ERROR_RSRC_IN_USE;
if (pool_is_free_index (im->esp_backends, backend_idx))
return VNET_API_ERROR_INVALID_VALUE;
/* disable current backend */
if (im->esp_current_backend != ~0)
{
ipsec_esp_backend_t *cb = pool_elt_at_index (im->esp_backends,
im->esp_current_backend);
if (cb->enable_disable_cb)
{
if ((cb->enable_disable_cb) (0) != 0)
return -1;
}
}
ipsec_esp_backend_t *b = pool_elt_at_index (im->esp_backends, backend_idx);
im->esp_current_backend = backend_idx;
im->esp4_encrypt_node_index = b->esp4_encrypt_node_index;
im->esp4_decrypt_node_index = b->esp4_decrypt_node_index;
im->esp4_encrypt_next_index = b->esp4_encrypt_next_index;
im->esp4_decrypt_next_index = b->esp4_decrypt_next_index;
im->esp6_encrypt_node_index = b->esp6_encrypt_node_index;
im->esp6_decrypt_node_index = b->esp6_decrypt_node_index;
im->esp6_encrypt_next_index = b->esp6_encrypt_next_index;
im->esp6_decrypt_next_index = b->esp6_decrypt_next_index;
im->esp4_decrypt_tun_node_index = b->esp4_decrypt_tun_node_index;
im->esp4_decrypt_tun_next_index = b->esp4_decrypt_tun_next_index;
im->esp6_decrypt_tun_node_index = b->esp6_decrypt_tun_node_index;
im->esp6_decrypt_tun_next_index = b->esp6_decrypt_tun_next_index;
im->esp4_encrypt_tun_node_index = b->esp4_encrypt_tun_node_index;
im->esp6_encrypt_tun_node_index = b->esp6_encrypt_tun_node_index;
if (b->enable_disable_cb)
{
if ((b->enable_disable_cb) (1) != 0)
return -1;
}
return 0;
}
void
ipsec_set_async_mode (u32 is_enabled)
{
ipsec_main_t *im = &ipsec_main;
ipsec_sa_t *sa;
/* lock all SAs before change im->async_mode */
pool_foreach (sa, im->sad, (
{
fib_node_lock (&sa->node);
}));
im->async_mode = is_enabled;
/* change SA crypto op data before unlock them */
pool_foreach (sa, im->sad, (
{
sa->crypto_op_data = is_enabled ?
sa->async_op_data.data : sa->sync_op_data.data;
fib_node_unlock (&sa->node);
}));
}
static void
crypto_engine_backend_register_post_node (vlib_main_t * vm)
{
esp_async_post_next_t *eit;
esp_async_post_next_t *dit;
eit = &esp_encrypt_async_next;
eit->esp4_post_next =
vnet_crypto_register_post_node (vm, "esp4-encrypt-post");
eit->esp6_post_next =
vnet_crypto_register_post_node (vm, "esp6-encrypt-post");
eit->esp4_tun_post_next =
vnet_crypto_register_post_node (vm, "esp4-encrypt-tun-post");
eit->esp6_tun_post_next =
vnet_crypto_register_post_node (vm, "esp6-encrypt-tun-post");
dit = &esp_decrypt_async_next;
dit->esp4_post_next =
vnet_crypto_register_post_node (vm, "esp4-decrypt-post");
dit->esp6_post_next =
vnet_crypto_register_post_node (vm, "esp6-decrypt-post");
dit->esp4_tun_post_next =
vnet_crypto_register_post_node (vm, "esp4-decrypt-tun-post");
dit->esp6_tun_post_next =
vnet_crypto_register_post_node (vm, "esp6-decrypt-tun-post");
}
static clib_error_t *
ipsec_init (vlib_main_t * vm)
{
clib_error_t *error;
ipsec_main_t *im = &ipsec_main;
ipsec_main_crypto_alg_t *a;
/* Backend registration requires the feature arcs to be set up */
if ((error = vlib_call_init_function (vm, vnet_feature_init)))
return (error);
im->vnet_main = vnet_get_main ();
im->vlib_main = vm;
im->spd_index_by_spd_id = hash_create (0, sizeof (uword));
im->sa_index_by_sa_id = hash_create (0, sizeof (uword));
im->spd_index_by_sw_if_index = hash_create (0, sizeof (uword));
vlib_node_t *node = vlib_get_node_by_name (vm, (u8 *) "error-drop");
ASSERT (node);
im->error_drop_node_index = node->index;
im->ah_current_backend = ~0;
im->esp_current_backend = ~0;
u32 idx = ipsec_register_ah_backend (vm, im, "crypto engine backend",
"ah4-encrypt",
"ah4-decrypt",
"ah6-encrypt",
"ah6-decrypt",
ipsec_check_ah_support,
NULL);
im->ah_default_backend = idx;
int rv = ipsec_select_ah_backend (im, idx);
ASSERT (0 == rv);
(void) (rv); // avoid warning
idx = ipsec_register_esp_backend (vm, im, "crypto engine backend",
"esp4-encrypt",
"esp4-encrypt-tun",
"esp4-decrypt",
"esp4-decrypt-tun",
"esp6-encrypt",
"esp6-encrypt-tun",
"esp6-decrypt",
"esp6-decrypt-tun",
ipsec_check_esp_support,
NULL, crypto_dispatch_enable_disable);
im->esp_default_backend = idx;
rv = ipsec_select_esp_backend (im, idx);
ASSERT (0 == rv);
(void) (rv); // avoid warning
if ((error = vlib_call_init_function (vm, ipsec_cli_init)))
return error;
vec_validate (im->crypto_algs, IPSEC_CRYPTO_N_ALG - 1);
a = im->crypto_algs + IPSEC_CRYPTO_ALG_NONE;
a->enc_op_id = VNET_CRYPTO_OP_NONE;
a->dec_op_id = VNET_CRYPTO_OP_NONE;
a->alg = VNET_CRYPTO_ALG_NONE;
a->iv_size = 0;
a->block_size = 1;
a = im->crypto_algs + IPSEC_CRYPTO_ALG_DES_CBC;
a->enc_op_id = VNET_CRYPTO_OP_DES_CBC_ENC;
a->dec_op_id = VNET_CRYPTO_OP_DES_CBC_DEC;
a->alg = VNET_CRYPTO_ALG_DES_CBC;
a->iv_size = a->block_size = 8;
a = im->crypto_algs + IPSEC_CRYPTO_ALG_3DES_CBC;
a->enc_op_id = VNET_CRYPTO_OP_3DES_CBC_ENC;
a->dec_op_id = VNET_CRYPTO_OP_3DES_CBC_DEC;
a->alg = VNET_CRYPTO_ALG_3DES_CBC;
a->iv_size = a->block_size = 8;
a = im->crypto_algs + IPSEC_CRYPTO_ALG_AES_CBC_128;
a->enc_op_id = VNET_CRYPTO_OP_AES_128_CBC_ENC;
a->dec_op_id = VNET_CRYPTO_OP_AES_128_CBC_DEC;
a->alg = VNET_CRYPTO_ALG_AES_128_CBC;
a->iv_size = a->block_size = 16;
a = im->crypto_algs + IPSEC_CRYPTO_ALG_AES_CBC_192;
a->enc_op_id = VNET_CRYPTO_OP_AES_192_CBC_ENC;
a->dec_op_id = VNET_CRYPTO_OP_AES_192_CBC_DEC;
a->alg = VNET_CRYPTO_ALG_AES_192_CBC;
a->iv_size = a->block_size = 16;
a = im->crypto_algs + IPSEC_CRYPTO_ALG_AES_CBC_256;
a->enc_op_id = VNET_CRYPTO_OP_AES_256_CBC_ENC;
a->dec_op_id = VNET_CRYPTO_OP_AES_256_CBC_DEC;
a->alg = VNET_CRYPTO_ALG_AES_256_CBC;
a->iv_size = a->block_size = 16;
a = im->crypto_algs + IPSEC_CRYPTO_ALG_AES_GCM_128;
a->enc_op_id = VNET_CRYPTO_OP_AES_128_GCM_ENC;
a->dec_op_id = VNET_CRYPTO_OP_AES_128_GCM_DEC;
a->alg = VNET_CRYPTO_ALG_AES_128_GCM;
a->iv_size = 8;
a->block_size = 16;
a->icv_size = 16;
a = im->crypto_algs + IPSEC_CRYPTO_ALG_AES_GCM_192;
a->enc_op_id = VNET_CRYPTO_OP_AES_192_GCM_ENC;
a->dec_op_id = VNET_CRYPTO_OP_AES_192_GCM_DEC;
a->alg = VNET_CRYPTO_ALG_AES_192_GCM;
a->iv_size = 8;
a->block_size = 16;
a->icv_size = 16;
a = im->crypto_algs + IPSEC_CRYPTO_ALG_AES_GCM_256;
a->enc_op_id = VNET_CRYPTO_OP_AES_256_GCM_ENC;
a->dec_op_id = VNET_CRYPTO_OP_AES_256_GCM_DEC;
a->alg = VNET_CRYPTO_ALG_AES_256_GCM;
a->iv_size = 8;
a->block_size = 16;
a->icv_size = 16;
vec_validate (im->integ_algs, IPSEC_INTEG_N_ALG - 1);
ipsec_main_integ_alg_t *i;
i = &im->integ_algs[IPSEC_INTEG_ALG_MD5_96];
i->op_id = VNET_CRYPTO_OP_MD5_HMAC;
i->alg = VNET_CRYPTO_ALG_HMAC_MD5;
i->icv_size = 12;
i = &im->integ_algs[IPSEC_INTEG_ALG_SHA1_96];
i->op_id = VNET_CRYPTO_OP_SHA1_HMAC;
i->alg = VNET_CRYPTO_ALG_HMAC_SHA1;
i->icv_size = 12;
i = &im->integ_algs[IPSEC_INTEG_ALG_SHA_256_96];
i->op_id = VNET_CRYPTO_OP_SHA1_HMAC;
i->alg = VNET_CRYPTO_ALG_HMAC_SHA256;
i->icv_size = 12;
i = &im->integ_algs[IPSEC_INTEG_ALG_SHA_256_128];
i->op_id = VNET_CRYPTO_OP_SHA256_HMAC;
i->alg = VNET_CRYPTO_ALG_HMAC_SHA256;
i->icv_size = 16;
i = &im->integ_algs[IPSEC_INTEG_ALG_SHA_384_192];
i->op_id = VNET_CRYPTO_OP_SHA384_HMAC;
i->alg = VNET_CRYPTO_ALG_HMAC_SHA384;
i->icv_size = 24;
i = &im->integ_algs[IPSEC_INTEG_ALG_SHA_512_256];
i->op_id = VNET_CRYPTO_OP_SHA512_HMAC;
i->alg = VNET_CRYPTO_ALG_HMAC_SHA512;
i->icv_size = 32;
vec_validate_aligned (im->ptd, vlib_num_workers (), CLIB_CACHE_LINE_BYTES);
im->ah4_enc_fq_index =
vlib_frame_queue_main_init (ah4_encrypt_node.index, 0);
im->ah4_dec_fq_index =
vlib_frame_queue_main_init (ah4_decrypt_node.index, 0);
im->ah6_enc_fq_index =
vlib_frame_queue_main_init (ah6_encrypt_node.index, 0);
im->ah6_dec_fq_index =
vlib_frame_queue_main_init (ah6_decrypt_node.index, 0);
im->esp4_enc_fq_index =
vlib_frame_queue_main_init (esp4_encrypt_node.index, 0);
im->esp4_dec_fq_index =
vlib_frame_queue_main_init (esp4_decrypt_node.index, 0);
im->esp6_enc_fq_index =
vlib_frame_queue_main_init (esp6_encrypt_node.index, 0);
im->esp6_dec_fq_index =
vlib_frame_queue_main_init (esp6_decrypt_node.index, 0);
im->esp4_enc_tun_fq_index =
vlib_frame_queue_main_init (esp4_encrypt_tun_node.index, 0);
im->esp6_enc_tun_fq_index =
vlib_frame_queue_main_init (esp6_encrypt_tun_node.index, 0);
im->esp4_dec_tun_fq_index =
vlib_frame_queue_main_init (esp4_decrypt_tun_node.index, 0);
im->esp6_dec_tun_fq_index =
vlib_frame_queue_main_init (esp6_decrypt_tun_node.index, 0);
im->async_mode = 0;
crypto_engine_backend_register_post_node (vm);
return 0;
}
VLIB_INIT_FUNCTION (ipsec_init);
/*
* fd.io coding-style-patch-verification: ON
*
* Local Variables:
* eval: (c-set-style "gnu")
* End:
*/