| /* |
| * Copyright (c) 2015 Cisco and/or its affiliates. |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at: |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| #include <vnet/classify/vnet_classify.h> |
| #include <vnet/classify/input_acl.h> |
| #include <vnet/ip/ip.h> |
| #include <vnet/api_errno.h> /* for API error numbers */ |
| #include <vnet/l2/l2_classify.h> /* for L2_INPUT_CLASSIFY_NEXT_xxx */ |
| #include <vnet/fib/fib_table.h> |
| |
| vnet_classify_main_t vnet_classify_main; |
| |
| #if VALIDATION_SCAFFOLDING |
| /* Validation scaffolding */ |
| void mv (vnet_classify_table_t * t) |
| { |
| void * oldheap; |
| |
| oldheap = clib_mem_set_heap (t->mheap); |
| clib_mem_validate(); |
| clib_mem_set_heap (oldheap); |
| } |
| |
| void rogue (vnet_classify_table_t * t) |
| { |
| int i, j, k; |
| vnet_classify_entry_t * v, * save_v; |
| u32 active_elements = 0; |
| vnet_classify_bucket_t * b; |
| |
| for (i = 0; i < t->nbuckets; i++) |
| { |
| b = &t->buckets [i]; |
| if (b->offset == 0) |
| continue; |
| save_v = vnet_classify_get_entry (t, b->offset); |
| for (j = 0; j < (1<<b->log2_pages); j++) |
| { |
| for (k = 0; k < t->entries_per_page; k++) |
| { |
| v = vnet_classify_entry_at_index |
| (t, save_v, j*t->entries_per_page + k); |
| |
| if (vnet_classify_entry_is_busy (v)) |
| active_elements++; |
| } |
| } |
| } |
| |
| if (active_elements != t->active_elements) |
| clib_warning ("found %u expected %u elts", active_elements, |
| t->active_elements); |
| } |
| #else |
| void mv (vnet_classify_table_t * t) { } |
| void rogue (vnet_classify_table_t * t) { } |
| #endif |
| |
| void vnet_classify_register_unformat_l2_next_index_fn (unformat_function_t * fn) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| |
| vec_add1 (cm->unformat_l2_next_index_fns, fn); |
| } |
| |
| void vnet_classify_register_unformat_ip_next_index_fn (unformat_function_t * fn) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| |
| vec_add1 (cm->unformat_ip_next_index_fns, fn); |
| } |
| |
| void |
| vnet_classify_register_unformat_acl_next_index_fn (unformat_function_t * fn) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| |
| vec_add1 (cm->unformat_acl_next_index_fns, fn); |
| } |
| |
| void |
| vnet_classify_register_unformat_policer_next_index_fn (unformat_function_t * fn) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| |
| vec_add1 (cm->unformat_policer_next_index_fns, fn); |
| } |
| |
| void vnet_classify_register_unformat_opaque_index_fn (unformat_function_t * fn) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| |
| vec_add1 (cm->unformat_opaque_index_fns, fn); |
| } |
| |
| vnet_classify_table_t * |
| vnet_classify_new_table (vnet_classify_main_t *cm, |
| u8 * mask, u32 nbuckets, u32 memory_size, |
| u32 skip_n_vectors, |
| u32 match_n_vectors) |
| { |
| vnet_classify_table_t * t; |
| void * oldheap; |
| |
| nbuckets = 1 << (max_log2 (nbuckets)); |
| |
| pool_get_aligned (cm->tables, t, CLIB_CACHE_LINE_BYTES); |
| memset(t, 0, sizeof (*t)); |
| |
| vec_validate_aligned (t->mask, match_n_vectors - 1, sizeof(u32x4)); |
| clib_memcpy (t->mask, mask, match_n_vectors * sizeof (u32x4)); |
| |
| t->next_table_index = ~0; |
| t->nbuckets = nbuckets; |
| t->log2_nbuckets = max_log2 (nbuckets); |
| t->match_n_vectors = match_n_vectors; |
| t->skip_n_vectors = skip_n_vectors; |
| t->entries_per_page = 2; |
| |
| t->mheap = mheap_alloc (0 /* use VM */, memory_size); |
| |
| vec_validate_aligned (t->buckets, nbuckets - 1, CLIB_CACHE_LINE_BYTES); |
| oldheap = clib_mem_set_heap (t->mheap); |
| |
| t->writer_lock = clib_mem_alloc_aligned (CLIB_CACHE_LINE_BYTES, |
| CLIB_CACHE_LINE_BYTES); |
| t->writer_lock[0] = 0; |
| |
| clib_mem_set_heap (oldheap); |
| return (t); |
| } |
| |
| void vnet_classify_delete_table_index (vnet_classify_main_t *cm, |
| u32 table_index, int del_chain) |
| { |
| vnet_classify_table_t * t; |
| |
| /* Tolerate multiple frees, up to a point */ |
| if (pool_is_free_index (cm->tables, table_index)) |
| return; |
| |
| t = pool_elt_at_index (cm->tables, table_index); |
| if (del_chain && t->next_table_index != ~0) |
| /* Recursively delete the entire chain */ |
| vnet_classify_delete_table_index (cm, t->next_table_index, del_chain); |
| |
| vec_free (t->mask); |
| vec_free (t->buckets); |
| mheap_free (t->mheap); |
| |
| pool_put (cm->tables, t); |
| } |
| |
| static vnet_classify_entry_t * |
| vnet_classify_entry_alloc (vnet_classify_table_t * t, u32 log2_pages) |
| { |
| vnet_classify_entry_t * rv = 0; |
| #define _(size) \ |
| vnet_classify_entry_##size##_t * rv##size = 0; |
| foreach_size_in_u32x4; |
| #undef _ |
| |
| void * oldheap; |
| |
| ASSERT (t->writer_lock[0]); |
| if (log2_pages >= vec_len (t->freelists) || t->freelists [log2_pages] == 0) |
| { |
| oldheap = clib_mem_set_heap (t->mheap); |
| |
| vec_validate (t->freelists, log2_pages); |
| |
| switch(t->match_n_vectors) |
| { |
| /* Euchre the vector allocator into allocating the right sizes */ |
| #define _(size) \ |
| case size: \ |
| vec_validate_aligned \ |
| (rv##size, ((1<<log2_pages)*t->entries_per_page) - 1, \ |
| CLIB_CACHE_LINE_BYTES); \ |
| rv = (vnet_classify_entry_t *) rv##size; \ |
| break; |
| foreach_size_in_u32x4; |
| #undef _ |
| |
| default: |
| abort(); |
| } |
| |
| clib_mem_set_heap (oldheap); |
| goto initialize; |
| } |
| rv = t->freelists[log2_pages]; |
| t->freelists[log2_pages] = rv->next_free; |
| |
| initialize: |
| ASSERT(rv); |
| ASSERT (vec_len(rv) == (1<<log2_pages)*t->entries_per_page); |
| |
| switch (t->match_n_vectors) |
| { |
| #define _(size) \ |
| case size: \ |
| if(vec_len(rv)) \ |
| memset (rv, 0xff, sizeof (*rv##size) * vec_len(rv)); \ |
| break; |
| foreach_size_in_u32x4; |
| #undef _ |
| |
| default: |
| abort(); |
| } |
| |
| return rv; |
| } |
| |
| static void |
| vnet_classify_entry_free (vnet_classify_table_t * t, |
| vnet_classify_entry_t * v) |
| { |
| u32 free_list_index; |
| |
| ASSERT (t->writer_lock[0]); |
| |
| free_list_index = min_log2(vec_len(v)/t->entries_per_page); |
| |
| ASSERT(vec_len (t->freelists) > free_list_index); |
| |
| v->next_free = t->freelists[free_list_index]; |
| t->freelists[free_list_index] = v; |
| } |
| |
| static inline void make_working_copy |
| (vnet_classify_table_t * t, vnet_classify_bucket_t * b) |
| { |
| vnet_classify_entry_t * v; |
| vnet_classify_bucket_t working_bucket __attribute__((aligned (8))); |
| void * oldheap; |
| vnet_classify_entry_t * working_copy; |
| #define _(size) \ |
| vnet_classify_entry_##size##_t * working_copy##size = 0; |
| foreach_size_in_u32x4; |
| #undef _ |
| u32 cpu_number = os_get_cpu_number(); |
| |
| if (cpu_number >= vec_len (t->working_copies)) |
| { |
| oldheap = clib_mem_set_heap (t->mheap); |
| vec_validate (t->working_copies, cpu_number); |
| clib_mem_set_heap (oldheap); |
| } |
| |
| /* |
| * working_copies are per-cpu so that near-simultaneous |
| * updates from multiple threads will not result in sporadic, spurious |
| * lookup failures. |
| */ |
| working_copy = t->working_copies[cpu_number]; |
| |
| t->saved_bucket.as_u64 = b->as_u64; |
| oldheap = clib_mem_set_heap (t->mheap); |
| |
| if ((1<<b->log2_pages)*t->entries_per_page > vec_len (working_copy)) |
| { |
| switch(t->match_n_vectors) |
| { |
| /* Euchre the vector allocator into allocating the right sizes */ |
| #define _(size) \ |
| case size: \ |
| working_copy##size = (void *) working_copy; \ |
| vec_validate_aligned \ |
| (working_copy##size, \ |
| ((1<<b->log2_pages)*t->entries_per_page) - 1, \ |
| CLIB_CACHE_LINE_BYTES); \ |
| working_copy = (void *) working_copy##size; \ |
| break; |
| foreach_size_in_u32x4; |
| #undef _ |
| |
| default: |
| abort(); |
| } |
| t->working_copies[cpu_number] = working_copy; |
| } |
| |
| _vec_len(working_copy) = (1<<b->log2_pages)*t->entries_per_page; |
| clib_mem_set_heap (oldheap); |
| |
| v = vnet_classify_get_entry (t, b->offset); |
| |
| switch(t->match_n_vectors) |
| { |
| #define _(size) \ |
| case size: \ |
| clib_memcpy (working_copy, v, \ |
| sizeof (vnet_classify_entry_##size##_t) \ |
| * (1<<b->log2_pages) \ |
| * (t->entries_per_page)); \ |
| break; |
| foreach_size_in_u32x4 ; |
| #undef _ |
| |
| default: |
| abort(); |
| } |
| |
| working_bucket.as_u64 = b->as_u64; |
| working_bucket.offset = vnet_classify_get_offset (t, working_copy); |
| CLIB_MEMORY_BARRIER(); |
| b->as_u64 = working_bucket.as_u64; |
| t->working_copies[cpu_number] = working_copy; |
| } |
| |
| static vnet_classify_entry_t * |
| split_and_rehash (vnet_classify_table_t * t, |
| vnet_classify_entry_t * old_values, |
| u32 new_log2_pages) |
| { |
| vnet_classify_entry_t * new_values, * v, * new_v; |
| int i, j, k; |
| |
| new_values = vnet_classify_entry_alloc (t, new_log2_pages); |
| |
| for (i = 0; i < (vec_len (old_values)/t->entries_per_page); i++) |
| { |
| u64 new_hash; |
| |
| for (j = 0; j < t->entries_per_page; j++) |
| { |
| v = vnet_classify_entry_at_index |
| (t, old_values, i * t->entries_per_page + j); |
| |
| if (vnet_classify_entry_is_busy (v)) |
| { |
| /* Hack so we can use the packet hash routine */ |
| u8 * key_minus_skip; |
| key_minus_skip = (u8 *) v->key; |
| key_minus_skip -= t->skip_n_vectors * sizeof (u32x4); |
| |
| new_hash = vnet_classify_hash_packet (t, key_minus_skip); |
| new_hash >>= t->log2_nbuckets; |
| new_hash &= (1<<new_log2_pages) - 1; |
| |
| for (k = 0; k < t->entries_per_page; k++) |
| { |
| new_v = vnet_classify_entry_at_index (t, new_values, |
| new_hash + k); |
| |
| if (vnet_classify_entry_is_free (new_v)) |
| { |
| clib_memcpy (new_v, v, sizeof (vnet_classify_entry_t) |
| + (t->match_n_vectors * sizeof (u32x4))); |
| new_v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); |
| goto doublebreak; |
| } |
| } |
| /* Crap. Tell caller to try again */ |
| vnet_classify_entry_free (t, new_values); |
| return 0; |
| } |
| doublebreak: |
| ; |
| } |
| } |
| return new_values; |
| } |
| |
| int vnet_classify_add_del (vnet_classify_table_t * t, |
| vnet_classify_entry_t * add_v, |
| int is_add) |
| { |
| u32 bucket_index; |
| vnet_classify_bucket_t * b, tmp_b; |
| vnet_classify_entry_t * v, * new_v, * save_new_v, * working_copy, * save_v; |
| u32 value_index; |
| int rv = 0; |
| int i; |
| u64 hash, new_hash; |
| u32 new_log2_pages; |
| u32 cpu_number = os_get_cpu_number(); |
| u8 * key_minus_skip; |
| |
| ASSERT ((add_v->flags & VNET_CLASSIFY_ENTRY_FREE) == 0); |
| |
| key_minus_skip = (u8 *) add_v->key; |
| key_minus_skip -= t->skip_n_vectors * sizeof (u32x4); |
| |
| hash = vnet_classify_hash_packet (t, key_minus_skip); |
| |
| bucket_index = hash & (t->nbuckets-1); |
| b = &t->buckets[bucket_index]; |
| |
| hash >>= t->log2_nbuckets; |
| |
| while (__sync_lock_test_and_set (t->writer_lock, 1)) |
| ; |
| |
| /* First elt in the bucket? */ |
| if (b->offset == 0) |
| { |
| if (is_add == 0) |
| { |
| rv = -1; |
| goto unlock; |
| } |
| |
| v = vnet_classify_entry_alloc (t, 0 /* new_log2_pages */); |
| clib_memcpy (v, add_v, sizeof (vnet_classify_entry_t) + |
| t->match_n_vectors * sizeof (u32x4)); |
| v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); |
| |
| tmp_b.as_u64 = 0; |
| tmp_b.offset = vnet_classify_get_offset (t, v); |
| |
| b->as_u64 = tmp_b.as_u64; |
| t->active_elements ++; |
| |
| goto unlock; |
| } |
| |
| make_working_copy (t, b); |
| |
| save_v = vnet_classify_get_entry (t, t->saved_bucket.offset); |
| value_index = hash & ((1<<t->saved_bucket.log2_pages)-1); |
| |
| if (is_add) |
| { |
| /* |
| * For obvious (in hindsight) reasons, see if we're supposed to |
| * replace an existing key, then look for an empty slot. |
| */ |
| |
| for (i = 0; i < t->entries_per_page; i++) |
| { |
| v = vnet_classify_entry_at_index (t, save_v, value_index + i); |
| |
| if (!memcmp (v->key, add_v->key, t->match_n_vectors * sizeof (u32x4))) |
| { |
| clib_memcpy (v, add_v, sizeof (vnet_classify_entry_t) + |
| t->match_n_vectors * sizeof(u32x4)); |
| v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); |
| |
| CLIB_MEMORY_BARRIER(); |
| /* Restore the previous (k,v) pairs */ |
| b->as_u64 = t->saved_bucket.as_u64; |
| goto unlock; |
| } |
| } |
| for (i = 0; i < t->entries_per_page; i++) |
| { |
| v = vnet_classify_entry_at_index (t, save_v, value_index + i); |
| |
| if (vnet_classify_entry_is_free (v)) |
| { |
| clib_memcpy (v, add_v, sizeof (vnet_classify_entry_t) + |
| t->match_n_vectors * sizeof(u32x4)); |
| v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); |
| CLIB_MEMORY_BARRIER(); |
| b->as_u64 = t->saved_bucket.as_u64; |
| t->active_elements ++; |
| goto unlock; |
| } |
| } |
| /* no room at the inn... split case... */ |
| } |
| else |
| { |
| for (i = 0; i < t->entries_per_page; i++) |
| { |
| v = vnet_classify_entry_at_index (t, save_v, value_index + i); |
| |
| if (!memcmp (v->key, add_v->key, t->match_n_vectors * sizeof (u32x4))) |
| { |
| memset (v, 0xff, sizeof (vnet_classify_entry_t) + |
| t->match_n_vectors * sizeof(u32x4)); |
| v->flags |= VNET_CLASSIFY_ENTRY_FREE; |
| CLIB_MEMORY_BARRIER(); |
| b->as_u64 = t->saved_bucket.as_u64; |
| t->active_elements --; |
| goto unlock; |
| } |
| } |
| rv = -3; |
| b->as_u64 = t->saved_bucket.as_u64; |
| goto unlock; |
| } |
| |
| new_log2_pages = t->saved_bucket.log2_pages + 1; |
| |
| expand_again: |
| working_copy = t->working_copies[cpu_number]; |
| new_v = split_and_rehash (t, working_copy, new_log2_pages); |
| |
| if (new_v == 0) |
| { |
| new_log2_pages++; |
| goto expand_again; |
| } |
| |
| /* Try to add the new entry */ |
| save_new_v = new_v; |
| |
| key_minus_skip = (u8 *) add_v->key; |
| key_minus_skip -= t->skip_n_vectors * sizeof (u32x4); |
| |
| new_hash = vnet_classify_hash_packet_inline (t, key_minus_skip); |
| new_hash >>= t->log2_nbuckets; |
| new_hash &= (1<<min_log2((vec_len(new_v)/t->entries_per_page))) - 1; |
| |
| for (i = 0; i < t->entries_per_page; i++) |
| { |
| new_v = vnet_classify_entry_at_index (t, save_new_v, new_hash + i); |
| |
| if (vnet_classify_entry_is_free (new_v)) |
| { |
| clib_memcpy (new_v, add_v, sizeof (vnet_classify_entry_t) + |
| t->match_n_vectors * sizeof(u32x4)); |
| new_v->flags &= ~(VNET_CLASSIFY_ENTRY_FREE); |
| goto expand_ok; |
| } |
| } |
| /* Crap. Try again */ |
| new_log2_pages++; |
| vnet_classify_entry_free (t, save_new_v); |
| goto expand_again; |
| |
| expand_ok: |
| tmp_b.log2_pages = min_log2 (vec_len (save_new_v)/t->entries_per_page); |
| tmp_b.offset = vnet_classify_get_offset (t, save_new_v); |
| CLIB_MEMORY_BARRIER(); |
| b->as_u64 = tmp_b.as_u64; |
| t->active_elements ++; |
| v = vnet_classify_get_entry (t, t->saved_bucket.offset); |
| vnet_classify_entry_free (t, v); |
| |
| unlock: |
| CLIB_MEMORY_BARRIER(); |
| t->writer_lock[0] = 0; |
| |
| return rv; |
| } |
| |
| typedef CLIB_PACKED(struct { |
| ethernet_header_t eh; |
| ip4_header_t ip; |
| }) classify_data_or_mask_t; |
| |
| u64 vnet_classify_hash_packet (vnet_classify_table_t * t, u8 * h) |
| { |
| return vnet_classify_hash_packet_inline (t, h); |
| } |
| |
| vnet_classify_entry_t * |
| vnet_classify_find_entry (vnet_classify_table_t * t, |
| u8 * h, u64 hash, f64 now) |
| { |
| return vnet_classify_find_entry_inline (t, h, hash, now); |
| } |
| |
| static u8 * format_classify_entry (u8 * s, va_list * args) |
| { |
| vnet_classify_table_t * t = va_arg (*args, vnet_classify_table_t *); |
| vnet_classify_entry_t * e = va_arg (*args, vnet_classify_entry_t *); |
| |
| s = format |
| (s, "[%u]: next_index %d advance %d opaque %d action %d metadata %d\n", |
| vnet_classify_get_offset (t, e), e->next_index, e->advance, |
| e->opaque_index, e->action, e->metadata); |
| |
| |
| s = format (s, " k: %U\n", format_hex_bytes, e->key, |
| t->match_n_vectors * sizeof(u32x4)); |
| |
| if (vnet_classify_entry_is_busy (e)) |
| s = format (s, " hits %lld, last_heard %.2f\n", |
| e->hits, e->last_heard); |
| else |
| s = format (s, " entry is free\n"); |
| return s; |
| } |
| |
| u8 * format_classify_table (u8 * s, va_list * args) |
| { |
| vnet_classify_table_t * t = va_arg (*args, vnet_classify_table_t *); |
| int verbose = va_arg (*args, int); |
| vnet_classify_bucket_t * b; |
| vnet_classify_entry_t * v, * save_v; |
| int i, j, k; |
| u64 active_elements = 0; |
| |
| for (i = 0; i < t->nbuckets; i++) |
| { |
| b = &t->buckets [i]; |
| if (b->offset == 0) |
| { |
| if (verbose > 1) |
| s = format (s, "[%d]: empty\n", i); |
| continue; |
| } |
| |
| if (verbose) |
| { |
| s = format (s, "[%d]: heap offset %d, len %d\n", i, |
| b->offset, (1<<b->log2_pages)); |
| } |
| |
| save_v = vnet_classify_get_entry (t, b->offset); |
| for (j = 0; j < (1<<b->log2_pages); j++) |
| { |
| for (k = 0; k < t->entries_per_page; k++) |
| { |
| |
| v = vnet_classify_entry_at_index (t, save_v, |
| j*t->entries_per_page + k); |
| |
| if (vnet_classify_entry_is_free (v)) |
| { |
| if (verbose > 1) |
| s = format (s, " %d: empty\n", |
| j * t->entries_per_page + k); |
| continue; |
| } |
| if (verbose) |
| { |
| s = format (s, " %d: %U\n", |
| j * t->entries_per_page + k, |
| format_classify_entry, t, v); |
| } |
| active_elements++; |
| } |
| } |
| } |
| |
| s = format (s, " %lld active elements\n", active_elements); |
| s = format (s, " %d free lists\n", vec_len (t->freelists)); |
| return s; |
| } |
| |
| int vnet_classify_add_del_table (vnet_classify_main_t * cm, |
| u8 * mask, |
| u32 nbuckets, |
| u32 memory_size, |
| u32 skip, |
| u32 match, |
| u32 next_table_index, |
| u32 miss_next_index, |
| u32 * table_index, |
| u8 current_data_flag, |
| i16 current_data_offset, |
| int is_add, |
| int del_chain) |
| { |
| vnet_classify_table_t * t; |
| |
| if (is_add) |
| { |
| if (*table_index == ~0) /* add */ |
| { |
| if (memory_size == 0) |
| return VNET_API_ERROR_INVALID_MEMORY_SIZE; |
| |
| if (nbuckets == 0) |
| return VNET_API_ERROR_INVALID_VALUE; |
| |
| t = vnet_classify_new_table (cm, mask, nbuckets, memory_size, |
| skip, match); |
| t->next_table_index = next_table_index; |
| t->miss_next_index = miss_next_index; |
| t->current_data_flag = current_data_flag; |
| t->current_data_offset = current_data_offset; |
| *table_index = t - cm->tables; |
| } |
| else /* update */ |
| { |
| vnet_classify_main_t *cm = &vnet_classify_main; |
| t = pool_elt_at_index (cm->tables, *table_index); |
| |
| t->next_table_index = next_table_index; |
| } |
| return 0; |
| } |
| |
| vnet_classify_delete_table_index (cm, *table_index, del_chain); |
| return 0; |
| } |
| |
| #define foreach_tcp_proto_field \ |
| _(src) \ |
| _(dst) |
| |
| #define foreach_udp_proto_field \ |
| _(src_port) \ |
| _(dst_port) |
| |
| #define foreach_ip4_proto_field \ |
| _(src_address) \ |
| _(dst_address) \ |
| _(tos) \ |
| _(length) \ |
| _(fragment_id) \ |
| _(ttl) \ |
| _(protocol) \ |
| _(checksum) |
| |
| uword unformat_tcp_mask (unformat_input_t * input, va_list * args) |
| { |
| u8 ** maskp = va_arg (*args, u8 **); |
| u8 * mask = 0; |
| u8 found_something = 0; |
| tcp_header_t * tcp; |
| |
| #define _(a) u8 a=0; |
| foreach_tcp_proto_field; |
| #undef _ |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (0) ; |
| #define _(a) else if (unformat (input, #a)) a=1; |
| foreach_tcp_proto_field |
| #undef _ |
| else |
| break; |
| } |
| |
| #define _(a) found_something += a; |
| foreach_tcp_proto_field; |
| #undef _ |
| |
| if (found_something == 0) |
| return 0; |
| |
| vec_validate (mask, sizeof (*tcp) - 1); |
| |
| tcp = (tcp_header_t *) mask; |
| |
| #define _(a) if (a) memset (&tcp->a, 0xff, sizeof (tcp->a)); |
| foreach_tcp_proto_field; |
| #undef _ |
| |
| *maskp = mask; |
| return 1; |
| } |
| |
| uword unformat_udp_mask (unformat_input_t * input, va_list * args) |
| { |
| u8 ** maskp = va_arg (*args, u8 **); |
| u8 * mask = 0; |
| u8 found_something = 0; |
| udp_header_t * udp; |
| |
| #define _(a) u8 a=0; |
| foreach_udp_proto_field; |
| #undef _ |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (0) ; |
| #define _(a) else if (unformat (input, #a)) a=1; |
| foreach_udp_proto_field |
| #undef _ |
| else |
| break; |
| } |
| |
| #define _(a) found_something += a; |
| foreach_udp_proto_field; |
| #undef _ |
| |
| if (found_something == 0) |
| return 0; |
| |
| vec_validate (mask, sizeof (*udp) - 1); |
| |
| udp = (udp_header_t *) mask; |
| |
| #define _(a) if (a) memset (&udp->a, 0xff, sizeof (udp->a)); |
| foreach_udp_proto_field; |
| #undef _ |
| |
| *maskp = mask; |
| return 1; |
| } |
| |
| typedef struct { |
| u16 src_port, dst_port; |
| } tcpudp_header_t; |
| |
| uword unformat_l4_mask (unformat_input_t * input, va_list * args) |
| { |
| u8 ** maskp = va_arg (*args, u8 **); |
| u16 src_port = 0, dst_port = 0; |
| tcpudp_header_t * tcpudp; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (unformat (input, "tcp %U", unformat_tcp_mask, maskp)) |
| return 1; |
| else if (unformat (input, "udp %U", unformat_udp_mask, maskp)) |
| return 1; |
| else if (unformat (input, "src_port")) |
| src_port = 0xFFFF; |
| else if (unformat (input, "dst_port")) |
| dst_port = 0xFFFF; |
| else |
| return 0; |
| } |
| |
| if (!src_port && !dst_port) |
| return 0; |
| |
| u8 * mask = 0; |
| vec_validate (mask, sizeof (tcpudp_header_t) - 1); |
| |
| tcpudp = (tcpudp_header_t *) mask; |
| tcpudp->src_port = src_port; |
| tcpudp->dst_port = dst_port; |
| |
| *maskp = mask; |
| |
| return 1; |
| } |
| |
| uword unformat_ip4_mask (unformat_input_t * input, va_list * args) |
| { |
| u8 ** maskp = va_arg (*args, u8 **); |
| u8 * mask = 0; |
| u8 found_something = 0; |
| ip4_header_t * ip; |
| |
| #define _(a) u8 a=0; |
| foreach_ip4_proto_field; |
| #undef _ |
| u8 version = 0; |
| u8 hdr_length = 0; |
| |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (unformat (input, "version")) |
| version = 1; |
| else if (unformat (input, "hdr_length")) |
| hdr_length = 1; |
| else if (unformat (input, "src")) |
| src_address = 1; |
| else if (unformat (input, "dst")) |
| dst_address = 1; |
| else if (unformat (input, "proto")) |
| protocol = 1; |
| |
| #define _(a) else if (unformat (input, #a)) a=1; |
| foreach_ip4_proto_field |
| #undef _ |
| else |
| break; |
| } |
| |
| #define _(a) found_something += a; |
| foreach_ip4_proto_field; |
| #undef _ |
| |
| if (found_something == 0) |
| return 0; |
| |
| vec_validate (mask, sizeof (*ip) - 1); |
| |
| ip = (ip4_header_t *) mask; |
| |
| #define _(a) if (a) memset (&ip->a, 0xff, sizeof (ip->a)); |
| foreach_ip4_proto_field; |
| #undef _ |
| |
| ip->ip_version_and_header_length = 0; |
| |
| if (version) |
| ip->ip_version_and_header_length |= 0xF0; |
| |
| if (hdr_length) |
| ip->ip_version_and_header_length |= 0x0F; |
| |
| *maskp = mask; |
| return 1; |
| } |
| |
| #define foreach_ip6_proto_field \ |
| _(src_address) \ |
| _(dst_address) \ |
| _(payload_length) \ |
| _(hop_limit) \ |
| _(protocol) |
| |
| uword unformat_ip6_mask (unformat_input_t * input, va_list * args) |
| { |
| u8 ** maskp = va_arg (*args, u8 **); |
| u8 * mask = 0; |
| u8 found_something = 0; |
| ip6_header_t * ip; |
| u32 ip_version_traffic_class_and_flow_label; |
| |
| #define _(a) u8 a=0; |
| foreach_ip6_proto_field; |
| #undef _ |
| u8 version = 0; |
| u8 traffic_class = 0; |
| u8 flow_label = 0; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (unformat (input, "version")) |
| version = 1; |
| else if (unformat (input, "traffic-class")) |
| traffic_class = 1; |
| else if (unformat (input, "flow-label")) |
| flow_label = 1; |
| else if (unformat (input, "src")) |
| src_address = 1; |
| else if (unformat (input, "dst")) |
| dst_address = 1; |
| else if (unformat (input, "proto")) |
| protocol = 1; |
| |
| #define _(a) else if (unformat (input, #a)) a=1; |
| foreach_ip6_proto_field |
| #undef _ |
| else |
| break; |
| } |
| |
| #define _(a) found_something += a; |
| foreach_ip6_proto_field; |
| #undef _ |
| |
| if (found_something == 0) |
| return 0; |
| |
| vec_validate (mask, sizeof (*ip) - 1); |
| |
| ip = (ip6_header_t *) mask; |
| |
| #define _(a) if (a) memset (&ip->a, 0xff, sizeof (ip->a)); |
| foreach_ip6_proto_field; |
| #undef _ |
| |
| ip_version_traffic_class_and_flow_label = 0; |
| |
| if (version) |
| ip_version_traffic_class_and_flow_label |= 0xF0000000; |
| |
| if (traffic_class) |
| ip_version_traffic_class_and_flow_label |= 0x0FF00000; |
| |
| if (flow_label) |
| ip_version_traffic_class_and_flow_label |= 0x000FFFFF; |
| |
| ip->ip_version_traffic_class_and_flow_label = |
| clib_host_to_net_u32 (ip_version_traffic_class_and_flow_label); |
| |
| *maskp = mask; |
| return 1; |
| } |
| |
| uword unformat_l3_mask (unformat_input_t * input, va_list * args) |
| { |
| u8 ** maskp = va_arg (*args, u8 **); |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { |
| if (unformat (input, "ip4 %U", unformat_ip4_mask, maskp)) |
| return 1; |
| else if (unformat (input, "ip6 %U", unformat_ip6_mask, maskp)) |
| return 1; |
| else |
| break; |
| } |
| return 0; |
| } |
| |
| uword unformat_l2_mask (unformat_input_t * input, va_list * args) |
| { |
| u8 ** maskp = va_arg (*args, u8 **); |
| u8 * mask = 0; |
| u8 src = 0; |
| u8 dst = 0; |
| u8 proto = 0; |
| u8 tag1 = 0; |
| u8 tag2 = 0; |
| u8 ignore_tag1 = 0; |
| u8 ignore_tag2 = 0; |
| u8 cos1 = 0; |
| u8 cos2 = 0; |
| u8 dot1q = 0; |
| u8 dot1ad = 0; |
| int len = 14; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { |
| if (unformat (input, "src")) |
| src = 1; |
| else if (unformat (input, "dst")) |
| dst = 1; |
| else if (unformat (input, "proto")) |
| proto = 1; |
| else if (unformat (input, "tag1")) |
| tag1 = 1; |
| else if (unformat (input, "tag2")) |
| tag2 = 1; |
| else if (unformat (input, "ignore-tag1")) |
| ignore_tag1 = 1; |
| else if (unformat (input, "ignore-tag2")) |
| ignore_tag2 = 1; |
| else if (unformat (input, "cos1")) |
| cos1 = 1; |
| else if (unformat (input, "cos2")) |
| cos2 = 1; |
| else if (unformat (input, "dot1q")) |
| dot1q = 1; |
| else if (unformat (input, "dot1ad")) |
| dot1ad = 1; |
| else |
| break; |
| } |
| if ((src + dst + proto + tag1 + tag2 + dot1q + dot1ad + |
| ignore_tag1 + ignore_tag2 + cos1 + cos2) == 0) |
| return 0; |
| |
| if (tag1 || ignore_tag1 || cos1 || dot1q) |
| len = 18; |
| if (tag2 || ignore_tag2 || cos2 || dot1ad) |
| len = 22; |
| |
| vec_validate (mask, len-1); |
| |
| if (dst) |
| memset (mask, 0xff, 6); |
| |
| if (src) |
| memset (mask + 6, 0xff, 6); |
| |
| if (tag2 || dot1ad) |
| { |
| /* inner vlan tag */ |
| if (tag2) |
| { |
| mask[19] = 0xff; |
| mask[18] = 0x0f; |
| } |
| if (cos2) |
| mask[18] |= 0xe0; |
| if (proto) |
| mask[21] = mask [20] = 0xff; |
| if (tag1) |
| { |
| mask [15] = 0xff; |
| mask [14] = 0x0f; |
| } |
| if (cos1) |
| mask[14] |= 0xe0; |
| *maskp = mask; |
| return 1; |
| } |
| if (tag1 | dot1q) |
| { |
| if (tag1) |
| { |
| mask [15] = 0xff; |
| mask [14] = 0x0f; |
| } |
| if (cos1) |
| mask[14] |= 0xe0; |
| if (proto) |
| mask[16] = mask [17] = 0xff; |
| *maskp = mask; |
| return 1; |
| } |
| if (cos2) |
| mask[18] |= 0xe0; |
| if (cos1) |
| mask[14] |= 0xe0; |
| if (proto) |
| mask[12] = mask [13] = 0xff; |
| |
| *maskp = mask; |
| return 1; |
| } |
| |
| uword unformat_classify_mask (unformat_input_t * input, va_list * args) |
| { |
| u8 ** maskp = va_arg (*args, u8 **); |
| u32 * skipp = va_arg (*args, u32 *); |
| u32 * matchp = va_arg (*args, u32 *); |
| u32 match; |
| u8 * mask = 0; |
| u8 * l2 = 0; |
| u8 * l3 = 0; |
| u8 * l4 = 0; |
| int i; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { |
| if (unformat (input, "hex %U", unformat_hex_string, &mask)) |
| ; |
| else if (unformat (input, "l2 %U", unformat_l2_mask, &l2)) |
| ; |
| else if (unformat (input, "l3 %U", unformat_l3_mask, &l3)) |
| ; |
| else if (unformat (input, "l4 %U", unformat_l4_mask, &l4)) |
| ; |
| else |
| break; |
| } |
| |
| if (l4 && !l3) { |
| vec_free (mask); |
| vec_free (l2); |
| vec_free (l4); |
| return 0; |
| } |
| |
| if (mask || l2 || l3 || l4) |
| { |
| if (l2 || l3 || l4) |
| { |
| /* "With a free Ethernet header in every package" */ |
| if (l2 == 0) |
| vec_validate (l2, 13); |
| mask = l2; |
| if (l3) |
| { |
| vec_append (mask, l3); |
| vec_free (l3); |
| } |
| if (l4) |
| { |
| vec_append (mask, l4); |
| vec_free (l4); |
| } |
| } |
| |
| /* Scan forward looking for the first significant mask octet */ |
| for (i = 0; i < vec_len (mask); i++) |
| if (mask[i]) |
| break; |
| |
| /* compute (skip, match) params */ |
| *skipp = i / sizeof(u32x4); |
| vec_delete (mask, *skipp * sizeof(u32x4), 0); |
| |
| /* Pad mask to an even multiple of the vector size */ |
| while (vec_len (mask) % sizeof (u32x4)) |
| vec_add1 (mask, 0); |
| |
| match = vec_len (mask) / sizeof (u32x4); |
| |
| for (i = match*sizeof(u32x4); i > 0; i-= sizeof(u32x4)) |
| { |
| u64 *tmp = (u64 *)(mask + (i-sizeof(u32x4))); |
| if (*tmp || *(tmp+1)) |
| break; |
| match--; |
| } |
| if (match == 0) |
| clib_warning ("BUG: match 0"); |
| |
| _vec_len (mask) = match * sizeof(u32x4); |
| |
| *matchp = match; |
| *maskp = mask; |
| |
| return 1; |
| } |
| |
| return 0; |
| } |
| |
| #define foreach_l2_input_next \ |
| _(drop, DROP) \ |
| _(ethernet, ETHERNET_INPUT) \ |
| _(ip4, IP4_INPUT) \ |
| _(ip6, IP6_INPUT) \ |
| _(li, LI) |
| |
| uword unformat_l2_input_next_index (unformat_input_t * input, va_list * args) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| u32 * miss_next_indexp = va_arg (*args, u32 *); |
| u32 next_index = 0; |
| u32 tmp; |
| int i; |
| |
| /* First try registered unformat fns, allowing override... */ |
| for (i = 0; i < vec_len (cm->unformat_l2_next_index_fns); i++) |
| { |
| if (unformat (input, "%U", cm->unformat_l2_next_index_fns[i], &tmp)) |
| { |
| next_index = tmp; |
| goto out; |
| } |
| } |
| |
| #define _(n,N) \ |
| if (unformat (input, #n)) { next_index = L2_INPUT_CLASSIFY_NEXT_##N; goto out;} |
| foreach_l2_input_next; |
| #undef _ |
| |
| if (unformat (input, "%d", &tmp)) |
| { |
| next_index = tmp; |
| goto out; |
| } |
| |
| return 0; |
| |
| out: |
| *miss_next_indexp = next_index; |
| return 1; |
| } |
| |
| #define foreach_l2_output_next \ |
| _(drop, DROP) |
| |
| uword unformat_l2_output_next_index (unformat_input_t * input, va_list * args) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| u32 * miss_next_indexp = va_arg (*args, u32 *); |
| u32 next_index = 0; |
| u32 tmp; |
| int i; |
| |
| /* First try registered unformat fns, allowing override... */ |
| for (i = 0; i < vec_len (cm->unformat_l2_next_index_fns); i++) |
| { |
| if (unformat (input, "%U", cm->unformat_l2_next_index_fns[i], &tmp)) |
| { |
| next_index = tmp; |
| goto out; |
| } |
| } |
| |
| #define _(n,N) \ |
| if (unformat (input, #n)) { next_index = L2_OUTPUT_CLASSIFY_NEXT_##N; goto out;} |
| foreach_l2_output_next; |
| #undef _ |
| |
| if (unformat (input, "%d", &tmp)) |
| { |
| next_index = tmp; |
| goto out; |
| } |
| |
| return 0; |
| |
| out: |
| *miss_next_indexp = next_index; |
| return 1; |
| } |
| |
| #define foreach_ip_next \ |
| _(drop, DROP) \ |
| _(rewrite, REWRITE) |
| |
| uword unformat_ip_next_index (unformat_input_t * input, va_list * args) |
| { |
| u32 * miss_next_indexp = va_arg (*args, u32 *); |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| u32 next_index = 0; |
| u32 tmp; |
| int i; |
| |
| /* First try registered unformat fns, allowing override... */ |
| for (i = 0; i < vec_len (cm->unformat_ip_next_index_fns); i++) |
| { |
| if (unformat (input, "%U", cm->unformat_ip_next_index_fns[i], &tmp)) |
| { |
| next_index = tmp; |
| goto out; |
| } |
| } |
| |
| #define _(n,N) \ |
| if (unformat (input, #n)) { next_index = IP_LOOKUP_NEXT_##N; goto out;} |
| foreach_ip_next; |
| #undef _ |
| |
| if (unformat (input, "%d", &tmp)) |
| { |
| next_index = tmp; |
| goto out; |
| } |
| |
| return 0; |
| |
| out: |
| *miss_next_indexp = next_index; |
| return 1; |
| } |
| |
| #define foreach_acl_next \ |
| _(deny, DENY) |
| |
| uword unformat_acl_next_index (unformat_input_t * input, va_list * args) |
| { |
| u32 * next_indexp = va_arg (*args, u32 *); |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| u32 next_index = 0; |
| u32 tmp; |
| int i; |
| |
| /* First try registered unformat fns, allowing override... */ |
| for (i = 0; i < vec_len (cm->unformat_acl_next_index_fns); i++) |
| { |
| if (unformat (input, "%U", cm->unformat_acl_next_index_fns[i], &tmp)) |
| { |
| next_index = tmp; |
| goto out; |
| } |
| } |
| |
| #define _(n,N) \ |
| if (unformat (input, #n)) { next_index = ACL_NEXT_INDEX_##N; goto out;} |
| foreach_acl_next; |
| #undef _ |
| |
| if (unformat (input, "permit")) |
| { |
| next_index = ~0; |
| goto out; |
| } |
| else if (unformat (input, "%d", &tmp)) |
| { |
| next_index = tmp; |
| goto out; |
| } |
| |
| return 0; |
| |
| out: |
| *next_indexp = next_index; |
| return 1; |
| } |
| |
| uword unformat_policer_next_index (unformat_input_t * input, va_list * args) |
| { |
| u32 * next_indexp = va_arg (*args, u32 *); |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| u32 next_index = 0; |
| u32 tmp; |
| int i; |
| |
| /* First try registered unformat fns, allowing override... */ |
| for (i = 0; i < vec_len (cm->unformat_policer_next_index_fns); i++) |
| { |
| if (unformat (input, "%U", cm->unformat_policer_next_index_fns[i], &tmp)) |
| { |
| next_index = tmp; |
| goto out; |
| } |
| } |
| |
| if (unformat (input, "%d", &tmp)) |
| { |
| next_index = tmp; |
| goto out; |
| } |
| |
| return 0; |
| |
| out: |
| *next_indexp = next_index; |
| return 1; |
| } |
| |
| static clib_error_t * |
| classify_table_command_fn (vlib_main_t * vm, |
| unformat_input_t * input, |
| vlib_cli_command_t * cmd) |
| { |
| u32 nbuckets = 2; |
| u32 skip = ~0; |
| u32 match = ~0; |
| int is_add = 1; |
| int del_chain = 0; |
| u32 table_index = ~0; |
| u32 next_table_index = ~0; |
| u32 miss_next_index = ~0; |
| u32 memory_size = 2<<20; |
| u32 tmp; |
| u32 current_data_flag = 0; |
| int current_data_offset = 0; |
| |
| u8 * mask = 0; |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| int rv; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { |
| if (unformat (input, "del")) |
| is_add = 0; |
| else if (unformat (input, "del-chain")) |
| { |
| is_add = 0; |
| del_chain = 1; |
| } |
| else if (unformat (input, "buckets %d", &nbuckets)) |
| ; |
| else if (unformat (input, "skip %d", &skip)) |
| ; |
| else if (unformat (input, "match %d", &match)) |
| ; |
| else if (unformat (input, "table %d", &table_index)) |
| ; |
| else if (unformat (input, "mask %U", unformat_classify_mask, |
| &mask, &skip, &match)) |
| ; |
| else if (unformat (input, "memory-size %uM", &tmp)) |
| memory_size = tmp<<20; |
| else if (unformat (input, "memory-size %uG", &tmp)) |
| memory_size = tmp<<30; |
| else if (unformat (input, "next-table %d", &next_table_index)) |
| ; |
| else if (unformat (input, "miss-next %U", unformat_ip_next_index, |
| &miss_next_index)) |
| ; |
| else if (unformat (input, "l2-input-miss-next %U", unformat_l2_input_next_index, |
| &miss_next_index)) |
| ; |
| else if (unformat (input, "l2-output-miss-next %U", unformat_l2_output_next_index, |
| &miss_next_index)) |
| ; |
| else if (unformat (input, "acl-miss-next %U", unformat_acl_next_index, |
| &miss_next_index)) |
| ; |
| else if (unformat (input, "current-data-flag %d", ¤t_data_flag)) |
| ; |
| else if (unformat (input, "current-data-offset %d", ¤t_data_offset)) |
| ; |
| |
| else |
| break; |
| } |
| |
| if (is_add && mask == 0 && table_index == ~0) |
| return clib_error_return (0, "Mask required"); |
| |
| if (is_add && skip == ~0 && table_index == ~0) |
| return clib_error_return (0, "skip count required"); |
| |
| if (is_add && match == ~0 && table_index == ~0) |
| return clib_error_return (0, "match count required"); |
| |
| if (!is_add && table_index == ~0) |
| return clib_error_return (0, "table index required for delete"); |
| |
| rv = vnet_classify_add_del_table (cm, mask, nbuckets, memory_size, |
| skip, match, next_table_index, miss_next_index, &table_index, |
| current_data_flag, current_data_offset, is_add, del_chain); |
| switch (rv) |
| { |
| case 0: |
| break; |
| |
| default: |
| return clib_error_return (0, "vnet_classify_add_del_table returned %d", |
| rv); |
| } |
| return 0; |
| } |
| |
| VLIB_CLI_COMMAND (classify_table, static) = { |
| .path = "classify table", |
| .short_help = |
| "classify table [miss-next|l2-miss_next|acl-miss-next <next_index>]" |
| "\n mask <mask-value> buckets <nn> [skip <n>] [match <n>]" |
| "\n [current-data-flag <n>] [current-data-offset <n>] [table <n>]" |
| "\n [del] [del-chain]", |
| .function = classify_table_command_fn, |
| }; |
| |
| static u8 * format_vnet_classify_table (u8 * s, va_list * args) |
| { |
| vnet_classify_main_t * cm = va_arg (*args, vnet_classify_main_t *); |
| int verbose = va_arg (*args, int); |
| u32 index = va_arg (*args, u32); |
| vnet_classify_table_t * t; |
| |
| if (index == ~0) |
| { |
| s = format (s, "%10s%10s%10s%10s", "TableIdx", "Sessions", "NextTbl", |
| "NextNode", verbose ? "Details" : ""); |
| return s; |
| } |
| |
| t = pool_elt_at_index (cm->tables, index); |
| s = format (s, "%10u%10d%10d%10d", index, t->active_elements, |
| t->next_table_index, t->miss_next_index); |
| |
| s = format (s, "\n Heap: %U", format_mheap, t->mheap, 0 /*verbose*/); |
| |
| s = format (s, "\n nbuckets %d, skip %d match %d flag %d offset %d", |
| t->nbuckets, t->skip_n_vectors, t->match_n_vectors, |
| t->current_data_flag, t->current_data_offset); |
| s = format (s, "\n mask %U", format_hex_bytes, t->mask, |
| t->match_n_vectors * sizeof (u32x4)); |
| |
| if (verbose == 0) |
| return s; |
| |
| s = format (s, "\n%U", format_classify_table, t, verbose); |
| |
| return s; |
| } |
| |
| static clib_error_t * |
| show_classify_tables_command_fn (vlib_main_t * vm, |
| unformat_input_t * input, |
| vlib_cli_command_t * cmd) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| vnet_classify_table_t * t; |
| u32 match_index = ~0; |
| u32 * indices = 0; |
| int verbose = 0; |
| int i; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (unformat (input, "index %d", &match_index)) |
| ; |
| else if (unformat (input, "verbose %d", &verbose)) |
| ; |
| else if (unformat (input, "verbose")) |
| verbose = 1; |
| else |
| break; |
| } |
| |
| pool_foreach (t, cm->tables, |
| ({ |
| if (match_index == ~0 || (match_index == t - cm->tables)) |
| vec_add1 (indices, t - cm->tables); |
| })); |
| |
| if (vec_len(indices)) |
| { |
| vlib_cli_output (vm, "%U", format_vnet_classify_table, cm, verbose, |
| ~0 /* hdr */); |
| for (i = 0; i < vec_len (indices); i++) |
| vlib_cli_output (vm, "%U", format_vnet_classify_table, cm, |
| verbose, indices[i]); |
| } |
| else |
| vlib_cli_output (vm, "No classifier tables configured"); |
| |
| vec_free (indices); |
| |
| return 0; |
| } |
| |
| VLIB_CLI_COMMAND (show_classify_table_command, static) = { |
| .path = "show classify tables", |
| .short_help = "show classify tables [index <nn>]", |
| .function = show_classify_tables_command_fn, |
| }; |
| |
| uword unformat_l4_match (unformat_input_t * input, va_list * args) |
| { |
| u8 ** matchp = va_arg (*args, u8 **); |
| |
| u8 * proto_header = 0; |
| int src_port = 0; |
| int dst_port = 0; |
| |
| tcpudp_header_t h; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (unformat (input, "src_port %d", &src_port)) |
| ; |
| else if (unformat (input, "dst_port %d", &dst_port)) |
| ; |
| else |
| return 0; |
| } |
| |
| h.src_port = clib_host_to_net_u16(src_port); |
| h.dst_port = clib_host_to_net_u16(dst_port); |
| vec_validate(proto_header, sizeof(h)-1); |
| memcpy(proto_header, &h, sizeof(h)); |
| |
| *matchp = proto_header; |
| |
| return 1; |
| } |
| |
| uword unformat_ip4_match (unformat_input_t * input, va_list * args) |
| { |
| u8 ** matchp = va_arg (*args, u8 **); |
| u8 * match = 0; |
| ip4_header_t * ip; |
| int version = 0; |
| u32 version_val; |
| int hdr_length = 0; |
| u32 hdr_length_val; |
| int src = 0, dst = 0; |
| ip4_address_t src_val, dst_val; |
| int proto = 0; |
| u32 proto_val; |
| int tos = 0; |
| u32 tos_val; |
| int length = 0; |
| u32 length_val; |
| int fragment_id = 0; |
| u32 fragment_id_val; |
| int ttl = 0; |
| int ttl_val; |
| int checksum = 0; |
| u32 checksum_val; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (unformat (input, "version %d", &version_val)) |
| version = 1; |
| else if (unformat (input, "hdr_length %d", &hdr_length_val)) |
| hdr_length = 1; |
| else if (unformat (input, "src %U", unformat_ip4_address, &src_val)) |
| src = 1; |
| else if (unformat (input, "dst %U", unformat_ip4_address, &dst_val)) |
| dst = 1; |
| else if (unformat (input, "proto %d", &proto_val)) |
| proto = 1; |
| else if (unformat (input, "tos %d", &tos_val)) |
| tos = 1; |
| else if (unformat (input, "length %d", &length_val)) |
| length = 1; |
| else if (unformat (input, "fragment_id %d", &fragment_id_val)) |
| fragment_id = 1; |
| else if (unformat (input, "ttl %d", &ttl_val)) |
| ttl = 1; |
| else if (unformat (input, "checksum %d", &checksum_val)) |
| checksum = 1; |
| else |
| break; |
| } |
| |
| if (version + hdr_length + src + dst + proto + tos + length + fragment_id |
| + ttl + checksum == 0) |
| return 0; |
| |
| /* |
| * Aligned because we use the real comparison functions |
| */ |
| vec_validate_aligned (match, sizeof (*ip) - 1, sizeof(u32x4)); |
| |
| ip = (ip4_header_t *) match; |
| |
| /* These are realistically matched in practice */ |
| if (src) |
| ip->src_address.as_u32 = src_val.as_u32; |
| |
| if (dst) |
| ip->dst_address.as_u32 = dst_val.as_u32; |
| |
| if (proto) |
| ip->protocol = proto_val; |
| |
| |
| /* These are not, but they're included for completeness */ |
| if (version) |
| ip->ip_version_and_header_length |= (version_val & 0xF)<<4; |
| |
| if (hdr_length) |
| ip->ip_version_and_header_length |= (hdr_length_val & 0xF); |
| |
| if (tos) |
| ip->tos = tos_val; |
| |
| if (length) |
| ip->length = clib_host_to_net_u16 (length_val); |
| |
| if (ttl) |
| ip->ttl = ttl_val; |
| |
| if (checksum) |
| ip->checksum = clib_host_to_net_u16 (checksum_val); |
| |
| *matchp = match; |
| return 1; |
| } |
| |
| uword unformat_ip6_match (unformat_input_t * input, va_list * args) |
| { |
| u8 ** matchp = va_arg (*args, u8 **); |
| u8 * match = 0; |
| ip6_header_t * ip; |
| int version = 0; |
| u32 version_val; |
| u8 traffic_class = 0; |
| u32 traffic_class_val; |
| u8 flow_label = 0; |
| u8 flow_label_val; |
| int src = 0, dst = 0; |
| ip6_address_t src_val, dst_val; |
| int proto = 0; |
| u32 proto_val; |
| int payload_length = 0; |
| u32 payload_length_val; |
| int hop_limit = 0; |
| int hop_limit_val; |
| u32 ip_version_traffic_class_and_flow_label; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (unformat (input, "version %d", &version_val)) |
| version = 1; |
| else if (unformat (input, "traffic_class %d", &traffic_class_val)) |
| traffic_class = 1; |
| else if (unformat (input, "flow_label %d", &flow_label_val)) |
| flow_label = 1; |
| else if (unformat (input, "src %U", unformat_ip6_address, &src_val)) |
| src = 1; |
| else if (unformat (input, "dst %U", unformat_ip6_address, &dst_val)) |
| dst = 1; |
| else if (unformat (input, "proto %d", &proto_val)) |
| proto = 1; |
| else if (unformat (input, "payload_length %d", &payload_length_val)) |
| payload_length = 1; |
| else if (unformat (input, "hop_limit %d", &hop_limit_val)) |
| hop_limit = 1; |
| else |
| break; |
| } |
| |
| if (version + traffic_class + flow_label + src + dst + proto + |
| payload_length + hop_limit == 0) |
| return 0; |
| |
| /* |
| * Aligned because we use the real comparison functions |
| */ |
| vec_validate_aligned (match, sizeof (*ip) - 1, sizeof(u32x4)); |
| |
| ip = (ip6_header_t *) match; |
| |
| if (src) |
| clib_memcpy (&ip->src_address, &src_val, sizeof (ip->src_address)); |
| |
| if (dst) |
| clib_memcpy (&ip->dst_address, &dst_val, sizeof (ip->dst_address)); |
| |
| if (proto) |
| ip->protocol = proto_val; |
| |
| ip_version_traffic_class_and_flow_label = 0; |
| |
| if (version) |
| ip_version_traffic_class_and_flow_label |= (version_val & 0xF) << 28; |
| |
| if (traffic_class) |
| ip_version_traffic_class_and_flow_label |= (traffic_class_val & 0xFF) << 20; |
| |
| if (flow_label) |
| ip_version_traffic_class_and_flow_label |= (flow_label_val & 0xFFFFF); |
| |
| ip->ip_version_traffic_class_and_flow_label = |
| clib_host_to_net_u32 (ip_version_traffic_class_and_flow_label); |
| |
| if (payload_length) |
| ip->payload_length = clib_host_to_net_u16 (payload_length_val); |
| |
| if (hop_limit) |
| ip->hop_limit = hop_limit_val; |
| |
| *matchp = match; |
| return 1; |
| } |
| |
| uword unformat_l3_match (unformat_input_t * input, va_list * args) |
| { |
| u8 ** matchp = va_arg (*args, u8 **); |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { |
| if (unformat (input, "ip4 %U", unformat_ip4_match, matchp)) |
| return 1; |
| else if (unformat (input, "ip6 %U", unformat_ip6_match, matchp)) |
| return 1; |
| /* $$$$ add mpls */ |
| else |
| break; |
| } |
| return 0; |
| } |
| |
| uword unformat_vlan_tag (unformat_input_t * input, va_list * args) |
| { |
| u8 * tagp = va_arg (*args, u8 *); |
| u32 tag; |
| |
| if (unformat(input, "%d", &tag)) |
| { |
| tagp[0] = (tag>>8) & 0x0F; |
| tagp[1] = tag & 0xFF; |
| return 1; |
| } |
| |
| return 0; |
| } |
| |
| uword unformat_l2_match (unformat_input_t * input, va_list * args) |
| { |
| u8 ** matchp = va_arg (*args, u8 **); |
| u8 * match = 0; |
| u8 src = 0; |
| u8 src_val[6]; |
| u8 dst = 0; |
| u8 dst_val[6]; |
| u8 proto = 0; |
| u16 proto_val; |
| u8 tag1 = 0; |
| u8 tag1_val [2]; |
| u8 tag2 = 0; |
| u8 tag2_val [2]; |
| int len = 14; |
| u8 ignore_tag1 = 0; |
| u8 ignore_tag2 = 0; |
| u8 cos1 = 0; |
| u8 cos2 = 0; |
| u32 cos1_val = 0; |
| u32 cos2_val = 0; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { |
| if (unformat (input, "src %U", unformat_ethernet_address, &src_val)) |
| src = 1; |
| else if (unformat (input, "dst %U", unformat_ethernet_address, &dst_val)) |
| dst = 1; |
| else if (unformat (input, "proto %U", |
| unformat_ethernet_type_host_byte_order, &proto_val)) |
| proto = 1; |
| else if (unformat (input, "tag1 %U", unformat_vlan_tag, tag1_val)) |
| tag1 = 1; |
| else if (unformat (input, "tag2 %U", unformat_vlan_tag, tag2_val)) |
| tag2 = 1; |
| else if (unformat (input, "ignore-tag1")) |
| ignore_tag1 = 1; |
| else if (unformat (input, "ignore-tag2")) |
| ignore_tag2 = 1; |
| else if (unformat (input, "cos1 %d", &cos1_val)) |
| cos1 = 1; |
| else if (unformat (input, "cos2 %d", &cos2_val)) |
| cos2 = 1; |
| else |
| break; |
| } |
| if ((src + dst + proto + tag1 + tag2 + |
| ignore_tag1 + ignore_tag2 + cos1 + cos2) == 0) |
| return 0; |
| |
| if (tag1 || ignore_tag1 || cos1) |
| len = 18; |
| if (tag2 || ignore_tag2 || cos2) |
| len = 22; |
| |
| vec_validate_aligned (match, len-1, sizeof(u32x4)); |
| |
| if (dst) |
| clib_memcpy (match, dst_val, 6); |
| |
| if (src) |
| clib_memcpy (match + 6, src_val, 6); |
| |
| if (tag2) |
| { |
| /* inner vlan tag */ |
| match[19] = tag2_val[1]; |
| match[18] = tag2_val[0]; |
| if (cos2) |
| match [18] |= (cos2_val & 0x7) << 5; |
| if (proto) |
| { |
| match[21] = proto_val & 0xff; |
| match[20] = proto_val >> 8; |
| } |
| if (tag1) |
| { |
| match [15] = tag1_val[1]; |
| match [14] = tag1_val[0]; |
| } |
| if (cos1) |
| match [14] |= (cos1_val & 0x7) << 5; |
| *matchp = match; |
| return 1; |
| } |
| if (tag1) |
| { |
| match [15] = tag1_val[1]; |
| match [14] = tag1_val[0]; |
| if (proto) |
| { |
| match[17] = proto_val & 0xff; |
| match[16] = proto_val >> 8; |
| } |
| if (cos1) |
| match [14] |= (cos1_val & 0x7) << 5; |
| |
| *matchp = match; |
| return 1; |
| } |
| if (cos2) |
| match [18] |= (cos2_val & 0x7) << 5; |
| if (cos1) |
| match [14] |= (cos1_val & 0x7) << 5; |
| if (proto) |
| { |
| match[13] = proto_val & 0xff; |
| match[12] = proto_val >> 8; |
| } |
| |
| *matchp = match; |
| return 1; |
| } |
| |
| |
| uword unformat_classify_match (unformat_input_t * input, va_list * args) |
| { |
| vnet_classify_main_t * cm = va_arg (*args, vnet_classify_main_t *); |
| u8 ** matchp = va_arg (*args, u8 **); |
| u32 table_index = va_arg (*args, u32); |
| vnet_classify_table_t * t; |
| |
| u8 * match = 0; |
| u8 * l2 = 0; |
| u8 * l3 = 0; |
| u8 * l4 = 0; |
| |
| if (pool_is_free_index (cm->tables, table_index)) |
| return 0; |
| |
| t = pool_elt_at_index (cm->tables, table_index); |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { |
| if (unformat (input, "hex %U", unformat_hex_string, &match)) |
| ; |
| else if (unformat (input, "l2 %U", unformat_l2_match, &l2)) |
| ; |
| else if (unformat (input, "l3 %U", unformat_l3_match, &l3)) |
| ; |
| else if (unformat (input, "l4 %U", unformat_l4_match, &l4)) |
| ; |
| else |
| break; |
| } |
| |
| if (l4 && !l3) { |
| vec_free (match); |
| vec_free (l2); |
| vec_free (l4); |
| return 0; |
| } |
| |
| if (match || l2 || l3 || l4) |
| { |
| if (l2 || l3 || l4) |
| { |
| /* "Win a free Ethernet header in every packet" */ |
| if (l2 == 0) |
| vec_validate_aligned (l2, 13, sizeof(u32x4)); |
| match = l2; |
| if (l3) |
| { |
| vec_append_aligned (match, l3, sizeof(u32x4)); |
| vec_free (l3); |
| } |
| if (l4) |
| { |
| vec_append_aligned (match, l4, sizeof(u32x4)); |
| vec_free (l4); |
| } |
| } |
| |
| /* Make sure the vector is big enough even if key is all 0's */ |
| vec_validate_aligned |
| (match, ((t->match_n_vectors + t->skip_n_vectors) * sizeof(u32x4)) - 1, |
| sizeof(u32x4)); |
| |
| /* Set size, include skipped vectors*/ |
| _vec_len (match) = (t->match_n_vectors+t->skip_n_vectors) * sizeof(u32x4); |
| |
| *matchp = match; |
| |
| return 1; |
| } |
| |
| return 0; |
| } |
| |
| int vnet_classify_add_del_session (vnet_classify_main_t * cm, |
| u32 table_index, |
| u8 * match, |
| u32 hit_next_index, |
| u32 opaque_index, |
| i32 advance, |
| u8 action, |
| u32 metadata, |
| int is_add) |
| { |
| vnet_classify_table_t * t; |
| vnet_classify_entry_5_t _max_e __attribute__((aligned (16))); |
| vnet_classify_entry_t * e; |
| int i, rv; |
| |
| if (pool_is_free_index (cm->tables, table_index)) |
| return VNET_API_ERROR_NO_SUCH_TABLE; |
| |
| t = pool_elt_at_index (cm->tables, table_index); |
| |
| e = (vnet_classify_entry_t *)&_max_e; |
| e->next_index = hit_next_index; |
| e->opaque_index = opaque_index; |
| e->advance = advance; |
| e->hits = 0; |
| e->last_heard = 0; |
| e->flags = 0; |
| e->action = action; |
| if (e->action == CLASSIFY_ACTION_SET_IP4_FIB_INDEX) |
| e->metadata = fib_table_find_or_create_and_lock (FIB_PROTOCOL_IP4, metadata); |
| else if (e->action == CLASSIFY_ACTION_SET_IP6_FIB_INDEX) |
| e->metadata = fib_table_find_or_create_and_lock (FIB_PROTOCOL_IP6, metadata); |
| |
| /* Copy key data, honoring skip_n_vectors */ |
| clib_memcpy (&e->key, match + t->skip_n_vectors * sizeof (u32x4), |
| t->match_n_vectors * sizeof (u32x4)); |
| |
| /* Clear don't-care bits; likely when dynamically creating sessions */ |
| for (i = 0; i < t->match_n_vectors; i++) |
| e->key[i] &= t->mask[i]; |
| |
| rv = vnet_classify_add_del (t, e, is_add); |
| if (rv) |
| return VNET_API_ERROR_NO_SUCH_ENTRY; |
| return 0; |
| } |
| |
| static clib_error_t * |
| classify_session_command_fn (vlib_main_t * vm, |
| unformat_input_t * input, |
| vlib_cli_command_t * cmd) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| int is_add = 1; |
| u32 table_index = ~0; |
| u32 hit_next_index = ~0; |
| u64 opaque_index = ~0; |
| u8 * match = 0; |
| i32 advance = 0; |
| u32 action = 0; |
| u32 metadata = 0; |
| int i, rv; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (unformat (input, "del")) |
| is_add = 0; |
| else if (unformat (input, "hit-next %U", unformat_ip_next_index, |
| &hit_next_index)) |
| ; |
| else if (unformat (input, "l2-input-hit-next %U", unformat_l2_input_next_index, |
| &hit_next_index)) |
| ; |
| else if (unformat (input, "l2-output-hit-next %U", unformat_l2_output_next_index, |
| &hit_next_index)) |
| ; |
| else if (unformat (input, "acl-hit-next %U", unformat_acl_next_index, |
| &hit_next_index)) |
| ; |
| else if (unformat (input, "policer-hit-next %U", |
| unformat_policer_next_index, &hit_next_index)) |
| ; |
| else if (unformat (input, "opaque-index %lld", &opaque_index)) |
| ; |
| else if (unformat (input, "match %U", unformat_classify_match, |
| cm, &match, table_index)) |
| ; |
| else if (unformat (input, "advance %d", &advance)) |
| ; |
| else if (unformat (input, "table-index %d", &table_index)) |
| ; |
| else if (unformat (input, "action set-ip4-fib-id %d", &metadata)) |
| action = 1; |
| else if (unformat (input, "action set-ip6-fib-id %d", &metadata)) |
| action = 2; |
| else |
| { |
| /* Try registered opaque-index unformat fns */ |
| for (i = 0; i < vec_len (cm->unformat_opaque_index_fns); i++) |
| { |
| if (unformat (input, "%U", cm->unformat_opaque_index_fns[i], |
| &opaque_index)) |
| goto found_opaque; |
| } |
| break; |
| } |
| found_opaque: |
| ; |
| } |
| |
| if (table_index == ~0) |
| return clib_error_return (0, "Table index required"); |
| |
| if (is_add && match == 0) |
| return clib_error_return (0, "Match value required"); |
| |
| rv = vnet_classify_add_del_session (cm, table_index, match, |
| hit_next_index, |
| opaque_index, advance, |
| action, metadata, is_add); |
| |
| switch(rv) |
| { |
| case 0: |
| break; |
| |
| default: |
| return clib_error_return (0, "vnet_classify_add_del_session returned %d", |
| rv); |
| } |
| |
| return 0; |
| } |
| |
| VLIB_CLI_COMMAND (classify_session_command, static) = { |
| .path = "classify session", |
| .short_help = |
| "classify session [hit-next|l2-hit-next|" |
| "acl-hit-next <next_index>|policer-hit-next <policer_name>]" |
| "\n table-index <nn> match [hex] [l2] [l3 ip4] [opaque-index <index>]" |
| "\n [action set-ip4-fib-id <n>] [action set-ip6-fib-id <n>] [del]", |
| .function = classify_session_command_fn, |
| }; |
| |
| static uword |
| unformat_opaque_sw_if_index (unformat_input_t * input, va_list * args) |
| { |
| u64 * opaquep = va_arg (*args, u64 *); |
| u32 sw_if_index; |
| |
| if (unformat (input, "opaque-sw_if_index %U", unformat_vnet_sw_interface, |
| vnet_get_main(), &sw_if_index)) |
| { |
| *opaquep = sw_if_index; |
| return 1; |
| } |
| return 0; |
| } |
| |
| static uword |
| unformat_ip_next_node (unformat_input_t * input, va_list * args) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| u32 * next_indexp = va_arg (*args, u32 *); |
| u32 node_index; |
| u32 next_index = ~0; |
| |
| if (unformat (input, "ip6-node %U", unformat_vlib_node, |
| cm->vlib_main, &node_index)) |
| { |
| next_index = vlib_node_add_next (cm->vlib_main, |
| ip6_classify_node.index, node_index); |
| } |
| else if (unformat (input, "ip4-node %U", unformat_vlib_node, |
| cm->vlib_main, &node_index)) |
| { |
| next_index = vlib_node_add_next (cm->vlib_main, |
| ip4_classify_node.index, node_index); |
| } |
| else |
| return 0; |
| |
| *next_indexp = next_index; |
| return 1; |
| } |
| |
| static uword |
| unformat_acl_next_node (unformat_input_t * input, va_list * args) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| u32 * next_indexp = va_arg (*args, u32 *); |
| u32 node_index; |
| u32 next_index; |
| |
| if (unformat (input, "ip6-node %U", unformat_vlib_node, |
| cm->vlib_main, &node_index)) |
| { |
| next_index = vlib_node_add_next (cm->vlib_main, |
| ip6_inacl_node.index, node_index); |
| } |
| else if (unformat (input, "ip4-node %U", unformat_vlib_node, |
| cm->vlib_main, &node_index)) |
| { |
| next_index = vlib_node_add_next (cm->vlib_main, |
| ip4_inacl_node.index, node_index); |
| } |
| else |
| return 0; |
| |
| *next_indexp = next_index; |
| return 1; |
| } |
| |
| static uword |
| unformat_l2_input_next_node (unformat_input_t * input, va_list * args) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| u32 * next_indexp = va_arg (*args, u32 *); |
| u32 node_index; |
| u32 next_index; |
| |
| if (unformat (input, "input-node %U", unformat_vlib_node, |
| cm->vlib_main, &node_index)) |
| { |
| next_index = vlib_node_add_next |
| (cm->vlib_main, l2_input_classify_node.index, node_index); |
| |
| *next_indexp = next_index; |
| return 1; |
| } |
| return 0; |
| } |
| |
| static uword |
| unformat_l2_output_next_node (unformat_input_t * input, va_list * args) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| u32 * next_indexp = va_arg (*args, u32 *); |
| u32 node_index; |
| u32 next_index; |
| |
| if (unformat (input, "output-node %U", unformat_vlib_node, |
| cm->vlib_main, &node_index)) |
| { |
| next_index = vlib_node_add_next |
| (cm->vlib_main, l2_output_classify_node.index, node_index); |
| |
| *next_indexp = next_index; |
| return 1; |
| } |
| return 0; |
| } |
| |
| static clib_error_t * |
| vnet_classify_init (vlib_main_t * vm) |
| { |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| |
| cm->vlib_main = vm; |
| cm->vnet_main = vnet_get_main(); |
| |
| vnet_classify_register_unformat_opaque_index_fn |
| (unformat_opaque_sw_if_index); |
| |
| vnet_classify_register_unformat_ip_next_index_fn |
| (unformat_ip_next_node); |
| |
| vnet_classify_register_unformat_l2_next_index_fn |
| (unformat_l2_input_next_node); |
| |
| vnet_classify_register_unformat_l2_next_index_fn |
| (unformat_l2_output_next_node); |
| |
| vnet_classify_register_unformat_acl_next_index_fn |
| (unformat_acl_next_node); |
| |
| return 0; |
| } |
| |
| VLIB_INIT_FUNCTION (vnet_classify_init); |
| |
| #define TEST_CODE 1 |
| |
| #if TEST_CODE > 0 |
| static clib_error_t * |
| test_classify_command_fn (vlib_main_t * vm, |
| unformat_input_t * input, |
| vlib_cli_command_t * cmd) |
| { |
| u32 buckets = 2; |
| u32 sessions = 10; |
| int i, rv; |
| vnet_classify_table_t * t = 0; |
| classify_data_or_mask_t * mask; |
| classify_data_or_mask_t * data; |
| u8 *mp = 0, *dp = 0; |
| vnet_classify_main_t * cm = &vnet_classify_main; |
| vnet_classify_entry_t * e; |
| int is_add = 1; |
| u32 tmp; |
| u32 table_index = ~0; |
| ip4_address_t src; |
| u32 deleted = 0; |
| u32 memory_size = 64<<20; |
| |
| /* Default starting address 1.0.0.10 */ |
| src.as_u32 = clib_net_to_host_u32 (0x0100000A); |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { |
| if (unformat (input, "sessions %d", &sessions)) |
| ; |
| else if (unformat (input, "src %U", unformat_ip4_address, &src)) |
| ; |
| else if (unformat (input, "buckets %d", &buckets)) |
| ; |
| else if (unformat (input, "memory-size %uM", &tmp)) |
| memory_size = tmp<<20; |
| else if (unformat (input, "memory-size %uG", &tmp)) |
| memory_size = tmp<<30; |
| else if (unformat (input, "del")) |
| is_add = 0; |
| else if (unformat (input, "table %d", &table_index)) |
| ; |
| else |
| break; |
| } |
| |
| vec_validate_aligned (mp, 3 * sizeof(u32x4), sizeof(u32x4)); |
| vec_validate_aligned (dp, 3 * sizeof(u32x4), sizeof(u32x4)); |
| |
| mask = (classify_data_or_mask_t *) mp; |
| data = (classify_data_or_mask_t *) dp; |
| |
| data->ip.src_address.as_u32 = src.as_u32; |
| |
| /* Mask on src address */ |
| memset (&mask->ip.src_address, 0xff, 4); |
| |
| buckets = 1<<max_log2(buckets); |
| |
| if (table_index != ~0) |
| { |
| if (pool_is_free_index (cm->tables, table_index)) |
| { |
| vlib_cli_output (vm, "No such table %d", table_index); |
| goto out; |
| } |
| t = pool_elt_at_index (cm->tables, table_index); |
| } |
| |
| if (is_add) |
| { |
| if (t == 0) |
| { |
| t = vnet_classify_new_table (cm, (u8 *)mask, buckets, |
| memory_size, |
| 0 /* skip */, |
| 3 /* vectors to match */); |
| t->miss_next_index = IP_LOOKUP_NEXT_DROP; |
| vlib_cli_output (vm, "Create table %d", t - cm->tables); |
| } |
| |
| vlib_cli_output (vm, "Add %d sessions to %d buckets...", |
| sessions, buckets); |
| |
| for (i = 0; i < sessions; i++) |
| { |
| rv = vnet_classify_add_del_session (cm, t - cm->tables, (u8 *) data, |
| IP_LOOKUP_NEXT_DROP, |
| i+100 /* opaque_index */, |
| 0 /* advance */, 0, 0, |
| 1 /* is_add */); |
| |
| if (rv != 0) |
| clib_warning ("add: returned %d", rv); |
| |
| tmp = clib_net_to_host_u32 (data->ip.src_address.as_u32) + 1; |
| data->ip.src_address.as_u32 = clib_net_to_host_u32 (tmp); |
| } |
| goto out; |
| } |
| |
| if (t == 0) |
| { |
| vlib_cli_output (vm, "Must specify table index to delete sessions"); |
| goto out; |
| } |
| |
| vlib_cli_output (vm, "Try to delete %d sessions...", sessions); |
| |
| for (i = 0; i < sessions; i++) |
| { |
| u8 * key_minus_skip; |
| u64 hash; |
| |
| hash = vnet_classify_hash_packet (t, (u8 *) data); |
| |
| e = vnet_classify_find_entry (t, (u8 *) data, hash, 0 /* time_now */); |
| /* Previous delete, perhaps... */ |
| if (e == 0) |
| continue; |
| ASSERT (e->opaque_index == (i+100)); |
| |
| key_minus_skip = (u8 *)e->key; |
| key_minus_skip -= t->skip_n_vectors * sizeof (u32x4); |
| |
| rv = vnet_classify_add_del_session (cm, t - cm->tables, key_minus_skip, |
| IP_LOOKUP_NEXT_DROP, |
| i+100 /* opaque_index */, |
| 0 /* advance */, 0, 0, |
| 0 /* is_add */); |
| if (rv != 0) |
| clib_warning ("del: returned %d", rv); |
| |
| tmp = clib_net_to_host_u32 (data->ip.src_address.as_u32) + 1; |
| data->ip.src_address.as_u32 = clib_net_to_host_u32 (tmp); |
| deleted++; |
| } |
| |
| vlib_cli_output (vm, "Deleted %d sessions...", deleted); |
| |
| out: |
| vec_free (mp); |
| vec_free (dp); |
| |
| return 0; |
| } |
| |
| VLIB_CLI_COMMAND (test_classify_command, static) = { |
| .path = "test classify", |
| .short_help = |
| "test classify [src <ip>] [sessions <nn>] [buckets <nn>] [table <nn>] [del]", |
| .function = test_classify_command_fn, |
| }; |
| #endif /* TEST_CODE */ |