| /* |
| * Copyright (c) 2015 Cisco and/or its affiliates. |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at: |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| /* |
| * buffer.c: allocate/free network buffers. |
| * |
| * Copyright (c) 2008 Eliot Dresselhaus |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining |
| * a copy of this software and associated documentation files (the |
| * "Software"), to deal in the Software without restriction, including |
| * without limitation the rights to use, copy, modify, merge, publish, |
| * distribute, sublicense, and/or sell copies of the Software, and to |
| * permit persons to whom the Software is furnished to do so, subject to |
| * the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be |
| * included in all copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
| * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND |
| * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE |
| * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION |
| * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION |
| * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. |
| */ |
| |
| /** |
| * @file |
| * |
| * Allocate/free network buffers. |
| */ |
| |
| #include <vppinfra/linux/sysfs.h> |
| #include <vlib/vlib.h> |
| #include <vlib/unix/unix.h> |
| #include <vlib/stats/stats.h> |
| |
| #define VLIB_BUFFER_DEFAULT_BUFFERS_PER_NUMA 16384 |
| #define VLIB_BUFFER_DEFAULT_BUFFERS_PER_NUMA_UNPRIV 8192 |
| |
| #ifdef CLIB_HAVE_VEC128 |
| /* Assumptions by vlib_buffer_free_inline: */ |
| STATIC_ASSERT_FITS_IN (vlib_buffer_t, flags, 16); |
| STATIC_ASSERT_FITS_IN (vlib_buffer_t, ref_count, 16); |
| STATIC_ASSERT_FITS_IN (vlib_buffer_t, buffer_pool_index, 16); |
| #endif |
| |
| /* Make sure that buffer template size is not accidentally changed */ |
| STATIC_ASSERT_OFFSET_OF (vlib_buffer_t, template_end, 64); |
| |
| u16 __vlib_buffer_external_hdr_size = 0; |
| |
| uword |
| vlib_buffer_length_in_chain_slow_path (vlib_main_t * vm, |
| vlib_buffer_t * b_first) |
| { |
| vlib_buffer_t *b = b_first; |
| uword l_first = b_first->current_length; |
| uword l = 0; |
| while (b->flags & VLIB_BUFFER_NEXT_PRESENT) |
| { |
| b = vlib_get_buffer (vm, b->next_buffer); |
| l += b->current_length; |
| } |
| b_first->total_length_not_including_first_buffer = l; |
| b_first->flags |= VLIB_BUFFER_TOTAL_LENGTH_VALID; |
| return l + l_first; |
| } |
| |
| u8 * |
| format_vlib_buffer_no_chain (u8 * s, va_list * args) |
| { |
| vlib_buffer_t *b = va_arg (*args, vlib_buffer_t *); |
| u32 indent = format_get_indent (s); |
| u8 *a = 0; |
| |
| #define _(bit, name, v) \ |
| if (v && (b->flags & VLIB_BUFFER_##name)) \ |
| a = format (a, "%s ", v); |
| foreach_vlib_buffer_flag |
| #undef _ |
| s = format (s, "current data %d, length %d, buffer-pool %d, " |
| "ref-count %u", b->current_data, b->current_length, |
| b->buffer_pool_index, b->ref_count); |
| |
| if (b->flags & VLIB_BUFFER_NEXT_PRESENT) |
| s = format (s, ", totlen-nifb %d", |
| b->total_length_not_including_first_buffer); |
| |
| if (b->flags & VLIB_BUFFER_IS_TRACED) |
| s = format (s, ", trace handle 0x%x", b->trace_handle); |
| |
| if (a) |
| s = format (s, "\n%U%v", format_white_space, indent, a); |
| vec_free (a); |
| |
| return s; |
| } |
| |
| u8 * |
| format_vlib_buffer (u8 * s, va_list * args) |
| { |
| vlib_main_t *vm = vlib_get_main (); |
| vlib_buffer_t *b = va_arg (*args, vlib_buffer_t *); |
| u32 indent = format_get_indent (s); |
| |
| s = format (s, "%U", format_vlib_buffer_no_chain, b); |
| |
| while (b->flags & VLIB_BUFFER_NEXT_PRESENT) |
| { |
| u32 next_buffer = b->next_buffer; |
| b = vlib_get_buffer (vm, next_buffer); |
| |
| s = |
| format (s, "\n%Unext-buffer 0x%x, segment length %d, ref-count %u", |
| format_white_space, indent, next_buffer, b->current_length, |
| b->ref_count); |
| } |
| |
| return s; |
| } |
| |
| u8 * |
| format_vlib_buffer_and_data (u8 * s, va_list * args) |
| { |
| vlib_buffer_t *b = va_arg (*args, vlib_buffer_t *); |
| |
| s = format (s, "%U, %U", |
| format_vlib_buffer, b, |
| format_hex_bytes, vlib_buffer_get_current (b), 64); |
| |
| return s; |
| } |
| |
| static u8 * |
| format_vlib_buffer_known_state (u8 * s, va_list * args) |
| { |
| vlib_buffer_known_state_t state = va_arg (*args, vlib_buffer_known_state_t); |
| char *t; |
| |
| switch (state) |
| { |
| case VLIB_BUFFER_UNKNOWN: |
| t = "unknown"; |
| break; |
| |
| case VLIB_BUFFER_KNOWN_ALLOCATED: |
| t = "known-allocated"; |
| break; |
| |
| case VLIB_BUFFER_KNOWN_FREE: |
| t = "known-free"; |
| break; |
| |
| default: |
| t = "invalid"; |
| break; |
| } |
| |
| return format (s, "%s", t); |
| } |
| |
| u8 * |
| format_vlib_buffer_contents (u8 * s, va_list * va) |
| { |
| vlib_main_t *vm = va_arg (*va, vlib_main_t *); |
| vlib_buffer_t *b = va_arg (*va, vlib_buffer_t *); |
| |
| while (1) |
| { |
| vec_add (s, vlib_buffer_get_current (b), b->current_length); |
| if (!(b->flags & VLIB_BUFFER_NEXT_PRESENT)) |
| break; |
| b = vlib_get_buffer (vm, b->next_buffer); |
| } |
| |
| return s; |
| } |
| |
| static u8 * |
| vlib_validate_buffer_helper (vlib_main_t * vm, |
| u32 bi, |
| uword follow_buffer_next, uword ** unique_hash) |
| { |
| vlib_buffer_main_t *bm = vm->buffer_main; |
| vlib_buffer_t *b = vlib_get_buffer (vm, bi); |
| |
| if (vec_len (bm->buffer_pools) <= b->buffer_pool_index) |
| return format (0, "unknown buffer pool 0x%x", b->buffer_pool_index); |
| |
| if ((signed) b->current_data < (signed) -VLIB_BUFFER_PRE_DATA_SIZE) |
| return format (0, "current data %d before pre-data", b->current_data); |
| |
| if (b->current_data + b->current_length > |
| vlib_buffer_get_default_data_size (vm)) |
| return format (0, "%d-%d beyond end of buffer %d", b->current_data, |
| b->current_length, vlib_buffer_get_default_data_size (vm)); |
| |
| if (follow_buffer_next && (b->flags & VLIB_BUFFER_NEXT_PRESENT)) |
| { |
| vlib_buffer_known_state_t k; |
| u8 *msg, *result; |
| |
| k = vlib_buffer_is_known (vm, b->next_buffer); |
| if (k != VLIB_BUFFER_KNOWN_ALLOCATED) |
| return format (0, "next 0x%x: %U", |
| b->next_buffer, format_vlib_buffer_known_state, k); |
| |
| if (unique_hash) |
| { |
| if (hash_get (*unique_hash, b->next_buffer)) |
| return format (0, "duplicate buffer 0x%x", b->next_buffer); |
| |
| hash_set1 (*unique_hash, b->next_buffer); |
| } |
| |
| msg = vlib_validate_buffer (vm, b->next_buffer, follow_buffer_next); |
| if (msg) |
| { |
| result = format (0, "next 0x%x: %v", b->next_buffer, msg); |
| vec_free (msg); |
| return result; |
| } |
| } |
| |
| return 0; |
| } |
| |
| u8 * |
| vlib_validate_buffer (vlib_main_t * vm, u32 bi, uword follow_buffer_next) |
| { |
| return vlib_validate_buffer_helper (vm, bi, follow_buffer_next, |
| /* unique_hash */ 0); |
| } |
| |
| u8 * |
| vlib_validate_buffers (vlib_main_t * vm, |
| u32 * buffers, |
| uword next_buffer_stride, |
| uword n_buffers, |
| vlib_buffer_known_state_t known_state, |
| uword follow_buffer_next) |
| { |
| uword i, *hash; |
| u32 bi, *b = buffers; |
| vlib_buffer_known_state_t k; |
| u8 *msg = 0, *result = 0; |
| |
| hash = hash_create (0, 0); |
| for (i = 0; i < n_buffers; i++) |
| { |
| bi = b[0]; |
| b += next_buffer_stride; |
| |
| /* Buffer is not unique. */ |
| if (hash_get (hash, bi)) |
| { |
| msg = format (0, "not unique"); |
| goto done; |
| } |
| |
| k = vlib_buffer_is_known (vm, bi); |
| if (k != known_state) |
| { |
| msg = format (0, "is %U; expected %U", |
| format_vlib_buffer_known_state, k, |
| format_vlib_buffer_known_state, known_state); |
| goto done; |
| } |
| |
| msg = vlib_validate_buffer_helper (vm, bi, follow_buffer_next, &hash); |
| if (msg) |
| goto done; |
| |
| hash_set1 (hash, bi); |
| } |
| |
| done: |
| if (msg) |
| { |
| result = format (0, "0x%x: %v", bi, msg); |
| vec_free (msg); |
| } |
| hash_free (hash); |
| return result; |
| } |
| |
| /* When debugging validate that given buffers are either known allocated |
| or known free. */ |
| void |
| vlib_buffer_validate_alloc_free (vlib_main_t * vm, |
| u32 * buffers, |
| uword n_buffers, |
| vlib_buffer_known_state_t expected_state) |
| { |
| vlib_buffer_main_t *bm = vm->buffer_main; |
| u32 *b; |
| uword i, bi, is_free; |
| |
| if (CLIB_DEBUG == 0) |
| return; |
| |
| is_free = expected_state == VLIB_BUFFER_KNOWN_ALLOCATED; |
| b = buffers; |
| for (i = 0; i < n_buffers; i++) |
| { |
| vlib_buffer_known_state_t known; |
| |
| bi = b[0]; |
| b += 1; |
| known = vlib_buffer_is_known (vm, bi); |
| |
| if (known == VLIB_BUFFER_UNKNOWN && |
| expected_state == VLIB_BUFFER_KNOWN_FREE) |
| known = VLIB_BUFFER_KNOWN_FREE; |
| |
| if (known != expected_state) |
| { |
| clib_panic ("%s %U buffer 0x%x", is_free ? "freeing" : "allocating", |
| format_vlib_buffer_known_state, known, bi); |
| } |
| |
| clib_spinlock_lock (&bm->buffer_known_hash_lockp); |
| hash_set (bm->buffer_known_hash, bi, is_free ? VLIB_BUFFER_KNOWN_FREE : |
| VLIB_BUFFER_KNOWN_ALLOCATED); |
| clib_spinlock_unlock (&bm->buffer_known_hash_lockp); |
| } |
| } |
| |
| void |
| vlib_packet_template_init (vlib_main_t * vm, |
| vlib_packet_template_t * t, |
| void *packet_data, |
| uword n_packet_data_bytes, |
| uword min_n_buffers_each_alloc, char *fmt, ...) |
| { |
| va_list va; |
| |
| va_start (va, fmt); |
| t->name = va_format (0, fmt, &va); |
| va_end (va); |
| |
| vlib_worker_thread_barrier_sync (vm); |
| |
| clib_memset (t, 0, sizeof (t[0])); |
| |
| vec_add (t->packet_data, packet_data, n_packet_data_bytes); |
| t->min_n_buffers_each_alloc = min_n_buffers_each_alloc; |
| vlib_worker_thread_barrier_release (vm); |
| } |
| |
| void * |
| vlib_packet_template_get_packet (vlib_main_t * vm, |
| vlib_packet_template_t * t, u32 * bi_result) |
| { |
| u32 bi; |
| vlib_buffer_t *b; |
| |
| if (vlib_buffer_alloc (vm, &bi, 1) != 1) |
| return 0; |
| |
| *bi_result = bi; |
| |
| b = vlib_get_buffer (vm, bi); |
| clib_memcpy_fast (vlib_buffer_get_current (b), |
| t->packet_data, vec_len (t->packet_data)); |
| b->current_length = vec_len (t->packet_data); |
| |
| return b->data; |
| } |
| |
| /* Append given data to end of buffer, possibly allocating new buffers. */ |
| int |
| vlib_buffer_add_data (vlib_main_t * vm, u32 * buffer_index, void *data, |
| u32 n_data_bytes) |
| { |
| u32 n_buffer_bytes, n_left, n_left_this_buffer, bi; |
| vlib_buffer_t *b; |
| void *d; |
| |
| bi = *buffer_index; |
| if (bi == ~0 && 1 != vlib_buffer_alloc (vm, &bi, 1)) |
| goto out_of_buffers; |
| |
| d = data; |
| n_left = n_data_bytes; |
| n_buffer_bytes = vlib_buffer_get_default_data_size (vm); |
| |
| b = vlib_get_buffer (vm, bi); |
| b->flags &= ~VLIB_BUFFER_TOTAL_LENGTH_VALID; |
| |
| /* Get to the end of the chain before we try to append data... */ |
| while (b->flags & VLIB_BUFFER_NEXT_PRESENT) |
| b = vlib_get_buffer (vm, b->next_buffer); |
| |
| while (1) |
| { |
| u32 n; |
| |
| ASSERT (n_buffer_bytes >= b->current_length); |
| n_left_this_buffer = |
| n_buffer_bytes - (b->current_data + b->current_length); |
| n = clib_min (n_left_this_buffer, n_left); |
| clib_memcpy_fast (vlib_buffer_get_current (b) + b->current_length, d, |
| n); |
| b->current_length += n; |
| n_left -= n; |
| if (n_left == 0) |
| break; |
| |
| d += n; |
| if (1 != vlib_buffer_alloc (vm, &b->next_buffer, 1)) |
| goto out_of_buffers; |
| |
| b->flags |= VLIB_BUFFER_NEXT_PRESENT; |
| |
| b = vlib_get_buffer (vm, b->next_buffer); |
| } |
| |
| *buffer_index = bi; |
| return 0; |
| |
| out_of_buffers: |
| clib_warning ("out of buffers"); |
| return 1; |
| } |
| |
| u16 |
| vlib_buffer_chain_append_data_with_alloc (vlib_main_t * vm, |
| vlib_buffer_t * first, |
| vlib_buffer_t ** last, void *data, |
| u16 data_len) |
| { |
| vlib_buffer_t *l = *last; |
| u32 n_buffer_bytes = vlib_buffer_get_default_data_size (vm); |
| u16 copied = 0; |
| ASSERT (n_buffer_bytes >= l->current_length + l->current_data); |
| while (data_len) |
| { |
| u16 max = n_buffer_bytes - l->current_length - l->current_data; |
| if (max == 0) |
| { |
| if (1 != vlib_buffer_alloc_from_pool (vm, &l->next_buffer, 1, |
| first->buffer_pool_index)) |
| return copied; |
| *last = l = vlib_buffer_chain_buffer (vm, l, l->next_buffer); |
| max = n_buffer_bytes - l->current_length - l->current_data; |
| } |
| |
| u16 len = (data_len > max) ? max : data_len; |
| clib_memcpy_fast (vlib_buffer_get_current (l) + l->current_length, |
| data + copied, len); |
| vlib_buffer_chain_increase_length (first, l, len); |
| data_len -= len; |
| copied += len; |
| } |
| return copied; |
| } |
| |
| static uword |
| vlib_buffer_alloc_size (uword ext_hdr_size, uword data_size) |
| { |
| uword alloc_size = ext_hdr_size + sizeof (vlib_buffer_t) + data_size; |
| alloc_size = CLIB_CACHE_LINE_ROUND (alloc_size); |
| |
| /* in case when we have even number of cachelines, we add one more for |
| * better cache occupancy */ |
| alloc_size |= CLIB_CACHE_LINE_BYTES; |
| |
| return alloc_size; |
| } |
| |
| u8 |
| vlib_buffer_pool_create (vlib_main_t * vm, char *name, u32 data_size, |
| u32 physmem_map_index) |
| { |
| vlib_buffer_main_t *bm = vm->buffer_main; |
| vlib_buffer_pool_t *bp; |
| vlib_physmem_map_t *m = vlib_physmem_get_map (vm, physmem_map_index); |
| uword start = pointer_to_uword (m->base); |
| uword size = (uword) m->n_pages << m->log2_page_size; |
| uword i, j; |
| u32 alloc_size, n_alloc_per_page; |
| |
| if (vec_len (bm->buffer_pools) >= 255) |
| return ~0; |
| |
| vec_add2_aligned (bm->buffer_pools, bp, 1, CLIB_CACHE_LINE_BYTES); |
| |
| if (bm->buffer_mem_size == 0) |
| { |
| bm->buffer_mem_start = start; |
| bm->buffer_mem_size = size; |
| } |
| else if (start < bm->buffer_mem_start) |
| { |
| bm->buffer_mem_size += bm->buffer_mem_start - start; |
| bm->buffer_mem_start = start; |
| if (size > bm->buffer_mem_size) |
| bm->buffer_mem_size = size; |
| } |
| else if (start > bm->buffer_mem_start) |
| { |
| uword new_size = start - bm->buffer_mem_start + size; |
| if (new_size > bm->buffer_mem_size) |
| bm->buffer_mem_size = new_size; |
| } |
| |
| if ((u64) bm->buffer_mem_size > |
| ((u64) 1 << (32 + CLIB_LOG2_CACHE_LINE_BYTES))) |
| { |
| clib_panic ("buffer memory size out of range!"); |
| } |
| |
| bp->start = start; |
| bp->size = size; |
| bp->index = bp - bm->buffer_pools; |
| bp->buffer_template.buffer_pool_index = bp->index; |
| bp->buffer_template.ref_count = 1; |
| bp->physmem_map_index = physmem_map_index; |
| bp->name = format (0, "%s%c", name, 0); |
| bp->data_size = data_size; |
| bp->numa_node = m->numa_node; |
| |
| vec_validate_aligned (bp->threads, vlib_get_n_threads () - 1, |
| CLIB_CACHE_LINE_BYTES); |
| |
| alloc_size = vlib_buffer_alloc_size (bm->ext_hdr_size, data_size); |
| n_alloc_per_page = (1ULL << m->log2_page_size) / alloc_size; |
| |
| /* preallocate buffer indices memory */ |
| bp->n_buffers = m->n_pages * n_alloc_per_page; |
| bp->buffers = clib_mem_alloc_aligned (bp->n_buffers * sizeof (u32), |
| CLIB_CACHE_LINE_BYTES); |
| |
| clib_spinlock_init (&bp->lock); |
| |
| for (j = 0; j < m->n_pages; j++) |
| for (i = 0; i < n_alloc_per_page; i++) |
| { |
| u8 *p; |
| u32 bi; |
| |
| p = m->base + (j << m->log2_page_size) + i * alloc_size; |
| p += bm->ext_hdr_size; |
| |
| /* |
| * Waste 1 buffer (maximum) so that 0 is never a valid buffer index. |
| * Allows various places to ASSERT (bi != 0). Much easier |
| * than debugging downstream crashes in successor nodes. |
| */ |
| if (p == m->base) |
| continue; |
| |
| vlib_buffer_copy_template ((vlib_buffer_t *) p, &bp->buffer_template); |
| |
| bi = vlib_get_buffer_index (vm, (vlib_buffer_t *) p); |
| |
| bp->buffers[bp->n_avail++] = bi; |
| |
| vlib_get_buffer (vm, bi); |
| } |
| |
| return bp->index; |
| } |
| |
| static u8 * |
| format_vlib_buffer_pool (u8 * s, va_list * va) |
| { |
| vlib_main_t *vm = va_arg (*va, vlib_main_t *); |
| vlib_buffer_pool_t *bp = va_arg (*va, vlib_buffer_pool_t *); |
| vlib_buffer_pool_thread_t *bpt; |
| u32 cached = 0; |
| |
| if (!bp) |
| return format (s, "%-20s%=6s%=6s%=6s%=11s%=6s%=8s%=8s%=8s", |
| "Pool Name", "Index", "NUMA", "Size", "Data Size", |
| "Total", "Avail", "Cached", "Used"); |
| |
| /* *INDENT-OFF* */ |
| vec_foreach (bpt, bp->threads) |
| cached += bpt->n_cached; |
| /* *INDENT-ON* */ |
| |
| s = format (s, "%-20s%=6d%=6d%=6u%=11u%=6u%=8u%=8u%=8u", |
| bp->name, bp->index, bp->numa_node, bp->data_size + |
| sizeof (vlib_buffer_t) + vm->buffer_main->ext_hdr_size, |
| bp->data_size, bp->n_buffers, bp->n_avail, cached, |
| bp->n_buffers - bp->n_avail - cached); |
| |
| return s; |
| } |
| |
| u8 * |
| format_vlib_buffer_pool_all (u8 *s, va_list *va) |
| { |
| vlib_main_t *vm = va_arg (*va, vlib_main_t *); |
| vlib_buffer_main_t *bm = vm->buffer_main; |
| vlib_buffer_pool_t *bp; |
| |
| s = format (s, "%U", format_vlib_buffer_pool, vm, 0); |
| |
| vec_foreach (bp, bm->buffer_pools) |
| s = format (s, "\n%U", format_vlib_buffer_pool, vm, bp); |
| |
| return s; |
| } |
| |
| static clib_error_t * |
| show_buffers (vlib_main_t *vm, unformat_input_t *input, |
| vlib_cli_command_t *cmd) |
| { |
| vlib_cli_output (vm, "%U", format_vlib_buffer_pool_all, vm); |
| return 0; |
| } |
| |
| /* *INDENT-OFF* */ |
| VLIB_CLI_COMMAND (show_buffers_command, static) = { |
| .path = "show buffers", |
| .short_help = "Show packet buffer allocation", |
| .function = show_buffers, |
| }; |
| /* *INDENT-ON* */ |
| |
| clib_error_t * |
| vlib_buffer_worker_init (vlib_main_t * vm) |
| { |
| vlib_buffer_main_t *bm = vm->buffer_main; |
| vlib_buffer_pool_t *bp; |
| |
| /* *INDENT-OFF* */ |
| vec_foreach (bp, bm->buffer_pools) |
| { |
| clib_spinlock_lock (&bp->lock); |
| vec_validate_aligned (bp->threads, vlib_get_n_threads () - 1, |
| CLIB_CACHE_LINE_BYTES); |
| clib_spinlock_unlock (&bp->lock); |
| } |
| /* *INDENT-ON* */ |
| |
| return 0; |
| } |
| |
| VLIB_WORKER_INIT_FUNCTION (vlib_buffer_worker_init); |
| |
| static clib_error_t * |
| vlib_buffer_main_init_numa_alloc (struct vlib_main_t *vm, u32 numa_node, |
| u32 * physmem_map_index, |
| clib_mem_page_sz_t log2_page_size, |
| u8 unpriv) |
| { |
| vlib_buffer_main_t *bm = vm->buffer_main; |
| u32 buffers_per_numa = bm->buffers_per_numa; |
| clib_error_t *error; |
| u32 buffer_size; |
| uword n_pages, pagesize; |
| u8 *name = 0; |
| |
| ASSERT (log2_page_size != CLIB_MEM_PAGE_SZ_UNKNOWN); |
| |
| pagesize = clib_mem_page_bytes (log2_page_size); |
| buffer_size = vlib_buffer_alloc_size (bm->ext_hdr_size, |
| vlib_buffer_get_default_data_size |
| (vm)); |
| if (buffer_size > pagesize) |
| return clib_error_return (0, "buffer size (%llu) is greater than page " |
| "size (%llu)", buffer_size, pagesize); |
| |
| if (buffers_per_numa == 0) |
| buffers_per_numa = unpriv ? VLIB_BUFFER_DEFAULT_BUFFERS_PER_NUMA_UNPRIV : |
| VLIB_BUFFER_DEFAULT_BUFFERS_PER_NUMA; |
| |
| name = format (0, "buffers-numa-%d%c", numa_node, 0); |
| n_pages = (buffers_per_numa - 1) / (pagesize / buffer_size) + 1; |
| error = vlib_physmem_shared_map_create (vm, (char *) name, |
| n_pages * pagesize, |
| min_log2 (pagesize), numa_node, |
| physmem_map_index); |
| vec_free (name); |
| return error; |
| } |
| |
| static clib_error_t * |
| vlib_buffer_main_init_numa_node (struct vlib_main_t *vm, u32 numa_node, |
| u8 * index) |
| { |
| vlib_buffer_main_t *bm = vm->buffer_main; |
| u32 physmem_map_index; |
| clib_error_t *error; |
| u8 *name = 0; |
| |
| if (bm->log2_page_size == CLIB_MEM_PAGE_SZ_UNKNOWN) |
| { |
| error = vlib_buffer_main_init_numa_alloc (vm, numa_node, |
| &physmem_map_index, |
| CLIB_MEM_PAGE_SZ_DEFAULT_HUGE, |
| 0 /* unpriv */ ); |
| if (!error) |
| goto buffer_pool_create; |
| |
| /* If alloc failed, retry without hugepages */ |
| vlib_log_warn (bm->log_default, |
| "numa[%u] falling back to non-hugepage backed " |
| "buffer pool (%U)", numa_node, format_clib_error, error); |
| clib_error_free (error); |
| |
| error = vlib_buffer_main_init_numa_alloc (vm, numa_node, |
| &physmem_map_index, |
| CLIB_MEM_PAGE_SZ_DEFAULT, |
| 1 /* unpriv */ ); |
| } |
| else |
| error = vlib_buffer_main_init_numa_alloc (vm, numa_node, |
| &physmem_map_index, |
| bm->log2_page_size, |
| 0 /* unpriv */ ); |
| if (error) |
| return error; |
| |
| buffer_pool_create: |
| name = format (name, "default-numa-%d%c", numa_node, 0); |
| *index = vlib_buffer_pool_create (vm, (char *) name, |
| vlib_buffer_get_default_data_size (vm), |
| physmem_map_index); |
| |
| if (*index == (u8) ~ 0) |
| error = clib_error_return (0, "maximum number of buffer pools reached"); |
| vec_free (name); |
| |
| |
| return error; |
| } |
| |
| void |
| vlib_buffer_main_alloc (vlib_main_t * vm) |
| { |
| vlib_buffer_main_t *bm; |
| |
| if (vm->buffer_main) |
| return; |
| |
| vm->buffer_main = bm = clib_mem_alloc (sizeof (bm[0])); |
| clib_memset (vm->buffer_main, 0, sizeof (bm[0])); |
| bm->default_data_size = VLIB_BUFFER_DEFAULT_DATA_SIZE; |
| } |
| |
| static u32 |
| buffer_get_cached (vlib_buffer_pool_t * bp) |
| { |
| u32 cached = 0; |
| vlib_buffer_pool_thread_t *bpt; |
| |
| clib_spinlock_lock (&bp->lock); |
| |
| /* *INDENT-OFF* */ |
| vec_foreach (bpt, bp->threads) |
| cached += bpt->n_cached; |
| /* *INDENT-ON* */ |
| |
| clib_spinlock_unlock (&bp->lock); |
| |
| return cached; |
| } |
| |
| static vlib_buffer_pool_t * |
| buffer_get_by_index (vlib_buffer_main_t * bm, u32 index) |
| { |
| vlib_buffer_pool_t *bp; |
| if (!bm->buffer_pools || vec_len (bm->buffer_pools) < index) |
| return 0; |
| bp = vec_elt_at_index (bm->buffer_pools, index); |
| |
| return bp; |
| } |
| |
| static void |
| buffer_gauges_collect_used_fn (vlib_stats_collector_data_t *d) |
| { |
| vlib_main_t *vm = vlib_get_main (); |
| vlib_buffer_pool_t *bp = |
| buffer_get_by_index (vm->buffer_main, d->private_data); |
| if (!bp) |
| return; |
| |
| d->entry->value = bp->n_buffers - bp->n_avail - buffer_get_cached (bp); |
| } |
| |
| static void |
| buffer_gauges_collect_available_fn (vlib_stats_collector_data_t *d) |
| { |
| vlib_main_t *vm = vlib_get_main (); |
| vlib_buffer_pool_t *bp = |
| buffer_get_by_index (vm->buffer_main, d->private_data); |
| if (!bp) |
| return; |
| |
| d->entry->value = bp->n_avail; |
| } |
| |
| static void |
| buffer_gauges_collect_cached_fn (vlib_stats_collector_data_t *d) |
| { |
| vlib_main_t *vm = vlib_get_main (); |
| vlib_buffer_pool_t *bp = |
| buffer_get_by_index (vm->buffer_main, d->private_data); |
| if (!bp) |
| return; |
| |
| d->entry->value = buffer_get_cached (bp); |
| } |
| |
| clib_error_t * |
| vlib_buffer_main_init (struct vlib_main_t * vm) |
| { |
| vlib_buffer_main_t *bm; |
| clib_error_t *err; |
| clib_bitmap_t *bmp = 0, *bmp_has_memory = 0; |
| u32 numa_node; |
| vlib_buffer_pool_t *bp; |
| u8 *name = 0, first_valid_buffer_pool_index = ~0; |
| |
| vlib_buffer_main_alloc (vm); |
| |
| bm = vm->buffer_main; |
| bm->log_default = vlib_log_register_class ("buffer", 0); |
| bm->ext_hdr_size = __vlib_buffer_external_hdr_size; |
| |
| clib_spinlock_init (&bm->buffer_known_hash_lockp); |
| |
| if ((err = clib_sysfs_read ("/sys/devices/system/node/online", "%U", |
| unformat_bitmap_list, &bmp))) |
| clib_error_free (err); |
| |
| if ((err = clib_sysfs_read ("/sys/devices/system/node/has_memory", "%U", |
| unformat_bitmap_list, &bmp_has_memory))) |
| clib_error_free (err); |
| |
| if (bmp && bmp_has_memory) |
| bmp = clib_bitmap_and (bmp, bmp_has_memory); |
| |
| /* no info from sysfs, assuming that only numa 0 exists */ |
| if (bmp == 0) |
| bmp = clib_bitmap_set (bmp, 0, 1); |
| |
| if (clib_bitmap_last_set (bmp) >= VLIB_BUFFER_MAX_NUMA_NODES) |
| clib_panic ("system have more than %u NUMA nodes", |
| VLIB_BUFFER_MAX_NUMA_NODES); |
| |
| /* *INDENT-OFF* */ |
| clib_bitmap_foreach (numa_node, bmp) |
| { |
| u8 *index = bm->default_buffer_pool_index_for_numa + numa_node; |
| index[0] = ~0; |
| if ((err = vlib_buffer_main_init_numa_node (vm, numa_node, index))) |
| { |
| clib_error_report (err); |
| clib_error_free (err); |
| continue; |
| } |
| |
| if (first_valid_buffer_pool_index == 0xff) |
| first_valid_buffer_pool_index = index[0]; |
| } |
| /* *INDENT-ON* */ |
| |
| if (first_valid_buffer_pool_index == (u8) ~ 0) |
| { |
| err = clib_error_return (0, "failed to allocate buffer pool(s)"); |
| goto done; |
| } |
| |
| /* *INDENT-OFF* */ |
| clib_bitmap_foreach (numa_node, bmp) |
| { |
| if (bm->default_buffer_pool_index_for_numa[numa_node] == (u8) ~0) |
| bm->default_buffer_pool_index_for_numa[numa_node] = |
| first_valid_buffer_pool_index; |
| } |
| /* *INDENT-ON* */ |
| |
| vec_foreach (bp, bm->buffer_pools) |
| { |
| vlib_stats_collector_reg_t reg = { .private_data = bp - bm->buffer_pools }; |
| if (bp->n_buffers == 0) |
| continue; |
| |
| reg.entry_index = |
| vlib_stats_add_gauge ("/buffer-pools/%s/cached", bp->name); |
| reg.collect_fn = buffer_gauges_collect_cached_fn; |
| vlib_stats_register_collector_fn (®); |
| |
| reg.entry_index = vlib_stats_add_gauge ("/buffer-pools/%s/used", bp->name); |
| reg.collect_fn = buffer_gauges_collect_used_fn; |
| vlib_stats_register_collector_fn (®); |
| |
| reg.entry_index = |
| vlib_stats_add_gauge ("/buffer-pools/%s/available", bp->name); |
| reg.collect_fn = buffer_gauges_collect_available_fn; |
| vlib_stats_register_collector_fn (®); |
| } |
| |
| done: |
| vec_free (bmp); |
| vec_free (bmp_has_memory); |
| vec_free (name); |
| return err; |
| } |
| |
| static clib_error_t * |
| vlib_buffers_configure (vlib_main_t * vm, unformat_input_t * input) |
| { |
| vlib_buffer_main_t *bm; |
| |
| vlib_buffer_main_alloc (vm); |
| |
| bm = vm->buffer_main; |
| bm->log2_page_size = CLIB_MEM_PAGE_SZ_UNKNOWN; |
| |
| while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (unformat (input, "buffers-per-numa %u", &bm->buffers_per_numa)) |
| ; |
| else if (unformat (input, "page-size %U", unformat_log2_page_size, |
| &bm->log2_page_size)) |
| ; |
| else if (unformat (input, "default data-size %u", |
| &bm->default_data_size)) |
| ; |
| else |
| return unformat_parse_error (input); |
| } |
| |
| unformat_free (input); |
| return 0; |
| } |
| |
| VLIB_EARLY_CONFIG_FUNCTION (vlib_buffers_configure, "buffers"); |
| |
| #if VLIB_BUFFER_ALLOC_FAULT_INJECTOR > 0 |
| u32 |
| vlib_buffer_alloc_may_fail (vlib_main_t * vm, u32 n_buffers) |
| { |
| f64 r; |
| |
| r = random_f64 (&vm->buffer_alloc_success_seed); |
| |
| /* Fail this request? */ |
| if (r > vm->buffer_alloc_success_rate) |
| n_buffers--; |
| /* 5% chance of returning nothing at all */ |
| if (r > vm->buffer_alloc_success_rate && r > 0.95) |
| n_buffers = 0; |
| |
| return n_buffers; |
| } |
| #endif |
| |
| __clib_export int |
| vlib_buffer_set_alloc_free_callback ( |
| vlib_main_t *vm, vlib_buffer_alloc_free_callback_t *alloc_callback_fn, |
| vlib_buffer_alloc_free_callback_t *free_callback_fn) |
| { |
| vlib_buffer_main_t *bm = vm->buffer_main; |
| if ((alloc_callback_fn && bm->alloc_callback_fn) || |
| (free_callback_fn && bm->free_callback_fn)) |
| return 1; |
| bm->alloc_callback_fn = alloc_callback_fn; |
| bm->free_callback_fn = free_callback_fn; |
| return 0; |
| } |
| |
| /** @endcond */ |
| /* |
| * fd.io coding-style-patch-verification: ON |
| * |
| * Local Variables: |
| * eval: (c-set-style "gnu") |
| * End: |
| */ |