blob: 7d27a0dc169184ae342f812ed00b93f0ace27aca [file] [log] [blame]
/* SPDX-License-Identifier: Apache-2.0
* Copyright(c) 2021 Cisco Systems, Inc.
* Copyright(c) 2024 Tom Jones <thj@freebsd.org>
*/
#define _GNU_SOURCE
#include <stdlib.h>
#include <sys/types.h>
#include <sys/stat.h>
#include <unistd.h>
#include <sys/memrange.h>
#include <sys/mount.h>
#include <sys/mman.h>
#include <fcntl.h>
#include <vppinfra/clib.h>
#include <vppinfra/mem.h>
#include <vppinfra/lock.h>
#include <vppinfra/time.h>
#include <vppinfra/bitmap.h>
#include <vppinfra/format.h>
#include <vppinfra/clib_error.h>
#ifndef F_FBSD_SPECIFIC_BASE
#define F_FBSD_SPECIFIC_BASE 1024
#endif
#ifndef F_ADD_SEALS
#define F_ADD_SEALS (F_FBSD_SPECIFIC_BASE + 9)
#define F_GET_SEALS (F_FBSD_SPECIFIC_BASE + 10)
#define F_SEAL_SEAL 0x0001 /* prevent further seals from being set */
#define F_SEAL_SHRINK 0x0002 /* prevent file from shrinking */
#define F_SEAL_GROW 0x0004 /* prevent file from growing */
#define F_SEAL_WRITE 0x0008 /* prevent writes */
#endif
#ifndef MFD_HUGETLB
#define MFD_HUGETLB 0x0004U
#endif
#ifndef MAP_HUGE_SHIFT
#define MAP_HUGE_SHIFT 26
#endif
#ifndef MFD_HUGE_SHIFT
#define MFD_HUGE_SHIFT 26
#endif
#ifndef MAP_FIXED_NOREPLACE
#define MAP_FIXED_NOREPLACE MAP_FIXED
#endif
static void
map_lock ()
{
while (clib_atomic_test_and_set (&clib_mem_main.map_lock))
CLIB_PAUSE ();
}
static void
map_unlock ()
{
clib_atomic_release (&clib_mem_main.map_lock);
}
void
clib_mem_main_init (void)
{
clib_mem_main_t *mm = &clib_mem_main;
long sysconf_page_size;
uword page_size;
void *va;
if (mm->log2_page_sz != CLIB_MEM_PAGE_SZ_UNKNOWN)
return;
/* system page size */
sysconf_page_size = sysconf (_SC_PAGESIZE);
if (sysconf_page_size < 0)
{
clib_panic ("Could not determine the page size");
}
page_size = sysconf_page_size;
mm->log2_page_sz = min_log2 (page_size);
mm->log2_default_hugepage_sz = min_log2 (page_size);
mm->log2_sys_default_hugepage_sz = mm->log2_default_hugepage_sz;
/* numa nodes */
va = mmap (0, page_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANONYMOUS,
-1, 0);
if (va == MAP_FAILED)
return;
if (mlock (va, page_size))
goto done;
/*
* TODO: In linux/mem.c we can move pages to numa domains, this isn't an
* option in FreeBSD yet.
*/
done:
munmap (va, page_size);
}
__clib_export u64
clib_mem_get_fd_page_size (int fd)
{
struct stat st = { 0 };
if (fstat (fd, &st) == -1)
return 0;
return st.st_blksize;
}
__clib_export clib_mem_page_sz_t
clib_mem_get_fd_log2_page_size (int fd)
{
uword page_size = clib_mem_get_fd_page_size (fd);
return page_size ? min_log2 (page_size) : CLIB_MEM_PAGE_SZ_UNKNOWN;
}
__clib_export void
clib_mem_vm_randomize_va (uword *requested_va,
clib_mem_page_sz_t log2_page_size)
{
/* TODO: Not yet implemented */
}
__clib_export int
clib_mem_vm_create_fd (clib_mem_page_sz_t log2_page_size, char *fmt, ...)
{
clib_mem_main_t *mm = &clib_mem_main;
int fd;
unsigned int memfd_flags;
va_list va;
u8 *s = 0;
if (log2_page_size == mm->log2_page_sz)
log2_page_size = CLIB_MEM_PAGE_SZ_DEFAULT;
else if (log2_page_size == mm->log2_sys_default_hugepage_sz)
log2_page_size = CLIB_MEM_PAGE_SZ_DEFAULT_HUGE;
switch (log2_page_size)
{
case CLIB_MEM_PAGE_SZ_UNKNOWN:
return CLIB_MEM_ERROR;
case CLIB_MEM_PAGE_SZ_DEFAULT:
memfd_flags = MFD_ALLOW_SEALING;
break;
case CLIB_MEM_PAGE_SZ_DEFAULT_HUGE:
memfd_flags = MFD_HUGETLB;
break;
default:
memfd_flags = MFD_HUGETLB | log2_page_size << MFD_HUGE_SHIFT;
}
va_start (va, fmt);
s = va_format (0, fmt, &va);
va_end (va);
/* memfd_create maximum string size is 249 chars without trailing zero */
if (vec_len (s) > 249)
vec_set_len (s, 249);
vec_add1 (s, 0);
fd = memfd_create ((char *) s, memfd_flags);
if (fd == -1)
{
vec_reset_length (mm->error);
mm->error = clib_error_return_unix (mm->error, "memfd_create");
vec_free (s);
return CLIB_MEM_ERROR;
}
vec_free (s);
if ((memfd_flags & MFD_ALLOW_SEALING) &&
((fcntl (fd, F_ADD_SEALS, F_SEAL_SHRINK)) == -1))
{
vec_reset_length (mm->error);
mm->error = clib_error_return_unix (mm->error, "fcntl (F_ADD_SEALS)");
close (fd);
return CLIB_MEM_ERROR;
}
return fd;
}
uword
clib_mem_vm_reserve (uword start, uword size, clib_mem_page_sz_t log2_page_sz)
{
clib_mem_main_t *mm = &clib_mem_main;
uword pagesize = 1ULL << log2_page_sz;
uword sys_page_sz = 1ULL << mm->log2_page_sz;
uword n_bytes;
void *base = 0, *p;
size = round_pow2 (size, pagesize);
/* in adition of requested reservation, we also rserve one system page
* (typically 4K) adjacent to the start off reservation */
if (start)
{
/* start address is provided, so we just need to make sure we are not
* replacing existing map */
if (start & pow2_mask (log2_page_sz))
return ~0;
base = (void *) start - sys_page_sz;
base = mmap (base, size + sys_page_sz, PROT_NONE,
MAP_PRIVATE | MAP_ANONYMOUS | MAP_FIXED_NOREPLACE, -1, 0);
return (base == MAP_FAILED) ? ~0 : start;
}
/* to make sure that we get reservation aligned to page_size we need to
* request one additional page as mmap will return us address which is
* aligned only to system page size */
base =
mmap (0, size + pagesize, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
if (base == MAP_FAILED)
return ~0;
/* return additional space at the end of allocation */
p = base + size + pagesize;
n_bytes = (uword) p & pow2_mask (log2_page_sz);
if (n_bytes)
{
p -= n_bytes;
munmap (p, n_bytes);
}
/* return additional space at the start of allocation */
n_bytes = pagesize - sys_page_sz - n_bytes;
if (n_bytes)
{
munmap (base, n_bytes);
base += n_bytes;
}
return (uword) base + sys_page_sz;
}
__clib_export clib_mem_vm_map_hdr_t *
clib_mem_vm_get_next_map_hdr (clib_mem_vm_map_hdr_t *hdr)
{
/* TODO: Not yet implemented */
return NULL;
}
void *
clib_mem_vm_map_internal (void *base, clib_mem_page_sz_t log2_page_sz,
uword size, int fd, uword offset, char *name)
{
clib_mem_main_t *mm = &clib_mem_main;
clib_mem_vm_map_hdr_t *hdr;
uword sys_page_sz = 1ULL << mm->log2_page_sz;
int mmap_flags = MAP_FIXED, is_huge = 0;
if (fd != -1)
{
mmap_flags |= MAP_SHARED;
log2_page_sz = clib_mem_get_fd_log2_page_size (fd);
if (log2_page_sz > mm->log2_page_sz)
is_huge = 1;
}
else
{
mmap_flags |= MAP_PRIVATE | MAP_ANONYMOUS;
if (log2_page_sz == mm->log2_page_sz)
log2_page_sz = CLIB_MEM_PAGE_SZ_DEFAULT;
switch (log2_page_sz)
{
case CLIB_MEM_PAGE_SZ_UNKNOWN:
/* will fail later */
break;
case CLIB_MEM_PAGE_SZ_DEFAULT:
log2_page_sz = mm->log2_page_sz;
break;
case CLIB_MEM_PAGE_SZ_DEFAULT_HUGE:
/* We shouldn't be selecting HUGETLB on FreeBSD */
log2_page_sz = CLIB_MEM_PAGE_SZ_UNKNOWN;
break;
default:
log2_page_sz = mm->log2_page_sz;
break;
}
}
size = round_pow2 (size, 1ULL << log2_page_sz);
base = (void *) clib_mem_vm_reserve ((uword) base, size, log2_page_sz);
if (base == (void *) ~0)
return CLIB_MEM_VM_MAP_FAILED;
base = mmap (base, size, PROT_READ | PROT_WRITE, mmap_flags, fd, offset);
if (base == MAP_FAILED)
return CLIB_MEM_VM_MAP_FAILED;
if (is_huge && (mlock (base, size) != 0))
{
munmap (base, size);
return CLIB_MEM_VM_MAP_FAILED;
}
hdr = mmap (base - sys_page_sz, sys_page_sz, PROT_READ | PROT_WRITE,
MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
if (hdr != base - sys_page_sz)
{
munmap (base, size);
return CLIB_MEM_VM_MAP_FAILED;
}
map_lock ();
if (mm->last_map)
{
mprotect (mm->last_map, sys_page_sz, PROT_READ | PROT_WRITE);
mm->last_map->next = hdr;
mprotect (mm->last_map, sys_page_sz, PROT_NONE);
}
else
mm->first_map = hdr;
clib_mem_unpoison (hdr, sys_page_sz);
hdr->next = 0;
hdr->prev = mm->last_map;
snprintf (hdr->name, CLIB_VM_MAP_HDR_NAME_MAX_LEN - 1, "%s", (char *) name);
mm->last_map = hdr;
hdr->base_addr = (uword) base;
hdr->log2_page_sz = log2_page_sz;
hdr->num_pages = size >> log2_page_sz;
hdr->fd = fd;
hdr->name[CLIB_VM_MAP_HDR_NAME_MAX_LEN - 1] = 0;
mprotect (hdr, sys_page_sz, PROT_NONE);
map_unlock ();
clib_mem_unpoison (base, size);
return base;
}
__clib_export int
clib_mem_vm_unmap (void *base)
{
clib_mem_main_t *mm = &clib_mem_main;
uword size, sys_page_sz = 1ULL << mm->log2_page_sz;
clib_mem_vm_map_hdr_t *hdr = base - sys_page_sz;
;
map_lock ();
if (mprotect (hdr, sys_page_sz, PROT_READ | PROT_WRITE) != 0)
goto out;
size = hdr->num_pages << hdr->log2_page_sz;
if (munmap ((void *) hdr->base_addr, size) != 0)
goto out;
if (hdr->next)
{
mprotect (hdr->next, sys_page_sz, PROT_READ | PROT_WRITE);
hdr->next->prev = hdr->prev;
mprotect (hdr->next, sys_page_sz, PROT_NONE);
}
else
mm->last_map = hdr->prev;
if (hdr->prev)
{
mprotect (hdr->prev, sys_page_sz, PROT_READ | PROT_WRITE);
hdr->prev->next = hdr->next;
mprotect (hdr->prev, sys_page_sz, PROT_NONE);
}
else
mm->first_map = hdr->next;
map_unlock ();
if (munmap (hdr, sys_page_sz) != 0)
return CLIB_MEM_ERROR;
return 0;
out:
map_unlock ();
return CLIB_MEM_ERROR;
}
__clib_export void
clib_mem_get_page_stats (void *start, clib_mem_page_sz_t log2_page_size,
uword n_pages, clib_mem_page_stats_t *stats)
{
int i, *status = 0;
void **ptr = 0;
log2_page_size = clib_mem_log2_page_size_validate (log2_page_size);
vec_validate (status, n_pages - 1);
vec_validate (ptr, n_pages - 1);
for (i = 0; i < n_pages; i++)
ptr[i] = start + (i << log2_page_size);
clib_memset (stats, 0, sizeof (clib_mem_page_stats_t));
stats->total = n_pages;
stats->log2_page_sz = log2_page_size;
/*
* TODO: Until FreeBSD has support for tracking pages in NUMA domains just
* return that all are unknown for the statsistics.
*/
stats->unknown = n_pages;
vec_free (status);
vec_free (ptr);
}
__clib_export u64 *
clib_mem_vm_get_paddr (void *mem, clib_mem_page_sz_t log2_page_size,
int n_pages)
{
struct mem_extract meme;
int pagesize = sysconf (_SC_PAGESIZE);
int fd;
int i;
u64 *r = 0;
log2_page_size = clib_mem_log2_page_size_validate (log2_page_size);
if ((fd = open ((char *) "/dev/mem", O_RDONLY)) == -1)
return 0;
for (i = 0; i < n_pages; i++)
{
meme.me_vaddr = pointer_to_uword (mem) + (((u64) i) << log2_page_size);
if (ioctl (fd, MEM_EXTRACT_PADDR, &meme) == -1)
goto done;
vec_add1 (r, meme.me_paddr * pagesize);
}
done:
close (fd);
if (vec_len (r) != n_pages)
{
vec_free (r);
return 0;
}
return r;
}
__clib_export int
clib_mem_set_numa_affinity (u8 numa_node, int force)
{
/* TODO: Not yet implemented */
return CLIB_MEM_ERROR;
}
__clib_export int
clib_mem_set_default_numa_affinity ()
{
/* TODO: Not yet implemented */
return 0;
}