| /* |
| * Copyright (c) 2017 Cisco and/or its affiliates. |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at: |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #define _GNU_SOURCE |
| #include <stdlib.h> |
| #include <sys/types.h> |
| #include <sys/stat.h> |
| #include <unistd.h> |
| #include <sys/mount.h> |
| #include <sys/mman.h> |
| #include <fcntl.h> |
| #include <linux/mempolicy.h> |
| #include <linux/memfd.h> |
| |
| #include <vppinfra/clib.h> |
| #include <vppinfra/mem.h> |
| #include <vppinfra/time.h> |
| #include <vppinfra/format.h> |
| #include <vppinfra/clib_error.h> |
| #include <vppinfra/linux/syscall.h> |
| #include <vppinfra/linux/sysfs.h> |
| |
| #ifndef F_LINUX_SPECIFIC_BASE |
| #define F_LINUX_SPECIFIC_BASE 1024 |
| #endif |
| |
| #ifndef F_ADD_SEALS |
| #define F_ADD_SEALS (F_LINUX_SPECIFIC_BASE + 9) |
| #define F_GET_SEALS (F_LINUX_SPECIFIC_BASE + 10) |
| |
| #define F_SEAL_SEAL 0x0001 /* prevent further seals from being set */ |
| #define F_SEAL_SHRINK 0x0002 /* prevent file from shrinking */ |
| #define F_SEAL_GROW 0x0004 /* prevent file from growing */ |
| #define F_SEAL_WRITE 0x0008 /* prevent writes */ |
| #endif |
| |
| #ifndef MFD_HUGETLB |
| #define MFD_HUGETLB 0x0004U |
| #endif |
| |
| #ifndef MAP_HUGE_SHIFT |
| #define MAP_HUGE_SHIFT 26 |
| #endif |
| |
| #ifndef MAP_FIXED_NOREPLACE |
| #define MAP_FIXED_NOREPLACE 0x100000 |
| #endif |
| |
| uword |
| clib_mem_get_default_hugepage_size (void) |
| { |
| unformat_input_t input; |
| static u32 size = 0; |
| int fd; |
| |
| if (size) |
| goto done; |
| |
| /* |
| * If the kernel doesn't support hugepages, /proc/meminfo won't |
| * say anything about it. Use the regular page size as a default. |
| */ |
| size = clib_mem_get_page_size () / 1024; |
| |
| if ((fd = open ("/proc/meminfo", 0)) == -1) |
| return 0; |
| |
| unformat_init_clib_file (&input, fd); |
| |
| while (unformat_check_input (&input) != UNFORMAT_END_OF_INPUT) |
| { |
| if (unformat (&input, "Hugepagesize:%_%u kB", &size)) |
| ; |
| else |
| unformat_skip_line (&input); |
| } |
| unformat_free (&input); |
| close (fd); |
| done: |
| return 1024ULL * size; |
| } |
| |
| static clib_mem_page_sz_t |
| legacy_get_log2_default_hugepage_size (void) |
| { |
| clib_mem_page_sz_t log2_page_size = CLIB_MEM_PAGE_SZ_UNKNOWN; |
| FILE *fp; |
| char tmp[33] = { }; |
| |
| if ((fp = fopen ("/proc/meminfo", "r")) == NULL) |
| return CLIB_MEM_PAGE_SZ_UNKNOWN; |
| |
| while (fscanf (fp, "%32s", tmp) > 0) |
| if (strncmp ("Hugepagesize:", tmp, 13) == 0) |
| { |
| u32 size; |
| if (fscanf (fp, "%u", &size) > 0) |
| log2_page_size = 10 + min_log2 (size); |
| break; |
| } |
| |
| fclose (fp); |
| return log2_page_size; |
| } |
| |
| void |
| clib_mem_main_init () |
| { |
| clib_mem_main_t *mm = &clib_mem_main; |
| uword page_size; |
| void *va; |
| int fd; |
| |
| if (mm->log2_page_sz != CLIB_MEM_PAGE_SZ_UNKNOWN) |
| return; |
| |
| /* system page size */ |
| page_size = sysconf (_SC_PAGESIZE); |
| mm->log2_page_sz = min_log2 (page_size); |
| |
| /* default system hugeppage size */ |
| if ((fd = memfd_create ("test", MFD_HUGETLB)) != -1) |
| { |
| mm->log2_default_hugepage_sz = clib_mem_get_fd_log2_page_size (fd); |
| close (fd); |
| } |
| else /* likely kernel older than 4.14 */ |
| mm->log2_default_hugepage_sz = legacy_get_log2_default_hugepage_size (); |
| |
| /* numa nodes */ |
| va = mmap (0, page_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | |
| MAP_ANONYMOUS, -1, 0); |
| if (va == MAP_FAILED) |
| return; |
| |
| if (mlock (va, page_size)) |
| goto done; |
| |
| for (int i = 0; i < CLIB_MAX_NUMAS; i++) |
| { |
| int status; |
| if (move_pages (0, 1, &va, &i, &status, 0) == 0) |
| mm->numa_node_bitmap |= 1ULL << i; |
| } |
| |
| done: |
| munmap (va, page_size); |
| } |
| |
| u64 |
| clib_mem_get_fd_page_size (int fd) |
| { |
| struct stat st = { 0 }; |
| if (fstat (fd, &st) == -1) |
| return 0; |
| return st.st_blksize; |
| } |
| |
| clib_mem_page_sz_t |
| clib_mem_get_fd_log2_page_size (int fd) |
| { |
| uword page_size = clib_mem_get_fd_page_size (fd); |
| return page_size ? min_log2 (page_size) : CLIB_MEM_PAGE_SZ_UNKNOWN; |
| } |
| |
| void |
| clib_mem_vm_randomize_va (uword * requested_va, |
| clib_mem_page_sz_t log2_page_size) |
| { |
| u8 bit_mask = 15; |
| |
| if (log2_page_size <= 12) |
| bit_mask = 15; |
| else if (log2_page_size > 12 && log2_page_size <= 16) |
| bit_mask = 3; |
| else |
| bit_mask = 0; |
| |
| *requested_va += |
| (clib_cpu_time_now () & bit_mask) * (1ull << log2_page_size); |
| } |
| |
| clib_error_t * |
| clib_mem_create_fd (char *name, int *fdp) |
| { |
| int fd; |
| |
| ASSERT (name); |
| |
| if ((fd = memfd_create (name, MFD_ALLOW_SEALING)) == -1) |
| return clib_error_return_unix (0, "memfd_create"); |
| |
| if ((fcntl (fd, F_ADD_SEALS, F_SEAL_SHRINK)) == -1) |
| { |
| close (fd); |
| return clib_error_return_unix (0, "fcntl (F_ADD_SEALS)"); |
| } |
| |
| *fdp = fd; |
| return 0; |
| } |
| |
| clib_error_t * |
| clib_mem_create_hugetlb_fd (char *name, int *fdp) |
| { |
| clib_error_t *err = 0; |
| int fd = -1; |
| static int memfd_hugetlb_supported = 1; |
| char *mount_dir; |
| char template[] = "/tmp/hugepage_mount.XXXXXX"; |
| u8 *filename; |
| |
| ASSERT (name); |
| |
| if (memfd_hugetlb_supported) |
| { |
| if ((fd = memfd_create (name, MFD_HUGETLB)) != -1) |
| goto done; |
| |
| /* avoid further tries if memfd MFD_HUGETLB is not supported */ |
| if (errno == EINVAL && strnlen (name, 256) <= 249) |
| memfd_hugetlb_supported = 0; |
| } |
| |
| mount_dir = mkdtemp (template); |
| if (mount_dir == 0) |
| return clib_error_return_unix (0, "mkdtemp \'%s\'", template); |
| |
| if (mount ("none", (char *) mount_dir, "hugetlbfs", 0, NULL)) |
| { |
| rmdir ((char *) mount_dir); |
| err = clib_error_return_unix (0, "mount hugetlb directory '%s'", |
| mount_dir); |
| } |
| |
| filename = format (0, "%s/%s%c", mount_dir, name, 0); |
| fd = open ((char *) filename, O_CREAT | O_RDWR, 0755); |
| umount2 ((char *) mount_dir, MNT_DETACH); |
| rmdir ((char *) mount_dir); |
| |
| if (fd == -1) |
| err = clib_error_return_unix (0, "open"); |
| |
| done: |
| if (fd != -1) |
| fdp[0] = fd; |
| return err; |
| } |
| |
| clib_error_t * |
| clib_mem_vm_ext_alloc (clib_mem_vm_alloc_t * a) |
| { |
| int fd = -1; |
| clib_error_t *err = 0; |
| void *addr = 0; |
| u8 *filename = 0; |
| int mmap_flags = 0; |
| int log2_page_size; |
| int n_pages; |
| int old_mpol = -1; |
| long unsigned int old_mask[16] = { 0 }; |
| |
| /* save old numa mem policy if needed */ |
| if (a->flags & (CLIB_MEM_VM_F_NUMA_PREFER | CLIB_MEM_VM_F_NUMA_FORCE)) |
| { |
| int rv; |
| rv = get_mempolicy (&old_mpol, old_mask, sizeof (old_mask) * 8 + 1, |
| 0, 0); |
| |
| if (rv == -1) |
| { |
| if (a->numa_node != 0 && (a->flags & CLIB_MEM_VM_F_NUMA_FORCE) != 0) |
| { |
| err = clib_error_return_unix (0, "get_mempolicy"); |
| goto error; |
| } |
| else |
| old_mpol = -1; |
| } |
| } |
| |
| if (a->flags & CLIB_MEM_VM_F_LOCKED) |
| mmap_flags |= MAP_LOCKED; |
| |
| /* if we are creating shared segment, we need file descriptor */ |
| if (a->flags & CLIB_MEM_VM_F_SHARED) |
| { |
| mmap_flags |= MAP_SHARED; |
| /* if hugepages are needed we need to create mount point */ |
| if (a->flags & CLIB_MEM_VM_F_HUGETLB) |
| { |
| if ((err = clib_mem_create_hugetlb_fd (a->name, &fd))) |
| goto error; |
| |
| mmap_flags |= MAP_LOCKED; |
| } |
| else |
| { |
| if ((err = clib_mem_create_fd (a->name, &fd))) |
| goto error; |
| } |
| |
| log2_page_size = clib_mem_get_fd_log2_page_size (fd); |
| if (log2_page_size == 0) |
| { |
| err = clib_error_return_unix (0, "cannot determine page size"); |
| goto error; |
| } |
| |
| if (a->requested_va) |
| { |
| clib_mem_vm_randomize_va (&a->requested_va, log2_page_size); |
| mmap_flags |= MAP_FIXED; |
| } |
| } |
| else /* not CLIB_MEM_VM_F_SHARED */ |
| { |
| mmap_flags |= MAP_PRIVATE | MAP_ANONYMOUS; |
| if (a->flags & CLIB_MEM_VM_F_HUGETLB) |
| { |
| mmap_flags |= MAP_HUGETLB; |
| log2_page_size = 21; |
| } |
| else |
| { |
| log2_page_size = min_log2 (sysconf (_SC_PAGESIZE)); |
| } |
| } |
| |
| n_pages = ((a->size - 1) >> log2_page_size) + 1; |
| |
| if (a->flags & CLIB_MEM_VM_F_HUGETLB_PREALLOC) |
| { |
| err = clib_sysfs_prealloc_hugepages (a->numa_node, log2_page_size, |
| n_pages); |
| if (err) |
| goto error; |
| |
| } |
| |
| if (fd != -1) |
| if ((ftruncate (fd, (u64) n_pages * (1 << log2_page_size))) == -1) |
| { |
| err = clib_error_return_unix (0, "ftruncate"); |
| goto error; |
| } |
| |
| if (old_mpol != -1) |
| { |
| int rv; |
| long unsigned int mask[16] = { 0 }; |
| mask[0] = 1 << a->numa_node; |
| rv = set_mempolicy (MPOL_BIND, mask, sizeof (mask) * 8 + 1); |
| if (rv == -1 && a->numa_node != 0 && |
| (a->flags & CLIB_MEM_VM_F_NUMA_FORCE) != 0) |
| { |
| err = clib_error_return_unix (0, "set_mempolicy"); |
| goto error; |
| } |
| } |
| |
| addr = mmap (uword_to_pointer (a->requested_va, void *), a->size, |
| (PROT_READ | PROT_WRITE), mmap_flags, fd, 0); |
| if (addr == MAP_FAILED) |
| { |
| err = clib_error_return_unix (0, "mmap"); |
| goto error; |
| } |
| |
| /* re-apply old numa memory policy */ |
| if (old_mpol != -1 && |
| set_mempolicy (old_mpol, old_mask, sizeof (old_mask) * 8 + 1) == -1) |
| { |
| err = clib_error_return_unix (0, "set_mempolicy"); |
| goto error; |
| } |
| |
| a->log2_page_size = log2_page_size; |
| a->n_pages = n_pages; |
| a->addr = addr; |
| a->fd = fd; |
| CLIB_MEM_UNPOISON (addr, a->size); |
| goto done; |
| |
| error: |
| if (fd != -1) |
| close (fd); |
| |
| done: |
| vec_free (filename); |
| return err; |
| } |
| |
| void |
| clib_mem_vm_ext_free (clib_mem_vm_alloc_t * a) |
| { |
| if (a != 0) |
| { |
| clib_mem_vm_free (a->addr, 1ull << a->log2_page_size); |
| if (a->fd != -1) |
| close (a->fd); |
| } |
| } |
| |
| uword |
| clib_mem_vm_reserve (uword start, uword size, clib_mem_page_sz_t log2_page_sz) |
| { |
| clib_mem_main_t *mm = &clib_mem_main; |
| uword pagesize = 1ULL << log2_page_sz; |
| uword sys_page_sz = 1ULL << mm->log2_page_sz; |
| uword n_bytes; |
| void *base = 0, *p; |
| |
| size = round_pow2 (size, pagesize); |
| |
| /* in adition of requested reservation, we also rserve one system page |
| * (typically 4K) adjacent to the start off reservation */ |
| |
| if (start) |
| { |
| /* start address is provided, so we just need to make sure we are not |
| * replacing existing map */ |
| if (start & pow2_mask (log2_page_sz)) |
| return ~0; |
| |
| base = (void *) start - sys_page_sz; |
| base = mmap (base, size + sys_page_sz, PROT_NONE, |
| MAP_PRIVATE | MAP_ANONYMOUS | MAP_FIXED_NOREPLACE, -1, 0); |
| return (base == MAP_FAILED) ? ~0 : start; |
| } |
| |
| /* to make sure that we get reservation aligned to page_size we need to |
| * request one additional page as mmap will return us address which is |
| * aligned only to system page size */ |
| base = mmap (0, size + pagesize, PROT_NONE, |
| MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); |
| |
| if (base == MAP_FAILED) |
| return ~0; |
| |
| /* return additional space at the end of allocation */ |
| p = base + size + pagesize; |
| n_bytes = (uword) p & pow2_mask (log2_page_sz); |
| if (n_bytes) |
| { |
| p -= n_bytes; |
| munmap (p, n_bytes); |
| } |
| |
| /* return additional space at the start of allocation */ |
| n_bytes = pagesize - sys_page_sz - n_bytes; |
| if (n_bytes) |
| { |
| munmap (base, n_bytes); |
| base += n_bytes; |
| } |
| |
| return (uword) base + sys_page_sz; |
| } |
| |
| clib_mem_vm_map_hdr_t * |
| clib_mem_vm_get_next_map_hdr (clib_mem_vm_map_hdr_t * hdr) |
| { |
| clib_mem_main_t *mm = &clib_mem_main; |
| uword sys_page_sz = 1 << mm->log2_page_sz; |
| clib_mem_vm_map_hdr_t *next; |
| if (hdr == 0) |
| { |
| hdr = mm->first_map; |
| if (hdr) |
| mprotect (hdr, sys_page_sz, PROT_READ); |
| return hdr; |
| } |
| next = hdr->next; |
| mprotect (hdr, sys_page_sz, PROT_NONE); |
| if (next) |
| mprotect (next, sys_page_sz, PROT_READ); |
| return next; |
| } |
| |
| void * |
| clib_mem_vm_map_internal (void *base, clib_mem_page_sz_t log2_page_sz, |
| uword size, int fd, uword offset, char *name) |
| { |
| clib_mem_main_t *mm = &clib_mem_main; |
| clib_mem_vm_map_hdr_t *hdr; |
| uword sys_page_sz = 1 << mm->log2_page_sz; |
| int mmap_flags = MAP_FIXED, is_huge = 0; |
| |
| if (fd != -1) |
| { |
| mmap_flags |= MAP_SHARED; |
| log2_page_sz = clib_mem_get_fd_log2_page_size (fd); |
| if (log2_page_sz > mm->log2_page_sz) |
| is_huge = 1; |
| } |
| else |
| { |
| mmap_flags |= MAP_PRIVATE | MAP_ANONYMOUS; |
| |
| if (log2_page_sz == mm->log2_page_sz) |
| log2_page_sz = CLIB_MEM_PAGE_SZ_DEFAULT; |
| |
| switch (log2_page_sz) |
| { |
| case CLIB_MEM_PAGE_SZ_UNKNOWN: |
| /* will fail later */ |
| break; |
| case CLIB_MEM_PAGE_SZ_DEFAULT: |
| log2_page_sz = mm->log2_page_sz; |
| break; |
| case CLIB_MEM_PAGE_SZ_DEFAULT_HUGE: |
| mmap_flags |= MAP_HUGETLB; |
| log2_page_sz = mm->log2_default_hugepage_sz; |
| is_huge = 1; |
| break; |
| default: |
| mmap_flags |= MAP_HUGETLB; |
| mmap_flags |= log2_page_sz << MAP_HUGE_SHIFT; |
| is_huge = 1; |
| } |
| } |
| |
| if (log2_page_sz == CLIB_MEM_PAGE_SZ_UNKNOWN) |
| return CLIB_MEM_VM_MAP_FAILED; |
| |
| size = round_pow2 (size, 1 << log2_page_sz); |
| |
| base = (void *) clib_mem_vm_reserve ((uword) base, size, log2_page_sz); |
| |
| if (base == (void *) ~0) |
| return CLIB_MEM_VM_MAP_FAILED; |
| |
| base = mmap (base, size, PROT_READ | PROT_WRITE, mmap_flags, fd, offset); |
| |
| if (base == MAP_FAILED) |
| return CLIB_MEM_VM_MAP_FAILED; |
| |
| if (is_huge && (mlock (base, size) != 0)) |
| { |
| munmap (base, size); |
| return CLIB_MEM_VM_MAP_FAILED; |
| } |
| |
| hdr = mmap (base - sys_page_sz, sys_page_sz, PROT_READ | PROT_WRITE, |
| MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0); |
| |
| if (hdr != base - sys_page_sz) |
| { |
| munmap (base, size); |
| return CLIB_MEM_VM_MAP_FAILED; |
| } |
| |
| if (mm->last_map) |
| { |
| mprotect (mm->last_map, sys_page_sz, PROT_READ | PROT_WRITE); |
| mm->last_map->next = hdr; |
| mprotect (mm->last_map, sys_page_sz, PROT_NONE); |
| } |
| else |
| mm->first_map = hdr; |
| |
| hdr->next = 0; |
| hdr->prev = mm->last_map; |
| mm->last_map = hdr; |
| |
| hdr->base_addr = (uword) base; |
| hdr->log2_page_sz = log2_page_sz; |
| hdr->num_pages = size >> log2_page_sz; |
| hdr->fd = fd; |
| snprintf (hdr->name, CLIB_VM_MAP_HDR_NAME_MAX_LEN - 1, "%s", (char *) name); |
| hdr->name[CLIB_VM_MAP_HDR_NAME_MAX_LEN - 1] = 0; |
| mprotect (hdr, sys_page_sz, PROT_NONE); |
| |
| CLIB_MEM_UNPOISON (base, size); |
| return base; |
| } |
| |
| int |
| clib_mem_vm_unmap (void *base) |
| { |
| clib_mem_main_t *mm = &clib_mem_main; |
| uword size, sys_page_sz = 1 << mm->log2_page_sz; |
| clib_mem_vm_map_hdr_t *hdr = base - sys_page_sz;; |
| |
| if (mprotect (hdr, sys_page_sz, PROT_READ | PROT_WRITE) != 0) |
| return -1; |
| |
| size = hdr->num_pages << hdr->log2_page_sz; |
| if (munmap ((void *) hdr->base_addr, size) != 0) |
| return -1; |
| |
| if (hdr->next) |
| { |
| mprotect (hdr->next, sys_page_sz, PROT_READ | PROT_WRITE); |
| hdr->next->prev = hdr->prev; |
| mprotect (hdr->next, sys_page_sz, PROT_NONE); |
| } |
| else |
| mm->last_map = hdr->prev; |
| |
| if (hdr->prev) |
| { |
| mprotect (hdr->prev, sys_page_sz, PROT_READ | PROT_WRITE); |
| hdr->prev->next = hdr->next; |
| mprotect (hdr->prev, sys_page_sz, PROT_NONE); |
| } |
| else |
| mm->first_map = hdr->next; |
| |
| if (munmap (hdr, sys_page_sz) != 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| void |
| clib_mem_get_page_stats (void *start, clib_mem_page_sz_t log2_page_size, |
| uword n_pages, clib_mem_page_stats_t * stats) |
| { |
| int i, *status = 0; |
| void **ptr = 0; |
| |
| log2_page_size = clib_mem_log2_page_size_validate (log2_page_size); |
| |
| vec_validate (status, n_pages - 1); |
| vec_validate (ptr, n_pages - 1); |
| |
| for (i = 0; i < n_pages; i++) |
| ptr[i] = start + (i << log2_page_size); |
| |
| clib_memset (stats, 0, sizeof (clib_mem_page_stats_t)); |
| |
| if (move_pages (0, n_pages, ptr, 0, status, 0) != 0) |
| { |
| stats->unknown = n_pages; |
| return; |
| } |
| |
| for (i = 0; i < n_pages; i++) |
| { |
| if (status[i] >= 0 && status[i] < CLIB_MAX_NUMAS) |
| { |
| stats->mapped++; |
| stats->per_numa[status[i]]++; |
| } |
| else if (status[i] == -EFAULT) |
| stats->not_mapped++; |
| else |
| stats->unknown++; |
| } |
| } |
| |
| |
| u64 * |
| clib_mem_vm_get_paddr (void *mem, clib_mem_page_sz_t log2_page_size, |
| int n_pages) |
| { |
| int pagesize = sysconf (_SC_PAGESIZE); |
| int fd; |
| int i; |
| u64 *r = 0; |
| |
| log2_page_size = clib_mem_log2_page_size_validate (log2_page_size); |
| |
| if ((fd = open ((char *) "/proc/self/pagemap", O_RDONLY)) == -1) |
| return 0; |
| |
| for (i = 0; i < n_pages; i++) |
| { |
| u64 seek, pagemap = 0; |
| uword vaddr = pointer_to_uword (mem) + (((u64) i) << log2_page_size); |
| seek = ((u64) vaddr / pagesize) * sizeof (u64); |
| if (lseek (fd, seek, SEEK_SET) != seek) |
| goto done; |
| |
| if (read (fd, &pagemap, sizeof (pagemap)) != (sizeof (pagemap))) |
| goto done; |
| |
| if ((pagemap & (1ULL << 63)) == 0) |
| goto done; |
| |
| pagemap &= pow2_mask (55); |
| vec_add1 (r, pagemap * pagesize); |
| } |
| |
| done: |
| close (fd); |
| if (vec_len (r) != n_pages) |
| { |
| vec_free (r); |
| return 0; |
| } |
| return r; |
| } |
| |
| clib_error_t * |
| clib_mem_vm_ext_map (clib_mem_vm_map_t * a) |
| { |
| long unsigned int old_mask[16] = { 0 }; |
| int mmap_flags = MAP_SHARED; |
| clib_error_t *err = 0; |
| int old_mpol = -1; |
| void *addr; |
| int rv; |
| |
| if (a->numa_node) |
| { |
| rv = get_mempolicy (&old_mpol, old_mask, sizeof (old_mask) * 8 + 1, 0, |
| 0); |
| |
| if (rv == -1) |
| { |
| err = clib_error_return_unix (0, "get_mempolicy"); |
| goto done; |
| } |
| } |
| |
| if (a->requested_va) |
| mmap_flags |= MAP_FIXED; |
| |
| if (old_mpol != -1) |
| { |
| long unsigned int mask[16] = { 0 }; |
| mask[0] = 1 << a->numa_node; |
| rv = set_mempolicy (MPOL_BIND, mask, sizeof (mask) * 8 + 1); |
| if (rv == -1) |
| { |
| err = clib_error_return_unix (0, "set_mempolicy"); |
| goto done; |
| } |
| } |
| |
| addr = (void *) mmap (uword_to_pointer (a->requested_va, void *), a->size, |
| PROT_READ | PROT_WRITE, mmap_flags, a->fd, 0); |
| |
| if (addr == MAP_FAILED) |
| return clib_error_return_unix (0, "mmap"); |
| |
| /* re-apply old numa memory policy */ |
| if (old_mpol != -1 && |
| set_mempolicy (old_mpol, old_mask, sizeof (old_mask) * 8 + 1) == -1) |
| { |
| err = clib_error_return_unix (0, "set_mempolicy"); |
| goto done; |
| } |
| |
| a->addr = addr; |
| CLIB_MEM_UNPOISON (addr, a->size); |
| |
| done: |
| return err; |
| } |
| |
| /* |
| * fd.io coding-style-patch-verification: ON |
| * |
| * Local Variables: |
| * eval: (c-set-style "gnu") |
| * End: |
| */ |